1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright 2008 Sun Microsystems, Inc. All rights reserved. 23 * Use is subject to license terms. 24 */ 25 26 27 /* 28 * PIM-DR layer of DR driver. Provides interface between user 29 * level applications and the PSM-DR layer. 30 */ 31 32 #include <sys/note.h> 33 #include <sys/debug.h> 34 #include <sys/types.h> 35 #include <sys/errno.h> 36 #include <sys/cred.h> 37 #include <sys/dditypes.h> 38 #include <sys/devops.h> 39 #include <sys/modctl.h> 40 #include <sys/poll.h> 41 #include <sys/conf.h> 42 #include <sys/ddi.h> 43 #include <sys/sunddi.h> 44 #include <sys/sunndi.h> 45 #include <sys/stat.h> 46 #include <sys/kmem.h> 47 #include <sys/processor.h> 48 #include <sys/cpuvar.h> 49 #include <sys/mem_config.h> 50 51 #include <sys/autoconf.h> 52 #include <sys/cmn_err.h> 53 54 #include <sys/ddi_impldefs.h> 55 #include <sys/promif.h> 56 #include <sys/machsystm.h> 57 58 #include <sys/dr.h> 59 #include <sys/drmach.h> 60 #include <sys/dr_util.h> 61 62 extern int nulldev(); 63 extern int nodev(); 64 extern struct memlist *phys_install; 65 66 #ifdef DEBUG 67 uint_t dr_debug = 0; /* dr.h for bit values */ 68 #endif /* DEBUG */ 69 70 /* 71 * NOTE: state_str, nt_str and SBD_CMD_STR are only used in a debug 72 * kernel. They are, however, referenced during both debug and non-debug 73 * compiles. 74 */ 75 76 static char *state_str[] = { 77 "EMPTY", "OCCUPIED", "CONNECTED", "UNCONFIGURED", 78 "PARTIAL", "CONFIGURED", "RELEASE", "UNREFERENCED", 79 "FATAL" 80 }; 81 82 #define SBD_CMD_STR(c) \ 83 (((c) == SBD_CMD_ASSIGN) ? "ASSIGN" : \ 84 ((c) == SBD_CMD_UNASSIGN) ? "UNASSIGN" : \ 85 ((c) == SBD_CMD_POWERON) ? "POWERON" : \ 86 ((c) == SBD_CMD_POWEROFF) ? "POWEROFF" : \ 87 ((c) == SBD_CMD_TEST) ? "TEST" : \ 88 ((c) == SBD_CMD_CONNECT) ? "CONNECT" : \ 89 ((c) == SBD_CMD_DISCONNECT) ? "DISCONNECT" : \ 90 ((c) == SBD_CMD_CONFIGURE) ? "CONFIGURE" : \ 91 ((c) == SBD_CMD_UNCONFIGURE) ? "UNCONFIGURE" : \ 92 ((c) == SBD_CMD_GETNCM) ? "GETNCM" : \ 93 ((c) == SBD_CMD_PASSTHRU) ? "PASSTHRU" : \ 94 ((c) == SBD_CMD_STATUS) ? "STATUS" : "unknown") 95 96 #define DR_GET_BOARD_DEVUNIT(sb, ut, un) (&((sb)->b_dev[NIX(ut)][un])) 97 98 #define DR_MAKE_MINOR(i, b) (((i) << 16) | (b)) 99 #define DR_MINOR2INST(m) (((m) >> 16) & 0xffff) 100 #define DR_MINOR2BNUM(m) ((m) & 0xffff) 101 102 /* for the DR*INTERNAL_ERROR macros. see sys/dr.h. */ 103 static char *dr_ie_fmt = "dr.c %d"; 104 105 /* struct for drmach device name to sbd_comp_type_t mapping */ 106 typedef struct { 107 char *s_devtype; 108 sbd_comp_type_t s_nodetype; 109 } dr_devname_t; 110 111 /* struct to map starfire device attributes - name:sbd_comp_type_t */ 112 static dr_devname_t dr_devattr[] = { 113 { DRMACH_DEVTYPE_MEM, SBD_COMP_MEM }, 114 { DRMACH_DEVTYPE_CPU, SBD_COMP_CPU }, 115 { DRMACH_DEVTYPE_PCI, SBD_COMP_IO }, 116 #if defined(DRMACH_DEVTYPE_SBUS) 117 { DRMACH_DEVTYPE_SBUS, SBD_COMP_IO }, 118 #endif 119 #if defined(DRMACH_DEVTYPE_WCI) 120 { DRMACH_DEVTYPE_WCI, SBD_COMP_IO }, 121 #endif 122 /* last s_devtype must be NULL, s_nodetype must be SBD_COMP_UNKNOWN */ 123 { NULL, SBD_COMP_UNKNOWN } 124 }; 125 126 /* 127 * Per instance soft-state structure. 128 */ 129 typedef struct dr_softstate { 130 dev_info_t *dip; 131 dr_board_t *boards; 132 kmutex_t i_lock; 133 int dr_initialized; 134 } dr_softstate_t; 135 136 /* 137 * dr Global data elements 138 */ 139 struct dr_global { 140 dr_softstate_t *softsp; /* pointer to initialize soft state */ 141 kmutex_t lock; 142 } dr_g; 143 144 dr_unsafe_devs_t dr_unsafe_devs; 145 146 /* 147 * Table of known passthru commands. 148 */ 149 150 struct { 151 char *pt_name; 152 int (*pt_func)(dr_handle_t *); 153 } pt_arr[] = { 154 "quiesce", dr_pt_test_suspend, 155 }; 156 157 int dr_modunload_okay = 0; /* set to non-zero to allow unload */ 158 159 /* 160 * State transition table. States valid transitions for "board" state. 161 * Recall that non-zero return value terminates operation, however 162 * the herrno value is what really indicates an error , if any. 163 */ 164 static int 165 _cmd2index(int c) 166 { 167 /* 168 * Translate DR CMD to index into dr_state_transition. 169 */ 170 switch (c) { 171 case SBD_CMD_CONNECT: return (0); 172 case SBD_CMD_DISCONNECT: return (1); 173 case SBD_CMD_CONFIGURE: return (2); 174 case SBD_CMD_UNCONFIGURE: return (3); 175 case SBD_CMD_ASSIGN: return (4); 176 case SBD_CMD_UNASSIGN: return (5); 177 case SBD_CMD_POWERON: return (6); 178 case SBD_CMD_POWEROFF: return (7); 179 case SBD_CMD_TEST: return (8); 180 default: return (-1); 181 } 182 } 183 184 #define CMD2INDEX(c) _cmd2index(c) 185 186 static struct dr_state_trans { 187 int x_cmd; 188 struct { 189 int x_rv; /* return value of pre_op */ 190 int x_err; /* error, if any */ 191 } x_op[DR_STATE_MAX]; 192 } dr_state_transition[] = { 193 { SBD_CMD_CONNECT, 194 { 195 { 0, 0 }, /* empty */ 196 { 0, 0 }, /* occupied */ 197 { -1, ESBD_STATE }, /* connected */ 198 { -1, ESBD_STATE }, /* unconfigured */ 199 { -1, ESBD_STATE }, /* partial */ 200 { -1, ESBD_STATE }, /* configured */ 201 { -1, ESBD_STATE }, /* release */ 202 { -1, ESBD_STATE }, /* unreferenced */ 203 { -1, ESBD_FATAL_STATE }, /* fatal */ 204 } 205 }, 206 { SBD_CMD_DISCONNECT, 207 { 208 { -1, ESBD_STATE }, /* empty */ 209 { 0, 0 }, /* occupied */ 210 { 0, 0 }, /* connected */ 211 { 0, 0 }, /* unconfigured */ 212 { -1, ESBD_STATE }, /* partial */ 213 { -1, ESBD_STATE }, /* configured */ 214 { -1, ESBD_STATE }, /* release */ 215 { -1, ESBD_STATE }, /* unreferenced */ 216 { -1, ESBD_FATAL_STATE }, /* fatal */ 217 } 218 }, 219 { SBD_CMD_CONFIGURE, 220 { 221 { -1, ESBD_STATE }, /* empty */ 222 { -1, ESBD_STATE }, /* occupied */ 223 { 0, 0 }, /* connected */ 224 { 0, 0 }, /* unconfigured */ 225 { 0, 0 }, /* partial */ 226 { 0, 0 }, /* configured */ 227 { -1, ESBD_STATE }, /* release */ 228 { -1, ESBD_STATE }, /* unreferenced */ 229 { -1, ESBD_FATAL_STATE }, /* fatal */ 230 } 231 }, 232 { SBD_CMD_UNCONFIGURE, 233 { 234 { -1, ESBD_STATE }, /* empty */ 235 { -1, ESBD_STATE }, /* occupied */ 236 { -1, ESBD_STATE }, /* connected */ 237 { -1, ESBD_STATE }, /* unconfigured */ 238 { 0, 0 }, /* partial */ 239 { 0, 0 }, /* configured */ 240 { 0, 0 }, /* release */ 241 { 0, 0 }, /* unreferenced */ 242 { -1, ESBD_FATAL_STATE }, /* fatal */ 243 } 244 }, 245 { SBD_CMD_ASSIGN, 246 { 247 { 0, 0 }, /* empty */ 248 { 0, 0 }, /* occupied */ 249 { -1, ESBD_STATE }, /* connected */ 250 { -1, ESBD_STATE }, /* unconfigured */ 251 { -1, ESBD_STATE }, /* partial */ 252 { -1, ESBD_STATE }, /* configured */ 253 { -1, ESBD_STATE }, /* release */ 254 { -1, ESBD_STATE }, /* unreferenced */ 255 { -1, ESBD_FATAL_STATE }, /* fatal */ 256 } 257 }, 258 { SBD_CMD_UNASSIGN, 259 { 260 { 0, 0 }, /* empty */ 261 { 0, 0 }, /* occupied */ 262 { -1, ESBD_STATE }, /* connected */ 263 { -1, ESBD_STATE }, /* unconfigured */ 264 { -1, ESBD_STATE }, /* partial */ 265 { -1, ESBD_STATE }, /* configured */ 266 { -1, ESBD_STATE }, /* release */ 267 { -1, ESBD_STATE }, /* unreferenced */ 268 { -1, ESBD_FATAL_STATE }, /* fatal */ 269 } 270 }, 271 { SBD_CMD_POWERON, 272 { 273 { 0, 0 }, /* empty */ 274 { 0, 0 }, /* occupied */ 275 { -1, ESBD_STATE }, /* connected */ 276 { -1, ESBD_STATE }, /* unconfigured */ 277 { -1, ESBD_STATE }, /* partial */ 278 { -1, ESBD_STATE }, /* configured */ 279 { -1, ESBD_STATE }, /* release */ 280 { -1, ESBD_STATE }, /* unreferenced */ 281 { -1, ESBD_FATAL_STATE }, /* fatal */ 282 } 283 }, 284 { SBD_CMD_POWEROFF, 285 { 286 { 0, 0 }, /* empty */ 287 { 0, 0 }, /* occupied */ 288 { -1, ESBD_STATE }, /* connected */ 289 { -1, ESBD_STATE }, /* unconfigured */ 290 { -1, ESBD_STATE }, /* partial */ 291 { -1, ESBD_STATE }, /* configured */ 292 { -1, ESBD_STATE }, /* release */ 293 { -1, ESBD_STATE }, /* unreferenced */ 294 { -1, ESBD_FATAL_STATE }, /* fatal */ 295 } 296 }, 297 { SBD_CMD_TEST, 298 { 299 { 0, 0 }, /* empty */ 300 { 0, 0 }, /* occupied */ 301 { -1, ESBD_STATE }, /* connected */ 302 { -1, ESBD_STATE }, /* unconfigured */ 303 { -1, ESBD_STATE }, /* partial */ 304 { -1, ESBD_STATE }, /* configured */ 305 { -1, ESBD_STATE }, /* release */ 306 { -1, ESBD_STATE }, /* unreferenced */ 307 { -1, ESBD_FATAL_STATE }, /* fatal */ 308 } 309 }, 310 }; 311 312 /* 313 * Global R/W lock to synchronize access across 314 * multiple boards. Users wanting multi-board access 315 * must grab WRITE lock, others must grab READ lock. 316 */ 317 krwlock_t dr_grwlock; 318 319 /* 320 * Head of the boardlist used as a reference point for 321 * locating board structs. 322 * TODO: eliminate dr_boardlist 323 */ 324 dr_board_t *dr_boardlist; 325 326 /* 327 * DR support functions. 328 */ 329 static dr_devset_t dr_dev2devset(sbd_comp_id_t *cid); 330 static int dr_check_transition(dr_board_t *bp, 331 dr_devset_t *devsetp, 332 struct dr_state_trans *transp, 333 int cmd); 334 static int dr_check_unit_attached(dr_common_unit_t *dp); 335 static sbd_error_t *dr_init_devlists(dr_board_t *bp); 336 static void dr_board_discovery(dr_board_t *bp); 337 static int dr_board_init(dr_board_t *bp, dev_info_t *dip, 338 int bd); 339 static void dr_board_destroy(dr_board_t *bp); 340 static void dr_board_transition(dr_board_t *bp, dr_state_t st); 341 342 /* 343 * DR driver (DDI) entry points. 344 */ 345 static int dr_getinfo(dev_info_t *dip, ddi_info_cmd_t cmd, 346 void *arg, void **result); 347 static int dr_detach(dev_info_t *dip, ddi_detach_cmd_t cmd); 348 static int dr_attach(dev_info_t *dip, ddi_attach_cmd_t cmd); 349 static int dr_probe(dev_info_t *dip); 350 static int dr_ioctl(dev_t dev, int cmd, intptr_t arg, int mode, 351 cred_t *cred_p, int *rval_p); 352 static int dr_close(dev_t dev, int flag, int otyp, cred_t *cred_p); 353 static int dr_open(dev_t *dev, int flag, int otyp, cred_t *cred_p); 354 355 /* 356 * DR command processing operations. 357 */ 358 359 static int dr_copyin_iocmd(dr_handle_t *hp); 360 static int dr_copyout_iocmd(dr_handle_t *hp); 361 static int dr_copyout_errs(dr_handle_t *hp); 362 static int dr_pre_op(dr_handle_t *hp); 363 static int dr_post_op(dr_handle_t *hp); 364 static int dr_exec_op(dr_handle_t *hp); 365 static void dr_assign_board(dr_handle_t *hp); 366 static void dr_unassign_board(dr_handle_t *hp); 367 static void dr_connect(dr_handle_t *hp); 368 static int dr_disconnect(dr_handle_t *hp); 369 static void dr_dev_configure(dr_handle_t *hp); 370 static void dr_dev_release(dr_handle_t *hp); 371 static int dr_dev_unconfigure(dr_handle_t *hp); 372 static void dr_dev_cancel(dr_handle_t *hp); 373 static int dr_dev_status(dr_handle_t *hp); 374 static int dr_get_ncm(dr_handle_t *hp); 375 static int dr_pt_ioctl(dr_handle_t *hp); 376 static void dr_poweron_board(dr_handle_t *hp); 377 static void dr_poweroff_board(dr_handle_t *hp); 378 static void dr_test_board(dr_handle_t *hp); 379 380 381 382 /* 383 * Autoconfiguration data structures 384 */ 385 386 struct cb_ops dr_cb_ops = { 387 dr_open, /* open */ 388 dr_close, /* close */ 389 nodev, /* strategy */ 390 nodev, /* print */ 391 nodev, /* dump */ 392 nodev, /* read */ 393 nodev, /* write */ 394 dr_ioctl, /* ioctl */ 395 nodev, /* devmap */ 396 nodev, /* mmap */ 397 nodev, /* segmap */ 398 nochpoll, /* chpoll */ 399 ddi_prop_op, /* cb_prop_op */ 400 NULL, /* struct streamtab */ 401 D_NEW | D_MP | D_MTSAFE, /* compatibility flags */ 402 CB_REV, /* Rev */ 403 nodev, /* cb_aread */ 404 nodev /* cb_awrite */ 405 }; 406 407 struct dev_ops dr_dev_ops = { 408 DEVO_REV, /* build version */ 409 0, /* dev ref count */ 410 dr_getinfo, /* getinfo */ 411 nulldev, /* identify */ 412 dr_probe, /* probe */ 413 dr_attach, /* attach */ 414 dr_detach, /* detach */ 415 nodev, /* reset */ 416 &dr_cb_ops, /* cb_ops */ 417 (struct bus_ops *)NULL, /* bus ops */ 418 NULL, /* power */ 419 ddi_quiesce_not_needed, /* quiesce */ 420 }; 421 422 extern struct mod_ops mod_driverops; 423 424 static struct modldrv modldrv = { 425 &mod_driverops, 426 "Dynamic Reconfiguration", 427 &dr_dev_ops 428 }; 429 430 static struct modlinkage modlinkage = { 431 MODREV_1, 432 (void *)&modldrv, 433 NULL 434 }; 435 436 /* 437 * Driver entry points. 438 */ 439 int 440 _init(void) 441 { 442 int err; 443 444 /* 445 * If you need to support multiple nodes (instances), then 446 * whatever the maximum number of supported nodes is would 447 * need to passed as the third parameter to ddi_soft_state_init(). 448 * Alternative would be to dynamically fini and re-init the 449 * soft state structure each time a node is attached. 450 */ 451 err = ddi_soft_state_init((void **)&dr_g.softsp, 452 sizeof (dr_softstate_t), 1); 453 if (err) 454 return (err); 455 456 mutex_init(&dr_g.lock, NULL, MUTEX_DRIVER, NULL); 457 rw_init(&dr_grwlock, NULL, RW_DEFAULT, NULL); 458 459 return (mod_install(&modlinkage)); 460 } 461 462 int 463 _fini(void) 464 { 465 int err; 466 467 if ((err = mod_remove(&modlinkage)) != 0) 468 return (err); 469 470 mutex_destroy(&dr_g.lock); 471 rw_destroy(&dr_grwlock); 472 473 ddi_soft_state_fini((void **)&dr_g.softsp); 474 475 return (0); 476 } 477 478 int 479 _info(struct modinfo *modinfop) 480 { 481 return (mod_info(&modlinkage, modinfop)); 482 } 483 484 /*ARGSUSED1*/ 485 static int 486 dr_open(dev_t *dev, int flag, int otyp, cred_t *cred_p) 487 { 488 int instance; 489 dr_softstate_t *softsp; 490 dr_board_t *bp; 491 /* 492 * Don't open unless we've attached. 493 */ 494 instance = DR_MINOR2INST(getminor(*dev)); 495 softsp = ddi_get_soft_state(dr_g.softsp, instance); 496 if (softsp == NULL) 497 return (ENXIO); 498 499 mutex_enter(&softsp->i_lock); 500 if (!softsp->dr_initialized) { 501 int bd; 502 int rv = 0; 503 504 bp = softsp->boards; 505 506 /* initialize each array element */ 507 for (bd = 0; bd < MAX_BOARDS; bd++, bp++) { 508 rv = dr_board_init(bp, softsp->dip, bd); 509 if (rv) 510 break; 511 } 512 513 if (rv == 0) { 514 softsp->dr_initialized = 1; 515 } else { 516 /* destroy elements initialized thus far */ 517 while (--bp >= softsp->boards) 518 dr_board_destroy(bp); 519 520 521 /* TODO: should this be another errno val ? */ 522 mutex_exit(&softsp->i_lock); 523 return (ENXIO); 524 } 525 } 526 mutex_exit(&softsp->i_lock); 527 528 bp = &softsp->boards[DR_MINOR2BNUM(getminor(*dev))]; 529 530 /* 531 * prevent opening of a dyn-ap for a board 532 * that does not exist 533 */ 534 if (!bp->b_assigned) { 535 if (drmach_board_lookup(bp->b_num, &bp->b_id) != 0) 536 return (ENODEV); 537 } 538 539 return (0); 540 } 541 542 /*ARGSUSED*/ 543 static int 544 dr_close(dev_t dev, int flag, int otyp, cred_t *cred_p) 545 { 546 return (0); 547 } 548 549 /* 550 * Enable/disable DR features. 551 */ 552 int dr_enable = 1; 553 554 /*ARGSUSED3*/ 555 static int 556 dr_ioctl(dev_t dev, int cmd, intptr_t arg, int mode, 557 cred_t *cred_p, int *rval_p) 558 { 559 static int dr_dev_type_to_nt(char *); 560 561 int rv = 0; 562 int instance; 563 int bd; 564 dr_handle_t *hp; 565 dr_softstate_t *softsp; 566 static fn_t f = "dr_ioctl"; 567 568 PR_ALL("%s...\n", f); 569 570 instance = DR_MINOR2INST(getminor(dev)); 571 softsp = ddi_get_soft_state(dr_g.softsp, instance); 572 if (softsp == NULL) { 573 cmn_err(CE_WARN, "dr%d: module not yet attached", instance); 574 return (ENXIO); 575 } 576 577 if (!dr_enable) { 578 switch (cmd) { 579 case SBD_CMD_STATUS: 580 case SBD_CMD_GETNCM: 581 case SBD_CMD_PASSTHRU: 582 break; 583 default: 584 return (ENOTSUP); 585 } 586 } 587 588 bd = DR_MINOR2BNUM(getminor(dev)); 589 if (bd >= MAX_BOARDS) 590 return (ENXIO); 591 592 /* get and initialize storage for new handle */ 593 hp = GETSTRUCT(dr_handle_t, 1); 594 hp->h_bd = &softsp->boards[bd]; 595 hp->h_err = NULL; 596 hp->h_dev = getminor(dev); 597 hp->h_cmd = cmd; 598 hp->h_mode = mode; 599 hp->h_iap = (sbd_ioctl_arg_t *)arg; 600 601 /* copy sbd command into handle */ 602 rv = dr_copyin_iocmd(hp); 603 if (rv) { 604 FREESTRUCT(hp, dr_handle_t, 1); 605 return (EINVAL); 606 } 607 608 /* translate canonical name to component type */ 609 if (hp->h_sbdcmd.cmd_cm.c_id.c_name[0] != '\0') { 610 hp->h_sbdcmd.cmd_cm.c_id.c_type = 611 dr_dev_type_to_nt(hp->h_sbdcmd.cmd_cm.c_id.c_name); 612 613 PR_ALL("%s: c_name = %s, c_type = %d\n", 614 f, 615 hp->h_sbdcmd.cmd_cm.c_id.c_name, 616 hp->h_sbdcmd.cmd_cm.c_id.c_type); 617 } else { 618 /*EMPTY*/ 619 PR_ALL("%s: c_name is NULL\n", f); 620 } 621 622 /* determine scope of operation */ 623 hp->h_devset = dr_dev2devset(&hp->h_sbdcmd.cmd_cm.c_id); 624 625 switch (hp->h_cmd) { 626 case SBD_CMD_STATUS: 627 case SBD_CMD_GETNCM: 628 /* no locks needed for these commands */ 629 break; 630 631 default: 632 rw_enter(&dr_grwlock, RW_WRITER); 633 mutex_enter(&hp->h_bd->b_lock); 634 635 /* 636 * If we're dealing with memory at all, then we have 637 * to keep the "exclusive" global lock held. This is 638 * necessary since we will probably need to look at 639 * multiple board structs. Otherwise, we only have 640 * to deal with the board in question and so can drop 641 * the global lock to "shared". 642 */ 643 rv = DEVSET_IN_SET(hp->h_devset, SBD_COMP_MEM, DEVSET_ANYUNIT); 644 if (rv == 0) 645 rw_downgrade(&dr_grwlock); 646 break; 647 } 648 rv = 0; 649 650 if (rv == 0) 651 rv = dr_pre_op(hp); 652 if (rv == 0) 653 rv = dr_exec_op(hp); 654 if (rv == 0) 655 rv = dr_post_op(hp); 656 657 if (rv == -1) 658 rv = EIO; 659 660 if (hp->h_err != NULL) 661 if (!(rv = dr_copyout_errs(hp))) 662 rv = EIO; 663 664 /* undo locking, if any, done before dr_pre_op */ 665 switch (hp->h_cmd) { 666 case SBD_CMD_STATUS: 667 case SBD_CMD_GETNCM: 668 break; 669 670 case SBD_CMD_ASSIGN: 671 case SBD_CMD_UNASSIGN: 672 case SBD_CMD_POWERON: 673 case SBD_CMD_POWEROFF: 674 case SBD_CMD_CONNECT: 675 case SBD_CMD_CONFIGURE: 676 case SBD_CMD_UNCONFIGURE: 677 case SBD_CMD_DISCONNECT: 678 /* Board changed state. Log a sysevent. */ 679 if (rv == 0) 680 (void) drmach_log_sysevent(hp->h_bd->b_num, "", 681 SE_SLEEP, 1); 682 /* Fall through */ 683 684 default: 685 mutex_exit(&hp->h_bd->b_lock); 686 rw_exit(&dr_grwlock); 687 } 688 689 if (hp->h_opts.size != 0) 690 FREESTRUCT(hp->h_opts.copts, char, hp->h_opts.size); 691 692 FREESTRUCT(hp, dr_handle_t, 1); 693 694 return (rv); 695 } 696 697 /*ARGSUSED*/ 698 static int 699 dr_probe(dev_info_t *dip) 700 { 701 return (DDI_PROBE_SUCCESS); 702 } 703 704 static int 705 dr_attach(dev_info_t *dip, ddi_attach_cmd_t cmd) 706 { 707 int rv, rv2; 708 int bd; 709 int instance; 710 sbd_error_t *err; 711 dr_softstate_t *softsp; 712 713 instance = ddi_get_instance(dip); 714 715 switch (cmd) { 716 717 case DDI_ATTACH: 718 719 rw_enter(&dr_grwlock, RW_WRITER); 720 721 rv = ddi_soft_state_zalloc(dr_g.softsp, instance); 722 if (rv != DDI_SUCCESS) { 723 cmn_err(CE_WARN, "dr%d: failed to alloc soft-state", 724 instance); 725 return (DDI_FAILURE); 726 } 727 728 /* initialize softstate structure */ 729 softsp = ddi_get_soft_state(dr_g.softsp, instance); 730 softsp->dip = dip; 731 732 mutex_init(&softsp->i_lock, NULL, MUTEX_DRIVER, NULL); 733 734 /* allocate board array (aka boardlist) */ 735 softsp->boards = GETSTRUCT(dr_board_t, MAX_BOARDS); 736 737 /* TODO: eliminate dr_boardlist */ 738 dr_boardlist = softsp->boards; 739 740 /* initialize each array element */ 741 rv = DDI_SUCCESS; 742 for (bd = 0; bd < MAX_BOARDS; bd++) { 743 dr_board_t *bp = &softsp->boards[bd]; 744 char *p, *name; 745 int l, minor_num; 746 747 /* 748 * initialized board attachment point path 749 * (relative to pseudo) in a form immediately 750 * reusable as an cfgadm command argument. 751 * TODO: clean this up 752 */ 753 p = bp->b_path; 754 l = sizeof (bp->b_path); 755 (void) snprintf(p, l, "dr@%d:", instance); 756 while (*p != '\0') { 757 l--; 758 p++; 759 } 760 761 name = p; 762 err = drmach_board_name(bd, p, l); 763 if (err) { 764 sbd_err_clear(&err); 765 rv = DDI_FAILURE; 766 break; 767 } 768 769 minor_num = DR_MAKE_MINOR(instance, bd); 770 rv = ddi_create_minor_node(dip, name, S_IFCHR, 771 minor_num, DDI_NT_SBD_ATTACHMENT_POINT, NULL); 772 if (rv != DDI_SUCCESS) 773 rv = DDI_FAILURE; 774 } 775 776 if (rv == DDI_SUCCESS) { 777 /* 778 * Announce the node's presence. 779 */ 780 ddi_report_dev(dip); 781 } else { 782 ddi_remove_minor_node(dip, NULL); 783 } 784 /* 785 * Init registered unsafe devs. 786 */ 787 dr_unsafe_devs.devnames = NULL; 788 rv2 = ddi_prop_lookup_string_array(DDI_DEV_T_ANY, dip, 789 DDI_PROP_DONTPASS | DDI_PROP_NOTPROM, 790 "unsupported-io-drivers", &dr_unsafe_devs.devnames, 791 &dr_unsafe_devs.ndevs); 792 793 if (rv2 != DDI_PROP_SUCCESS) 794 dr_unsafe_devs.ndevs = 0; 795 796 rw_exit(&dr_grwlock); 797 return (rv); 798 799 default: 800 return (DDI_FAILURE); 801 } 802 803 /*NOTREACHED*/ 804 } 805 806 static int 807 dr_detach(dev_info_t *dip, ddi_detach_cmd_t cmd) 808 { 809 int instance; 810 dr_softstate_t *softsp; 811 812 switch (cmd) { 813 case DDI_DETACH: 814 if (!dr_modunload_okay) 815 return (DDI_FAILURE); 816 817 rw_enter(&dr_grwlock, RW_WRITER); 818 819 instance = ddi_get_instance(dip); 820 softsp = ddi_get_soft_state(dr_g.softsp, instance); 821 822 /* TODO: eliminate dr_boardlist */ 823 ASSERT(softsp->boards == dr_boardlist); 824 825 /* remove all minor nodes */ 826 ddi_remove_minor_node(dip, NULL); 827 828 if (softsp->dr_initialized) { 829 int bd; 830 831 for (bd = 0; bd < MAX_BOARDS; bd++) 832 dr_board_destroy(&softsp->boards[bd]); 833 } 834 835 FREESTRUCT(softsp->boards, dr_board_t, MAX_BOARDS); 836 mutex_destroy(&softsp->i_lock); 837 ddi_soft_state_free(dr_g.softsp, instance); 838 839 rw_exit(&dr_grwlock); 840 return (DDI_SUCCESS); 841 842 default: 843 return (DDI_FAILURE); 844 } 845 /*NOTREACHED*/ 846 } 847 848 static int 849 dr_getinfo(dev_info_t *dip, ddi_info_cmd_t cmd, void *arg, void **result) 850 { 851 _NOTE(ARGUNUSED(dip)) 852 853 dev_t dev = (dev_t)arg; 854 int instance, error; 855 dr_softstate_t *softsp; 856 857 *result = NULL; 858 error = DDI_SUCCESS; 859 instance = DR_MINOR2INST(getminor(dev)); 860 861 switch (cmd) { 862 case DDI_INFO_DEVT2DEVINFO: 863 softsp = ddi_get_soft_state(dr_g.softsp, instance); 864 if (softsp == NULL) 865 return (DDI_FAILURE); 866 *result = (void *)softsp->dip; 867 break; 868 869 case DDI_INFO_DEVT2INSTANCE: 870 *result = (void *)(uintptr_t)instance; 871 break; 872 873 default: 874 error = DDI_FAILURE; 875 break; 876 } 877 878 return (error); 879 } 880 881 /* 882 * DR operations. 883 */ 884 885 static int 886 dr_copyin_iocmd(dr_handle_t *hp) 887 { 888 static fn_t f = "dr_copyin_iocmd"; 889 sbd_cmd_t *scp = &hp->h_sbdcmd; 890 891 if (hp->h_iap == NULL) 892 return (EINVAL); 893 894 bzero((caddr_t)scp, sizeof (sbd_cmd_t)); 895 896 #ifdef _MULTI_DATAMODEL 897 if (ddi_model_convert_from(hp->h_mode & FMODELS) == DDI_MODEL_ILP32) { 898 sbd_cmd32_t scmd32; 899 900 bzero((caddr_t)&scmd32, sizeof (sbd_cmd32_t)); 901 902 if (ddi_copyin((void *)hp->h_iap, (void *)&scmd32, 903 sizeof (sbd_cmd32_t), hp->h_mode)) { 904 cmn_err(CE_WARN, 905 "%s: (32bit) failed to copyin " 906 "sbdcmd-struct", f); 907 return (EFAULT); 908 } 909 scp->cmd_cm.c_id.c_type = scmd32.cmd_cm.c_id.c_type; 910 scp->cmd_cm.c_id.c_unit = scmd32.cmd_cm.c_id.c_unit; 911 bcopy(&scmd32.cmd_cm.c_id.c_name[0], 912 &scp->cmd_cm.c_id.c_name[0], OBP_MAXPROPNAME); 913 scp->cmd_cm.c_flags = scmd32.cmd_cm.c_flags; 914 scp->cmd_cm.c_len = scmd32.cmd_cm.c_len; 915 scp->cmd_cm.c_opts = (caddr_t)(uintptr_t)scmd32.cmd_cm.c_opts; 916 917 switch (hp->h_cmd) { 918 case SBD_CMD_STATUS: 919 scp->cmd_stat.s_nbytes = scmd32.cmd_stat.s_nbytes; 920 scp->cmd_stat.s_statp = 921 (caddr_t)(uintptr_t)scmd32.cmd_stat.s_statp; 922 break; 923 default: 924 break; 925 926 } 927 } else 928 #endif /* _MULTI_DATAMODEL */ 929 if (ddi_copyin((void *)hp->h_iap, (void *)scp, 930 sizeof (sbd_cmd_t), hp->h_mode) != 0) { 931 cmn_err(CE_WARN, 932 "%s: failed to copyin sbdcmd-struct", f); 933 return (EFAULT); 934 } 935 936 if ((hp->h_opts.size = scp->cmd_cm.c_len) != 0) { 937 hp->h_opts.copts = GETSTRUCT(char, scp->cmd_cm.c_len + 1); 938 ++hp->h_opts.size; 939 if (ddi_copyin((void *)scp->cmd_cm.c_opts, 940 (void *)hp->h_opts.copts, 941 scp->cmd_cm.c_len, hp->h_mode) != 0) { 942 cmn_err(CE_WARN, "%s: failed to copyin options", f); 943 return (EFAULT); 944 } 945 } 946 return (0); 947 } 948 949 static int 950 dr_copyout_iocmd(dr_handle_t *hp) 951 { 952 static fn_t f = "dr_copyout_iocmd"; 953 sbd_cmd_t *scp = &hp->h_sbdcmd; 954 955 if (hp->h_iap == NULL) 956 return (EINVAL); 957 958 #ifdef _MULTI_DATAMODEL 959 if (ddi_model_convert_from(hp->h_mode & FMODELS) == DDI_MODEL_ILP32) { 960 sbd_cmd32_t scmd32; 961 962 scmd32.cmd_cm.c_id.c_type = scp->cmd_cm.c_id.c_type; 963 scmd32.cmd_cm.c_id.c_unit = scp->cmd_cm.c_id.c_unit; 964 bcopy(&scp->cmd_cm.c_id.c_name[0], 965 &scmd32.cmd_cm.c_id.c_name[0], OBP_MAXPROPNAME); 966 967 scmd32.cmd_cm.c_flags = scp->cmd_cm.c_flags; 968 scmd32.cmd_cm.c_len = scp->cmd_cm.c_len; 969 scmd32.cmd_cm.c_opts = (caddr32_t)(uintptr_t)scp->cmd_cm.c_opts; 970 971 switch (hp->h_cmd) { 972 case SBD_CMD_GETNCM: 973 scmd32.cmd_getncm.g_ncm = scp->cmd_getncm.g_ncm; 974 break; 975 default: 976 break; 977 } 978 979 if (ddi_copyout((void *)&scmd32, (void *)hp->h_iap, 980 sizeof (sbd_cmd32_t), hp->h_mode)) { 981 cmn_err(CE_WARN, 982 "%s: (32bit) failed to copyout " 983 "sbdcmd-struct", f); 984 return (EFAULT); 985 } 986 } else 987 #endif /* _MULTI_DATAMODEL */ 988 if (ddi_copyout((void *)scp, (void *)hp->h_iap, 989 sizeof (sbd_cmd_t), hp->h_mode) != 0) { 990 cmn_err(CE_WARN, 991 "%s: failed to copyout sbdcmd-struct", f); 992 return (EFAULT); 993 } 994 995 return (0); 996 } 997 998 static int 999 dr_copyout_errs(dr_handle_t *hp) 1000 { 1001 static fn_t f = "dr_copyout_errs"; 1002 1003 if (hp->h_err == NULL) 1004 return (0); 1005 1006 if (hp->h_err->e_code) { 1007 PR_ALL("%s: error %d %s", 1008 f, hp->h_err->e_code, hp->h_err->e_rsc); 1009 } 1010 1011 #ifdef _MULTI_DATAMODEL 1012 if (ddi_model_convert_from(hp->h_mode & FMODELS) == DDI_MODEL_ILP32) { 1013 sbd_error32_t *serr32p; 1014 1015 serr32p = GETSTRUCT(sbd_error32_t, 1); 1016 1017 serr32p->e_code = hp->h_err->e_code; 1018 bcopy(&hp->h_err->e_rsc[0], &serr32p->e_rsc[0], 1019 MAXPATHLEN); 1020 if (ddi_copyout((void *)serr32p, 1021 (void *)&((sbd_ioctl_arg32_t *)hp->h_iap)->i_err, 1022 sizeof (sbd_error32_t), hp->h_mode)) { 1023 cmn_err(CE_WARN, 1024 "%s: (32bit) failed to copyout", f); 1025 return (EFAULT); 1026 } 1027 FREESTRUCT(serr32p, sbd_error32_t, 1); 1028 } else 1029 #endif /* _MULTI_DATAMODEL */ 1030 if (ddi_copyout((void *)hp->h_err, 1031 (void *)&hp->h_iap->i_err, 1032 sizeof (sbd_error_t), hp->h_mode)) { 1033 cmn_err(CE_WARN, 1034 "%s: failed to copyout", f); 1035 return (EFAULT); 1036 } 1037 1038 sbd_err_clear(&hp->h_err); 1039 1040 return (0); 1041 1042 } 1043 1044 /* 1045 * pre-op entry point must sbd_err_set_c(), if needed. 1046 * Return value of non-zero indicates failure. 1047 */ 1048 static int 1049 dr_pre_op(dr_handle_t *hp) 1050 { 1051 int rv = 0, t; 1052 int cmd, serr = 0; 1053 dr_devset_t devset; 1054 dr_board_t *bp = hp->h_bd; 1055 dr_handle_t *shp = hp; 1056 static fn_t f = "dr_pre_op"; 1057 1058 cmd = hp->h_cmd; 1059 devset = shp->h_devset; 1060 1061 PR_ALL("%s (cmd = %s)...\n", f, SBD_CMD_STR(cmd)); 1062 1063 hp->h_err = drmach_pre_op(cmd, bp->b_id, &hp->h_opts); 1064 if (hp->h_err != NULL) { 1065 PR_ALL("drmach_pre_op failed for cmd %s(%d)\n", 1066 SBD_CMD_STR(cmd), cmd); 1067 return (-1); 1068 } 1069 1070 /* 1071 * Check for valid state transitions. 1072 */ 1073 if ((t = CMD2INDEX(cmd)) != -1) { 1074 struct dr_state_trans *transp; 1075 int state_err; 1076 1077 transp = &dr_state_transition[t]; 1078 ASSERT(transp->x_cmd == cmd); 1079 1080 state_err = dr_check_transition(bp, &devset, transp, cmd); 1081 1082 if (state_err < 0) { 1083 /* 1084 * Invalidate device. 1085 */ 1086 dr_op_err(CE_IGNORE, hp, ESBD_INVAL, NULL); 1087 serr = -1; 1088 PR_ALL("%s: invalid devset (0x%x)\n", 1089 f, (uint_t)devset); 1090 } else if (state_err != 0) { 1091 /* 1092 * State transition is not a valid one. 1093 */ 1094 dr_op_err(CE_IGNORE, hp, 1095 transp->x_op[state_err].x_err, NULL); 1096 1097 serr = transp->x_op[state_err].x_rv; 1098 1099 PR_ALL("%s: invalid state %s(%d) for cmd %s(%d)\n", 1100 f, state_str[state_err], state_err, 1101 SBD_CMD_STR(cmd), cmd); 1102 } else { 1103 shp->h_devset = devset; 1104 } 1105 } 1106 1107 if (serr) { 1108 rv = -1; 1109 } 1110 1111 return (rv); 1112 } 1113 1114 static int 1115 dr_post_op(dr_handle_t *hp) 1116 { 1117 int rv = 0; 1118 int cmd; 1119 dr_board_t *bp = hp->h_bd; 1120 static fn_t f = "dr_post_op"; 1121 1122 cmd = hp->h_cmd; 1123 1124 PR_ALL("%s (cmd = %s)...\n", f, SBD_CMD_STR(cmd)); 1125 1126 /* errors should have been caught by now */ 1127 ASSERT(hp->h_err == NULL); 1128 1129 hp->h_err = drmach_post_op(cmd, bp->b_id, &hp->h_opts); 1130 if (hp->h_err != NULL) { 1131 PR_ALL("drmach_post_op failed for cmd %s(%d)\n", 1132 SBD_CMD_STR(cmd), cmd); 1133 return (-1); 1134 } 1135 1136 switch (cmd) { 1137 case SBD_CMD_CONFIGURE: 1138 case SBD_CMD_UNCONFIGURE: 1139 case SBD_CMD_CONNECT: 1140 case SBD_CMD_DISCONNECT: 1141 case SBD_CMD_GETNCM: 1142 case SBD_CMD_STATUS: 1143 break; 1144 1145 default: 1146 break; 1147 } 1148 1149 return (rv); 1150 } 1151 1152 static int 1153 dr_exec_op(dr_handle_t *hp) 1154 { 1155 int rv = 0; 1156 static fn_t f = "dr_exec_op"; 1157 1158 /* errors should have been caught by now */ 1159 ASSERT(hp->h_err == NULL); 1160 1161 switch (hp->h_cmd) { 1162 case SBD_CMD_ASSIGN: 1163 dr_assign_board(hp); 1164 break; 1165 1166 case SBD_CMD_UNASSIGN: 1167 dr_unassign_board(hp); 1168 break; 1169 1170 case SBD_CMD_POWEROFF: 1171 dr_poweroff_board(hp); 1172 break; 1173 1174 case SBD_CMD_POWERON: 1175 dr_poweron_board(hp); 1176 break; 1177 1178 case SBD_CMD_TEST: 1179 dr_test_board(hp); 1180 break; 1181 1182 case SBD_CMD_CONNECT: 1183 dr_connect(hp); 1184 break; 1185 1186 case SBD_CMD_CONFIGURE: 1187 dr_dev_configure(hp); 1188 break; 1189 1190 case SBD_CMD_UNCONFIGURE: 1191 dr_dev_release(hp); 1192 if (hp->h_err == NULL) 1193 rv = dr_dev_unconfigure(hp); 1194 else 1195 dr_dev_cancel(hp); 1196 break; 1197 1198 case SBD_CMD_DISCONNECT: 1199 rv = dr_disconnect(hp); 1200 break; 1201 1202 case SBD_CMD_STATUS: 1203 rv = dr_dev_status(hp); 1204 break; 1205 1206 case SBD_CMD_GETNCM: 1207 hp->h_sbdcmd.cmd_getncm.g_ncm = dr_get_ncm(hp); 1208 rv = dr_copyout_iocmd(hp); 1209 break; 1210 1211 case SBD_CMD_PASSTHRU: 1212 rv = dr_pt_ioctl(hp); 1213 break; 1214 1215 default: 1216 cmn_err(CE_WARN, 1217 "%s: unknown command (%d)", 1218 f, hp->h_cmd); 1219 break; 1220 } 1221 1222 if (hp->h_err != NULL) { 1223 rv = -1; 1224 } 1225 1226 return (rv); 1227 } 1228 1229 static void 1230 dr_assign_board(dr_handle_t *hp) 1231 { 1232 dr_board_t *bp = hp->h_bd; 1233 1234 hp->h_err = drmach_board_assign(bp->b_num, &bp->b_id); 1235 if (hp->h_err == NULL) { 1236 bp->b_assigned = 1; 1237 } 1238 } 1239 1240 static void 1241 dr_unassign_board(dr_handle_t *hp) 1242 { 1243 dr_board_t *bp = hp->h_bd; 1244 1245 /* 1246 * Block out status during unassign. 1247 * Not doing cv_wait_sig here as starfire SSP software 1248 * ignores unassign failure and removes board from 1249 * domain mask causing system panic. 1250 * TODO: Change cv_wait to cv_wait_sig when SSP software 1251 * handles unassign failure. 1252 */ 1253 dr_lock_status(bp); 1254 1255 hp->h_err = drmach_board_unassign(bp->b_id); 1256 if (hp->h_err == NULL) { 1257 /* 1258 * clear drmachid_t handle; not valid after board unassign 1259 */ 1260 bp->b_id = 0; 1261 bp->b_assigned = 0; 1262 } 1263 1264 dr_unlock_status(bp); 1265 } 1266 1267 static void 1268 dr_poweron_board(dr_handle_t *hp) 1269 { 1270 dr_board_t *bp = hp->h_bd; 1271 1272 hp->h_err = drmach_board_poweron(bp->b_id); 1273 } 1274 1275 static void 1276 dr_poweroff_board(dr_handle_t *hp) 1277 { 1278 dr_board_t *bp = hp->h_bd; 1279 1280 hp->h_err = drmach_board_poweroff(bp->b_id); 1281 } 1282 1283 static void 1284 dr_test_board(dr_handle_t *hp) 1285 { 1286 dr_board_t *bp = hp->h_bd; 1287 hp->h_err = drmach_board_test(bp->b_id, &hp->h_opts, 1288 dr_cmd_flags(hp) & SBD_FLAG_FORCE); 1289 } 1290 1291 /* 1292 * Create and populate the component nodes for a board. Assumes that the 1293 * devlists for the board have been initialized. 1294 */ 1295 static void 1296 dr_make_comp_nodes(dr_board_t *bp) { 1297 1298 int i; 1299 1300 /* 1301 * Make nodes for the individual components on the board. 1302 * First we need to initialize memory unit data structures of board 1303 * structure. 1304 */ 1305 for (i = 0; i < MAX_MEM_UNITS_PER_BOARD; i++) { 1306 dr_mem_unit_t *mp; 1307 1308 mp = dr_get_mem_unit(bp, i); 1309 dr_init_mem_unit(mp); 1310 } 1311 1312 /* 1313 * Initialize cpu unit data structures. 1314 */ 1315 for (i = 0; i < MAX_CPU_UNITS_PER_BOARD; i++) { 1316 dr_cpu_unit_t *cp; 1317 1318 cp = dr_get_cpu_unit(bp, i); 1319 dr_init_cpu_unit(cp); 1320 } 1321 1322 /* 1323 * Initialize io unit data structures. 1324 */ 1325 for (i = 0; i < MAX_IO_UNITS_PER_BOARD; i++) { 1326 dr_io_unit_t *ip; 1327 1328 ip = dr_get_io_unit(bp, i); 1329 dr_init_io_unit(ip); 1330 } 1331 1332 dr_board_transition(bp, DR_STATE_CONNECTED); 1333 1334 bp->b_rstate = SBD_STAT_CONNECTED; 1335 bp->b_ostate = SBD_STAT_UNCONFIGURED; 1336 bp->b_cond = SBD_COND_OK; 1337 (void) drv_getparm(TIME, (void *)&bp->b_time); 1338 1339 } 1340 1341 /* 1342 * Only do work if called to operate on an entire board 1343 * which doesn't already have components present. 1344 */ 1345 static void 1346 dr_connect(dr_handle_t *hp) 1347 { 1348 dr_board_t *bp = hp->h_bd; 1349 static fn_t f = "dr_connect"; 1350 1351 PR_ALL("%s...\n", f); 1352 1353 if (DR_DEVS_PRESENT(bp)) { 1354 /* 1355 * Board already has devices present. 1356 */ 1357 PR_ALL("%s: devices already present (0x%lx)\n", 1358 f, DR_DEVS_PRESENT(bp)); 1359 return; 1360 } 1361 1362 hp->h_err = drmach_board_connect(bp->b_id, &hp->h_opts); 1363 if (hp->h_err) 1364 return; 1365 1366 hp->h_err = dr_init_devlists(bp); 1367 if (hp->h_err) 1368 return; 1369 else if (bp->b_ndev == 0) { 1370 dr_op_err(CE_WARN, hp, ESBD_EMPTY_BD, bp->b_path); 1371 return; 1372 } else { 1373 dr_make_comp_nodes(bp); 1374 return; 1375 } 1376 /*NOTREACHED*/ 1377 } 1378 1379 static int 1380 dr_disconnect(dr_handle_t *hp) 1381 { 1382 int i; 1383 dr_devset_t devset; 1384 dr_board_t *bp = hp->h_bd; 1385 static fn_t f = "dr_disconnect"; 1386 1387 PR_ALL("%s...\n", f); 1388 1389 /* 1390 * Only devices which are present, but 1391 * unattached can be disconnected. 1392 */ 1393 devset = hp->h_devset & DR_DEVS_PRESENT(bp) & 1394 DR_DEVS_UNATTACHED(bp); 1395 1396 if ((devset == 0) && DR_DEVS_PRESENT(bp)) { 1397 dr_op_err(CE_IGNORE, hp, ESBD_EMPTY_BD, bp->b_path); 1398 return (0); 1399 } 1400 1401 /* 1402 * Block out status during disconnect. 1403 */ 1404 mutex_enter(&bp->b_slock); 1405 while (bp->b_sflags & DR_BSLOCK) { 1406 if (cv_wait_sig(&bp->b_scv, &bp->b_slock) == 0) { 1407 mutex_exit(&bp->b_slock); 1408 return (EINTR); 1409 } 1410 } 1411 bp->b_sflags |= DR_BSLOCK; 1412 mutex_exit(&bp->b_slock); 1413 1414 hp->h_err = drmach_board_disconnect(bp->b_id, &hp->h_opts); 1415 1416 DR_DEVS_DISCONNECT(bp, devset); 1417 1418 ASSERT((DR_DEVS_ATTACHED(bp) & devset) == 0); 1419 1420 /* 1421 * Update per-device state transitions. 1422 */ 1423 for (i = 0; i < MAX_CPU_UNITS_PER_BOARD; i++) { 1424 dr_cpu_unit_t *cp; 1425 1426 if (!DEVSET_IN_SET(devset, SBD_COMP_CPU, i)) 1427 continue; 1428 1429 cp = dr_get_cpu_unit(bp, i); 1430 if (dr_disconnect_cpu(cp) == 0) 1431 dr_device_transition(&cp->sbc_cm, DR_STATE_EMPTY); 1432 else if (cp->sbc_cm.sbdev_error != NULL) 1433 DRERR_SET_C(&hp->h_err, &cp->sbc_cm.sbdev_error); 1434 1435 ASSERT(cp->sbc_cm.sbdev_error == NULL); 1436 } 1437 1438 for (i = 0; i < MAX_MEM_UNITS_PER_BOARD; i++) { 1439 dr_mem_unit_t *mp; 1440 1441 if (!DEVSET_IN_SET(devset, SBD_COMP_MEM, i)) 1442 continue; 1443 1444 mp = dr_get_mem_unit(bp, i); 1445 if (dr_disconnect_mem(mp) == 0) 1446 dr_device_transition(&mp->sbm_cm, DR_STATE_EMPTY); 1447 else if (mp->sbm_cm.sbdev_error != NULL) 1448 DRERR_SET_C(&hp->h_err, &mp->sbm_cm.sbdev_error); 1449 1450 ASSERT(mp->sbm_cm.sbdev_error == NULL); 1451 } 1452 1453 for (i = 0; i < MAX_IO_UNITS_PER_BOARD; i++) { 1454 dr_io_unit_t *ip; 1455 1456 if (!DEVSET_IN_SET(devset, SBD_COMP_IO, i)) 1457 continue; 1458 1459 ip = dr_get_io_unit(bp, i); 1460 if (dr_disconnect_io(ip) == 0) 1461 dr_device_transition(&ip->sbi_cm, DR_STATE_EMPTY); 1462 else if (ip->sbi_cm.sbdev_error != NULL) 1463 DRERR_SET_C(&hp->h_err, &ip->sbi_cm.sbdev_error); 1464 1465 ASSERT(ip->sbi_cm.sbdev_error == NULL); 1466 } 1467 if (hp->h_err) { 1468 /* 1469 * For certain errors, drmach_board_disconnect will mark 1470 * the board as unusable; in these cases the devtree must 1471 * be purged so that status calls will succeed. 1472 * XXX 1473 * This implementation checks for discrete error codes - 1474 * someday, the i/f to drmach_board_disconnect should be 1475 * changed to avoid the e_code testing. 1476 */ 1477 if ((hp->h_err->e_code == ESTC_MBXRPLY) || 1478 (hp->h_err->e_code == ESTC_MBXRQST) || 1479 (hp->h_err->e_code == ESTC_SMS_ERR_UNRECOVERABLE) || 1480 (hp->h_err->e_code == ESTC_SMS_ERR_RECOVERABLE) || 1481 (hp->h_err->e_code == ESTC_DEPROBE) || 1482 (hp->h_err->e_code == EOPL_DEPROBE)) { 1483 bp->b_ostate = SBD_STAT_UNCONFIGURED; 1484 bp->b_busy = 0; 1485 (void) drv_getparm(TIME, (void *)&bp->b_time); 1486 1487 if (drmach_board_deprobe(bp->b_id)) 1488 goto disconnect_done; 1489 else 1490 bp->b_ndev = 0; 1491 } 1492 1493 /* 1494 * If the disconnect failed in a recoverable way, 1495 * more work is required. 1496 * XXX 1497 * This implementation checks for discrete error codes - 1498 * someday, the i/f to drmach_board_disconnect should be 1499 * changed to avoid the e_code testing. 1500 */ 1501 if ((hp->h_err->e_code == ESTC_MBXRQST) || 1502 (hp->h_err->e_code == ESTC_SMS_ERR_RECOVERABLE) || 1503 (hp->h_err->e_code == ESTC_DEPROBE) || 1504 (hp->h_err->e_code == EOPL_DEPROBE)) { 1505 /* 1506 * With this failure, the board has been deprobed 1507 * by IKP, and reprobed. We've already gotten rid 1508 * of the old devtree, now we need to reconstruct it 1509 * based on the new IKP probe 1510 */ 1511 if (dr_init_devlists(bp) || (bp->b_ndev == 0)) 1512 goto disconnect_done; 1513 1514 dr_make_comp_nodes(bp); 1515 } 1516 } 1517 /* 1518 * Once all the components on a board have been disconnect 1519 * the board's state can transition to disconnected and 1520 * we can allow the deprobe to take place. 1521 */ 1522 if (hp->h_err == NULL && DR_DEVS_PRESENT(bp) == 0) { 1523 dr_board_transition(bp, DR_STATE_OCCUPIED); 1524 bp->b_rstate = SBD_STAT_DISCONNECTED; 1525 bp->b_ostate = SBD_STAT_UNCONFIGURED; 1526 bp->b_busy = 0; 1527 (void) drv_getparm(TIME, (void *)&bp->b_time); 1528 1529 hp->h_err = drmach_board_deprobe(bp->b_id); 1530 1531 if (hp->h_err == NULL) { 1532 bp->b_ndev = 0; 1533 dr_board_transition(bp, DR_STATE_EMPTY); 1534 bp->b_rstate = SBD_STAT_EMPTY; 1535 (void) drv_getparm(TIME, (void *)&bp->b_time); 1536 } 1537 } 1538 1539 disconnect_done: 1540 dr_unlock_status(bp); 1541 1542 return (0); 1543 } 1544 1545 /* 1546 * Check if a particular device is a valid target of the current 1547 * operation. Return 1 if it is a valid target, and 0 otherwise. 1548 */ 1549 static int 1550 dr_dev_is_target(dr_dev_unit_t *dp, int present_only, uint_t uset) 1551 { 1552 dr_common_unit_t *cp; 1553 int is_present; 1554 int is_attached; 1555 1556 cp = &dp->du_common; 1557 1558 /* check if the user requested this device */ 1559 if ((uset & (1 << cp->sbdev_unum)) == 0) { 1560 return (0); 1561 } 1562 1563 is_present = DR_DEV_IS_PRESENT(cp) ? 1 : 0; 1564 is_attached = DR_DEV_IS_ATTACHED(cp) ? 1 : 0; 1565 1566 /* 1567 * If the present_only flag is set, a valid target 1568 * must be present but not attached. Otherwise, it 1569 * must be both present and attached. 1570 */ 1571 if (is_present && (present_only ^ is_attached)) { 1572 /* sanity check */ 1573 ASSERT(cp->sbdev_id != (drmachid_t)0); 1574 1575 return (1); 1576 } 1577 1578 return (0); 1579 } 1580 1581 static void 1582 dr_dev_make_list(dr_handle_t *hp, sbd_comp_type_t type, int present_only, 1583 dr_common_unit_t ***devlist, int *devnum) 1584 { 1585 dr_board_t *bp = hp->h_bd; 1586 int unum; 1587 int nunits; 1588 uint_t uset; 1589 int len; 1590 dr_common_unit_t **list, **wp; 1591 1592 switch (type) { 1593 case SBD_COMP_CPU: 1594 nunits = MAX_CPU_UNITS_PER_BOARD; 1595 break; 1596 case SBD_COMP_MEM: 1597 nunits = MAX_MEM_UNITS_PER_BOARD; 1598 break; 1599 case SBD_COMP_IO: 1600 nunits = MAX_IO_UNITS_PER_BOARD; 1601 break; 1602 default: 1603 /* catch this in debug kernels */ 1604 ASSERT(0); 1605 break; 1606 } 1607 1608 /* allocate list storage. */ 1609 len = sizeof (dr_common_unit_t *) * (nunits + 1); 1610 list = kmem_zalloc(len, KM_SLEEP); 1611 1612 /* record length of storage in first element */ 1613 *list++ = (dr_common_unit_t *)(uintptr_t)len; 1614 1615 /* get bit array signifying which units are to be involved */ 1616 uset = DEVSET_GET_UNITSET(hp->h_devset, type); 1617 1618 /* 1619 * Adjust the loop count for CPU devices since all cores 1620 * in a CMP will be examined in a single iteration. 1621 */ 1622 if (type == SBD_COMP_CPU) { 1623 nunits = MAX_CMP_UNITS_PER_BOARD; 1624 } 1625 1626 /* populate list */ 1627 for (wp = list, unum = 0; unum < nunits; unum++) { 1628 1629 dr_dev_unit_t *dp; 1630 int core; 1631 int cunum; 1632 1633 dp = DR_GET_BOARD_DEVUNIT(bp, type, unum); 1634 if (dr_dev_is_target(dp, present_only, uset)) { 1635 *wp++ = &dp->du_common; 1636 } 1637 1638 /* further processing is only required for CPUs */ 1639 if (type != SBD_COMP_CPU) { 1640 continue; 1641 } 1642 1643 /* 1644 * Add any additional cores from the current CPU 1645 * device. This is to ensure that all the cores 1646 * are grouped together in the device list, and 1647 * consequently sequenced together during the actual 1648 * operation. 1649 */ 1650 for (core = 1; core < MAX_CORES_PER_CMP; core++) { 1651 1652 cunum = DR_CMP_CORE_UNUM(unum, core); 1653 dp = DR_GET_BOARD_DEVUNIT(bp, type, cunum); 1654 1655 if (dr_dev_is_target(dp, present_only, uset)) { 1656 *wp++ = &dp->du_common; 1657 } 1658 } 1659 } 1660 1661 /* calculate number of units in list, return result and list pointer */ 1662 *devnum = wp - list; 1663 *devlist = list; 1664 } 1665 1666 static void 1667 dr_dev_clean_up(dr_handle_t *hp, dr_common_unit_t **list, int devnum) 1668 { 1669 int len; 1670 int n = 0; 1671 dr_common_unit_t *cp, **rp = list; 1672 1673 /* 1674 * move first encountered unit error to handle if handle 1675 * does not yet have a recorded error. 1676 */ 1677 if (hp->h_err == NULL) { 1678 while (n++ < devnum) { 1679 cp = *rp++; 1680 if (cp->sbdev_error != NULL) { 1681 hp->h_err = cp->sbdev_error; 1682 cp->sbdev_error = NULL; 1683 break; 1684 } 1685 } 1686 } 1687 1688 /* free remaining unit errors */ 1689 while (n++ < devnum) { 1690 cp = *rp++; 1691 if (cp->sbdev_error != NULL) { 1692 sbd_err_clear(&cp->sbdev_error); 1693 cp->sbdev_error = NULL; 1694 } 1695 } 1696 1697 /* free list */ 1698 list -= 1; 1699 len = (int)(uintptr_t)list[0]; 1700 kmem_free(list, len); 1701 } 1702 1703 static int 1704 dr_dev_walk(dr_handle_t *hp, sbd_comp_type_t type, int present_only, 1705 int (*pre_op)(dr_handle_t *, dr_common_unit_t **, int), 1706 void (*op)(dr_handle_t *, dr_common_unit_t *), 1707 int (*post_op)(dr_handle_t *, dr_common_unit_t **, int), 1708 void (*board_op)(dr_handle_t *, dr_common_unit_t **, int)) 1709 { 1710 int devnum, rv; 1711 dr_common_unit_t **devlist; 1712 1713 dr_dev_make_list(hp, type, present_only, &devlist, &devnum); 1714 1715 rv = 0; 1716 if (devnum > 0) { 1717 rv = (*pre_op)(hp, devlist, devnum); 1718 if (rv == 0) { 1719 int n; 1720 1721 for (n = 0; n < devnum; n++) 1722 (*op)(hp, devlist[n]); 1723 1724 rv = (*post_op)(hp, devlist, devnum); 1725 1726 (*board_op)(hp, devlist, devnum); 1727 } 1728 } 1729 1730 dr_dev_clean_up(hp, devlist, devnum); 1731 return (rv); 1732 } 1733 1734 /*ARGSUSED*/ 1735 static int 1736 dr_dev_noop(dr_handle_t *hp, dr_common_unit_t **devlist, int devnum) 1737 { 1738 return (0); 1739 } 1740 1741 static void 1742 dr_attach_update_state(dr_handle_t *hp, 1743 dr_common_unit_t **devlist, int devnum) 1744 { 1745 dr_board_t *bp = hp->h_bd; 1746 int i; 1747 dr_devset_t devs_unattached, devs_present; 1748 static fn_t f = "dr_post_attach_devlist"; 1749 1750 for (i = 0; i < devnum; i++) { 1751 dr_common_unit_t *cp = devlist[i]; 1752 1753 if (dr_check_unit_attached(cp) == -1) { 1754 PR_ALL("%s: ERROR %s not attached\n", 1755 f, cp->sbdev_path); 1756 continue; 1757 } 1758 1759 DR_DEV_SET_ATTACHED(cp); 1760 1761 dr_device_transition(cp, DR_STATE_CONFIGURED); 1762 cp->sbdev_cond = SBD_COND_OK; 1763 } 1764 1765 devs_present = DR_DEVS_PRESENT(bp); 1766 devs_unattached = DR_DEVS_UNATTACHED(bp); 1767 1768 switch (bp->b_state) { 1769 case DR_STATE_CONNECTED: 1770 case DR_STATE_UNCONFIGURED: 1771 ASSERT(devs_present); 1772 1773 if (devs_unattached == 0) { 1774 /* 1775 * All devices finally attached. 1776 */ 1777 dr_board_transition(bp, DR_STATE_CONFIGURED); 1778 hp->h_bd->b_ostate = SBD_STAT_CONFIGURED; 1779 hp->h_bd->b_rstate = SBD_STAT_CONNECTED; 1780 hp->h_bd->b_cond = SBD_COND_OK; 1781 hp->h_bd->b_busy = 0; 1782 (void) drv_getparm(TIME, (void *)&hp->h_bd->b_time); 1783 } else if (devs_present != devs_unattached) { 1784 /* 1785 * Only some devices are fully attached. 1786 */ 1787 dr_board_transition(bp, DR_STATE_PARTIAL); 1788 hp->h_bd->b_rstate = SBD_STAT_CONNECTED; 1789 hp->h_bd->b_ostate = SBD_STAT_CONFIGURED; 1790 (void) drv_getparm(TIME, (void *)&hp->h_bd->b_time); 1791 } 1792 break; 1793 1794 case DR_STATE_PARTIAL: 1795 ASSERT(devs_present); 1796 /* 1797 * All devices finally attached. 1798 */ 1799 if (devs_unattached == 0) { 1800 dr_board_transition(bp, DR_STATE_CONFIGURED); 1801 hp->h_bd->b_rstate = SBD_STAT_CONNECTED; 1802 hp->h_bd->b_ostate = SBD_STAT_CONFIGURED; 1803 hp->h_bd->b_cond = SBD_COND_OK; 1804 hp->h_bd->b_busy = 0; 1805 (void) drv_getparm(TIME, (void *)&hp->h_bd->b_time); 1806 } 1807 break; 1808 1809 default: 1810 break; 1811 } 1812 } 1813 1814 static void 1815 dr_dev_configure(dr_handle_t *hp) 1816 { 1817 int rv; 1818 1819 rv = dr_dev_walk(hp, SBD_COMP_CPU, 1, 1820 dr_pre_attach_cpu, 1821 dr_attach_cpu, 1822 dr_post_attach_cpu, 1823 dr_attach_update_state); 1824 1825 if (rv >= 0) { 1826 rv = dr_dev_walk(hp, SBD_COMP_MEM, 1, 1827 dr_pre_attach_mem, 1828 dr_attach_mem, 1829 dr_post_attach_mem, 1830 dr_attach_update_state); 1831 } 1832 1833 if (rv >= 0) { 1834 (void) dr_dev_walk(hp, SBD_COMP_IO, 1, 1835 dr_pre_attach_io, 1836 dr_attach_io, 1837 dr_post_attach_io, 1838 dr_attach_update_state); 1839 } 1840 } 1841 1842 static void 1843 dr_release_update_state(dr_handle_t *hp, 1844 dr_common_unit_t **devlist, int devnum) 1845 { 1846 _NOTE(ARGUNUSED(devlist)) 1847 _NOTE(ARGUNUSED(devnum)) 1848 1849 dr_board_t *bp = hp->h_bd; 1850 1851 /* 1852 * If the entire board was released and all components 1853 * unreferenced then transfer it to the UNREFERENCED state. 1854 */ 1855 if ((bp->b_state != DR_STATE_RELEASE) && 1856 (DR_DEVS_RELEASED(bp) == DR_DEVS_ATTACHED(bp))) { 1857 dr_board_transition(bp, DR_STATE_RELEASE); 1858 hp->h_bd->b_busy = 1; 1859 } 1860 } 1861 1862 /* called by dr_release_done [below] and dr_release_mem_done [dr_mem.c] */ 1863 int 1864 dr_release_dev_done(dr_common_unit_t *cp) 1865 { 1866 if (cp->sbdev_state == DR_STATE_RELEASE) { 1867 ASSERT(DR_DEV_IS_RELEASED(cp)); 1868 1869 DR_DEV_SET_UNREFERENCED(cp); 1870 1871 dr_device_transition(cp, DR_STATE_UNREFERENCED); 1872 1873 return (0); 1874 } else { 1875 return (-1); 1876 } 1877 } 1878 1879 static void 1880 dr_release_done(dr_handle_t *hp, dr_common_unit_t *cp) 1881 { 1882 _NOTE(ARGUNUSED(hp)) 1883 1884 dr_board_t *bp; 1885 static fn_t f = "dr_release_done"; 1886 1887 PR_ALL("%s...\n", f); 1888 1889 /* get board pointer & sanity check */ 1890 bp = cp->sbdev_bp; 1891 ASSERT(bp == hp->h_bd); 1892 1893 /* 1894 * Transfer the device which just completed its release 1895 * to the UNREFERENCED state. 1896 */ 1897 switch (cp->sbdev_type) { 1898 case SBD_COMP_MEM: 1899 dr_release_mem_done(cp); 1900 break; 1901 1902 default: 1903 DR_DEV_SET_RELEASED(cp); 1904 1905 dr_device_transition(cp, DR_STATE_RELEASE); 1906 1907 (void) dr_release_dev_done(cp); 1908 break; 1909 } 1910 1911 /* 1912 * If we're not already in the RELEASE state for this 1913 * board and we now have released all that were previously 1914 * attached, then transfer the board to the RELEASE state. 1915 */ 1916 if ((bp->b_state == DR_STATE_RELEASE) && 1917 (DR_DEVS_RELEASED(bp) == DR_DEVS_UNREFERENCED(bp))) { 1918 dr_board_transition(bp, DR_STATE_UNREFERENCED); 1919 bp->b_busy = 1; 1920 (void) drv_getparm(TIME, (void *)&bp->b_time); 1921 } 1922 } 1923 1924 static void 1925 dr_dev_release_mem(dr_handle_t *hp, dr_common_unit_t *dv) 1926 { 1927 dr_release_mem(dv); 1928 dr_release_done(hp, dv); 1929 } 1930 1931 static void 1932 dr_dev_release(dr_handle_t *hp) 1933 { 1934 int rv; 1935 1936 hp->h_bd->b_busy = 1; 1937 1938 rv = dr_dev_walk(hp, SBD_COMP_CPU, 0, 1939 dr_pre_release_cpu, 1940 dr_release_done, 1941 dr_dev_noop, 1942 dr_release_update_state); 1943 1944 if (rv >= 0) { 1945 rv = dr_dev_walk(hp, SBD_COMP_MEM, 0, 1946 dr_pre_release_mem, 1947 dr_dev_release_mem, 1948 dr_dev_noop, 1949 dr_release_update_state); 1950 } 1951 1952 if (rv >= 0) { 1953 rv = dr_dev_walk(hp, SBD_COMP_IO, 0, 1954 dr_pre_release_io, 1955 dr_release_done, 1956 dr_dev_noop, 1957 dr_release_update_state); 1958 1959 } 1960 1961 if (rv < 0) 1962 hp->h_bd->b_busy = 0; 1963 /* else, b_busy will be cleared in dr_detach_update_state() */ 1964 } 1965 1966 static void 1967 dr_detach_update_state(dr_handle_t *hp, 1968 dr_common_unit_t **devlist, int devnum) 1969 { 1970 dr_board_t *bp = hp->h_bd; 1971 int i; 1972 dr_state_t bstate; 1973 static fn_t f = "dr_detach_update_state"; 1974 1975 for (i = 0; i < devnum; i++) { 1976 dr_common_unit_t *cp = devlist[i]; 1977 1978 if (dr_check_unit_attached(cp) >= 0) { 1979 /* 1980 * Device is still attached probably due 1981 * to an error. Need to keep track of it. 1982 */ 1983 PR_ALL("%s: ERROR %s not detached\n", 1984 f, cp->sbdev_path); 1985 1986 continue; 1987 } 1988 1989 DR_DEV_CLR_ATTACHED(cp); 1990 DR_DEV_CLR_RELEASED(cp); 1991 DR_DEV_CLR_UNREFERENCED(cp); 1992 dr_device_transition(cp, DR_STATE_UNCONFIGURED); 1993 } 1994 1995 bstate = bp->b_state; 1996 if (bstate != DR_STATE_UNCONFIGURED) { 1997 if (DR_DEVS_PRESENT(bp) == DR_DEVS_UNATTACHED(bp)) { 1998 /* 1999 * All devices are finally detached. 2000 */ 2001 dr_board_transition(bp, DR_STATE_UNCONFIGURED); 2002 hp->h_bd->b_ostate = SBD_STAT_UNCONFIGURED; 2003 (void) drv_getparm(TIME, (void *)&hp->h_bd->b_time); 2004 } else if ((bp->b_state != DR_STATE_PARTIAL) && 2005 (DR_DEVS_ATTACHED(bp) != 2006 DR_DEVS_PRESENT(bp))) { 2007 /* 2008 * Some devices remain attached. 2009 */ 2010 dr_board_transition(bp, DR_STATE_PARTIAL); 2011 (void) drv_getparm(TIME, (void *)&hp->h_bd->b_time); 2012 } 2013 2014 if ((hp->h_devset & DR_DEVS_UNATTACHED(bp)) == hp->h_devset) 2015 hp->h_bd->b_busy = 0; 2016 } 2017 } 2018 2019 static int 2020 dr_dev_unconfigure(dr_handle_t *hp) 2021 { 2022 dr_board_t *bp = hp->h_bd; 2023 2024 /* 2025 * Block out status during IO unconfig. 2026 */ 2027 mutex_enter(&bp->b_slock); 2028 while (bp->b_sflags & DR_BSLOCK) { 2029 if (cv_wait_sig(&bp->b_scv, &bp->b_slock) == 0) { 2030 mutex_exit(&bp->b_slock); 2031 return (EINTR); 2032 } 2033 } 2034 bp->b_sflags |= DR_BSLOCK; 2035 mutex_exit(&bp->b_slock); 2036 2037 (void) dr_dev_walk(hp, SBD_COMP_IO, 0, 2038 dr_pre_detach_io, 2039 dr_detach_io, 2040 dr_post_detach_io, 2041 dr_detach_update_state); 2042 2043 dr_unlock_status(bp); 2044 2045 (void) dr_dev_walk(hp, SBD_COMP_CPU, 0, 2046 dr_pre_detach_cpu, 2047 dr_detach_cpu, 2048 dr_post_detach_cpu, 2049 dr_detach_update_state); 2050 2051 (void) dr_dev_walk(hp, SBD_COMP_MEM, 0, 2052 dr_pre_detach_mem, 2053 dr_detach_mem, 2054 dr_post_detach_mem, 2055 dr_detach_update_state); 2056 2057 return (0); 2058 } 2059 2060 static void 2061 dr_dev_cancel(dr_handle_t *hp) 2062 { 2063 int i; 2064 dr_devset_t devset; 2065 dr_board_t *bp = hp->h_bd; 2066 static fn_t f = "dr_dev_cancel"; 2067 2068 PR_ALL("%s...\n", f); 2069 2070 /* 2071 * Only devices which have been "released" are 2072 * subject to cancellation. 2073 */ 2074 devset = hp->h_devset & DR_DEVS_RELEASED(bp); 2075 2076 /* 2077 * Nothing to do for CPUs or IO other than change back 2078 * their state. 2079 */ 2080 for (i = 0; i < MAX_CPU_UNITS_PER_BOARD; i++) { 2081 dr_cpu_unit_t *cp; 2082 dr_state_t nstate; 2083 2084 if (!DEVSET_IN_SET(devset, SBD_COMP_CPU, i)) 2085 continue; 2086 2087 cp = dr_get_cpu_unit(bp, i); 2088 if (dr_cancel_cpu(cp) == 0) 2089 nstate = DR_STATE_CONFIGURED; 2090 else 2091 nstate = DR_STATE_FATAL; 2092 2093 dr_device_transition(&cp->sbc_cm, nstate); 2094 } 2095 2096 for (i = 0; i < MAX_IO_UNITS_PER_BOARD; i++) { 2097 dr_io_unit_t *ip; 2098 2099 if (!DEVSET_IN_SET(devset, SBD_COMP_IO, i)) 2100 continue; 2101 ip = dr_get_io_unit(bp, i); 2102 dr_device_transition(&ip->sbi_cm, DR_STATE_CONFIGURED); 2103 } 2104 for (i = 0; i < MAX_MEM_UNITS_PER_BOARD; i++) { 2105 dr_mem_unit_t *mp; 2106 dr_state_t nstate; 2107 2108 if (!DEVSET_IN_SET(devset, SBD_COMP_MEM, i)) 2109 continue; 2110 2111 mp = dr_get_mem_unit(bp, i); 2112 if (dr_cancel_mem(mp) == 0) 2113 nstate = DR_STATE_CONFIGURED; 2114 else 2115 nstate = DR_STATE_FATAL; 2116 2117 dr_device_transition(&mp->sbm_cm, nstate); 2118 } 2119 2120 PR_ALL("%s: unreleasing devset (0x%x)\n", f, (uint_t)devset); 2121 2122 DR_DEVS_CANCEL(bp, devset); 2123 2124 if (DR_DEVS_RELEASED(bp) == 0) { 2125 dr_state_t new_state; 2126 /* 2127 * If the board no longer has any released devices 2128 * than transfer it back to the CONFIG/PARTIAL state. 2129 */ 2130 if (DR_DEVS_ATTACHED(bp) == DR_DEVS_PRESENT(bp)) 2131 new_state = DR_STATE_CONFIGURED; 2132 else 2133 new_state = DR_STATE_PARTIAL; 2134 if (bp->b_state != new_state) { 2135 dr_board_transition(bp, new_state); 2136 } 2137 hp->h_bd->b_ostate = SBD_STAT_CONFIGURED; 2138 hp->h_bd->b_busy = 0; 2139 (void) drv_getparm(TIME, (void *)&hp->h_bd->b_time); 2140 } 2141 } 2142 2143 static int 2144 dr_dev_status(dr_handle_t *hp) 2145 { 2146 int nstat, mode, ncm, sz, pbsz, pnstat; 2147 dr_handle_t *shp; 2148 dr_devset_t devset = 0; 2149 sbd_stat_t *dstatp = NULL; 2150 sbd_dev_stat_t *devstatp; 2151 dr_board_t *bp; 2152 drmach_status_t pstat; 2153 int rv = 0; 2154 2155 #ifdef _MULTI_DATAMODEL 2156 int sz32 = 0; 2157 #endif /* _MULTI_DATAMODEL */ 2158 2159 static fn_t f = "dr_status"; 2160 2161 PR_ALL("%s...\n", f); 2162 2163 mode = hp->h_mode; 2164 shp = hp; 2165 devset = shp->h_devset; 2166 bp = hp->h_bd; 2167 2168 /* 2169 * Block out disconnect, unassign, IO unconfigure and 2170 * devinfo branch creation during status. 2171 */ 2172 mutex_enter(&bp->b_slock); 2173 while (bp->b_sflags & DR_BSLOCK) { 2174 if (cv_wait_sig(&bp->b_scv, &bp->b_slock) == 0) { 2175 mutex_exit(&bp->b_slock); 2176 return (EINTR); 2177 } 2178 } 2179 bp->b_sflags |= DR_BSLOCK; 2180 mutex_exit(&bp->b_slock); 2181 2182 ncm = 1; 2183 if (hp->h_sbdcmd.cmd_cm.c_id.c_type == SBD_COMP_NONE) { 2184 if (dr_cmd_flags(hp) & SBD_FLAG_ALLCMP) { 2185 /* 2186 * Calculate the maximum number of components possible 2187 * for a board. This number will be used to size the 2188 * status scratch buffer used by board and component 2189 * status functions. 2190 * This buffer may differ in size from what is provided 2191 * by the plugin, since the known component set on the 2192 * board may change between the plugin's GETNCM call, and 2193 * the status call. Sizing will be adjusted to the plugin's 2194 * receptacle buffer at copyout time. 2195 */ 2196 ncm = MAX_CPU_UNITS_PER_BOARD + 2197 MAX_MEM_UNITS_PER_BOARD + 2198 MAX_IO_UNITS_PER_BOARD; 2199 2200 } else { 2201 /* 2202 * In the case of c_type == SBD_COMP_NONE, and 2203 * SBD_FLAG_ALLCMP not specified, only the board 2204 * info is to be returned, no components. 2205 */ 2206 ncm = 0; 2207 devset = 0; 2208 } 2209 } 2210 2211 sz = sizeof (sbd_stat_t); 2212 if (ncm > 1) 2213 sz += sizeof (sbd_dev_stat_t) * (ncm - 1); 2214 2215 2216 pbsz = (int)hp->h_sbdcmd.cmd_stat.s_nbytes; 2217 pnstat = (pbsz - sizeof (sbd_stat_t))/sizeof (sbd_dev_stat_t); 2218 2219 /* 2220 * s_nbytes describes the size of the preallocated user 2221 * buffer into which the application is execting to 2222 * receive the sbd_stat_t and sbd_dev_stat_t structures. 2223 */ 2224 2225 #ifdef _MULTI_DATAMODEL 2226 2227 /* 2228 * More buffer space is required for the 64bit to 32bit 2229 * conversion of data structures. 2230 */ 2231 if (ddi_model_convert_from(mode & FMODELS) == DDI_MODEL_ILP32) { 2232 sz32 = sizeof (sbd_stat32_t); 2233 if (ncm > 1) 2234 sz32 += sizeof (sbd_dev_stat32_t) * (ncm - 1); 2235 pnstat = (pbsz - sizeof (sbd_stat32_t))/ 2236 sizeof (sbd_dev_stat32_t); 2237 } 2238 2239 sz += sz32; 2240 #endif 2241 /* 2242 * Since one sbd_dev_stat_t is included in the sbd_stat_t, 2243 * increment the plugin's nstat count. 2244 */ 2245 ++pnstat; 2246 2247 if (bp->b_id == 0) { 2248 bzero(&pstat, sizeof (pstat)); 2249 } else { 2250 sbd_error_t *err; 2251 2252 err = drmach_status(bp->b_id, &pstat); 2253 if (err) { 2254 DRERR_SET_C(&hp->h_err, &err); 2255 rv = EIO; 2256 goto status_done; 2257 } 2258 } 2259 2260 dstatp = (sbd_stat_t *)GETSTRUCT(char, sz); 2261 2262 devstatp = &dstatp->s_stat[0]; 2263 2264 dstatp->s_board = bp->b_num; 2265 2266 /* 2267 * Detect transitions between empty and disconnected. 2268 */ 2269 if (!pstat.empty && (bp->b_rstate == SBD_STAT_EMPTY)) 2270 bp->b_rstate = SBD_STAT_DISCONNECTED; 2271 else if (pstat.empty && (bp->b_rstate == SBD_STAT_DISCONNECTED)) 2272 bp->b_rstate = SBD_STAT_EMPTY; 2273 2274 dstatp->s_rstate = bp->b_rstate; 2275 dstatp->s_ostate = bp->b_ostate; 2276 dstatp->s_cond = bp->b_cond = pstat.cond; 2277 dstatp->s_busy = bp->b_busy | pstat.busy; 2278 dstatp->s_time = bp->b_time; 2279 dstatp->s_power = pstat.powered; 2280 dstatp->s_assigned = bp->b_assigned = pstat.assigned; 2281 dstatp->s_nstat = nstat = 0; 2282 bcopy(&pstat.type[0], &dstatp->s_type[0], SBD_TYPE_LEN); 2283 bcopy(&pstat.info[0], &dstatp->s_info[0], SBD_MAX_INFO); 2284 2285 devset &= DR_DEVS_PRESENT(bp); 2286 if (devset == 0) { 2287 /* 2288 * No device chosen. 2289 */ 2290 PR_ALL("%s: no device present\n", f); 2291 } 2292 2293 if (DEVSET_IN_SET(devset, SBD_COMP_CPU, DEVSET_ANYUNIT)) 2294 if ((nstat = dr_cpu_status(hp, devset, devstatp)) > 0) { 2295 dstatp->s_nstat += nstat; 2296 devstatp += nstat; 2297 } 2298 2299 if (DEVSET_IN_SET(devset, SBD_COMP_MEM, DEVSET_ANYUNIT)) 2300 if ((nstat = dr_mem_status(hp, devset, devstatp)) > 0) { 2301 dstatp->s_nstat += nstat; 2302 devstatp += nstat; 2303 } 2304 2305 if (DEVSET_IN_SET(devset, SBD_COMP_IO, DEVSET_ANYUNIT)) 2306 if ((nstat = dr_io_status(hp, devset, devstatp)) > 0) { 2307 dstatp->s_nstat += nstat; 2308 devstatp += nstat; 2309 } 2310 2311 /* 2312 * Due to a possible change in number of components between 2313 * the time of plugin's GETNCM call and now, there may be 2314 * more or less components than the plugin's buffer can 2315 * hold. Adjust s_nstat accordingly. 2316 */ 2317 2318 dstatp->s_nstat = dstatp->s_nstat > pnstat ? pnstat : dstatp->s_nstat; 2319 2320 2321 #ifdef _MULTI_DATAMODEL 2322 if (ddi_model_convert_from(mode & FMODELS) == DDI_MODEL_ILP32) { 2323 int i, j; 2324 sbd_stat32_t *dstat32p; 2325 2326 dstat32p = (sbd_stat32_t *)devstatp; 2327 2328 /* Alignment Paranoia */ 2329 if ((ulong_t)dstat32p & 0x1) { 2330 PR_ALL("%s: alignment: sz=0x%lx dstat32p=0x%p\n", 2331 f, sizeof (sbd_stat32_t), dstat32p); 2332 DR_OP_INTERNAL_ERROR(hp); 2333 rv = EINVAL; 2334 goto status_done; 2335 } 2336 2337 /* paranoia: detect buffer overrun */ 2338 if ((caddr_t)&dstat32p->s_stat[dstatp->s_nstat] > 2339 ((caddr_t)dstatp) + sz) { 2340 DR_OP_INTERNAL_ERROR(hp); 2341 rv = EINVAL; 2342 goto status_done; 2343 } 2344 2345 /* copy sbd_stat_t structure members */ 2346 #define _SBD_STAT(t, m) dstat32p->m = (t)dstatp->m 2347 _SBD_STAT(int32_t, s_board); 2348 _SBD_STAT(int32_t, s_rstate); 2349 _SBD_STAT(int32_t, s_ostate); 2350 _SBD_STAT(int32_t, s_cond); 2351 _SBD_STAT(int32_t, s_busy); 2352 _SBD_STAT(time32_t, s_time); 2353 _SBD_STAT(uint32_t, s_power); 2354 _SBD_STAT(uint32_t, s_assigned); 2355 _SBD_STAT(int32_t, s_nstat); 2356 bcopy(&dstatp->s_type[0], &dstat32p->s_type[0], 2357 SBD_TYPE_LEN); 2358 bcopy(&dstatp->s_info[0], &dstat32p->s_info[0], 2359 SBD_MAX_INFO); 2360 #undef _SBD_STAT 2361 2362 for (i = 0; i < dstatp->s_nstat; i++) { 2363 sbd_dev_stat_t *dsp = &dstatp->s_stat[i]; 2364 sbd_dev_stat32_t *ds32p = &dstat32p->s_stat[i]; 2365 #define _SBD_DEV_STAT(t, m) ds32p->m = (t)dsp->m 2366 2367 /* copy sbd_cm_stat_t structure members */ 2368 _SBD_DEV_STAT(int32_t, ds_type); 2369 _SBD_DEV_STAT(int32_t, ds_unit); 2370 _SBD_DEV_STAT(int32_t, ds_ostate); 2371 _SBD_DEV_STAT(int32_t, ds_cond); 2372 _SBD_DEV_STAT(int32_t, ds_busy); 2373 _SBD_DEV_STAT(int32_t, ds_suspend); 2374 _SBD_DEV_STAT(time32_t, ds_time); 2375 bcopy(&dsp->ds_name[0], &ds32p->ds_name[0], 2376 OBP_MAXPROPNAME); 2377 2378 switch (dsp->ds_type) { 2379 case SBD_COMP_CPU: 2380 /* copy sbd_cpu_stat_t structure members */ 2381 _SBD_DEV_STAT(int32_t, d_cpu.cs_isbootproc); 2382 _SBD_DEV_STAT(int32_t, d_cpu.cs_cpuid); 2383 _SBD_DEV_STAT(int32_t, d_cpu.cs_speed); 2384 _SBD_DEV_STAT(int32_t, d_cpu.cs_ecache); 2385 break; 2386 2387 case SBD_COMP_MEM: 2388 /* copy sbd_mem_stat_t structure members */ 2389 _SBD_DEV_STAT(int32_t, d_mem.ms_interleave); 2390 _SBD_DEV_STAT(uint32_t, d_mem.ms_basepfn); 2391 _SBD_DEV_STAT(uint32_t, d_mem.ms_totpages); 2392 _SBD_DEV_STAT(uint32_t, d_mem.ms_detpages); 2393 _SBD_DEV_STAT(int32_t, d_mem.ms_pageslost); 2394 _SBD_DEV_STAT(uint32_t, d_mem.ms_managed_pages); 2395 _SBD_DEV_STAT(uint32_t, d_mem.ms_noreloc_pages); 2396 _SBD_DEV_STAT(uint32_t, d_mem.ms_noreloc_first); 2397 _SBD_DEV_STAT(uint32_t, d_mem.ms_noreloc_last); 2398 _SBD_DEV_STAT(int32_t, d_mem.ms_cage_enabled); 2399 _SBD_DEV_STAT(int32_t, d_mem.ms_peer_is_target); 2400 bcopy(&dsp->d_mem.ms_peer_ap_id[0], 2401 &ds32p->d_mem.ms_peer_ap_id[0], 2402 sizeof (ds32p->d_mem.ms_peer_ap_id)); 2403 break; 2404 2405 case SBD_COMP_IO: 2406 /* copy sbd_io_stat_t structure members */ 2407 _SBD_DEV_STAT(int32_t, d_io.is_referenced); 2408 _SBD_DEV_STAT(int32_t, d_io.is_unsafe_count); 2409 2410 for (j = 0; j < SBD_MAX_UNSAFE; j++) 2411 _SBD_DEV_STAT(int32_t, 2412 d_io.is_unsafe_list[j]); 2413 2414 bcopy(&dsp->d_io.is_pathname[0], 2415 &ds32p->d_io.is_pathname[0], MAXPATHLEN); 2416 break; 2417 2418 case SBD_COMP_CMP: 2419 /* copy sbd_cmp_stat_t structure members */ 2420 bcopy(&dsp->d_cmp.ps_cpuid[0], 2421 &ds32p->d_cmp.ps_cpuid[0], 2422 sizeof (ds32p->d_cmp.ps_cpuid)); 2423 _SBD_DEV_STAT(int32_t, d_cmp.ps_ncores); 2424 _SBD_DEV_STAT(int32_t, d_cmp.ps_speed); 2425 _SBD_DEV_STAT(int32_t, d_cmp.ps_ecache); 2426 break; 2427 2428 default: 2429 cmn_err(CE_WARN, "%s: unknown dev type (%d)", 2430 f, (int)dsp->ds_type); 2431 rv = EFAULT; 2432 goto status_done; 2433 } 2434 #undef _SBD_DEV_STAT 2435 } 2436 2437 2438 if (ddi_copyout((void *)dstat32p, 2439 hp->h_sbdcmd.cmd_stat.s_statp, pbsz, mode) != 0) { 2440 cmn_err(CE_WARN, 2441 "%s: failed to copyout status " 2442 "for board %d", f, bp->b_num); 2443 rv = EFAULT; 2444 goto status_done; 2445 } 2446 } else 2447 #endif /* _MULTI_DATAMODEL */ 2448 2449 if (ddi_copyout((void *)dstatp, hp->h_sbdcmd.cmd_stat.s_statp, 2450 pbsz, mode) != 0) { 2451 cmn_err(CE_WARN, 2452 "%s: failed to copyout status for board %d", 2453 f, bp->b_num); 2454 rv = EFAULT; 2455 goto status_done; 2456 } 2457 2458 status_done: 2459 if (dstatp != NULL) 2460 FREESTRUCT(dstatp, char, sz); 2461 2462 dr_unlock_status(bp); 2463 2464 return (rv); 2465 } 2466 2467 static int 2468 dr_get_ncm(dr_handle_t *hp) 2469 { 2470 int i; 2471 int ncm = 0; 2472 dr_devset_t devset; 2473 2474 devset = DR_DEVS_PRESENT(hp->h_bd); 2475 if (hp->h_sbdcmd.cmd_cm.c_id.c_type != SBD_COMP_NONE) 2476 devset &= DEVSET(hp->h_sbdcmd.cmd_cm.c_id.c_type, 2477 DEVSET_ANYUNIT); 2478 2479 /* 2480 * Handle CPUs first to deal with possible CMP 2481 * devices. If the CPU is a CMP, we need to only 2482 * increment ncm once even if there are multiple 2483 * cores for that CMP present in the devset. 2484 */ 2485 for (i = 0; i < MAX_CMP_UNITS_PER_BOARD; i++) { 2486 if (devset & DEVSET(SBD_COMP_CMP, i)) { 2487 ncm++; 2488 } 2489 } 2490 2491 /* eliminate the CPU information from the devset */ 2492 devset &= ~(DEVSET(SBD_COMP_CMP, DEVSET_ANYUNIT)); 2493 2494 for (i = 0; i < (sizeof (dr_devset_t) * 8); i++) { 2495 ncm += devset & 0x1; 2496 devset >>= 1; 2497 } 2498 2499 return (ncm); 2500 } 2501 2502 /* used by dr_mem.c */ 2503 /* TODO: eliminate dr_boardlist */ 2504 dr_board_t * 2505 dr_lookup_board(int board_num) 2506 { 2507 dr_board_t *bp; 2508 2509 ASSERT(board_num >= 0 && board_num < MAX_BOARDS); 2510 2511 bp = &dr_boardlist[board_num]; 2512 ASSERT(bp->b_num == board_num); 2513 2514 return (bp); 2515 } 2516 2517 static dr_dev_unit_t * 2518 dr_get_dev_unit(dr_board_t *bp, sbd_comp_type_t nt, int unit_num) 2519 { 2520 dr_dev_unit_t *dp; 2521 2522 dp = DR_GET_BOARD_DEVUNIT(bp, nt, unit_num); 2523 ASSERT(dp->du_common.sbdev_bp == bp); 2524 ASSERT(dp->du_common.sbdev_unum == unit_num); 2525 ASSERT(dp->du_common.sbdev_type == nt); 2526 2527 return (dp); 2528 } 2529 2530 dr_cpu_unit_t * 2531 dr_get_cpu_unit(dr_board_t *bp, int unit_num) 2532 { 2533 dr_dev_unit_t *dp; 2534 2535 ASSERT(unit_num >= 0 && unit_num < MAX_CPU_UNITS_PER_BOARD); 2536 2537 dp = dr_get_dev_unit(bp, SBD_COMP_CPU, unit_num); 2538 return (&dp->du_cpu); 2539 } 2540 2541 dr_mem_unit_t * 2542 dr_get_mem_unit(dr_board_t *bp, int unit_num) 2543 { 2544 dr_dev_unit_t *dp; 2545 2546 ASSERT(unit_num >= 0 && unit_num < MAX_MEM_UNITS_PER_BOARD); 2547 2548 dp = dr_get_dev_unit(bp, SBD_COMP_MEM, unit_num); 2549 return (&dp->du_mem); 2550 } 2551 2552 dr_io_unit_t * 2553 dr_get_io_unit(dr_board_t *bp, int unit_num) 2554 { 2555 dr_dev_unit_t *dp; 2556 2557 ASSERT(unit_num >= 0 && unit_num < MAX_IO_UNITS_PER_BOARD); 2558 2559 dp = dr_get_dev_unit(bp, SBD_COMP_IO, unit_num); 2560 return (&dp->du_io); 2561 } 2562 2563 dr_common_unit_t * 2564 dr_get_common_unit(dr_board_t *bp, sbd_comp_type_t nt, int unum) 2565 { 2566 dr_dev_unit_t *dp; 2567 2568 dp = dr_get_dev_unit(bp, nt, unum); 2569 return (&dp->du_common); 2570 } 2571 2572 static dr_devset_t 2573 dr_dev2devset(sbd_comp_id_t *cid) 2574 { 2575 static fn_t f = "dr_dev2devset"; 2576 2577 dr_devset_t devset; 2578 int unit = cid->c_unit; 2579 2580 switch (cid->c_type) { 2581 case SBD_COMP_NONE: 2582 devset = DEVSET(SBD_COMP_CPU, DEVSET_ANYUNIT); 2583 devset |= DEVSET(SBD_COMP_MEM, DEVSET_ANYUNIT); 2584 devset |= DEVSET(SBD_COMP_IO, DEVSET_ANYUNIT); 2585 PR_ALL("%s: COMP_NONE devset = 0x%lx\n", f, devset); 2586 break; 2587 2588 case SBD_COMP_CPU: 2589 if ((unit > MAX_CPU_UNITS_PER_BOARD) || (unit < 0)) { 2590 cmn_err(CE_WARN, 2591 "%s: invalid cpu unit# = %d", 2592 f, unit); 2593 devset = 0; 2594 } else { 2595 /* 2596 * Generate a devset that includes all the 2597 * cores of a CMP device. If this is not a 2598 * CMP, the extra cores will be eliminated 2599 * later since they are not present. This is 2600 * also true for CMP devices that do not have 2601 * all cores active. 2602 */ 2603 devset = DEVSET(SBD_COMP_CMP, unit); 2604 } 2605 2606 PR_ALL("%s: CPU devset = 0x%lx\n", f, devset); 2607 break; 2608 2609 case SBD_COMP_MEM: 2610 if (unit == SBD_NULL_UNIT) { 2611 unit = 0; 2612 cid->c_unit = 0; 2613 } 2614 2615 if ((unit > MAX_MEM_UNITS_PER_BOARD) || (unit < 0)) { 2616 cmn_err(CE_WARN, 2617 "%s: invalid mem unit# = %d", 2618 f, unit); 2619 devset = 0; 2620 } else 2621 devset = DEVSET(cid->c_type, unit); 2622 2623 PR_ALL("%s: MEM devset = 0x%lx\n", f, devset); 2624 break; 2625 2626 case SBD_COMP_IO: 2627 if ((unit > MAX_IO_UNITS_PER_BOARD) || (unit < 0)) { 2628 cmn_err(CE_WARN, 2629 "%s: invalid io unit# = %d", 2630 f, unit); 2631 devset = 0; 2632 } else 2633 devset = DEVSET(cid->c_type, unit); 2634 2635 PR_ALL("%s: IO devset = 0x%lx\n", f, devset); 2636 break; 2637 2638 default: 2639 case SBD_COMP_UNKNOWN: 2640 devset = 0; 2641 break; 2642 } 2643 2644 return (devset); 2645 } 2646 2647 /* 2648 * Converts a dynamic attachment point name to a SBD_COMP_* type. 2649 * Returns SDB_COMP_UNKNOWN if name is not recognized. 2650 */ 2651 static int 2652 dr_dev_type_to_nt(char *type) 2653 { 2654 int i; 2655 2656 for (i = 0; dr_devattr[i].s_nodetype != SBD_COMP_UNKNOWN; i++) 2657 if (strcmp(dr_devattr[i].s_devtype, type) == 0) 2658 break; 2659 2660 return (dr_devattr[i].s_nodetype); 2661 } 2662 2663 /* 2664 * Converts a SBD_COMP_* type to a dynamic attachment point name. 2665 * Return NULL if SBD_COMP_ type is not recognized. 2666 */ 2667 char * 2668 dr_nt_to_dev_type(int nt) 2669 { 2670 int i; 2671 2672 for (i = 0; dr_devattr[i].s_nodetype != SBD_COMP_UNKNOWN; i++) 2673 if (dr_devattr[i].s_nodetype == nt) 2674 break; 2675 2676 return (dr_devattr[i].s_devtype); 2677 } 2678 2679 2680 /* 2681 * State transition policy is that if there is some component for which 2682 * the state transition is valid, then let it through. The exception is 2683 * SBD_CMD_DISCONNECT. On disconnect, the state transition must be valid 2684 * for ALL components. 2685 * Returns the state that is in error, if any. 2686 */ 2687 static int 2688 dr_check_transition(dr_board_t *bp, dr_devset_t *devsetp, 2689 struct dr_state_trans *transp, int cmd) 2690 { 2691 int s, ut; 2692 int state_err = 0; 2693 dr_devset_t devset; 2694 dr_common_unit_t *cp; 2695 static fn_t f = "dr_check_transition"; 2696 2697 devset = *devsetp; 2698 2699 if (DEVSET_IN_SET(devset, SBD_COMP_CPU, DEVSET_ANYUNIT)) { 2700 for (ut = 0; ut < MAX_CPU_UNITS_PER_BOARD; ut++) { 2701 if (DEVSET_IN_SET(devset, SBD_COMP_CPU, ut) == 0) 2702 continue; 2703 2704 cp = dr_get_common_unit(bp, SBD_COMP_CPU, ut); 2705 s = (int)cp->sbdev_state; 2706 if (!DR_DEV_IS_PRESENT(cp)) { 2707 DEVSET_DEL(devset, SBD_COMP_CPU, ut); 2708 } else { 2709 if (transp->x_op[s].x_rv) { 2710 if (!state_err) 2711 state_err = s; 2712 DEVSET_DEL(devset, SBD_COMP_CPU, ut); 2713 } 2714 } 2715 } 2716 } 2717 if (DEVSET_IN_SET(devset, SBD_COMP_MEM, DEVSET_ANYUNIT)) { 2718 for (ut = 0; ut < MAX_MEM_UNITS_PER_BOARD; ut++) { 2719 if (DEVSET_IN_SET(devset, SBD_COMP_MEM, ut) == 0) 2720 continue; 2721 2722 cp = dr_get_common_unit(bp, SBD_COMP_MEM, ut); 2723 s = (int)cp->sbdev_state; 2724 if (!DR_DEV_IS_PRESENT(cp)) { 2725 DEVSET_DEL(devset, SBD_COMP_MEM, ut); 2726 } else { 2727 if (transp->x_op[s].x_rv) { 2728 if (!state_err) 2729 state_err = s; 2730 DEVSET_DEL(devset, SBD_COMP_MEM, ut); 2731 } 2732 } 2733 } 2734 } 2735 if (DEVSET_IN_SET(devset, SBD_COMP_IO, DEVSET_ANYUNIT)) { 2736 for (ut = 0; ut < MAX_IO_UNITS_PER_BOARD; ut++) { 2737 if (DEVSET_IN_SET(devset, SBD_COMP_IO, ut) == 0) 2738 continue; 2739 2740 cp = dr_get_common_unit(bp, SBD_COMP_IO, ut); 2741 s = (int)cp->sbdev_state; 2742 if (!DR_DEV_IS_PRESENT(cp)) { 2743 DEVSET_DEL(devset, SBD_COMP_IO, ut); 2744 } else { 2745 if (transp->x_op[s].x_rv) { 2746 if (!state_err) 2747 state_err = s; 2748 DEVSET_DEL(devset, SBD_COMP_IO, ut); 2749 } 2750 } 2751 } 2752 } 2753 2754 PR_ALL("%s: requested devset = 0x%x, final devset = 0x%x\n", 2755 f, (uint_t)*devsetp, (uint_t)devset); 2756 2757 *devsetp = devset; 2758 /* 2759 * If there are some remaining components for which 2760 * this state transition is valid, then allow them 2761 * through, otherwise if none are left then return 2762 * the state error. The exception is SBD_CMD_DISCONNECT. 2763 * On disconnect, the state transition must be valid for ALL 2764 * components. 2765 */ 2766 if (cmd == SBD_CMD_DISCONNECT) 2767 return (state_err); 2768 return (devset ? 0 : state_err); 2769 } 2770 2771 void 2772 dr_device_transition(dr_common_unit_t *cp, dr_state_t st) 2773 { 2774 PR_STATE("%s STATE %s(%d) -> %s(%d)\n", 2775 cp->sbdev_path, 2776 state_str[cp->sbdev_state], cp->sbdev_state, 2777 state_str[st], st); 2778 2779 cp->sbdev_state = st; 2780 if (st == DR_STATE_CONFIGURED) { 2781 cp->sbdev_ostate = SBD_STAT_CONFIGURED; 2782 if (cp->sbdev_bp->b_ostate != SBD_STAT_CONFIGURED) { 2783 cp->sbdev_bp->b_ostate = SBD_STAT_CONFIGURED; 2784 (void) drv_getparm(TIME, 2785 (void *) &cp->sbdev_bp->b_time); 2786 } 2787 } else 2788 cp->sbdev_ostate = SBD_STAT_UNCONFIGURED; 2789 2790 (void) drv_getparm(TIME, (void *) &cp->sbdev_time); 2791 } 2792 2793 static void 2794 dr_board_transition(dr_board_t *bp, dr_state_t st) 2795 { 2796 PR_STATE("BOARD %d STATE: %s(%d) -> %s(%d)\n", 2797 bp->b_num, 2798 state_str[bp->b_state], bp->b_state, 2799 state_str[st], st); 2800 2801 bp->b_state = st; 2802 } 2803 2804 void 2805 dr_op_err(int ce, dr_handle_t *hp, int code, char *fmt, ...) 2806 { 2807 sbd_error_t *err; 2808 va_list args; 2809 2810 va_start(args, fmt); 2811 err = drerr_new_v(code, fmt, args); 2812 va_end(args); 2813 2814 if (ce != CE_IGNORE) 2815 sbd_err_log(err, ce); 2816 2817 DRERR_SET_C(&hp->h_err, &err); 2818 } 2819 2820 void 2821 dr_dev_err(int ce, dr_common_unit_t *cp, int code) 2822 { 2823 sbd_error_t *err; 2824 2825 err = drerr_new(0, code, cp->sbdev_path, NULL); 2826 2827 if (ce != CE_IGNORE) 2828 sbd_err_log(err, ce); 2829 2830 DRERR_SET_C(&cp->sbdev_error, &err); 2831 } 2832 2833 /* 2834 * A callback routine. Called from the drmach layer as a result of 2835 * call to drmach_board_find_devices from dr_init_devlists. 2836 */ 2837 static sbd_error_t * 2838 dr_dev_found(void *data, const char *name, int unum, drmachid_t id) 2839 { 2840 dr_board_t *bp = data; 2841 dr_dev_unit_t *dp; 2842 int nt; 2843 static fn_t f = "dr_dev_found"; 2844 2845 PR_ALL("%s (board = %d, name = %s, unum = %d, id = %p)...\n", 2846 f, bp->b_num, name, unum, id); 2847 2848 nt = dr_dev_type_to_nt((char *)name); 2849 if (nt == SBD_COMP_UNKNOWN) { 2850 /* 2851 * this should not happen. When it does, it indicates 2852 * a missmatch in devices supported by the drmach layer 2853 * vs devices supported by this layer. 2854 */ 2855 return (DR_INTERNAL_ERROR()); 2856 } 2857 2858 dp = DR_GET_BOARD_DEVUNIT(bp, nt, unum); 2859 2860 /* sanity check */ 2861 ASSERT(dp->du_common.sbdev_bp == bp); 2862 ASSERT(dp->du_common.sbdev_unum == unum); 2863 ASSERT(dp->du_common.sbdev_type == nt); 2864 2865 /* render dynamic attachment point path of this unit */ 2866 (void) snprintf(dp->du_common.sbdev_path, 2867 sizeof (dp->du_common.sbdev_path), 2868 (nt == SBD_COMP_MEM ? "%s::%s" : "%s::%s%d"), 2869 bp->b_path, name, DR_UNUM2SBD_UNUM(unum, nt)); 2870 2871 dp->du_common.sbdev_id = id; 2872 DR_DEV_SET_PRESENT(&dp->du_common); 2873 2874 bp->b_ndev++; 2875 2876 return (NULL); 2877 } 2878 2879 static sbd_error_t * 2880 dr_init_devlists(dr_board_t *bp) 2881 { 2882 int i; 2883 sbd_error_t *err; 2884 dr_dev_unit_t *dp; 2885 static fn_t f = "dr_init_devlists"; 2886 2887 PR_ALL("%s (%s)...\n", f, bp->b_path); 2888 2889 /* sanity check */ 2890 ASSERT(bp->b_ndev == 0); 2891 2892 DR_DEVS_DISCONNECT(bp, (uint_t)-1); 2893 2894 /* 2895 * This routine builds the board's devlist and initializes 2896 * the common portion of the unit data structures. 2897 * Note: because the common portion is considered 2898 * uninitialized, the dr_get_*_unit() routines can not 2899 * be used. 2900 */ 2901 2902 /* 2903 * Clear out old entries, if any. 2904 */ 2905 for (i = 0; i < MAX_CPU_UNITS_PER_BOARD; i++) { 2906 dp = DR_GET_BOARD_DEVUNIT(bp, SBD_COMP_CPU, i); 2907 2908 bzero(dp, sizeof (*dp)); 2909 dp->du_common.sbdev_bp = bp; 2910 dp->du_common.sbdev_unum = i; 2911 dp->du_common.sbdev_type = SBD_COMP_CPU; 2912 } 2913 2914 for (i = 0; i < MAX_MEM_UNITS_PER_BOARD; i++) { 2915 dp = DR_GET_BOARD_DEVUNIT(bp, SBD_COMP_MEM, i); 2916 2917 bzero(dp, sizeof (*dp)); 2918 dp->du_common.sbdev_bp = bp; 2919 dp->du_common.sbdev_unum = i; 2920 dp->du_common.sbdev_type = SBD_COMP_MEM; 2921 } 2922 2923 for (i = 0; i < MAX_IO_UNITS_PER_BOARD; i++) { 2924 dp = DR_GET_BOARD_DEVUNIT(bp, SBD_COMP_IO, i); 2925 2926 bzero(dp, sizeof (*dp)); 2927 dp->du_common.sbdev_bp = bp; 2928 dp->du_common.sbdev_unum = i; 2929 dp->du_common.sbdev_type = SBD_COMP_IO; 2930 } 2931 2932 err = NULL; 2933 if (bp->b_id) { 2934 /* find devices on this board */ 2935 err = drmach_board_find_devices( 2936 bp->b_id, bp, dr_dev_found); 2937 } 2938 2939 return (err); 2940 } 2941 2942 /* 2943 * Return the unit number of the respective drmachid if 2944 * it's found to be attached. 2945 */ 2946 static int 2947 dr_check_unit_attached(dr_common_unit_t *cp) 2948 { 2949 int rv = 0; 2950 processorid_t cpuid; 2951 uint64_t basepa, endpa; 2952 struct memlist *ml; 2953 extern struct memlist *phys_install; 2954 sbd_error_t *err; 2955 int yes; 2956 static fn_t f = "dr_check_unit_attached"; 2957 2958 switch (cp->sbdev_type) { 2959 case SBD_COMP_CPU: 2960 err = drmach_cpu_get_id(cp->sbdev_id, &cpuid); 2961 if (err) { 2962 DRERR_SET_C(&cp->sbdev_error, &err); 2963 rv = -1; 2964 break; 2965 } 2966 mutex_enter(&cpu_lock); 2967 if (cpu_get(cpuid) == NULL) 2968 rv = -1; 2969 mutex_exit(&cpu_lock); 2970 break; 2971 2972 case SBD_COMP_MEM: 2973 err = drmach_mem_get_base_physaddr(cp->sbdev_id, &basepa); 2974 if (err) { 2975 DRERR_SET_C(&cp->sbdev_error, &err); 2976 rv = -1; 2977 break; 2978 } 2979 2980 /* 2981 * basepa may not be on a alignment boundary, make it so. 2982 */ 2983 err = drmach_mem_get_slice_size(cp->sbdev_id, &endpa); 2984 if (err) { 2985 DRERR_SET_C(&cp->sbdev_error, &err); 2986 rv = -1; 2987 break; 2988 } 2989 2990 basepa &= ~(endpa - 1); 2991 endpa += basepa; 2992 2993 /* 2994 * Check if base address is in phys_install. 2995 */ 2996 memlist_read_lock(); 2997 for (ml = phys_install; ml; ml = ml->next) 2998 if ((endpa <= ml->address) || 2999 (basepa >= (ml->address + ml->size))) 3000 continue; 3001 else 3002 break; 3003 memlist_read_unlock(); 3004 if (ml == NULL) 3005 rv = -1; 3006 break; 3007 3008 case SBD_COMP_IO: 3009 err = drmach_io_is_attached(cp->sbdev_id, &yes); 3010 if (err) { 3011 DRERR_SET_C(&cp->sbdev_error, &err); 3012 rv = -1; 3013 break; 3014 } else if (!yes) 3015 rv = -1; 3016 break; 3017 3018 default: 3019 PR_ALL("%s: unexpected nodetype(%d) for id 0x%p\n", 3020 f, cp->sbdev_type, cp->sbdev_id); 3021 rv = -1; 3022 break; 3023 } 3024 3025 return (rv); 3026 } 3027 3028 /* 3029 * See if drmach recognizes the passthru command. DRMACH expects the 3030 * id to identify the thing to which the command is being applied. Using 3031 * nonsense SBD terms, that information has been perversely encoded in the 3032 * c_id member of the sbd_cmd_t structure. This logic reads those tea 3033 * leaves, finds the associated drmach id, then calls drmach to process 3034 * the passthru command. 3035 */ 3036 static int 3037 dr_pt_try_drmach(dr_handle_t *hp) 3038 { 3039 dr_board_t *bp = hp->h_bd; 3040 sbd_comp_id_t *comp_id = &hp->h_sbdcmd.cmd_cm.c_id; 3041 drmachid_t id; 3042 3043 if (comp_id->c_type == SBD_COMP_NONE) { 3044 id = bp->b_id; 3045 } else { 3046 sbd_comp_type_t nt; 3047 3048 nt = dr_dev_type_to_nt(comp_id->c_name); 3049 if (nt == SBD_COMP_UNKNOWN) { 3050 dr_op_err(CE_IGNORE, hp, ESBD_INVAL, comp_id->c_name); 3051 id = 0; 3052 } else { 3053 /* pt command applied to dynamic attachment point */ 3054 dr_common_unit_t *cp; 3055 cp = dr_get_common_unit(bp, nt, comp_id->c_unit); 3056 id = cp->sbdev_id; 3057 } 3058 } 3059 3060 if (hp->h_err == NULL) 3061 hp->h_err = drmach_passthru(id, &hp->h_opts); 3062 3063 return (hp->h_err == NULL ? 0 : -1); 3064 } 3065 3066 static int 3067 dr_pt_ioctl(dr_handle_t *hp) 3068 { 3069 int cmd, rv, len; 3070 int32_t sz; 3071 int found; 3072 char *copts; 3073 static fn_t f = "dr_pt_ioctl"; 3074 3075 PR_ALL("%s...\n", f); 3076 3077 sz = hp->h_opts.size; 3078 copts = hp->h_opts.copts; 3079 3080 if (sz == 0 || copts == (char *)NULL) { 3081 cmn_err(CE_WARN, "%s: invalid passthru args", f); 3082 return (EINVAL); 3083 } 3084 3085 found = 0; 3086 for (cmd = 0; cmd < (sizeof (pt_arr) / sizeof (pt_arr[0])); cmd++) { 3087 len = strlen(pt_arr[cmd].pt_name); 3088 found = (strncmp(pt_arr[cmd].pt_name, copts, len) == 0); 3089 if (found) 3090 break; 3091 } 3092 3093 if (found) 3094 rv = (*pt_arr[cmd].pt_func)(hp); 3095 else 3096 rv = dr_pt_try_drmach(hp); 3097 3098 return (rv); 3099 } 3100 3101 /* 3102 * Called at driver load time to determine the state and condition 3103 * of an existing board in the system. 3104 */ 3105 static void 3106 dr_board_discovery(dr_board_t *bp) 3107 { 3108 int i; 3109 dr_devset_t devs_lost, devs_attached = 0; 3110 dr_cpu_unit_t *cp; 3111 dr_mem_unit_t *mp; 3112 dr_io_unit_t *ip; 3113 static fn_t f = "dr_board_discovery"; 3114 3115 if (DR_DEVS_PRESENT(bp) == 0) { 3116 PR_ALL("%s: board %d has no devices present\n", 3117 f, bp->b_num); 3118 return; 3119 } 3120 3121 /* 3122 * Check for existence of cpus. 3123 */ 3124 for (i = 0; i < MAX_CPU_UNITS_PER_BOARD; i++) { 3125 cp = dr_get_cpu_unit(bp, i); 3126 3127 if (!DR_DEV_IS_PRESENT(&cp->sbc_cm)) 3128 continue; 3129 3130 if (dr_check_unit_attached(&cp->sbc_cm) >= 0) { 3131 DR_DEV_SET_ATTACHED(&cp->sbc_cm); 3132 DEVSET_ADD(devs_attached, SBD_COMP_CPU, i); 3133 PR_ALL("%s: board %d, cpu-unit %d - attached\n", 3134 f, bp->b_num, i); 3135 } 3136 dr_init_cpu_unit(cp); 3137 } 3138 3139 /* 3140 * Check for existence of memory. 3141 */ 3142 for (i = 0; i < MAX_MEM_UNITS_PER_BOARD; i++) { 3143 mp = dr_get_mem_unit(bp, i); 3144 3145 if (!DR_DEV_IS_PRESENT(&mp->sbm_cm)) 3146 continue; 3147 3148 if (dr_check_unit_attached(&mp->sbm_cm) >= 0) { 3149 DR_DEV_SET_ATTACHED(&mp->sbm_cm); 3150 DEVSET_ADD(devs_attached, SBD_COMP_MEM, i); 3151 PR_ALL("%s: board %d, mem-unit %d - attached\n", 3152 f, bp->b_num, i); 3153 } 3154 dr_init_mem_unit(mp); 3155 } 3156 3157 /* 3158 * Check for i/o state. 3159 */ 3160 for (i = 0; i < MAX_IO_UNITS_PER_BOARD; i++) { 3161 ip = dr_get_io_unit(bp, i); 3162 3163 if (!DR_DEV_IS_PRESENT(&ip->sbi_cm)) 3164 continue; 3165 3166 if (dr_check_unit_attached(&ip->sbi_cm) >= 0) { 3167 /* 3168 * Found it! 3169 */ 3170 DR_DEV_SET_ATTACHED(&ip->sbi_cm); 3171 DEVSET_ADD(devs_attached, SBD_COMP_IO, i); 3172 PR_ALL("%s: board %d, io-unit %d - attached\n", 3173 f, bp->b_num, i); 3174 } 3175 dr_init_io_unit(ip); 3176 } 3177 3178 DR_DEVS_CONFIGURE(bp, devs_attached); 3179 if (devs_attached && ((devs_lost = DR_DEVS_UNATTACHED(bp)) != 0)) { 3180 int ut; 3181 /* 3182 * It is not legal on board discovery to have a 3183 * board that is only partially attached. A board 3184 * is either all attached or all connected. If a 3185 * board has at least one attached device, then 3186 * the the remaining devices, if any, must have 3187 * been lost or disconnected. These devices can 3188 * only be recovered by a full attach from scratch. 3189 * Note that devices previously in the unreferenced 3190 * state are subsequently lost until the next full 3191 * attach. This is necessary since the driver unload 3192 * that must have occurred would have wiped out the 3193 * information necessary to re-configure the device 3194 * back online, e.g. memlist. 3195 */ 3196 PR_ALL("%s: some devices LOST (0x%lx)...\n", f, devs_lost); 3197 3198 for (ut = 0; ut < MAX_CPU_UNITS_PER_BOARD; ut++) { 3199 if (!DEVSET_IN_SET(devs_lost, SBD_COMP_CPU, ut)) 3200 continue; 3201 3202 cp = dr_get_cpu_unit(bp, ut); 3203 dr_device_transition(&cp->sbc_cm, DR_STATE_EMPTY); 3204 } 3205 3206 for (ut = 0; ut < MAX_MEM_UNITS_PER_BOARD; ut++) { 3207 if (!DEVSET_IN_SET(devs_lost, SBD_COMP_MEM, ut)) 3208 continue; 3209 3210 mp = dr_get_mem_unit(bp, ut); 3211 dr_device_transition(&mp->sbm_cm, DR_STATE_EMPTY); 3212 } 3213 3214 for (ut = 0; ut < MAX_IO_UNITS_PER_BOARD; ut++) { 3215 if (!DEVSET_IN_SET(devs_lost, SBD_COMP_IO, ut)) 3216 continue; 3217 3218 ip = dr_get_io_unit(bp, ut); 3219 dr_device_transition(&ip->sbi_cm, DR_STATE_EMPTY); 3220 } 3221 3222 DR_DEVS_DISCONNECT(bp, devs_lost); 3223 } 3224 } 3225 3226 static int 3227 dr_board_init(dr_board_t *bp, dev_info_t *dip, int bd) 3228 { 3229 sbd_error_t *err; 3230 3231 mutex_init(&bp->b_lock, NULL, MUTEX_DRIVER, NULL); 3232 mutex_init(&bp->b_slock, NULL, MUTEX_DRIVER, NULL); 3233 cv_init(&bp->b_scv, NULL, CV_DRIVER, NULL); 3234 bp->b_rstate = SBD_STAT_EMPTY; 3235 bp->b_ostate = SBD_STAT_UNCONFIGURED; 3236 bp->b_cond = SBD_COND_UNKNOWN; 3237 (void) drv_getparm(TIME, (void *)&bp->b_time); 3238 3239 (void) drmach_board_lookup(bd, &bp->b_id); 3240 bp->b_num = bd; 3241 bp->b_dip = dip; 3242 3243 bp->b_dev[NIX(SBD_COMP_CPU)] = GETSTRUCT(dr_dev_unit_t, 3244 MAX_CPU_UNITS_PER_BOARD); 3245 3246 bp->b_dev[NIX(SBD_COMP_MEM)] = GETSTRUCT(dr_dev_unit_t, 3247 MAX_MEM_UNITS_PER_BOARD); 3248 3249 bp->b_dev[NIX(SBD_COMP_IO)] = GETSTRUCT(dr_dev_unit_t, 3250 MAX_IO_UNITS_PER_BOARD); 3251 3252 /* 3253 * Initialize the devlists 3254 */ 3255 err = dr_init_devlists(bp); 3256 if (err) { 3257 sbd_err_clear(&err); 3258 dr_board_destroy(bp); 3259 return (-1); 3260 } else if (bp->b_ndev == 0) { 3261 dr_board_transition(bp, DR_STATE_EMPTY); 3262 } else { 3263 /* 3264 * Couldn't have made it down here without 3265 * having found at least one device. 3266 */ 3267 ASSERT(DR_DEVS_PRESENT(bp) != 0); 3268 /* 3269 * Check the state of any possible devices on the 3270 * board. 3271 */ 3272 dr_board_discovery(bp); 3273 3274 bp->b_assigned = 1; 3275 3276 if (DR_DEVS_UNATTACHED(bp) == 0) { 3277 /* 3278 * The board has no unattached devices, therefore 3279 * by reason of insanity it must be configured! 3280 */ 3281 dr_board_transition(bp, DR_STATE_CONFIGURED); 3282 bp->b_ostate = SBD_STAT_CONFIGURED; 3283 bp->b_rstate = SBD_STAT_CONNECTED; 3284 bp->b_cond = SBD_COND_OK; 3285 (void) drv_getparm(TIME, (void *)&bp->b_time); 3286 } else if (DR_DEVS_ATTACHED(bp)) { 3287 dr_board_transition(bp, DR_STATE_PARTIAL); 3288 bp->b_ostate = SBD_STAT_CONFIGURED; 3289 bp->b_rstate = SBD_STAT_CONNECTED; 3290 bp->b_cond = SBD_COND_OK; 3291 (void) drv_getparm(TIME, (void *)&bp->b_time); 3292 } else { 3293 dr_board_transition(bp, DR_STATE_CONNECTED); 3294 bp->b_rstate = SBD_STAT_CONNECTED; 3295 (void) drv_getparm(TIME, (void *)&bp->b_time); 3296 } 3297 } 3298 3299 return (0); 3300 } 3301 3302 static void 3303 dr_board_destroy(dr_board_t *bp) 3304 { 3305 PR_ALL("dr_board_destroy: num %d, path %s\n", 3306 bp->b_num, bp->b_path); 3307 3308 dr_board_transition(bp, DR_STATE_EMPTY); 3309 bp->b_rstate = SBD_STAT_EMPTY; 3310 (void) drv_getparm(TIME, (void *)&bp->b_time); 3311 3312 /* 3313 * Free up MEM unit structs. 3314 */ 3315 FREESTRUCT(bp->b_dev[NIX(SBD_COMP_MEM)], 3316 dr_dev_unit_t, MAX_MEM_UNITS_PER_BOARD); 3317 bp->b_dev[NIX(SBD_COMP_MEM)] = NULL; 3318 /* 3319 * Free up CPU unit structs. 3320 */ 3321 FREESTRUCT(bp->b_dev[NIX(SBD_COMP_CPU)], 3322 dr_dev_unit_t, MAX_CPU_UNITS_PER_BOARD); 3323 bp->b_dev[NIX(SBD_COMP_CPU)] = NULL; 3324 /* 3325 * Free up IO unit structs. 3326 */ 3327 FREESTRUCT(bp->b_dev[NIX(SBD_COMP_IO)], 3328 dr_dev_unit_t, MAX_IO_UNITS_PER_BOARD); 3329 bp->b_dev[NIX(SBD_COMP_IO)] = NULL; 3330 3331 mutex_destroy(&bp->b_lock); 3332 mutex_destroy(&bp->b_slock); 3333 cv_destroy(&bp->b_scv); 3334 } 3335 3336 void 3337 dr_lock_status(dr_board_t *bp) 3338 { 3339 mutex_enter(&bp->b_slock); 3340 while (bp->b_sflags & DR_BSLOCK) 3341 cv_wait(&bp->b_scv, &bp->b_slock); 3342 bp->b_sflags |= DR_BSLOCK; 3343 mutex_exit(&bp->b_slock); 3344 } 3345 3346 void 3347 dr_unlock_status(dr_board_t *bp) 3348 { 3349 mutex_enter(&bp->b_slock); 3350 bp->b_sflags &= ~DR_BSLOCK; 3351 cv_signal(&bp->b_scv); 3352 mutex_exit(&bp->b_slock); 3353 } 3354 3355 /* 3356 * Extract flags passed via ioctl. 3357 */ 3358 int 3359 dr_cmd_flags(dr_handle_t *hp) 3360 { 3361 return (hp->h_sbdcmd.cmd_cm.c_flags); 3362 } 3363