1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright 2007 Sun Microsystems, Inc. All rights reserved. 23 * Use is subject to license terms. 24 */ 25 26 #pragma ident "%Z%%M% %I% %E% SMI" 27 28 /* 29 * memory support routines for sbd. 30 */ 31 32 #include <sys/debug.h> 33 #include <sys/types.h> 34 #include <sys/errno.h> 35 #include <sys/param.h> 36 #include <sys/dditypes.h> 37 #include <sys/kmem.h> 38 #include <sys/conf.h> 39 #include <sys/ddi.h> 40 #include <sys/sunddi.h> 41 #include <sys/sunndi.h> 42 #include <sys/ddi_impldefs.h> 43 #include <sys/sysmacros.h> 44 #include <sys/machsystm.h> 45 #include <sys/spitregs.h> 46 #include <sys/cpuvar.h> 47 #include <sys/cpu_module.h> 48 #include <sys/promif.h> 49 #include <sys/memlist_impl.h> 50 #include <sys/mem_cage.h> 51 #include <sys/lgrp.h> 52 #include <sys/platform_module.h> 53 #include <vm/seg_kmem.h> 54 55 #include <sys/sbdpriv.h> 56 57 #define _ptob64(p) ((uint64_t)(p) << PAGESHIFT) 58 #define _b64top(b) ((pgcnt_t)((b) >> PAGESHIFT)) 59 60 static int sbd_post_detach_mem_unit(sbd_mem_unit_t *mp, 61 sbderror_t *ep); 62 static int sbd_reserve_mem_spans(memhandle_t *mhp, 63 struct memlist *mlist); 64 static int sbd_check_boundaries(struct memlist *orig_memlist, 65 sbd_mem_unit_t *s_mp, 66 sbd_mem_unit_t *t_mp); 67 static int sbd_select_mem_target(sbd_handle_t *hp, 68 sbd_mem_unit_t *mp, struct memlist *ml); 69 static void sbd_init_mem_unit_data(sbd_mem_unit_t *mp, sbderror_t 70 *ep); 71 static int memlist_canfit(struct memlist *s_mlist, 72 struct memlist *t_mlist); 73 static void sbd_mem_cleanup(sbd_mem_unit_t *s_mp, 74 sbd_mem_unit_t *t_mp, sbderror_t *ep); 75 static void sbd_flush_ecache(uint64_t a, uint64_t b); 76 77 struct memlist * 78 sbd_get_memlist(sbd_mem_unit_t *mp, sbderror_t *ep) 79 { 80 struct memlist *mlist; 81 static fn_t f = "sbd_get_memlist"; 82 sbd_board_t *sbp = (sbd_board_t *)mp->sbm_cm.sbdev_sbp; 83 sbdp_handle_t *hdp; 84 sbd_handle_t *hp = MACHBD2HD(sbp); 85 86 PR_MEM("%s...\n", f); 87 88 /* 89 * Return cached memlist, if present. 90 * This memlist will be present following an 91 * unconfigure (a.k.a: detach) of this memunit. 92 * It should only be used in the case were a configure 93 * is bringing this memunit back in without going 94 * through the disconnect and connect states. 95 */ 96 if (mp->sbm_mlist) { 97 PR_MEM("%s: found cached memlist\n", f); 98 99 mlist = memlist_dup(mp->sbm_mlist); 100 } else { 101 /* attempt to construct a memlist using phys_install */ 102 103 /* 104 * NOTE: this code block assumes only one memunit per 105 * board. This is currently safe because the function 106 * sbd_init_mem_devlist() forces this assumption to be 107 * valid. 108 */ 109 110 /* round down to slice base address */ 111 /* build mlist from the lower layer */ 112 hdp = sbd_get_sbdp_handle(sbp, hp); 113 mlist = sbdp_get_memlist(hdp, mp->sbm_cm.sbdev_dip); 114 if (mlist == NULL) { 115 SBD_GET_PERR(hdp->h_err, ep); 116 PR_MEM("sbd:%s: failed to get memlist for " 117 "dip (0x%p) ecode %d errno %d", f, 118 (void *)mp->sbm_cm.sbdev_dip, 119 ep->e_code, ep->e_errno); 120 sbd_release_sbdp_handle(hdp); 121 return (NULL); 122 } 123 sbd_release_sbdp_handle(hdp); 124 } 125 126 PR_MEM("%s: memlist for mem-unit (%d.%d), dip 0x%p:\n", 127 f, sbp->sb_num, 128 mp->sbm_cm.sbdev_unum, 129 (void *)mp->sbm_cm.sbdev_dip); 130 SBD_MEMLIST_DUMP(mlist); 131 132 return (mlist); 133 } 134 135 int 136 sbd_pre_attach_mem(sbd_handle_t *hp, sbd_devlist_t devlist[], int devnum) 137 { 138 int err_flag = 0; 139 sbderror_t *ep = SBD_HD2ERR(hp); 140 sbd_board_t *sbp = SBDH2BD(hp->h_sbd); 141 int d, i; 142 sbdp_handle_t *hdp; 143 static fn_t f = "sbd_pre_attach_mem"; 144 145 PR_MEM("%s...\n", f); 146 147 SBD_SET_ERR(ep, 0); 148 hdp = sbd_get_sbdp_handle(sbp, hp); 149 150 for (d = 0; d < devnum; d++) { 151 sbd_mem_unit_t *mp; 152 int unit; 153 dev_info_t *dip; 154 sbd_istate_t state; 155 int rv; 156 157 /* sbd_get_devlist will not devlist element w/ dip of 0 */ 158 ASSERT(devlist[d].dv_dip != NULL); 159 160 dip = devlist[d].dv_dip; 161 unit = sbdp_get_unit_num(hdp, dip); 162 if (unit == -1) { 163 if (hp->h_flags & SBD_IOCTL_FLAG_FORCE) 164 continue; 165 else { 166 SBD_GET_PERR(hdp->h_err, ep); 167 err_flag = 1; 168 break; 169 } 170 } 171 172 mp = SBD_GET_BOARD_MEMUNIT(sbp, unit); 173 174 ASSERT(mp->sbm_cm.sbdev_sbp == sbp); 175 ASSERT(unit == mp->sbm_cm.sbdev_unum); 176 177 PR_MEM("sbd: OS attach mem-unit (%d.%d)\n", 178 sbp->sb_num, 179 mp->sbm_cm.sbdev_unum); 180 181 state = mp->sbm_cm.sbdev_state; 182 switch (state) { 183 case SBD_STATE_UNCONFIGURED: 184 /* use memlist cached by sbd_post_detach_mem_unit */ 185 if (mp->sbm_mlist != NULL) { 186 PR_MEM("%s: recovering from UNCONFIG" 187 " mem-unit (%d.%d)\n", 188 f, sbp->sb_num, 189 mp->sbm_cm.sbdev_unum); 190 191 PR_MEM("%s: re-configure cached memlist:\n", f); 192 SBD_MEMLIST_DUMP(mp->sbm_mlist); 193 194 /* 195 * kphysm del handle should have been freed 196 */ 197 ASSERT((mp->sbm_flags & SBD_MFLAG_RELOWNER) 198 == 0); 199 } else { 200 if (hp->h_flags & SBD_IOCTL_FLAG_FORCE) 201 continue; 202 else { 203 SBD_GET_PERR(hdp->h_err, ep); 204 err_flag = 1; 205 PR_MEM("%s: mem-unit (%d.%d)" 206 " unusable\n", 207 f, sbp->sb_num, 208 mp->sbm_cm.sbdev_unum); 209 break; 210 } 211 } 212 213 /*FALLTHROUGH*/ 214 215 case SBD_STATE_CONNECTED: 216 PR_MEM("%s: reprogramming mem hardware (board %d)\n", 217 f, sbp->sb_num); 218 219 for (i = 0; i < SBD_NUM_MC_PER_BOARD; i++) { 220 if (mp->sbm_dip[i] == NULL) 221 continue; 222 dip = mp->sbm_dip[i]; 223 224 PR_MEM("%s: enabling mc 0x%p on board %d\n", 225 f, (void *)dip, sbp->sb_num); 226 227 rv = sbdphw_enable_memctrl(hdp, dip); 228 if (rv < 0) { 229 SBD_GET_PERR(hdp->h_err, ep); 230 cmn_err(CE_WARN, 231 "%s: failed to program mem ctrlr %p on " 232 "board %d", f, (void *)mp->sbm_dip[i], 233 sbp->sb_num); 234 err_flag = 1; 235 } 236 } 237 break; 238 239 default: 240 cmn_err(CE_WARN, 241 "%s: unexpected state (%d) for mem-unit " 242 "(%d.%d)", f, state, sbp->sb_num, 243 mp->sbm_cm.sbdev_unum); 244 if (SBD_GET_ERR(ep) == 0) { 245 SBD_SET_ERR(ep, ESBD_STATE); 246 err_flag = 1; 247 } 248 break; 249 } 250 251 /* exit for loop if error encountered */ 252 if (err_flag) { 253 SBD_SET_ERRSTR(ep, 254 sbp->sb_mempath[mp->sbm_cm.sbdev_unum]); 255 break; 256 } 257 } 258 sbd_release_sbdp_handle(hdp); 259 260 return (err_flag ? -1 : 0); 261 } 262 263 int 264 sbd_post_attach_mem(sbd_handle_t *hp, sbd_devlist_t devlist[], int devnum) 265 { 266 int d; 267 sbdp_handle_t *hdp; 268 sbd_board_t *sbp = SBDH2BD(hp->h_sbd); 269 sbderror_t *ep = SBD_HD2ERR(hp); 270 static fn_t f = "sbd_post_attach_mem"; 271 272 PR_MEM("%s...\n", f); 273 hdp = sbd_get_sbdp_handle(sbp, hp); 274 275 for (d = 0; d < devnum; d++) { 276 sbd_mem_unit_t *mp; 277 dev_info_t *dip; 278 int unit; 279 struct memlist *mlist, *ml; 280 281 /* sbd_get_devlist will not devlist element w/ dip of 0 */ 282 ASSERT(devlist[d].dv_dip != NULL); 283 284 dip = devlist[d].dv_dip; 285 unit = sbdp_get_unit_num(hdp, dip); 286 if (unit == -1) { 287 if (hp->h_flags & SBD_IOCTL_FLAG_FORCE) 288 continue; 289 else { 290 SBD_GET_PERR(hdp->h_err, ep); 291 break; 292 } 293 } 294 295 mp = SBD_GET_BOARD_MEMUNIT(sbp, unit); 296 297 mlist = sbd_get_memlist(mp, ep); 298 if (mlist == NULL) { 299 cmn_err(CE_WARN, 300 "%s: no memlist for mem-unit (%d.%d)", 301 f, 302 sbp->sb_num, 303 mp->sbm_cm.sbdev_unum); 304 305 if (SBD_GET_ERR(ep) == 0) { 306 SBD_SET_ERR(ep, ESBD_MEMFAIL); 307 SBD_SET_ERRSTR(ep, sbp->sb_mempath[unit]); 308 } 309 310 continue; 311 } 312 313 /* 314 * Verify the memory really did successfully attach 315 * by checking for its existence in phys_install. 316 */ 317 318 memlist_read_lock(); 319 if (memlist_intersect(phys_install, mlist) == 0) { 320 memlist_read_unlock(); 321 322 cmn_err(CE_WARN, 323 "%s: mem-unit (%d.%d) memlist not in" 324 " phys_install", f, sbp->sb_num, 325 mp->sbm_cm.sbdev_unum); 326 327 if (SBD_GET_ERR(ep) == 0) { 328 SBD_SET_ERR(ep, ESBD_INTERNAL); 329 SBD_SET_ERRSTR(ep, sbp->sb_mempath[unit]); 330 } 331 332 memlist_delete(mlist); 333 continue; 334 } 335 memlist_read_unlock(); 336 337 for (ml = mlist; ml != NULL; ml = ml->next) { 338 sbdp_mem_add_span(hdp, ml->address, ml->size); 339 } 340 341 memlist_delete(mlist); 342 343 /* 344 * Destroy cached memlist, if any. 345 * There will be a cached memlist in sbm_mlist if 346 * this board is being configured directly after 347 * an unconfigure. 348 * To support this transition, sbd_post_detach_mem 349 * left a copy of the last known memlist in sbm_mlist. 350 * This memlist could differ from any derived from 351 * hardware if while this memunit was last configured 352 * the system detected and deleted bad pages from 353 * phys_install. The location of those bad pages 354 * will be reflected in the cached memlist. 355 */ 356 if (mp->sbm_mlist) { 357 memlist_delete(mp->sbm_mlist); 358 mp->sbm_mlist = NULL; 359 } 360 sbd_init_mem_unit_data(mp, ep); 361 } 362 363 sbd_release_sbdp_handle(hdp); 364 return (0); 365 } 366 367 int 368 sbd_pre_detach_mem(sbd_handle_t *hp, sbd_devlist_t devlist[], int devnum) 369 { 370 int d; 371 int unit; 372 sbdp_handle_t *hdp; 373 sbderror_t *ep = SBD_HD2ERR(hp); 374 sbd_board_t *sbp = SBDH2BD(hp->h_sbd); 375 dev_info_t *dip; 376 377 hdp = sbd_get_sbdp_handle(sbp, hp); 378 379 for (d = 0; d < devnum; d++) { 380 sbd_mem_unit_t *mp; 381 382 /* sbd_get_devlist will not devlist element w/ dip of 0 */ 383 ASSERT(devlist[d].dv_dip != NULL); 384 385 dip = devlist[d].dv_dip; 386 unit = sbdp_get_unit_num(hdp, dip); 387 if (unit == -1) { 388 if (hp->h_flags & SBD_IOCTL_FLAG_FORCE) 389 continue; 390 else { 391 SBD_GET_PERR(hdp->h_err, ep); 392 sbd_release_sbdp_handle(hdp); 393 return (-1); 394 } 395 } 396 397 mp = SBD_GET_BOARD_MEMUNIT(sbp, unit); 398 399 /* sanity check */ 400 ASSERT(mp->sbm_cm.sbdev_sbp == sbp); 401 ASSERT(unit == mp->sbm_cm.sbdev_unum); 402 403 PR_MEM("sbd: OS detach mem-unit (%d.%d)\n", 404 sbp->sb_num, mp->sbm_cm.sbdev_unum); 405 } 406 407 sbd_release_sbdp_handle(hdp); 408 return (0); 409 } 410 411 int 412 sbd_post_detach_mem(sbd_handle_t *hp, sbd_devlist_t devlist[], int devnum) 413 { 414 int d, rv; 415 sbdp_handle_t *hdp; 416 sbd_board_t *sbp; 417 sbd_mem_unit_t *s_mp, *t_mp; 418 static fn_t f = "sbd_post_detach_mem"; 419 420 PR_MEM("%s...\n", f); 421 422 sbp = SBDH2BD(hp->h_sbd); 423 424 hdp = sbd_get_sbdp_handle(sbp, hp); 425 426 427 rv = 0; 428 for (d = 0; d < devnum; d++) { 429 sbderror_t *ep; 430 dev_info_t *dip; 431 int unit; 432 433 /* sbd_get_devlist will not devlist element w/ dip of 0 */ 434 ASSERT(devlist[d].dv_dip != NULL); 435 436 ep = &devlist[d].dv_error; 437 if ((SBD_GET_ERR(SBD_HD2ERR(hp)) != 0) || 438 (sbd_set_err_in_hdl(hp, ep) == 0)) { 439 rv = -1; 440 } 441 442 dip = devlist[d].dv_dip; 443 unit = sbdp_get_unit_num(hdp, dip); 444 if (unit == -1) { 445 if (hp->h_flags & SBD_IOCTL_FLAG_FORCE) 446 continue; 447 else { 448 if (rv != -1) 449 SBD_GET_PERR(hdp->h_err, ep); 450 break; 451 } 452 } 453 454 s_mp = SBD_GET_BOARD_MEMUNIT(sbp, unit); 455 456 ASSERT(s_mp->sbm_cm.sbdev_sbp == sbp); 457 458 if (rv == -1) { 459 if (s_mp->sbm_flags & SBD_MFLAG_SOURCE) { 460 t_mp = s_mp->sbm_peer; 461 } else { 462 /* this is no target unit */ 463 t_mp = NULL; 464 } 465 466 sbd_mem_cleanup(s_mp, t_mp, ep); 467 } else if (sbd_post_detach_mem_unit(s_mp, ep)) 468 rv = -1; 469 } 470 471 sbd_release_sbdp_handle(hdp); 472 return (rv); 473 } 474 475 static void 476 sbd_add_memory_spans(sbd_board_t *sbp, struct memlist *ml) 477 { 478 sbdp_handle_t *hdp; 479 static fn_t f = "sbd_add_memory_spans"; 480 481 PR_MEM("%s...", f); 482 SBD_MEMLIST_DUMP(ml); 483 484 #ifdef DEBUG 485 memlist_read_lock(); 486 if (memlist_intersect(phys_install, ml)) { 487 PR_MEM("%s:WARNING: memlist intersects with phys_install\n", f); 488 } 489 memlist_read_unlock(); 490 #endif 491 hdp = sbd_get_sbdp_handle(NULL, NULL); 492 493 for (; ml; ml = ml->next) { 494 update_membounds_t umb; 495 pfn_t base; 496 pgcnt_t npgs; 497 int rv; 498 499 base = _b64top(ml->address); 500 npgs = _b64top(ml->size); 501 502 umb.u_board = sbp->sb_num; 503 umb.u_base = (uint64_t)base << MMU_PAGESHIFT; 504 umb.u_len = (uint64_t)npgs << MMU_PAGESHIFT; 505 506 lgrp_plat_config(LGRP_CONFIG_MEM_ADD, (uintptr_t)&umb); 507 rv = kphysm_add_memory_dynamic(base, npgs); 508 509 (void) sbdp_mem_add_span(hdp, ml->address, ml->size); 510 511 if (rv != KPHYSM_OK) { 512 cmn_err(CE_WARN, "sbd:%s:" 513 " unexpected kphysm_add_memory_dynamic" 514 " return value %d;" 515 " basepfn=0x%lx, npages=%ld\n", 516 f, rv, base, npgs); 517 518 continue; 519 } 520 rv = kcage_range_add(base, npgs, KCAGE_DOWN); 521 if (rv != 0) 522 continue; 523 } 524 sbd_release_sbdp_handle(hdp); 525 } 526 527 /* hack for test scripts. *** remove before code finalized *** */ 528 int sbd_last_target; 529 530 static int 531 sbd_post_detach_mem_unit(sbd_mem_unit_t *s_mp, sbderror_t *ep) 532 { 533 uint64_t sz; 534 uint64_t sm; 535 uint64_t t_basepa; 536 uint64_t tmp_basepa; 537 uint64_t s_basepa; 538 sbd_board_t *sbp; 539 sbdp_handle_t *hdp; 540 uint64_t s_nbytes; 541 uint64_t s_new_basepa; 542 sbd_mem_unit_t *t_mp, *x_mp; 543 struct memlist *ml; 544 int rv; 545 static fn_t f = "sbd_post_detach_mem_unit"; 546 sbd_handle_t *hp; 547 548 PR_MEM("%s...\n", f); 549 550 sbp = (sbd_board_t *)s_mp->sbm_cm.sbdev_sbp; 551 hp = MACHBD2HD(sbp); 552 hdp = sbd_get_sbdp_handle(sbp, hp); 553 554 if (sbdp_get_mem_alignment(hdp, s_mp->sbm_cm.sbdev_dip, &sz)) { 555 cmn_err(CE_WARN, 556 "sbd:%s: no alignment for mem-unit (%d.%d)", 557 f, sbp->sb_num, s_mp->sbm_cm.sbdev_unum); 558 SBD_GET_PERR(hdp->h_err, ep); 559 sbd_release_sbdp_handle(hdp); 560 return (-1); 561 } 562 sm = sz - 1; 563 564 /* s_mp->sbm_del_mlist could be NULL, meaning no deleted spans */ 565 PR_MEM("%s: brd %d: deleted memlist (EMPTY maybe okay):\n", 566 f, sbp->sb_num); 567 SBD_MEMLIST_DUMP(s_mp->sbm_del_mlist); 568 569 /* sanity check */ 570 ASSERT(s_mp->sbm_del_mlist == NULL || 571 (s_mp->sbm_flags & SBD_MFLAG_RELDONE) != 0); 572 573 if (s_mp->sbm_flags & SBD_MFLAG_SOURCE) { 574 t_mp = s_mp->sbm_peer; 575 576 ASSERT(t_mp != NULL); 577 ASSERT(t_mp->sbm_flags & SBD_MFLAG_TARGET); 578 ASSERT(t_mp->sbm_peer == s_mp); 579 580 ASSERT(t_mp->sbm_flags & SBD_MFLAG_RELDONE); 581 ASSERT(t_mp->sbm_del_mlist); 582 583 sbp = (sbd_board_t *)t_mp->sbm_cm.sbdev_sbp; 584 PR_MEM("%s: target brd %d: deleted memlist:\n", 585 f, sbp->sb_num); 586 SBD_MEMLIST_DUMP(t_mp->sbm_del_mlist); 587 } else { 588 /* this is no target unit */ 589 t_mp = NULL; 590 } 591 592 /* 593 * Verify the memory really did successfully detach 594 * by checking for its non-existence in phys_install. 595 */ 596 rv = 0; 597 memlist_read_lock(); 598 if (s_mp->sbm_flags & SBD_MFLAG_RELDONE) { 599 x_mp = s_mp; 600 rv = memlist_intersect(phys_install, x_mp->sbm_del_mlist); 601 } 602 if (rv == 0 && t_mp && (t_mp->sbm_flags & SBD_MFLAG_RELDONE)) { 603 x_mp = t_mp; 604 rv = memlist_intersect(phys_install, x_mp->sbm_del_mlist); 605 } 606 memlist_read_unlock(); 607 608 if (rv) { 609 sbp = (sbd_board_t *)x_mp->sbm_cm.sbdev_sbp; 610 611 cmn_err(CE_WARN, 612 "%s: %smem-unit (%d.%d) memlist still in phys_install", 613 f, 614 x_mp == t_mp ? "target " : "", 615 sbp->sb_num, 616 x_mp->sbm_cm.sbdev_unum); 617 SBD_SET_ERR(ep, ESBD_INTERNAL); 618 SBD_SET_ERRSTR(ep, sbp->sb_mempath[x_mp->sbm_cm.sbdev_unum]); 619 sbd_release_sbdp_handle(hdp); 620 return (-1); 621 } 622 623 s_basepa = _ptob64(s_mp->sbm_basepfn); 624 s_nbytes = _ptob64(s_mp->sbm_npages); 625 626 if (t_mp != NULL) { 627 t_basepa = _ptob64(t_mp->sbm_basepfn); 628 s_new_basepa = (s_basepa & ~ sm) + 629 _ptob64(t_mp->sbm_slice_offset); 630 631 /* 632 * We had to swap mem-units, so update 633 * memlists accordingly with new base 634 * addresses. 635 */ 636 for (ml = t_mp->sbm_mlist; ml; ml = ml->next) { 637 ml->address -= t_basepa; 638 ml->address += s_new_basepa; 639 } 640 641 /* 642 * There is no need to explicitly rename the target delete 643 * memlist, because sbm_del_mlist and sbm_mlist always 644 * point to the same memlist for a copy/rename operation. 645 */ 646 ASSERT(t_mp->sbm_del_mlist == t_mp->sbm_mlist); 647 648 PR_MEM("%s: renamed target memlist and delete memlist", f); 649 SBD_MEMLIST_DUMP(t_mp->sbm_mlist); 650 651 for (ml = s_mp->sbm_mlist; ml; ml = ml->next) { 652 ml->address -= s_basepa; 653 ml->address += t_basepa; 654 } 655 656 PR_MEM("%s: renamed source memlist", f); 657 SBD_MEMLIST_DUMP(s_mp->sbm_mlist); 658 659 #ifdef DEBUG 660 ASSERT(s_mp->sbm_mlist != s_mp->sbm_del_mlist); 661 /* 662 * Renaming s_mp->sbm_del_mlist is not necessary. This 663 * list is not used beyond this point, and in fact, is 664 * disposed of at the end of this function. 665 */ 666 for (ml = s_mp->sbm_del_mlist; ml; ml = ml->next) { 667 ml->address -= s_basepa; 668 ml->address += t_basepa; 669 } 670 671 PR_MEM("%s: renamed source delete memlist", f); 672 SBD_MEMLIST_DUMP(s_mp->sbm_del_mlist); 673 #endif 674 675 if (s_mp->sbm_flags & SBD_MFLAG_MEMUPSIZE) { 676 struct memlist *nl; 677 int mlret; 678 679 /* 680 * We had to perform a copy-rename from a 681 * small memory node to a big memory node. 682 * Need to add back the remaining memory on 683 * the big board that wasn't used by that 684 * from the small board during the copy. 685 * Subtract out the portion of the target memory 686 * node that was taken over by the source memory 687 * node. 688 */ 689 nl = memlist_dup(t_mp->sbm_mlist); 690 mlret = memlist_delete_span(s_basepa, s_nbytes, &nl); 691 PR_MEM("%s: mlret = %d\n", f, mlret); 692 693 sbp = (sbd_board_t *)t_mp->sbm_cm.sbdev_sbp; 694 PR_MEM("%s: adding back remaining portion" 695 " of mem-unit (%d.%d), memlist:\n", 696 f, sbp->sb_num, 697 t_mp->sbm_cm.sbdev_unum); 698 699 SBD_MEMLIST_DUMP(nl); 700 701 sbd_add_memory_spans(sbp, nl); 702 703 memlist_delete(nl); 704 } 705 } 706 707 708 if (t_mp != NULL) { 709 sbp = (sbd_board_t *)t_mp->sbm_cm.sbdev_sbp; 710 hdp->h_board = sbp->sb_num; 711 /* delete target's entire address space */ 712 tmp_basepa = t_basepa & ~ sm; 713 rv = sbdp_mem_del_span(hdp, tmp_basepa, sz); 714 ASSERT(rv == 0); 715 716 sbp = (sbd_board_t *)s_mp->sbm_cm.sbdev_sbp; 717 hdp->h_board = sbp->sb_num; 718 tmp_basepa = s_basepa & ~ sm; 719 sz = s_new_basepa & sm; 720 /* delete source board's vacant address space */ 721 rv = sbdp_mem_del_span(hdp, tmp_basepa, sz); 722 ASSERT(rv == 0); 723 } else { 724 sbp = (sbd_board_t *)s_mp->sbm_cm.sbdev_sbp; 725 hdp->h_board = sbp->sb_num; 726 tmp_basepa = s_basepa & ~ sm; 727 /* delete board's entire address space */ 728 rv = sbdp_mem_del_span(hdp, tmp_basepa, sz); 729 ASSERT(rv == 0); 730 } 731 732 #ifdef LINT 733 rv = rv; 734 #endif 735 736 sbd_mem_cleanup(s_mp, t_mp, ep); 737 738 sbp = (sbd_board_t *)s_mp->sbm_cm.sbdev_sbp; 739 PR_MEM("%s: board %d's memlist:", f, sbp->sb_num); 740 SBD_MEMLIST_DUMP(s_mp->sbm_mlist); 741 742 sbd_release_sbdp_handle(hdp); 743 return (0); 744 } 745 746 static void 747 sbd_mem_cleanup(sbd_mem_unit_t *s_mp, sbd_mem_unit_t *t_mp, sbderror_t *ep) 748 { 749 sbd_board_t *sbp; 750 751 /* clean up target mem unit */ 752 if (t_mp != NULL) { 753 sbp = (sbd_board_t *)t_mp->sbm_cm.sbdev_sbp; 754 755 ASSERT(t_mp->sbm_del_mlist == t_mp->sbm_mlist); 756 /* 757 * sbm_del_mlist and sbm_mlist point at the same list 758 * We only need to delete one and then set both pointers 759 * to NULL 760 */ 761 memlist_delete(t_mp->sbm_del_mlist); 762 763 t_mp->sbm_del_mlist = NULL; 764 t_mp->sbm_mlist = NULL; 765 t_mp->sbm_peer = NULL; 766 t_mp->sbm_flags = 0; 767 t_mp->sbm_cm.sbdev_busy = 0; 768 sbd_init_mem_unit_data(t_mp, ep); 769 770 /* 771 * now that copy/rename has completed, undo this 772 * work that was done in sbd_release_mem_done. 773 */ 774 /* 775 * If error don't set the target to configured 776 */ 777 if (SBD_GET_ERR(ep) == 0) { 778 SBD_DEV_CLR_UNREFERENCED(sbp, SBD_COMP_MEM, 0); 779 SBD_DEV_CLR_RELEASED(sbp, SBD_COMP_MEM, 0); 780 SBD_DEVICE_TRANSITION(sbp, SBD_COMP_MEM, 0, 781 SBD_STATE_CONFIGURED); 782 } 783 784 /* hack for test scripts. *** remove before code finalized *** */ 785 sbd_last_target = sbp->sb_num; 786 } 787 788 /* 789 * clean up (source) board's mem unit structure. 790 * NOTE: sbm_mlist is retained. It is referred to as the 791 * cached memlist. The cached memlist is used to re-attach 792 * (configure back in) this memunit from the unconfigured 793 * state. 794 */ 795 if (s_mp != NULL) { 796 sbp = (sbd_board_t *)s_mp->sbm_cm.sbdev_sbp; 797 798 /* 799 * Don't want to call memlist_delete for sbm_del_mlist, 800 * since that list points to the sbm_list 801 */ 802 s_mp->sbm_del_mlist = NULL; 803 s_mp->sbm_peer = NULL; 804 s_mp->sbm_flags = 0; 805 s_mp->sbm_cm.sbdev_busy = 0; 806 sbd_init_mem_unit_data(s_mp, ep); 807 } 808 } 809 810 /* 811 * Successful return from this function will have the memory 812 * handle in sbp->sb_dev[..mem-unit...].sbm_memhandle allocated 813 * and waiting. This routine's job is to select the memory that 814 * actually has to be released (detached) which may not necessarily 815 * be the same memory node that came in in devlist[], 816 * i.e. a copy-rename is needed. 817 */ 818 int 819 sbd_pre_release_mem(sbd_handle_t *hp, sbd_devlist_t devlist[], int devnum) 820 { 821 extern int kcage_on; 822 int d; 823 int err_flag = 0; 824 sbd_board_t *sbp = SBDH2BD(hp->h_sbd); 825 sbderror_t *ep = SBD_HD2ERR(hp); 826 sbderror_t *lep; 827 static fn_t f = "sbd_pre_release_mem"; 828 829 PR_MEM("%s...\n", f); 830 831 if (kcage_on == 0) { 832 /* 833 * Can't Detach memory if Cage is OFF. 834 */ 835 cmn_err(CE_WARN, "%s: kernel cage is disabled", f); 836 SBD_SET_ERR(ep, ESBD_KCAGE_OFF); 837 return (-1); 838 } 839 840 for (d = 0; d < devnum; d++) { 841 int rv; 842 memquery_t mq; 843 sbd_mem_unit_t *mp; 844 struct memlist *ml; 845 846 /* sbd_get_devlist will not devlist element w/ dip of 0 */ 847 ASSERT(devlist[d].dv_dip != NULL); 848 849 mp = SBD_GET_BOARD_MEMUNIT(sbp, d); 850 851 /* 852 * If all the mem unit is marked as failed then don't allow the 853 * operation 854 */ 855 if (mp->sbm_cm.sbdev_cond == SBD_COND_FAILED) { 856 SBD_SET_ERR(ep, ESBD_STATE); 857 SBD_SET_ERRSTR(ep, sbp->sb_mempath[d]); 858 err_flag = -1; 859 break; 860 } 861 862 ASSERT(d == mp->sbm_cm.sbdev_unum); 863 864 /* 865 * if interleave is set to across boards fail the op 866 */ 867 if (mp->sbm_interleave) { 868 SBD_SET_ERR(ep, ESBD_MEMINTLV); 869 SBD_SET_ERRSTR(ep, sbp->sb_mempath[d]); 870 err_flag = -1; 871 break; 872 } 873 874 lep = &devlist[d].dv_error; 875 if (SBD_GET_ERR(lep) != 0) { 876 err_flag = -1; 877 (void) sbd_set_err_in_hdl(hp, lep); 878 break; 879 } 880 881 if (mp->sbm_flags & SBD_MFLAG_RESERVED) { 882 /* 883 * Board is currently involved in a delete 884 * memory operation. Can't detach this guy until 885 * that operation completes. 886 */ 887 cmn_err(CE_WARN, 888 "%s: ineligible mem-unit (%d.%d) for detach", 889 f, sbp->sb_num, 890 mp->sbm_cm.sbdev_unum); 891 892 SBD_SET_ERR(lep, ESBD_INVAL); 893 SBD_SET_ERRSTR(lep, sbp->sb_mempath[d]); 894 (void) sbd_set_err_in_hdl(hp, lep); 895 err_flag = -1; 896 break; 897 } 898 899 /* 900 * Check whether the detaching memory requires a 901 * copy-rename. 902 */ 903 ASSERT(mp->sbm_npages != 0); 904 rv = kphysm_del_span_query( 905 mp->sbm_basepfn, mp->sbm_npages, &mq); 906 if (rv != KPHYSM_OK) { 907 cmn_err(CE_WARN, 908 "%s: unexpected kphysm_del_span_query" 909 " return value %d;" 910 " basepfn 0x%lx, npages 0x%lx," 911 " mem-unit (%d.%d), dip 0x%p", 912 f, 913 rv, 914 mp->sbm_basepfn, 915 mp->sbm_npages, 916 sbp->sb_num, 917 mp->sbm_cm.sbdev_unum, 918 (void *)mp->sbm_cm.sbdev_dip); 919 920 SBD_SET_ERR(lep, ESBD_INTERNAL); 921 SBD_SET_ERRSTR(lep, sbp->sb_mempath[d]); 922 (void) sbd_set_err_in_hdl(hp, lep); 923 err_flag = -1; 924 break; 925 } 926 927 if (mq.nonrelocatable != 0) { 928 if (!(hp->h_iap->i_flags & SBD_FLAG_QUIESCE_OKAY)) { 929 /* caller wasn't prompted for a suspend */ 930 SBD_SET_ERR(lep, ESBD_QUIESCE_REQD); 931 SBD_SET_ERRSTR(lep, sbp->sb_mempath[d]); 932 (void) sbd_set_err_in_hdl(hp, lep); 933 err_flag = 1; 934 break; 935 } 936 } 937 938 /* flags should be clean at this time */ 939 ASSERT(mp->sbm_flags == 0); 940 941 ASSERT(mp->sbm_del_mlist == NULL); /* should be null */ 942 943 if (mp->sbm_mlist != NULL) { 944 memlist_delete(mp->sbm_mlist); 945 mp->sbm_mlist = NULL; 946 } 947 948 ml = sbd_get_memlist(mp, lep); 949 (void) sbd_set_err_in_hdl(hp, lep); 950 if (ml == NULL) { 951 PR_MEM("%s: no memlist found for board %d\n", 952 f, sbp->sb_num); 953 err_flag = -1; 954 break; 955 } 956 957 /* allocate a kphysm handle */ 958 rv = kphysm_del_gethandle(&mp->sbm_memhandle); 959 if (rv != KPHYSM_OK) { 960 memlist_delete(ml); 961 962 cmn_err(CE_WARN, 963 "%s: unexpected kphysm_del_gethandle" 964 " return value %d", f, rv); 965 966 SBD_SET_ERR(lep, ESBD_INTERNAL); 967 SBD_SET_ERRSTR(lep, sbp->sb_mempath[d]); 968 (void) sbd_set_err_in_hdl(hp, lep); 969 err_flag = -1; 970 break; 971 } 972 mp->sbm_flags |= SBD_MFLAG_RELOWNER; 973 974 if ((mq.nonrelocatable != 0) || 975 sbd_reserve_mem_spans(&mp->sbm_memhandle, ml)) { 976 /* 977 * Either the detaching memory node contains 978 * non-reloc memory or we failed to reserve the 979 * detaching memory node (which did _not_ have 980 * any non-reloc memory, i.e. some non-reloc mem 981 * got onboard). 982 */ 983 984 if (sbd_select_mem_target(hp, mp, ml)) { 985 int rv; 986 987 /* 988 * We had no luck locating a target 989 * memory node to be the recipient of 990 * the non-reloc memory on the node 991 * we're trying to detach. 992 * Clean up be disposing the mem handle 993 * and the mem list. 994 */ 995 rv = kphysm_del_release(mp->sbm_memhandle); 996 if (rv != KPHYSM_OK) { 997 /* 998 * can do nothing but complain 999 * and hope helpful for debug 1000 */ 1001 cmn_err(CE_WARN, "sbd:%s: unexpected" 1002 " kphysm_del_release return" 1003 " value %d", 1004 f, rv); 1005 } 1006 mp->sbm_flags &= ~SBD_MFLAG_RELOWNER; 1007 1008 memlist_delete(ml); 1009 1010 /* make sure sbm_flags is clean */ 1011 ASSERT(mp->sbm_flags == 0); 1012 1013 cmn_err(CE_WARN, 1014 "%s: no available target for " 1015 "mem-unit (%d.%d)", 1016 f, sbp->sb_num, 1017 mp->sbm_cm.sbdev_unum); 1018 1019 SBD_SET_ERR(lep, ESBD_NO_TARGET); 1020 SBD_SET_ERRSTR(lep, 1021 sbp->sb_mempath[mp->sbm_cm.sbdev_unum]); 1022 (void) sbd_set_err_in_hdl(hp, lep); 1023 1024 err_flag = -1; 1025 break; 1026 } 1027 1028 /* 1029 * ml is not memlist_deleted here because 1030 * it has been assigned to mp->sbm_mlist 1031 * by sbd_select_mem_target. 1032 */ 1033 } else { 1034 /* no target needed to detach this board */ 1035 mp->sbm_flags |= SBD_MFLAG_RESERVED; 1036 mp->sbm_peer = NULL; 1037 mp->sbm_del_mlist = ml; 1038 mp->sbm_mlist = ml; 1039 mp->sbm_cm.sbdev_busy = 1; 1040 } 1041 #ifdef DEBUG 1042 ASSERT(mp->sbm_mlist != NULL); 1043 1044 if (mp->sbm_flags & SBD_MFLAG_SOURCE) { 1045 int src, targ; 1046 1047 sbp = (sbd_board_t *) 1048 mp->sbm_peer->sbm_cm.sbdev_sbp; 1049 targ = sbp->sb_num; 1050 sbp = (sbd_board_t *)mp->sbm_cm.sbdev_sbp; 1051 src = sbp->sb_num; 1052 PR_MEM("%s: release of board %d requires copy/rename;" 1053 " selected target board %d\n", 1054 f, src, targ); 1055 } else { 1056 sbp = (sbd_board_t *)mp->sbm_cm.sbdev_sbp; 1057 PR_MEM("%s: copy/rename not required to release" 1058 " board %d\n", f, sbp->sb_num); 1059 } 1060 1061 ASSERT(mp->sbm_flags & SBD_MFLAG_RELOWNER); 1062 ASSERT(mp->sbm_flags & SBD_MFLAG_RESERVED); 1063 #endif 1064 } 1065 1066 return (err_flag); 1067 } 1068 1069 void 1070 sbd_release_mem_done(sbd_handle_t *hp, int unit) 1071 { 1072 sbd_mem_unit_t *s_mp, *t_mp, *mp; 1073 sbderror_t *ep = SBD_HD2ERR(hp); 1074 sbd_board_t *sbp = SBDH2BD(hp->h_sbd); 1075 int rv; 1076 static fn_t f = "sbd_release_mem_done"; 1077 1078 s_mp = SBD_GET_BOARD_MEMUNIT(sbp, unit); 1079 sbp = (sbd_board_t *)s_mp->sbm_cm.sbdev_sbp; 1080 1081 /* 1082 * This unit will be flagged with SBD_MFLAG_SOURCE, if it 1083 * has a target unit. 1084 */ 1085 if (s_mp->sbm_flags & SBD_MFLAG_SOURCE) { 1086 t_mp = s_mp->sbm_peer; 1087 ASSERT(t_mp != NULL); 1088 ASSERT(t_mp->sbm_peer == s_mp); 1089 ASSERT(t_mp->sbm_flags & SBD_MFLAG_TARGET); 1090 ASSERT(t_mp->sbm_flags & SBD_MFLAG_RESERVED); 1091 } else { 1092 /* this is no target unit */ 1093 t_mp = NULL; 1094 } 1095 1096 /* free delete handle */ 1097 ASSERT(s_mp->sbm_flags & SBD_MFLAG_RELOWNER); 1098 ASSERT(s_mp->sbm_flags & SBD_MFLAG_RESERVED); 1099 1100 rv = kphysm_del_release(s_mp->sbm_memhandle); 1101 if (rv != KPHYSM_OK) { 1102 /* 1103 * can do nothing but complain 1104 * and hope helpful for debug 1105 */ 1106 cmn_err(CE_WARN, "sbd:%s: unexpected kphysm_del_release" 1107 " return value %d", f, rv); 1108 } 1109 s_mp->sbm_flags &= ~SBD_MFLAG_RELOWNER; 1110 1111 /* 1112 * If an error was encountered during release, clean up 1113 * the source (and target, if present) unit data. 1114 */ 1115 if (SBD_GET_ERR(ep) != 0) { 1116 1117 PR_MEM("%s: unit %d.%d: error %d noted\n", 1118 f, sbp->sb_num, 1119 s_mp->sbm_cm.sbdev_unum, 1120 SBD_GET_ERR(ep)); 1121 1122 sbd_mem_cleanup(s_mp, t_mp, ep); 1123 1124 /* bail out */ 1125 return; 1126 } 1127 1128 SBD_DEV_SET_RELEASED(sbp, SBD_COMP_MEM, unit); 1129 SBD_DEVICE_TRANSITION(sbp, SBD_COMP_MEM, unit, SBD_STATE_RELEASE); 1130 1131 if (t_mp != NULL) { 1132 sbp = (sbd_board_t *)t_mp->sbm_cm.sbdev_sbp; 1133 /* 1134 * the kphysm delete operation that drained the source 1135 * board also drained this target board. Since the source 1136 * board drain is now known to have succeeded, we know this 1137 * target board is drained too. 1138 */ 1139 SBD_DEV_SET_RELEASED(sbp, SBD_COMP_MEM, 1140 t_mp->sbm_cm.sbdev_unum); 1141 SBD_DEVICE_TRANSITION(sbp, SBD_COMP_MEM, 1142 t_mp->sbm_cm.sbdev_unum, 1143 SBD_STATE_RELEASE); 1144 1145 /* 1146 * NOTE: do not transition target's board state, 1147 * even if the mem-unit was the last configure 1148 * unit of the board. When copy/rename completes 1149 * this mem-unit will transitioned back to 1150 * the configured state. In the meantime, the 1151 * board's must remain as is. 1152 */ 1153 } 1154 1155 /* if board(s) had deleted memory, verify it is gone */ 1156 rv = 0; 1157 memlist_read_lock(); 1158 if (s_mp->sbm_del_mlist != NULL) { 1159 sbp = (sbd_board_t *)s_mp->sbm_cm.sbdev_sbp; 1160 mp = s_mp; 1161 rv = memlist_intersect(phys_install, mp->sbm_del_mlist); 1162 } 1163 if (rv == 0 && t_mp && t_mp->sbm_del_mlist != NULL) { 1164 sbp = (sbd_board_t *)t_mp->sbm_cm.sbdev_sbp; 1165 mp = t_mp; 1166 rv = memlist_intersect(phys_install, mp->sbm_del_mlist); 1167 } 1168 memlist_read_unlock(); 1169 if (rv) { 1170 cmn_err(CE_WARN, "sbd:%s: %smem-unit (%d.%d): " 1171 "deleted memory still found in phys_install", 1172 f, 1173 (mp == t_mp ? "target " : ""), 1174 sbp->sb_num, 1175 mp->sbm_cm.sbdev_unum); 1176 1177 SBD_SET_ERR(ep, ESBD_INTERNAL); 1178 SBD_SET_ERRSTR(ep, sbp->sb_mempath[mp->sbm_cm.sbdev_unum]); 1179 return; 1180 } 1181 1182 s_mp->sbm_flags |= SBD_MFLAG_RELDONE; 1183 if (t_mp != NULL) { 1184 t_mp->sbm_flags &= ~SBD_MFLAG_RESERVED; 1185 t_mp->sbm_flags |= SBD_MFLAG_RELDONE; 1186 } 1187 1188 sbp = (sbd_board_t *)s_mp->sbm_cm.sbdev_sbp; 1189 1190 SBD_DEV_SET_UNREFERENCED(sbp, SBD_COMP_MEM, unit); 1191 SBD_DEVICE_TRANSITION(sbp, SBD_COMP_MEM, unit, SBD_STATE_UNREFERENCED); 1192 1193 PR_MEM("%s: marking mem-unit (%d.%d) release DONE\n", 1194 f, sbp->sb_num, 1195 s_mp->sbm_cm.sbdev_unum); 1196 1197 s_mp->sbm_cm.sbdev_ostate = SBD_STAT_UNCONFIGURED; 1198 1199 if (t_mp != NULL) { 1200 sbp = (sbd_board_t *)t_mp->sbm_cm.sbdev_sbp; 1201 1202 SBD_DEV_SET_UNREFERENCED(sbp, SBD_COMP_MEM, 1203 t_mp->sbm_cm.sbdev_unum); 1204 SBD_DEVICE_TRANSITION(sbp, SBD_COMP_MEM, 1205 t_mp->sbm_cm.sbdev_unum, 1206 SBD_STATE_UNREFERENCED); 1207 1208 sbp = (sbd_board_t *)s_mp->sbm_cm.sbdev_sbp; 1209 1210 PR_MEM("%s: marking mem-unit (%d.%d) release DONE\n", 1211 f, sbp->sb_num, 1212 t_mp->sbm_cm.sbdev_unum); 1213 1214 t_mp->sbm_cm.sbdev_ostate = SBD_STAT_UNCONFIGURED; 1215 } 1216 } 1217 1218 int 1219 sbd_disconnect_mem(sbd_handle_t *hp, int unit) 1220 { 1221 static fn_t f = "sbd_disconnect_mem"; 1222 sbd_mem_unit_t *mp; 1223 sbd_board_t *sbp = SBDH2BD(hp->h_sbd); 1224 1225 mp = SBD_GET_BOARD_MEMUNIT(sbp, unit); 1226 1227 ASSERT(mp->sbm_cm.sbdev_state == SBD_STATE_CONNECTED || 1228 mp->sbm_cm.sbdev_state == SBD_STATE_UNCONFIGURED); 1229 1230 PR_MEM("%s...\n", f); 1231 1232 if (mp->sbm_del_mlist && mp->sbm_del_mlist != mp->sbm_mlist) 1233 memlist_delete(mp->sbm_del_mlist); 1234 mp->sbm_del_mlist = NULL; 1235 1236 if (mp->sbm_mlist) { 1237 memlist_delete(mp->sbm_mlist); 1238 mp->sbm_mlist = NULL; 1239 } 1240 1241 return (0); 1242 } 1243 1244 int 1245 sbd_cancel_mem(sbd_handle_t *hp, int unit) 1246 { 1247 sbd_mem_unit_t *s_mp, *t_mp; 1248 sbd_istate_t state; 1249 sbd_board_t *sbp = SBDH2BD(hp->h_sbd); 1250 sbd_board_t *tsbp; 1251 static fn_t f = "sbd_cancel_mem"; 1252 sbderror_t *ep = SBD_HD2ERR(hp); 1253 1254 s_mp = SBD_GET_BOARD_MEMUNIT(sbp, unit); 1255 1256 state = s_mp->sbm_cm.sbdev_state; 1257 1258 if (s_mp->sbm_flags & SBD_MFLAG_TARGET) { 1259 /* must cancel source board, not target board */ 1260 SBD_SET_ERR(ep, ESBD_INTERNAL); 1261 SBD_SET_ERRSTR(ep, sbp->sb_mempath[unit]); 1262 return (-1); 1263 } else if (s_mp->sbm_flags & SBD_MFLAG_SOURCE) { 1264 t_mp = s_mp->sbm_peer; 1265 tsbp = t_mp->sbm_cm.sbdev_sbp; 1266 ASSERT(t_mp != NULL); 1267 ASSERT(t_mp->sbm_peer == s_mp); 1268 1269 /* must always match the source board's state */ 1270 ASSERT(t_mp->sbm_cm.sbdev_state == state); 1271 } else { 1272 /* this is no target unit */ 1273 t_mp = NULL; 1274 } 1275 1276 switch (state) { 1277 case SBD_STATE_UNREFERENCED: /* state set by sbd_release_mem_done */ 1278 ASSERT((s_mp->sbm_flags & SBD_MFLAG_RELOWNER) == 0); 1279 1280 if (t_mp != NULL && t_mp->sbm_del_mlist != NULL) { 1281 PR_MEM("%s: undoing target board %d memory delete\n", 1282 f, tsbp->sb_num); 1283 sbd_add_memory_spans(tsbp, t_mp->sbm_del_mlist); 1284 SBD_DEV_CLR_UNREFERENCED(tsbp, SBD_COMP_MEM, 1285 t_mp->sbm_cm.sbdev_unum); 1286 } 1287 1288 if (s_mp->sbm_del_mlist != NULL) { 1289 PR_MEM("%s: undoing board %d memory delete\n", 1290 f, sbp->sb_num); 1291 sbd_add_memory_spans(sbp, s_mp->sbm_del_mlist); 1292 } 1293 1294 /*FALLTHROUGH*/ 1295 1296 case SBD_STATE_CONFIGURED: 1297 /* 1298 * we got here because of an error early in the release process 1299 * Just leave the memory as is and report the error 1300 */ 1301 1302 ASSERT((s_mp->sbm_flags & SBD_MFLAG_RELOWNER) == 0); 1303 1304 if (t_mp != NULL) { 1305 ASSERT(t_mp->sbm_del_mlist == t_mp->sbm_mlist); 1306 t_mp->sbm_del_mlist = NULL; 1307 1308 if (t_mp->sbm_mlist != NULL) { 1309 memlist_delete(t_mp->sbm_mlist); 1310 t_mp->sbm_mlist = NULL; 1311 } 1312 1313 t_mp->sbm_peer = NULL; 1314 t_mp->sbm_flags = 0; 1315 t_mp->sbm_cm.sbdev_busy = 0; 1316 sbd_init_mem_unit_data(t_mp, ep); 1317 1318 SBD_DEV_CLR_RELEASED(tsbp, SBD_COMP_MEM, 1319 t_mp->sbm_cm.sbdev_unum); 1320 1321 SBD_DEVICE_TRANSITION(tsbp, SBD_COMP_MEM, 1322 t_mp->sbm_cm.sbdev_unum, 1323 SBD_STATE_CONFIGURED); 1324 } 1325 1326 if (s_mp->sbm_del_mlist != s_mp->sbm_mlist) 1327 memlist_delete(s_mp->sbm_del_mlist); 1328 s_mp->sbm_del_mlist = NULL; 1329 1330 if (s_mp->sbm_mlist != NULL) { 1331 memlist_delete(s_mp->sbm_mlist); 1332 s_mp->sbm_mlist = NULL; 1333 } 1334 1335 s_mp->sbm_peer = NULL; 1336 s_mp->sbm_flags = 0; 1337 s_mp->sbm_cm.sbdev_busy = 0; 1338 sbd_init_mem_unit_data(s_mp, ep); 1339 1340 return (0); 1341 default: 1342 PR_MEM("%s: WARNING unexpected state (%d) for " 1343 "mem-unit %d.%d\n", 1344 f, 1345 (int)state, 1346 sbp->sb_num, 1347 s_mp->sbm_cm.sbdev_unum); 1348 1349 return (-1); 1350 } 1351 /*NOTREACHED*/ 1352 } 1353 1354 void 1355 sbd_init_mem_unit(sbd_board_t *sbp, int unit, sbderror_t *ep) 1356 { 1357 sbd_istate_t new_state; 1358 sbd_mem_unit_t *mp; 1359 dev_info_t *cur_mc_dip; 1360 int failed_mcs = 0, present_mcs = 0; 1361 sbd_cond_t mc_cond; 1362 int i; 1363 1364 mp = SBD_GET_BOARD_MEMUNIT(sbp, unit); 1365 1366 if (SBD_DEV_IS_ATTACHED(sbp, SBD_COMP_MEM, unit)) { 1367 new_state = SBD_STATE_CONFIGURED; 1368 } else if (SBD_DEV_IS_PRESENT(sbp, SBD_COMP_MEM, unit)) { 1369 new_state = SBD_STATE_CONNECTED; 1370 } else if (mp->sbm_cm.sbdev_dip != NULL) { 1371 new_state = SBD_STATE_OCCUPIED; 1372 } else { 1373 new_state = SBD_STATE_EMPTY; 1374 } 1375 1376 /* 1377 * Check all the possible memory nodes on the board. If all of them 1378 * have a failed status mark memory as failed. Otherwise mem is ok 1379 */ 1380 if (!sbp->sb_memaccess_ok) { 1381 mp->sbm_cm.sbdev_cond = SBD_COND_UNKNOWN; 1382 return; 1383 } 1384 1385 for (i = 0; i < SBD_NUM_MC_PER_BOARD; i++) { 1386 cur_mc_dip = mp->sbm_dip[i]; 1387 1388 if (cur_mc_dip == NULL) 1389 continue; 1390 1391 present_mcs |= (1 << i); 1392 1393 mc_cond = sbd_get_comp_cond(cur_mc_dip); 1394 if (mc_cond == SBD_COND_FAILED) { 1395 failed_mcs |= (1 << i); 1396 } 1397 } 1398 1399 if (failed_mcs == present_mcs) { 1400 /* 1401 * All mem nodes failed, therefore mark all mem 1402 * as failed 1403 */ 1404 mp->sbm_cm.sbdev_cond = SBD_COND_FAILED; 1405 } else { 1406 mp->sbm_cm.sbdev_cond = SBD_COND_OK; 1407 } 1408 1409 sbd_init_mem_unit_data(mp, ep); 1410 1411 /* 1412 * Any changes to this memory unit should be performed above 1413 * this call to ensure the unit is fully initialized 1414 * before transitioning to the new state. 1415 */ 1416 SBD_DEVICE_TRANSITION(sbp, SBD_COMP_MEM, unit, new_state); 1417 1418 } 1419 1420 static void 1421 sbd_init_mem_unit_data(sbd_mem_unit_t *mp, sbderror_t *ep) 1422 { 1423 uint64_t basepa; 1424 uint64_t sz; 1425 sbd_board_t *sbp = mp->sbm_cm.sbdev_sbp; 1426 sbdp_handle_t *hdp; 1427 static fn_t f = "sbd_init_mem_unit_data"; 1428 sbd_handle_t *hp = MACHBD2HD(sbp); 1429 1430 PR_MEM("%s...\n", f); 1431 1432 /* a little sanity checking */ 1433 ASSERT(mp->sbm_peer == NULL); 1434 ASSERT(mp->sbm_flags == 0); 1435 1436 hdp = sbd_get_sbdp_handle(sbp, hp); 1437 1438 /* get basepfn of mem unit */ 1439 if (sbdphw_get_base_physaddr(hdp, mp->sbm_cm.sbdev_dip, &basepa)) { 1440 cmn_err(CE_WARN, "sbd:%s: failed to get physaddr" 1441 " for mem-unit (%d.%d)", 1442 f, 1443 sbp->sb_num, 1444 mp->sbm_cm.sbdev_unum); 1445 SBD_GET_PERR(hdp->h_err, ep); 1446 sbd_release_sbdp_handle(hdp); 1447 return; 1448 } 1449 mp->sbm_basepfn = _b64top(basepa); 1450 1451 /* attempt to get number of pages from PDA */ 1452 mp->sbm_npages = sbdp_get_mem_size(hdp); 1453 1454 /* if didn't work, calculate using memlist */ 1455 if (mp->sbm_npages == 0) { 1456 struct memlist *ml, *mlist; 1457 mlist = sbd_get_memlist(mp, ep); 1458 for (ml = mlist; ml; ml = ml->next) 1459 mp->sbm_npages += btop(ml->size); 1460 memlist_delete(mlist); 1461 } 1462 1463 1464 if (sbdp_get_mem_alignment(hdp, mp->sbm_cm.sbdev_dip, &sz)) { 1465 cmn_err(CE_WARN, 1466 "sbd:%s: no alignment for mem-unit (%d.%d)", 1467 f, sbp->sb_num, mp->sbm_cm.sbdev_unum); 1468 SBD_GET_PERR(hdp->h_err, ep); 1469 sbd_release_sbdp_handle(hdp); 1470 return; 1471 } 1472 mp->sbm_alignment_mask = _b64top(sz); 1473 1474 1475 mp->sbm_interleave = sbdp_isinterleaved(hdp, 1476 mp->sbm_cm.sbdev_dip); 1477 1478 PR_MEM("%s: board %d (basepfn = 0x%lx, npgs = 0x%lx interleave %d)\n", 1479 f, sbp->sb_num, 1480 mp->sbm_basepfn, 1481 mp->sbm_npages, 1482 mp->sbm_interleave); 1483 1484 sbd_release_sbdp_handle(hdp); 1485 } 1486 1487 static int 1488 sbd_reserve_mem_spans(memhandle_t *mhp, struct memlist *ml) 1489 { 1490 int err; 1491 pfn_t base; 1492 pgcnt_t npgs; 1493 struct memlist *mc; 1494 static fn_t f = "sbd_reserve_mem_spans"; 1495 1496 PR_MEM("%s...\n", f); 1497 1498 /* 1499 * Walk the supplied memlist scheduling each span for removal 1500 * with kphysm_del_span. It is possible that a span may intersect 1501 * an area occupied by the cage. 1502 */ 1503 for (mc = ml; mc != NULL; mc = mc->next) { 1504 base = _b64top(mc->address); 1505 npgs = _b64top(mc->size); 1506 1507 err = kphysm_del_span(*mhp, base, npgs); 1508 if (err != KPHYSM_OK) { 1509 cmn_err(CE_WARN, "sbd:%s memory reserve failed." 1510 " unexpected kphysm_del_span return value %d;" 1511 " basepfn=0x%lx npages=%ld", 1512 f, err, base, npgs); 1513 return (-1); 1514 } 1515 } 1516 return (0); 1517 } 1518 1519 /* debug counters */ 1520 int sbd_smt_realigned; 1521 int sbd_smt_preference[4]; 1522 1523 #ifdef DEBUG 1524 uint_t sbd_ignore_board; /* if bit[bnum-1] set, board won't be candidate */ 1525 #endif 1526 1527 /* 1528 * Verify that there is no memory overlapping if copy-rename is 1529 * done with the selected target board. 1530 * 1531 * Returns 0 if OK, -1 otherwise. 1532 */ 1533 static int 1534 sbd_check_boundaries(struct memlist *orig_memlist, sbd_mem_unit_t *s_mp, 1535 sbd_mem_unit_t *t_mp) 1536 { 1537 struct memlist *new_memlist; 1538 int mlret; 1539 static fn_t f = "sbd_check_boundaries"; 1540 1541 new_memlist = memlist_dup(orig_memlist); 1542 if (new_memlist == NULL) { 1543 PR_MEM("%s: can't dup original memlist\n", f); 1544 return (-1); 1545 } 1546 1547 mlret = memlist_delete_span( 1548 _ptob64(s_mp->sbm_basepfn), 1549 _ptob64(s_mp->sbm_npages), 1550 &new_memlist); 1551 if (mlret != MEML_SPANOP_OK) { 1552 PR_MEM("%s: del s/s mlret = %d\n", f, mlret); 1553 goto check_done; 1554 } 1555 1556 mlret = memlist_delete_span( 1557 _ptob64(t_mp->sbm_basepfn), 1558 _ptob64(t_mp->sbm_npages), 1559 &new_memlist); 1560 if (mlret != MEML_SPANOP_OK) { 1561 PR_MEM("%s: del t/t mlret = %d\n", f, mlret); 1562 goto check_done; 1563 } 1564 1565 mlret = memlist_add_span( 1566 _ptob64(t_mp->sbm_basepfn), 1567 _ptob64(s_mp->sbm_npages), 1568 &new_memlist); 1569 if (mlret != MEML_SPANOP_OK) { 1570 PR_MEM("%s: add t/s mlret = %d\n", f, mlret); 1571 goto check_done; 1572 } 1573 1574 mlret = memlist_add_span( 1575 _ptob64(s_mp->sbm_basepfn), 1576 _ptob64(t_mp->sbm_npages), 1577 &new_memlist); 1578 if (mlret != MEML_SPANOP_OK) { 1579 PR_MEM("%s: add s/t mlret = %d\n", f, mlret); 1580 } 1581 1582 check_done: 1583 memlist_delete(new_memlist); 1584 1585 if (mlret == MEML_SPANOP_OK) 1586 return (0); 1587 else 1588 return (-1); 1589 } 1590 1591 /* 1592 * Find and reserve a copy/rename target board suitable for the 1593 * given source board. 1594 * All boards in the system are examined and categorized in relation to 1595 * their memory size versus the source board's memory size. Order of 1596 * preference is: 1597 * 1st: board has same memory size 1598 * 2nd: board has larger memory size 1599 * 3rd: board has smaller memory size 1600 * 4th: board has smaller memory size, available memory will be reduced. 1601 * Boards in category 3 and 4 will have their MC's reprogrammed to locate the 1602 * span to which the MC responds to address span that appropriately covers 1603 * the nonrelocatable span of the source board. 1604 */ 1605 static int 1606 sbd_select_mem_target(sbd_handle_t *hp, 1607 sbd_mem_unit_t *s_mp, struct memlist *s_ml) 1608 { 1609 uint64_t sz; 1610 pgcnt_t sm; 1611 int n_sets = 4; /* same, larger, smaller, clipped */ 1612 int preference; /* lower value is higher preference */ 1613 int n_units_per_set; 1614 int idx; 1615 sbd_mem_unit_t **sets; 1616 sbdp_handle_t *hdp; 1617 int t_bd; 1618 sbd_softstate_t *softsp; 1619 int t_unit; 1620 int max_boards; 1621 int rv; 1622 sbd_board_t *s_sbp, *t_sbp; 1623 sbd_mem_unit_t *t_mp, *c_mp; 1624 struct memlist *d_ml, *t_ml, *x_ml; 1625 memquery_t s_mq = {0}; 1626 static fn_t f = "sbd_select_mem_target"; 1627 1628 PR_MEM("%s...\n", f); 1629 1630 ASSERT(s_ml != NULL); 1631 1632 s_sbp = s_mp->sbm_cm.sbdev_sbp; 1633 1634 hdp = sbd_get_sbdp_handle(s_sbp, hp); 1635 1636 if (sbdp_get_mem_alignment(hdp, s_mp->sbm_cm.sbdev_dip, &sz)) { 1637 sbderror_t *ep = SBD_HD2ERR(hp); 1638 cmn_err(CE_WARN, 1639 "sbd:%s: no alignment for mem-unit (%d.%d)", 1640 f, s_sbp->sb_num, s_mp->sbm_cm.sbdev_unum); 1641 SBD_GET_PERR(hdp->h_err, ep); 1642 sbd_release_sbdp_handle(hdp); 1643 return (-1); 1644 } 1645 sm = sz - 1; 1646 sbd_release_sbdp_handle(hdp); 1647 1648 softsp = (sbd_softstate_t *)s_sbp->sb_softsp; 1649 1650 max_boards = softsp->max_boards; 1651 n_units_per_set = max_boards * MAX_MEM_UNITS_PER_BOARD; 1652 sets = GETSTRUCT(sbd_mem_unit_t *, n_units_per_set * n_sets); 1653 1654 /* 1655 * Make one pass through all memory units on all boards 1656 * and categorize them with respect to the source board. 1657 */ 1658 for (t_bd = 0; t_bd < max_boards; t_bd++) { 1659 /* 1660 * The board structs are a contiguous array 1661 * so we take advantage of that to find the 1662 * correct board struct pointer for a given 1663 * board number. 1664 */ 1665 t_sbp = (sbd_board_t *)softsp->sbd_boardlist; 1666 t_sbp += t_bd; 1667 1668 /* source board can not be its own target */ 1669 if (s_sbp->sb_num == t_sbp->sb_num) 1670 continue; 1671 1672 for (t_unit = 0; t_unit < MAX_MEM_UNITS_PER_BOARD; t_unit++) { 1673 1674 t_mp = SBD_GET_BOARD_MEMUNIT(t_sbp, t_unit); 1675 1676 /* this memory node must be attached */ 1677 if (!SBD_DEV_IS_ATTACHED(t_sbp, SBD_COMP_MEM, t_unit)) 1678 continue; 1679 1680 /* source unit can not be its own target */ 1681 if (s_mp == t_mp) { 1682 /* catch this in debug kernels */ 1683 ASSERT(0); 1684 continue; 1685 } 1686 1687 /* 1688 * this memory node must not already be reserved 1689 * by some other memory delete operation. 1690 */ 1691 if (t_mp->sbm_flags & SBD_MFLAG_RESERVED) 1692 continue; 1693 1694 /* 1695 * categorize the memory node 1696 * If this is a smaller memory node, create a 1697 * temporary, edited copy of the source board's 1698 * memlist containing only the span of the non- 1699 * relocatable pages. 1700 */ 1701 if (t_mp->sbm_npages == s_mp->sbm_npages) { 1702 preference = 0; 1703 t_mp->sbm_slice_offset = 0; 1704 } else if (t_mp->sbm_npages > s_mp->sbm_npages) { 1705 preference = 1; 1706 t_mp->sbm_slice_offset = 0; 1707 } else { 1708 /* 1709 * We do not allow other options right now 1710 */ 1711 continue; 1712 } 1713 1714 sbd_smt_preference[preference]++; 1715 1716 /* calculate index to start of preference set */ 1717 idx = n_units_per_set * preference; 1718 /* calculate offset to respective element */ 1719 idx += t_bd * MAX_MEM_UNITS_PER_BOARD + t_unit; 1720 1721 ASSERT(idx < n_units_per_set * n_sets); 1722 sets[idx] = t_mp; 1723 } 1724 } 1725 1726 /* 1727 * NOTE: this would be a good place to sort each candidate 1728 * set in to some desired order, e.g. memory size in ascending 1729 * order. Without an additional sorting step here, the order 1730 * within a set is ascending board number order. 1731 */ 1732 1733 c_mp = NULL; 1734 x_ml = NULL; 1735 t_ml = NULL; 1736 for (idx = 0; idx < n_units_per_set * n_sets; idx++) { 1737 memquery_t mq; 1738 1739 /* cleanup t_ml after previous pass */ 1740 if (t_ml != NULL) { 1741 memlist_delete(t_ml); 1742 t_ml = NULL; 1743 } 1744 1745 /* get candidate target board mem unit */ 1746 t_mp = sets[idx]; 1747 if (t_mp == NULL) 1748 continue; 1749 1750 t_sbp = t_mp->sbm_cm.sbdev_sbp; 1751 1752 /* get target board memlist */ 1753 t_ml = sbd_get_memlist(t_mp, SBD_HD2ERR(hp)); 1754 if (t_ml == NULL) { 1755 cmn_err(CE_WARN, "sbd:%s: no memlist for" 1756 " mem-unit %d, board %d", 1757 f, 1758 t_sbp->sb_num, 1759 t_mp->sbm_cm.sbdev_unum); 1760 1761 continue; 1762 } 1763 1764 /* get appropriate source board memlist */ 1765 if (t_mp->sbm_npages < s_mp->sbm_npages) { 1766 spgcnt_t excess; 1767 1768 /* 1769 * make a copy of the source board memlist 1770 * then edit it to remove the spans that 1771 * are outside the calculated span of 1772 * [pfn..s_mq.last_nonrelocatable]. 1773 */ 1774 if (x_ml != NULL) 1775 memlist_delete(x_ml); 1776 1777 x_ml = memlist_dup(s_ml); 1778 if (x_ml == NULL) { 1779 PR_MEM("%s: memlist_dup failed\n", f); 1780 /* TODO: should abort */ 1781 continue; 1782 } 1783 1784 /* trim off lower portion */ 1785 excess = t_mp->sbm_slice_offset; 1786 if (excess > 0) { 1787 int mlret; 1788 1789 mlret = memlist_delete_span( 1790 _ptob64(s_mp->sbm_basepfn), 1791 _ptob64(excess), 1792 &x_ml); 1793 PR_MEM("%s: mlret = %d\n", f, mlret); 1794 } 1795 1796 /* 1797 * Since this candidate target board is smaller 1798 * than the source board, s_mq must have been 1799 * initialized in previous loop while processing 1800 * this or some other candidate board. 1801 * FIXME: this is weak. 1802 */ 1803 ASSERT(s_mq.phys_pages != 0); 1804 1805 /* trim off upper portion */ 1806 excess = (s_mp->sbm_basepfn + s_mp->sbm_npages) 1807 - (s_mq.last_nonrelocatable + 1); 1808 if (excess > 0) { 1809 pfn_t p; 1810 int mlret; 1811 1812 p = s_mq.last_nonrelocatable + 1; 1813 p -= excess; 1814 1815 mlret = memlist_delete_span( 1816 _ptob64(p), 1817 _ptob64(excess), 1818 &x_ml); 1819 PR_MEM("%s: mlret = %d\n", f, mlret); 1820 } 1821 1822 PR_MEM("%s: brd %d: edited source memlist:\n", 1823 f, s_sbp->sb_num); 1824 SBD_MEMLIST_DUMP(x_ml); 1825 1826 #ifdef DEBUG 1827 /* sanity check memlist */ 1828 d_ml = x_ml; 1829 while (d_ml->next != NULL) 1830 d_ml = d_ml->next; 1831 ASSERT(x_ml->address == _ptob64(s_mp->sbm_basepfn) + 1832 _ptob64(t_mp->sbm_slice_offset)); 1833 ASSERT(d_ml->address + d_ml->size == 1834 _ptob64(s_mq.last_nonrelocatable + 1)); 1835 #endif 1836 1837 /* 1838 * x_ml now describes only the portion of the 1839 * source board that will be moved during the 1840 * copy/rename operation. 1841 */ 1842 d_ml = x_ml; 1843 } else { 1844 /* use original memlist; all spans will be moved */ 1845 d_ml = s_ml; 1846 } 1847 1848 /* verify target can support source memory spans. */ 1849 if (memlist_canfit(d_ml, t_ml) == 0) { 1850 PR_MEM("%s: source memlist won't" 1851 " fit in target memlist\n", f); 1852 PR_MEM("%s: source memlist:\n", f); 1853 SBD_MEMLIST_DUMP(d_ml); 1854 PR_MEM("%s: target memlist:\n", f); 1855 SBD_MEMLIST_DUMP(t_ml); 1856 1857 continue; 1858 } 1859 1860 /* NOTE: the value of d_ml is not used beyond this point */ 1861 1862 PR_MEM("%s: checking for no-reloc on board %d, " 1863 " basepfn=0x%lx, npages=%ld\n", 1864 f, 1865 t_sbp->sb_num, 1866 t_mp->sbm_basepfn, 1867 t_mp->sbm_npages); 1868 1869 rv = kphysm_del_span_query( 1870 t_mp->sbm_basepfn, t_mp->sbm_npages, &mq); 1871 if (rv != KPHYSM_OK) { 1872 PR_MEM("%s: kphysm_del_span_query:" 1873 " unexpected return value %d\n", f, rv); 1874 1875 continue; 1876 } 1877 1878 if (mq.nonrelocatable != 0) { 1879 PR_MEM("%s: candidate board %d has" 1880 " nonrelocatable span [0x%lx..0x%lx]\n", 1881 f, 1882 t_sbp->sb_num, 1883 mq.first_nonrelocatable, 1884 mq.last_nonrelocatable); 1885 1886 continue; 1887 } 1888 1889 #ifdef DEBUG 1890 /* 1891 * This is a debug tool for excluding certain boards 1892 * from being selected as a target board candidate. 1893 * sbd_ignore_board is only tested by this driver. 1894 * It must be set with adb, obp, /etc/system or your 1895 * favorite debugger. 1896 */ 1897 if (sbd_ignore_board & 1898 (1 << (t_sbp->sb_num - 1))) { 1899 PR_MEM("%s: sbd_ignore_board flag set," 1900 " ignoring board %d as candidate\n", 1901 f, t_sbp->sb_num); 1902 continue; 1903 } 1904 #endif 1905 1906 /* 1907 * Make sure there is no memory overlap if this 1908 * target board is used for copy-rename. 1909 */ 1910 if (sbd_check_boundaries(phys_install, s_mp, t_mp) != 0) 1911 continue; 1912 1913 /* 1914 * Reserve excess source board memory, if any. 1915 * 1916 * When the number of pages on the candidate target 1917 * board is less than the number of pages on the source, 1918 * then some spans (clearly) of the source board's address 1919 * space will not be covered by physical memory after the 1920 * copy/rename completes. The following code block 1921 * schedules those spans to be deleted. 1922 */ 1923 if (t_mp->sbm_npages < s_mp->sbm_npages) { 1924 pfn_t pfn; 1925 int mlret; 1926 1927 d_ml = memlist_dup(s_ml); 1928 if (d_ml == NULL) { 1929 PR_MEM("%s: cant dup src brd memlist\n", f); 1930 /* TODO: should abort */ 1931 continue; 1932 } 1933 1934 /* calculate base pfn relative to target board */ 1935 pfn = s_mp->sbm_basepfn & ~sm; 1936 pfn += t_mp->sbm_slice_offset; 1937 1938 /* remove span that will reside on candidate board */ 1939 mlret = memlist_delete_span( 1940 _ptob64(pfn), 1941 _ptob64(t_mp->sbm_npages), 1942 &d_ml); 1943 PR_MEM("%s: mlret = %d\n", f, mlret); 1944 1945 PR_MEM("%s: brd %d: reserving src brd memlist:\n", 1946 f, s_sbp->sb_num); 1947 SBD_MEMLIST_DUMP(d_ml); 1948 1949 /* reserve excess spans */ 1950 if (sbd_reserve_mem_spans( 1951 &s_mp->sbm_memhandle, d_ml) != 0) { 1952 1953 /* likely more non-reloc pages appeared */ 1954 /* TODO: restart from top? */ 1955 continue; 1956 } 1957 } else { 1958 /* no excess source board memory */ 1959 d_ml = NULL; 1960 } 1961 1962 s_mp->sbm_flags |= SBD_MFLAG_RESERVED; 1963 1964 /* 1965 * reserve all memory on target board. 1966 * NOTE: source board's memhandle is used. 1967 * 1968 * If this succeeds (eq 0), then target selection is 1969 * complete and all unwanted memory spans, both source and 1970 * target, have been reserved. Loop is terminated. 1971 */ 1972 if (sbd_reserve_mem_spans(&s_mp->sbm_memhandle, t_ml) == 0) { 1973 PR_MEM("%s: brd %d: target board memory reserved\n", 1974 f, t_sbp->sb_num); 1975 1976 /* a candidate target board is now reserved */ 1977 t_mp->sbm_flags |= SBD_MFLAG_RESERVED; 1978 c_mp = t_mp; 1979 1980 /* *** EXITING LOOP *** */ 1981 break; 1982 } 1983 1984 /* did not successfully reserve the target board. */ 1985 PR_MEM("%s: could not reserve target board %d\n", 1986 f, t_sbp->sb_num); 1987 1988 /* 1989 * NOTE: an undo of the sbd_reserve_mem_span work 1990 * will happen automatically when the memhandle 1991 * (s_mp->sbm_memhandle) is kphysm_del_release'd. 1992 */ 1993 1994 s_mp->sbm_flags &= ~SBD_MFLAG_RESERVED; 1995 } 1996 1997 /* clean up after memlist editing logic */ 1998 if (x_ml != NULL) 1999 memlist_delete(x_ml); 2000 2001 FREESTRUCT(sets, sbd_mem_unit_t *, n_units_per_set * n_sets); 2002 2003 /* 2004 * c_mp will be NULL when the entire sets[] array 2005 * has been searched without reserving a target board. 2006 */ 2007 if (c_mp == NULL) { 2008 PR_MEM("%s: brd %d: target selection failed.\n", 2009 f, s_sbp->sb_num); 2010 2011 if (t_ml != NULL) 2012 memlist_delete(t_ml); 2013 2014 return (-1); 2015 } 2016 2017 PR_MEM("%s: found target board %d for source board %d\n", 2018 f, 2019 t_sbp->sb_num, 2020 s_sbp->sb_num); 2021 2022 s_mp->sbm_peer = c_mp; 2023 s_mp->sbm_flags |= SBD_MFLAG_SOURCE; 2024 s_mp->sbm_del_mlist = d_ml; /* spans to be deleted, if any */ 2025 s_mp->sbm_mlist = s_ml; 2026 s_mp->sbm_cm.sbdev_busy = 1; 2027 2028 c_mp->sbm_peer = s_mp; 2029 c_mp->sbm_flags |= SBD_MFLAG_TARGET; 2030 c_mp->sbm_del_mlist = t_ml; /* spans to be deleted */ 2031 c_mp->sbm_mlist = t_ml; 2032 c_mp->sbm_cm.sbdev_busy = 1; 2033 2034 s_mp->sbm_flags &= ~SBD_MFLAG_MEMRESIZE; 2035 if (c_mp->sbm_npages > s_mp->sbm_npages) { 2036 s_mp->sbm_flags |= SBD_MFLAG_MEMUPSIZE; 2037 PR_MEM("%s: upsize (source pgs 0x%lx < target pgs 0x%lx)\n", 2038 f, s_mp->sbm_npages, c_mp->sbm_npages); 2039 } else if (c_mp->sbm_npages < s_mp->sbm_npages) { 2040 s_mp->sbm_flags |= SBD_MFLAG_MEMDOWNSIZE; 2041 PR_MEM("%s: downsize (source pgs 0x%lx > target pgs 0x%lx)\n", 2042 f, s_mp->sbm_npages, c_mp->sbm_npages); 2043 } 2044 2045 return (0); 2046 } 2047 2048 int 2049 sbd_move_memory(sbd_handle_t *hp, sbd_board_t *s_bp, sbd_board_t *t_bp) 2050 { 2051 int ret; 2052 sbdp_handle_t *hdp; 2053 sbderror_t *ep = SBD_HD2ERR(hp); 2054 2055 hdp = sbd_get_sbdp_handle(s_bp, hp); 2056 2057 ret = sbdp_move_memory(hdp, t_bp->sb_num); 2058 if (ret != 0) 2059 SBD_GET_PERR(hdp->h_err, ep); 2060 2061 sbd_release_sbdp_handle(hdp); 2062 2063 return (ret); 2064 } 2065 2066 /* 2067 * Memlist support. 2068 */ 2069 void 2070 memlist_delete(struct memlist *mlist) 2071 { 2072 sbdp_handle_t *hdp; 2073 2074 hdp = sbd_get_sbdp_handle(NULL, NULL); 2075 2076 (void) sbdp_del_memlist(hdp, mlist); 2077 2078 sbd_release_sbdp_handle(hdp); 2079 } 2080 2081 struct memlist * 2082 memlist_dup(struct memlist *mlist) 2083 { 2084 struct memlist *hl, *prev; 2085 2086 if (mlist == NULL) 2087 return (NULL); 2088 2089 prev = NULL; 2090 hl = NULL; 2091 for (; mlist; mlist = mlist->next) { 2092 struct memlist *mp; 2093 2094 mp = memlist_get_one(); 2095 if (mp == NULL) { 2096 if (hl != NULL) 2097 memlist_free_list(hl); 2098 hl = NULL; 2099 break; 2100 } 2101 mp->address = mlist->address; 2102 mp->size = mlist->size; 2103 mp->next = NULL; 2104 mp->prev = prev; 2105 2106 if (prev == NULL) 2107 hl = mp; 2108 else 2109 prev->next = mp; 2110 prev = mp; 2111 } 2112 2113 return (hl); 2114 } 2115 2116 void 2117 memlist_dump(struct memlist *mlist) 2118 { 2119 register struct memlist *ml; 2120 2121 if (mlist == NULL) { 2122 PR_MEM("memlist> EMPTY\n"); 2123 } else { 2124 for (ml = mlist; ml; ml = ml->next) 2125 PR_MEM("memlist> 0x%" PRIx64 " " 2126 "0x%" PRIx64 " \n", 2127 ml->address, ml->size); 2128 } 2129 } 2130 2131 int 2132 memlist_intersect(struct memlist *al, struct memlist *bl) 2133 { 2134 uint64_t astart, aend, bstart, bend; 2135 2136 if ((al == NULL) || (bl == NULL)) 2137 return (0); 2138 2139 aend = al->address + al->size; 2140 bstart = bl->address; 2141 bend = bl->address + bl->size; 2142 2143 while (al && bl) { 2144 while (al && (aend <= bstart)) 2145 if ((al = al->next) != NULL) 2146 aend = al->address + al->size; 2147 if (al == NULL) 2148 return (0); 2149 2150 if ((astart = al->address) <= bstart) 2151 return (1); 2152 2153 while (bl && (bend <= astart)) 2154 if ((bl = bl->next) != NULL) 2155 bend = bl->address + bl->size; 2156 if (bl == NULL) 2157 return (0); 2158 2159 if ((bstart = bl->address) <= astart) 2160 return (1); 2161 } 2162 2163 return (0); 2164 } 2165 2166 /* 2167 * Determine whether the source memlist (s_mlist) will 2168 * fit into the target memlist (t_mlist) in terms of 2169 * size and holes (i.e. based on same relative base address). 2170 */ 2171 static int 2172 memlist_canfit(struct memlist *s_mlist, struct memlist *t_mlist) 2173 { 2174 int rv = 0; 2175 uint64_t s_basepa, t_basepa; 2176 struct memlist *s_ml, *t_ml; 2177 2178 if ((s_mlist == NULL) || (t_mlist == NULL)) 2179 return (0); 2180 2181 /* 2182 * Base both memlists on common base address (0). 2183 */ 2184 s_basepa = s_mlist->address; 2185 t_basepa = t_mlist->address; 2186 2187 for (s_ml = s_mlist; s_ml; s_ml = s_ml->next) 2188 s_ml->address -= s_basepa; 2189 2190 for (t_ml = t_mlist; t_ml; t_ml = t_ml->next) 2191 t_ml->address -= t_basepa; 2192 2193 s_ml = s_mlist; 2194 for (t_ml = t_mlist; t_ml && s_ml; t_ml = t_ml->next) { 2195 uint64_t s_start, s_end; 2196 uint64_t t_start, t_end; 2197 2198 t_start = t_ml->address; 2199 t_end = t_start + t_ml->size; 2200 2201 for (; s_ml; s_ml = s_ml->next) { 2202 s_start = s_ml->address; 2203 s_end = s_start + s_ml->size; 2204 2205 if ((s_start < t_start) || (s_end > t_end)) 2206 break; 2207 } 2208 } 2209 /* 2210 * If we ran out of source memlist chunks that mean 2211 * we found a home for all of them. 2212 */ 2213 if (s_ml == NULL) 2214 rv = 1; 2215 2216 /* 2217 * Need to add base addresses back since memlists 2218 * are probably in use by caller. 2219 */ 2220 for (s_ml = s_mlist; s_ml; s_ml = s_ml->next) 2221 s_ml->address += s_basepa; 2222 2223 for (t_ml = t_mlist; t_ml; t_ml = t_ml->next) 2224 t_ml->address += t_basepa; 2225 2226 return (rv); 2227 } 2228 2229 void 2230 sbd_attach_mem(sbd_handle_t *hp, sbderror_t *ep) 2231 { 2232 sbd_mem_unit_t *mp; 2233 dev_info_t *dip; 2234 sbd_board_t *sbp = SBDH2BD(hp->h_sbd); 2235 sbdp_handle_t *hdp; 2236 int err, unit; 2237 struct memlist *ml, *mc; 2238 static fn_t f = "sbd_attach_mem"; 2239 int i; 2240 2241 PR_MEM("%s...\n", f); 2242 2243 /* 2244 * all four cpus have to be attached before 2245 * configuring mem 2246 */ 2247 for (i = 0; i < MAX_CPU_UNITS_PER_BOARD; i++) { 2248 sbd_cpu_unit_t *cpup; 2249 struct cpu *cp; 2250 2251 if (!SBD_DEV_IS_PRESENT(sbp, SBD_COMP_CPU, i)) 2252 continue; 2253 2254 if (!SBD_DEV_IS_ATTACHED(sbp, SBD_COMP_CPU, i)) 2255 goto error; 2256 2257 cpup = SBD_GET_BOARD_CPUUNIT(sbp, i); 2258 2259 if (cpup == NULL) 2260 goto error; 2261 2262 mutex_enter(&cpu_lock); 2263 cp = cpu_get(cpup->sbc_cpu_id); 2264 if (cp == NULL) { 2265 mutex_exit(&cpu_lock); 2266 cmn_err(CE_WARN, 2267 "sbd:%s: cpu_get failed for cpu %d", 2268 f, cpup->sbc_cpu_id); 2269 goto error; 2270 } 2271 if (cpu_is_poweredoff(cp)) { 2272 mutex_exit(&cpu_lock); 2273 goto error; 2274 } 2275 mutex_exit(&cpu_lock); 2276 continue; 2277 2278 error: 2279 SBD_SET_ERR(ep, ESBD_CPUONLINE); 2280 SBD_SET_ERRSTR(ep, sbp->sb_mempath[i]); 2281 (void) sbd_set_err_in_hdl(hp, ep); 2282 return; 2283 } 2284 2285 dip = *(sbp->sb_devlist[NIX(SBD_COMP_MEM)]); 2286 2287 hdp = sbd_get_sbdp_handle(sbp, hp); 2288 unit = sbdp_get_unit_num(hdp, dip); 2289 if (unit < 0) { 2290 SBD_GET_PERR(hdp->h_err, ep); 2291 sbd_release_sbdp_handle(hdp); 2292 return; 2293 } 2294 2295 ASSERT(sbp->sb_mempath[unit] != NULL); 2296 ASSERT(e_ddi_branch_held(dip)); 2297 2298 (void) ddi_pathname(dip, sbp->sb_mempath[unit]); 2299 2300 mp = SBD_GET_BOARD_MEMUNIT(sbp, unit); 2301 2302 ml = sbd_get_memlist(mp, ep); 2303 if (ml == NULL) { 2304 cmn_err(CE_WARN, 2305 "sbd:%s: failed to get memlist for " 2306 "board %d", f, sbp->sb_num); 2307 /* 2308 * Need to record an error and return. 2309 */ 2310 SBD_SET_ERR(ep, ESBD_MEMFAIL); 2311 SBD_SET_ERRSTR(ep, sbp->sb_mempath[unit]); 2312 sbd_release_sbdp_handle(hdp); 2313 return; 2314 } 2315 2316 SBD_MEMLIST_DUMP(ml); 2317 err = 0; 2318 for (mc = ml; mc; mc = mc->next) { 2319 update_membounds_t umb; 2320 pfn_t base; 2321 pgcnt_t npgs; 2322 2323 base = (pfn_t)(mc->address >> PAGESHIFT); 2324 npgs = (pgcnt_t)(mc->size >> PAGESHIFT); 2325 2326 umb.u_board = sbp->sb_num; 2327 umb.u_base = (uint64_t)base << MMU_PAGESHIFT; 2328 umb.u_len = (uint64_t)npgs << MMU_PAGESHIFT; 2329 2330 lgrp_plat_config(LGRP_CONFIG_MEM_ADD, (uintptr_t)&umb); 2331 err = kphysm_add_memory_dynamic(base, npgs); 2332 2333 if (err != KPHYSM_OK) { 2334 cmn_err(CE_WARN, 2335 "%s: kphysm_add_memory_dynamic fail %d", f, err); 2336 2337 /* translate kphysm error */ 2338 switch (err) { 2339 case KPHYSM_ERESOURCE: 2340 err = ESBD_NOMEM; 2341 break; 2342 2343 case KPHYSM_EFAULT: 2344 err = ESBD_FAULT; 2345 break; 2346 2347 default: 2348 err = ESBD_INVAL; 2349 break; 2350 } 2351 break; 2352 } 2353 2354 err = kcage_range_add(base, npgs, KCAGE_DOWN); 2355 if (err != 0) { 2356 cmn_err(CE_WARN, 2357 "%s: kcage_range_add fail %d", f, err); 2358 2359 /* Translate kcage error. */ 2360 switch (err) { 2361 case ENOMEM: 2362 err = ESBD_NOMEM; 2363 break; 2364 default: 2365 err = ESBD_INVAL; 2366 break; 2367 } 2368 break; 2369 } 2370 (void) sbdp_mem_add_span(hdp, mc->address, mc->size); 2371 } 2372 2373 if (err != 0) { 2374 SBD_SET_ERR(ep, err); 2375 SBD_SET_ERRSTR(ep, sbp->sb_mempath[unit]); 2376 } 2377 2378 memlist_delete(ml); 2379 sbd_release_sbdp_handle(hdp); 2380 2381 /* 2382 * Now attach all mem devinfo nodes to the device tree. 2383 */ 2384 for (i = 0; i < SBD_NUM_MC_PER_BOARD; i++) { 2385 if (mp->sbm_dip[i] == NULL) 2386 continue; 2387 ASSERT(e_ddi_branch_held(mp->sbm_dip[i])); 2388 if (e_ddi_branch_configure(mp->sbm_dip[i], NULL, 0) && 2389 SBD_GET_ERR(ep) == 0) { 2390 SBD_SET_ERR(ep, ESBD_INVAL); 2391 SBD_SET_ERRSTR(ep, sbp->sb_mempath[unit]); 2392 } 2393 } 2394 } 2395 2396 typedef struct { 2397 kcondvar_t cond; 2398 kmutex_t lock; 2399 int error; 2400 int done; 2401 } sbd_release_mem_sync_t; 2402 2403 /* 2404 * When we reach here the memory being drained should have 2405 * already been reserved in sbd_pre_release_mem(). 2406 * Our only task here is to kick off the "drain". 2407 * Returns -1 when error encountered or zero for success. 2408 */ 2409 int 2410 sbd_release_mem(sbd_handle_t *hp, dev_info_t *dip, int unit) 2411 { 2412 memhandle_t mh; 2413 int err; 2414 int cancel_flag = 0; 2415 int e_code = 0; 2416 sbd_board_t *sbp = SBDH2BD(hp->h_sbd); 2417 sbd_release_mem_sync_t rms; 2418 static fn_t f = "sbd_release_mem"; 2419 2420 /* 2421 * If this node has a scheduled memory delete operation, 2422 * it will have a memhandle. If it does have a memhandle (the 2423 * return value of sbd_get_memhandle is zero when true), 2424 * then perform the delete. 2425 */ 2426 2427 if ((cancel_flag = sbd_get_memhandle(hp, dip, &mh)) != 0) { 2428 cmn_err(CE_WARN, "%s: couldn't get the memhandle\n", f); 2429 return (cancel_flag); 2430 } 2431 2432 bzero((void *) &rms, sizeof (rms)); 2433 2434 mutex_init(&rms.lock, NULL, MUTEX_DRIVER, NULL); 2435 cv_init(&rms.cond, NULL, CV_DRIVER, NULL); 2436 2437 mutex_enter(&rms.lock); 2438 err = kphysm_del_start(mh, sbd_release_memory_done, (void *) &rms); 2439 if (err == KPHYSM_OK) { 2440 /* wait for completion */ 2441 while (!rms.done) { 2442 if (cancel_flag) { 2443 /* previously canceled */ 2444 cv_wait(&rms.cond, &rms.lock); 2445 } else if (cv_wait_sig(&rms.cond, &rms.lock) == 0) { 2446 /* interrupted: cancel and wait */ 2447 cancel_flag = -1; 2448 (void) kphysm_del_cancel(mh); 2449 } 2450 } 2451 /* get the result of the memory delete operation */ 2452 err = rms.error; 2453 } else { 2454 (void) kphysm_del_release(mh); 2455 } 2456 2457 mutex_exit(&rms.lock); 2458 2459 cv_destroy(&rms.cond); 2460 mutex_destroy(&rms.lock); 2461 2462 if (err != KPHYSM_OK) { 2463 switch (err) { 2464 case KPHYSM_ENOWORK: 2465 e_code = ESBD_NOERROR; 2466 break; 2467 2468 case KPHYSM_EHANDLE: 2469 case KPHYSM_ESEQUENCE: 2470 e_code = ESBD_INTERNAL; 2471 break; 2472 2473 case KPHYSM_ENOTVIABLE: 2474 e_code = ESBD_MEM_NOTVIABLE; 2475 break; 2476 2477 case KPHYSM_EREFUSED: 2478 e_code = ESBD_MEM_REFUSED; 2479 break; 2480 2481 case KPHYSM_ENONRELOC: 2482 e_code = ESBD_MEM_NONRELOC; 2483 break; 2484 2485 case KPHYSM_ECANCELLED: 2486 e_code = ESBD_MEM_CANCELLED; 2487 break; 2488 2489 case KPHYSM_ERESOURCE: 2490 e_code = ESBD_MEMFAIL; 2491 break; 2492 2493 default: 2494 cmn_err(CE_WARN, "sbd:%s:" 2495 " unexpected kphysm error code %d," 2496 " dip 0x%p", 2497 f, err, (void *)dip); 2498 2499 e_code = ESBD_IO; 2500 break; 2501 } 2502 2503 if (e_code != 0) { 2504 cancel_flag = -1; 2505 SBD_SET_ERR(SBD_HD2ERR(hp), e_code); 2506 SBD_SET_ERRSTR(SBD_HD2ERR(hp), sbp->sb_mempath[unit]); 2507 } 2508 } 2509 2510 return (cancel_flag); 2511 } 2512 2513 /* 2514 * Memory has been logically removed by the time this routine is called. 2515 */ 2516 void 2517 sbd_release_memory_done(void *arg, int error) 2518 { 2519 sbd_release_mem_sync_t *ds = arg; 2520 2521 mutex_enter(&ds->lock); 2522 ds->error = error; 2523 ds->done = 1; 2524 cv_signal(&ds->cond); 2525 mutex_exit(&ds->lock); 2526 } 2527 2528 /* 2529 * If detaching node contains memory that is "non-permanent" 2530 * then the memory adr's are simply cleared. If the memory 2531 * is non-relocatable, then do a copy-rename. 2532 */ 2533 int 2534 sbd_detach_memory(sbd_handle_t *hp, sbderror_t *ep, sbd_mem_unit_t *s_mp, 2535 int unit) 2536 { 2537 int rv; 2538 sbd_mem_unit_t *t_mp; 2539 sbd_istate_t state; 2540 sbdp_handle_t *hdp; 2541 sbd_board_t *sbp = (sbd_board_t *)s_mp->sbm_cm.sbdev_sbp; 2542 sbd_board_t *tbp; 2543 static fn_t f = "sbd_detach_memory"; 2544 2545 PR_MEM("%s...\n", f); 2546 2547 /* lookup target mem unit and target board structure, if any */ 2548 if (s_mp->sbm_flags & SBD_MFLAG_SOURCE) { 2549 t_mp = s_mp->sbm_peer; 2550 ASSERT(t_mp != NULL); 2551 ASSERT(t_mp->sbm_peer == s_mp); 2552 tbp = (sbd_board_t *)t_mp->sbm_cm.sbdev_sbp; 2553 } else { 2554 t_mp = NULL; 2555 } 2556 2557 /* verify mem unit's state is UNREFERENCED */ 2558 state = s_mp->sbm_cm.sbdev_state; 2559 if (state != SBD_STATE_UNREFERENCED) { 2560 cmn_err(CE_WARN, "%s: invalid state transition for" 2561 " mem-unit (%d.%d)", 2562 f, 2563 sbp->sb_num, 2564 s_mp->sbm_cm.sbdev_unum); 2565 SBD_SET_ERR(ep, ESBD_STATE); 2566 SBD_SET_ERRSTR(ep, sbp->sb_mempath[unit]); 2567 return (-1); 2568 } 2569 2570 /* verify target mem unit's state is UNREFERENCED, if any */ 2571 if (t_mp != NULL) { 2572 state = t_mp->sbm_cm.sbdev_state; 2573 if (state != SBD_STATE_UNREFERENCED) { 2574 cmn_err(CE_WARN, "%s: invalid state transition for" 2575 " target mem-unit (%d.%d)", 2576 f, 2577 tbp->sb_num, 2578 t_mp->sbm_cm.sbdev_unum); 2579 SBD_SET_ERR(ep, ESBD_STATE); 2580 SBD_SET_ERRSTR(ep, sbp->sb_mempath[unit]); 2581 return (-1); 2582 } 2583 } 2584 2585 /* 2586 * Displacement flush all ecaches in the system. 2587 * That's the fastest way to remove all cache references 2588 * to the detaching memory. 2589 */ 2590 xc_all(sbd_flush_ecache, 0, 0); 2591 2592 hdp = sbd_get_sbdp_handle(sbp, hp); 2593 2594 /* 2595 * If there is no target board (no copy/rename was needed), then 2596 * we're done! 2597 */ 2598 if (t_mp == NULL) { 2599 /* 2600 * Reprogram interconnect hardware and disable 2601 * memory controllers for memory node that's going away. 2602 */ 2603 2604 rv = sbdphw_disable_memctrl(hdp, s_mp->sbm_cm.sbdev_dip); 2605 if (rv) { 2606 cmn_err(CE_WARN, 2607 "%s: failed to deprogram mem-unit (%d.%d)," 2608 " dip 0x%p", 2609 f, 2610 sbp->sb_num, 2611 s_mp->sbm_cm.sbdev_unum, 2612 (void *)s_mp->sbm_cm.sbdev_dip); 2613 /* 2614 * Make sure we don't rewrite an sbdp error 2615 */ 2616 if (SBD_GET_ERR(ep) != 0) { 2617 SBD_SET_ERR(ep, ESBD_HW_PROGRAM); 2618 SBD_SET_ERRSTR(ep, sbp->sb_mempath[unit]); 2619 } 2620 } 2621 } else { 2622 rv = sbd_move_memory(hp, sbp, tbp); 2623 if (rv) { 2624 int i; 2625 2626 cmn_err(CE_WARN, "%s: failed to move memory" 2627 " from board %d to board %d", 2628 f, 2629 sbp->sb_num, 2630 tbp->sb_num); 2631 /* 2632 * Make sure we don't rewrite an sbdp error 2633 */ 2634 if (SBD_GET_ERR(ep) != 0) { 2635 SBD_SET_ERR(ep, ESBD_INTERNAL); 2636 SBD_SET_ERRSTR(ep, sbp->sb_mempath[unit]); 2637 } 2638 /* 2639 * If we failed here, it means that the target board's 2640 * memory has been unconfigured. We need to configure 2641 * it back 2642 */ 2643 for (i = 0; i < MAX_MEM_UNITS_PER_BOARD; i++) { 2644 int unit; 2645 dev_info_t *dip; 2646 dev_info_t **devlist; 2647 2648 2649 devlist = tbp->sb_devlist[NIX(SBD_COMP_MEM)]; 2650 dip = devlist[i]; 2651 sbd_reset_error_sbdph(hdp); 2652 unit = sbdp_get_unit_num(hdp, dip); 2653 2654 /* 2655 * We already saved the error that created 2656 * this mess. If we fail, make sure not 2657 * to overwrite the original error 2658 */ 2659 if (unit == -1) { 2660 continue; 2661 } 2662 if (sbd_cancel_mem(hp, unit) != 0) 2663 continue; 2664 2665 t_mp->sbm_flags = 0; 2666 /* 2667 * clean up 2668 */ 2669 sbd_mem_cleanup(s_mp, t_mp, ep); 2670 if (s_mp->sbm_mlist) { 2671 memlist_delete(s_mp->sbm_mlist); 2672 s_mp->sbm_mlist = NULL; 2673 } 2674 2675 SBD_DEVICE_TRANSITION(tbp, SBD_COMP_MEM, 2676 unit, SBD_STATE_CONFIGURED); 2677 } 2678 } 2679 2680 PR_MEM("%s: %s memory COPY-RENAME (board %d -> %d)\n", 2681 f, 2682 rv ? "FAILED" : "COMPLETED", 2683 sbp->sb_num, 2684 tbp->sb_num); 2685 } 2686 2687 if (rv == 0) { 2688 update_membounds_t umb; 2689 2690 umb.u_board = sbp->sb_num; 2691 umb.u_base = (uint64_t)-1; 2692 umb.u_len = (uint64_t)-1; 2693 2694 lgrp_plat_config(LGRP_CONFIG_MEM_DEL, (uintptr_t)&umb); 2695 } 2696 2697 sbd_release_sbdp_handle(hdp); 2698 return (rv); 2699 } 2700 2701 /*ARGSUSED*/ 2702 static void 2703 sbd_flush_ecache(uint64_t a, uint64_t b) 2704 { 2705 cpu_flush_ecache(); 2706 } 2707