1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright 2007 Sun Microsystems, Inc. All rights reserved. 23 * Use is subject to license terms. 24 */ 25 26 #pragma ident "%Z%%M% %I% %E% SMI" 27 28 /* 29 * Platform specific implementation code 30 */ 31 32 #define SUNDDI_IMPL 33 34 #include <sys/types.h> 35 #include <sys/promif.h> 36 #include <sys/prom_isa.h> 37 #include <sys/prom_plat.h> 38 #include <sys/mmu.h> 39 #include <vm/hat_sfmmu.h> 40 #include <sys/iommu.h> 41 #include <sys/scb.h> 42 #include <sys/cpuvar.h> 43 #include <sys/intreg.h> 44 #include <sys/pte.h> 45 #include <vm/hat.h> 46 #include <vm/page.h> 47 #include <vm/as.h> 48 #include <sys/cpr.h> 49 #include <sys/kmem.h> 50 #include <sys/clock.h> 51 #include <sys/kmem.h> 52 #include <sys/panic.h> 53 #include <vm/seg_kmem.h> 54 #include <sys/cpu_module.h> 55 #include <sys/callb.h> 56 #include <sys/machsystm.h> 57 #include <sys/vmsystm.h> 58 #include <sys/systm.h> 59 #include <sys/archsystm.h> 60 #include <sys/stack.h> 61 #include <sys/fs/ufs_fs.h> 62 #include <sys/memlist.h> 63 #include <sys/bootconf.h> 64 #include <sys/thread.h> 65 #include <vm/vm_dep.h> 66 67 extern void cpr_clear_bitmaps(void); 68 extern int cpr_setbit(pfn_t ppn, int mapflag); 69 extern int cpr_clrbit(pfn_t ppn, int mapflag); 70 extern pgcnt_t cpr_scan_kvseg(int mapflag, bitfunc_t bitfunc, struct seg *seg); 71 extern pgcnt_t cpr_count_seg_pages(int mapflag, bitfunc_t bitfunc); 72 extern void dtlb_wr_entry(uint_t, tte_t *, uint64_t *); 73 extern void itlb_wr_entry(uint_t, tte_t *, uint64_t *); 74 75 static int i_cpr_storage_desc_alloc(csd_t **, pgcnt_t *, csd_t **, int); 76 static void i_cpr_storage_desc_init(csd_t *, pgcnt_t, csd_t *); 77 static caddr_t i_cpr_storage_data_alloc(pgcnt_t, pgcnt_t *, int); 78 static int cpr_dump_sensitive(vnode_t *, csd_t *); 79 static void i_cpr_clear_entries(uint64_t, uint64_t); 80 static void i_cpr_xcall(xcfunc_t); 81 82 void i_cpr_storage_free(void); 83 84 extern void *i_cpr_data_page; 85 extern int cpr_test_mode; 86 extern int cpr_nbitmaps; 87 extern char cpr_default_path[]; 88 extern caddr_t textva, datava; 89 90 static struct cpr_map_info cpr_prom_retain[CPR_PROM_RETAIN_CNT]; 91 caddr_t cpr_vaddr = NULL; 92 93 static uint_t sensitive_pages_saved; 94 static uint_t sensitive_size_saved; 95 96 caddr_t i_cpr_storage_data_base; 97 caddr_t i_cpr_storage_data_end; 98 csd_t *i_cpr_storage_desc_base; 99 csd_t *i_cpr_storage_desc_end; /* one byte beyond last used descp */ 100 csd_t *i_cpr_storage_desc_last_used; /* last used descriptor */ 101 caddr_t sensitive_write_ptr; /* position for next storage write */ 102 103 size_t i_cpr_sensitive_bytes_dumped; 104 pgcnt_t i_cpr_sensitive_pgs_dumped; 105 pgcnt_t i_cpr_storage_data_sz; /* in pages */ 106 pgcnt_t i_cpr_storage_desc_pgcnt; /* in pages */ 107 108 ushort_t cpr_mach_type = CPR_MACHTYPE_4U; 109 static csu_md_t m_info; 110 111 112 #define MAX_STORAGE_RETRY 3 113 #define MAX_STORAGE_ALLOC_RETRY 3 114 #define INITIAL_ALLOC_PCNT 40 /* starting allocation percentage */ 115 #define INTEGRAL 100 /* to get 1% precision */ 116 117 #define EXTRA_RATE 2 /* add EXTRA_RATE% extra space */ 118 #define EXTRA_DESCS 10 119 120 #define CPR_NO_STORAGE_DESC 1 121 #define CPR_NO_STORAGE_DATA 2 122 123 #define CIF_SPLICE 0 124 #define CIF_UNLINK 1 125 126 127 /* 128 * CPR miscellaneous support routines 129 */ 130 #define cpr_open(path, mode, vpp) (vn_open(path, UIO_SYSSPACE, \ 131 mode, 0600, vpp, CRCREAT, 0)) 132 #define cpr_rdwr(rw, vp, basep, cnt) (vn_rdwr(rw, vp, (caddr_t)(basep), \ 133 cnt, 0LL, UIO_SYSSPACE, 0, (rlim64_t)MAXOFF_T, CRED(), \ 134 (ssize_t *)NULL)) 135 136 /* 137 * definitions for saving/restoring prom pages 138 */ 139 static void *ppage_buf; 140 static pgcnt_t ppage_count; 141 static pfn_t *pphys_list; 142 static size_t pphys_list_size; 143 144 typedef void (*tlb_rw_t)(uint_t, tte_t *, uint64_t *); 145 typedef void (*tlb_filter_t)(int, tte_t *, uint64_t, void *); 146 147 /* 148 * private struct for tlb handling 149 */ 150 struct cpr_trans_info { 151 sutlb_t *dst; 152 sutlb_t *tail; 153 tlb_rw_t reader; 154 tlb_rw_t writer; 155 tlb_filter_t filter; 156 int index; 157 uint64_t skip; /* assumes TLB <= 64 locked entries */ 158 }; 159 typedef struct cpr_trans_info cti_t; 160 161 162 /* 163 * special handling for tlb info 164 */ 165 #define WITHIN_OFW(va) \ 166 (((va) > (uint64_t)OFW_START_ADDR) && ((va) < (uint64_t)OFW_END_ADDR)) 167 168 #define WITHIN_NUCLEUS(va, base) \ 169 (((va) >= (base)) && \ 170 (((va) + MMU_PAGESIZE) <= ((base) + MMU_PAGESIZE4M))) 171 172 #define IS_BIGKTSB(va) \ 173 (enable_bigktsb && \ 174 ((va) >= (uint64_t)ktsb_base) && \ 175 ((va) < (uint64_t)(ktsb_base + ktsb_sz))) 176 177 178 /* 179 * WARNING: 180 * the text from this file is linked to follow cpr_resume_setup.o; 181 * only add text between here and i_cpr_end_jumpback when it needs 182 * to be called during resume before we switch back to the kernel 183 * trap table. all the text in this range must fit within a page. 184 */ 185 186 187 /* 188 * each time a machine is reset, the prom uses an inconsistent set of phys 189 * pages and the cif cookie may differ as well. so prior to restoring the 190 * original prom, we have to use to use the new/tmp prom's translations 191 * when requesting prom services. 192 * 193 * cif_handler starts out as the original prom cookie, and that gets used 194 * by client_handler() to jump into the prom. here we splice-in a wrapper 195 * routine by writing cif_handler; client_handler() will now jump to the 196 * wrapper which switches the %tba to the new/tmp prom's trap table then 197 * jumps to the new cookie. 198 */ 199 void 200 i_cpr_cif_setup(int action) 201 { 202 extern void *i_cpr_orig_cif, *cif_handler; 203 extern int i_cpr_cif_wrapper(void *); 204 205 /* 206 * save the original cookie and change the current cookie to the 207 * wrapper routine. later we just restore the original cookie. 208 */ 209 if (action == CIF_SPLICE) { 210 i_cpr_orig_cif = cif_handler; 211 cif_handler = (void *)i_cpr_cif_wrapper; 212 } else if (action == CIF_UNLINK) 213 cif_handler = i_cpr_orig_cif; 214 } 215 216 217 /* 218 * launch slave cpus into kernel text, pause them, 219 * and restore the original prom pages 220 */ 221 void 222 i_cpr_mp_setup(void) 223 { 224 extern void restart_other_cpu(int); 225 cpu_t *cp; 226 227 uint64_t kctx = kcontextreg; 228 229 /* 230 * Do not allow setting page size codes in MMU primary context 231 * register while using cif wrapper. This is needed to work 232 * around OBP incorrect handling of this MMU register. 233 */ 234 kcontextreg = 0; 235 236 /* 237 * reset cpu_ready_set so x_calls work properly 238 */ 239 CPUSET_ZERO(cpu_ready_set); 240 CPUSET_ADD(cpu_ready_set, getprocessorid()); 241 242 /* 243 * setup cif to use the cookie from the new/tmp prom 244 * and setup tmp handling for calling prom services. 245 */ 246 i_cpr_cif_setup(CIF_SPLICE); 247 248 /* 249 * at this point, only the nucleus and a few cpr pages are 250 * mapped in. once we switch to the kernel trap table, 251 * we can access the rest of kernel space. 252 */ 253 prom_set_traptable(&trap_table); 254 255 if (ncpus > 1) { 256 sfmmu_init_tsbs(); 257 258 mutex_enter(&cpu_lock); 259 /* 260 * All of the slave cpus are not ready at this time, 261 * yet the cpu structures have various cpu_flags set; 262 * clear cpu_flags and mutex_ready. 263 * Since we are coming up from a CPU suspend, the slave cpus 264 * are frozen. 265 */ 266 for (cp = CPU->cpu_next; cp != CPU; cp = cp->cpu_next) { 267 cp->cpu_flags = CPU_FROZEN; 268 cp->cpu_m.mutex_ready = 0; 269 } 270 271 for (cp = CPU->cpu_next; cp != CPU; cp = cp->cpu_next) 272 restart_other_cpu(cp->cpu_id); 273 274 pause_cpus(NULL); 275 mutex_exit(&cpu_lock); 276 277 i_cpr_xcall(i_cpr_clear_entries); 278 } else 279 i_cpr_clear_entries(0, 0); 280 281 /* 282 * now unlink the cif wrapper; WARNING: do not call any 283 * prom_xxx() routines until after prom pages are restored. 284 */ 285 i_cpr_cif_setup(CIF_UNLINK); 286 287 (void) i_cpr_prom_pages(CPR_PROM_RESTORE); 288 289 /* allow setting page size codes in MMU primary context register */ 290 kcontextreg = kctx; 291 } 292 293 294 /* 295 * end marker for jumpback page; 296 * this symbol is used to check the size of i_cpr_resume_setup() 297 * and the above text. For simplicity, the Makefile needs to 298 * link i_cpr_resume_setup.o and cpr_impl.o consecutively. 299 */ 300 void 301 i_cpr_end_jumpback(void) 302 { 303 } 304 305 306 /* 307 * scan tlb entries with reader; when valid entries are found, 308 * the filter routine will selectively save/clear them 309 */ 310 static void 311 i_cpr_scan_tlb(cti_t *ctip) 312 { 313 uint64_t va_tag; 314 int tlb_index; 315 tte_t tte; 316 317 for (tlb_index = ctip->index; tlb_index >= 0; tlb_index--) { 318 (*ctip->reader)((uint_t)tlb_index, &tte, &va_tag); 319 if (va_tag && TTE_IS_VALID(&tte)) 320 (*ctip->filter)(tlb_index, &tte, va_tag, ctip); 321 } 322 } 323 324 325 /* 326 * filter for locked tlb entries that reference the text/data nucleus 327 * and any bigktsb's; these will be reinstalled by cprboot on all cpus 328 */ 329 /* ARGSUSED */ 330 static void 331 i_cpr_lnb(int index, tte_t *ttep, uint64_t va_tag, void *ctrans) 332 { 333 cti_t *ctip; 334 335 /* 336 * record tlb data at ctip->dst; the target tlb index starts 337 * at the highest tlb offset and moves towards 0. the prom 338 * reserves both dtlb and itlb index 0. any selected entry 339 * also gets marked to prevent being flushed during resume 340 */ 341 if (TTE_IS_LOCKED(ttep) && (va_tag == (uint64_t)textva || 342 va_tag == (uint64_t)datava || IS_BIGKTSB(va_tag))) { 343 ctip = ctrans; 344 while ((1 << ctip->index) & ctip->skip) 345 ctip->index--; 346 ASSERT(ctip->index > 0); 347 ASSERT(ctip->dst < ctip->tail); 348 ctip->dst->tte.ll = ttep->ll; 349 ctip->dst->va_tag = va_tag; 350 ctip->dst->index = ctip->index--; 351 ctip->dst->tmp = 0; 352 ctip->dst++; 353 } 354 } 355 356 357 /* 358 * some tlb entries are stale, filter for unlocked entries 359 * within the prom virt range and clear them 360 */ 361 static void 362 i_cpr_ufw(int index, tte_t *ttep, uint64_t va_tag, void *ctrans) 363 { 364 sutlb_t clr; 365 cti_t *ctip; 366 367 if (!TTE_IS_LOCKED(ttep) && WITHIN_OFW(va_tag)) { 368 ctip = ctrans; 369 bzero(&clr, sizeof (clr)); 370 (*ctip->writer)((uint_t)index, &clr.tte, &clr.va_tag); 371 } 372 } 373 374 375 /* 376 * some of the entries installed by cprboot are needed only on a 377 * short-term basis and need to be flushed to avoid clogging the tlbs. 378 * scan the dtte/itte arrays for items marked as temporary and clear 379 * dtlb/itlb entries using wrfunc. 380 */ 381 static void 382 i_cpr_clear_tmp(sutlb_t *listp, int max, tlb_rw_t wrfunc) 383 { 384 sutlb_t clr, *tail; 385 386 bzero(&clr, sizeof (clr)); 387 for (tail = listp + max; listp < tail && listp->va_tag; listp++) { 388 if (listp->tmp) 389 (*wrfunc)((uint_t)listp->index, &clr.tte, &clr.va_tag); 390 } 391 } 392 393 394 /* ARGSUSED */ 395 static void 396 i_cpr_clear_entries(uint64_t arg1, uint64_t arg2) 397 { 398 extern void demap_all(void); 399 cti_t cti; 400 401 i_cpr_clear_tmp(m_info.dtte, CPR_MAX_TLB, dtlb_wr_entry); 402 i_cpr_clear_tmp(m_info.itte, CPR_MAX_TLB, itlb_wr_entry); 403 404 /* 405 * for newer cpus that implement DEMAP_ALL_TYPE, demap_all is 406 * a second label for vtag_flushall. the call is made using 407 * vtag_flushall() instead of demap_all() due to runtime and 408 * krtld results with both older and newer cpu modules. 409 */ 410 if (&demap_all != 0) { 411 vtag_flushall(); 412 return; 413 } 414 415 /* 416 * for older V9 cpus, scan tlbs and clear stale entries 417 */ 418 bzero(&cti, sizeof (cti)); 419 cti.filter = i_cpr_ufw; 420 421 cti.index = cpunodes[CPU->cpu_id].dtlb_size - 1; 422 cti.reader = dtlb_rd_entry; 423 cti.writer = dtlb_wr_entry; 424 i_cpr_scan_tlb(&cti); 425 426 cti.index = cpunodes[CPU->cpu_id].itlb_size - 1; 427 cti.reader = itlb_rd_entry; 428 cti.writer = itlb_wr_entry; 429 i_cpr_scan_tlb(&cti); 430 } 431 432 433 /* 434 * craft tlb info for tmp use during resume; this data gets used by 435 * cprboot to install tlb entries. we also mark each struct as tmp 436 * so those tlb entries will get flushed after switching to the kernel 437 * trap table. no data needs to be recorded for vaddr when it falls 438 * within the nucleus since we've already recorded nucleus ttes and 439 * a 8K tte would conflict with a 4MB tte. eg: the cpr module 440 * text/data may have been loaded into the text/data nucleus. 441 */ 442 static void 443 i_cpr_make_tte(cti_t *ctip, void *vaddr, caddr_t nbase) 444 { 445 pfn_t ppn; 446 uint_t rw; 447 448 if (WITHIN_NUCLEUS((caddr_t)vaddr, nbase)) 449 return; 450 451 while ((1 << ctip->index) & ctip->skip) 452 ctip->index--; 453 ASSERT(ctip->index > 0); 454 ASSERT(ctip->dst < ctip->tail); 455 456 /* 457 * without any global service available to lookup 458 * a tte by vaddr, we craft our own here: 459 */ 460 ppn = va_to_pfn(vaddr); 461 rw = (nbase == datava) ? TTE_HWWR_INT : 0; 462 ctip->dst->tte.tte_inthi = TTE_VALID_INT | TTE_PFN_INTHI(ppn); 463 ctip->dst->tte.tte_intlo = TTE_PFN_INTLO(ppn) | TTE_LCK_INT | 464 TTE_CP_INT | TTE_PRIV_INT | rw; 465 ctip->dst->va_tag = ((uintptr_t)vaddr & MMU_PAGEMASK); 466 ctip->dst->index = ctip->index--; 467 ctip->dst->tmp = 1; 468 ctip->dst++; 469 } 470 471 472 static void 473 i_cpr_xcall(xcfunc_t func) 474 { 475 uint_t pil, reset_pil; 476 477 pil = getpil(); 478 if (pil < XCALL_PIL) 479 reset_pil = 0; 480 else { 481 reset_pil = 1; 482 setpil(XCALL_PIL - 1); 483 } 484 xc_some(cpu_ready_set, func, 0, 0); 485 if (reset_pil) 486 setpil(pil); 487 } 488 489 490 /* 491 * restart paused slave cpus 492 */ 493 void 494 i_cpr_machdep_setup(void) 495 { 496 if (ncpus > 1) { 497 CPR_DEBUG(CPR_DEBUG1, "MP restarted...\n"); 498 mutex_enter(&cpu_lock); 499 start_cpus(); 500 mutex_exit(&cpu_lock); 501 } 502 } 503 504 505 /* 506 * Stop all interrupt activities in the system 507 */ 508 void 509 i_cpr_stop_intr(void) 510 { 511 (void) spl7(); 512 } 513 514 /* 515 * Set machine up to take interrupts 516 */ 517 void 518 i_cpr_enable_intr(void) 519 { 520 (void) spl0(); 521 } 522 523 524 /* 525 * record cpu nodes and ids 526 */ 527 static void 528 i_cpr_save_cpu_info(void) 529 { 530 struct sun4u_cpu_info *scip; 531 cpu_t *cp; 532 533 scip = m_info.sci; 534 cp = CPU; 535 do { 536 ASSERT(scip < &m_info.sci[NCPU]); 537 scip->cpu_id = cp->cpu_id; 538 scip->node = cpunodes[cp->cpu_id].nodeid; 539 scip++; 540 } while ((cp = cp->cpu_next) != CPU); 541 } 542 543 544 /* 545 * Write necessary machine dependent information to cpr state file, 546 * eg. sun4u mmu ctx secondary for the current running process (cpr) ... 547 */ 548 int 549 i_cpr_write_machdep(vnode_t *vp) 550 { 551 extern uint_t getpstate(), getwstate(); 552 extern uint_t i_cpr_tstack_size; 553 const char ustr[] = ": unix-tte 2drop false ;"; 554 uintptr_t tinfo; 555 label_t *ltp; 556 cmd_t cmach; 557 char *fmt; 558 int rc; 559 560 /* 561 * ustr[] is used as temporary forth words during 562 * slave startup sequence, see sfmmu_mp_startup() 563 */ 564 565 cmach.md_magic = (uint_t)CPR_MACHDEP_MAGIC; 566 cmach.md_size = sizeof (m_info) + sizeof (ustr); 567 568 if (rc = cpr_write(vp, (caddr_t)&cmach, sizeof (cmach))) { 569 cpr_err(CE_WARN, "Failed to write descriptor."); 570 return (rc); 571 } 572 573 /* 574 * m_info is now cleared in i_cpr_dump_setup() 575 */ 576 m_info.ksb = (uint32_t)STACK_BIAS; 577 m_info.kpstate = (uint16_t)getpstate(); 578 m_info.kwstate = (uint16_t)getwstate(); 579 CPR_DEBUG(CPR_DEBUG1, "stack bias 0x%x, pstate 0x%x, wstate 0x%x\n", 580 m_info.ksb, m_info.kpstate, m_info.kwstate); 581 582 ltp = &ttolwp(curthread)->lwp_qsav; 583 m_info.qsav_pc = (cpr_ext)ltp->val[0]; 584 m_info.qsav_sp = (cpr_ext)ltp->val[1]; 585 586 /* 587 * Set secondary context to INVALID_CONTEXT to force the HAT 588 * to re-setup the MMU registers and locked TTEs it needs for 589 * TLB miss handling. 590 */ 591 m_info.mmu_ctx_sec = INVALID_CONTEXT; 592 m_info.mmu_ctx_pri = KCONTEXT; 593 594 tinfo = (uintptr_t)curthread; 595 m_info.thrp = (cpr_ptr)tinfo; 596 597 tinfo = (uintptr_t)i_cpr_resume_setup; 598 m_info.func = (cpr_ptr)tinfo; 599 600 /* 601 * i_cpr_data_page is comprised of a 4K stack area and a few 602 * trailing data symbols; the page is shared by the prom and 603 * kernel during resume. the stack size is recorded here 604 * and used by cprboot to set %sp 605 */ 606 tinfo = (uintptr_t)&i_cpr_data_page; 607 m_info.tmp_stack = (cpr_ptr)tinfo; 608 m_info.tmp_stacksize = i_cpr_tstack_size; 609 610 m_info.test_mode = cpr_test_mode; 611 612 i_cpr_save_cpu_info(); 613 614 if (rc = cpr_write(vp, (caddr_t)&m_info, sizeof (m_info))) { 615 cpr_err(CE_WARN, "Failed to write machdep info."); 616 return (rc); 617 } 618 619 fmt = "error writing %s forth info"; 620 if (rc = cpr_write(vp, (caddr_t)ustr, sizeof (ustr))) 621 cpr_err(CE_WARN, fmt, "unix-tte"); 622 623 return (rc); 624 } 625 626 627 /* 628 * Save miscellaneous information which needs to be written to the 629 * state file. This information is required to re-initialize 630 * kernel/prom handshaking. 631 */ 632 void 633 i_cpr_save_machdep_info(void) 634 { 635 CPR_DEBUG(CPR_DEBUG5, "jumpback size = 0x%lx\n", 636 (uintptr_t)&i_cpr_end_jumpback - 637 (uintptr_t)i_cpr_resume_setup); 638 639 /* 640 * Verify the jumpback code all falls in one page. 641 */ 642 if (((uintptr_t)&i_cpr_end_jumpback & MMU_PAGEMASK) != 643 ((uintptr_t)i_cpr_resume_setup & MMU_PAGEMASK)) 644 cpr_err(CE_PANIC, "jumpback code exceeds one page."); 645 } 646 647 648 /* 649 * cpu0 should contain bootcpu info 650 */ 651 cpu_t * 652 i_cpr_bootcpu(void) 653 { 654 return (&cpu0); 655 } 656 657 processorid_t 658 i_cpr_bootcpuid(void) 659 { 660 return (0); 661 } 662 663 /* 664 * Return the virtual address of the mapping area 665 */ 666 caddr_t 667 i_cpr_map_setup(void) 668 { 669 /* 670 * Allocate a virtual memory range spanned by an hmeblk. 671 * This would be 8 hments or 64k bytes. Starting VA 672 * must be 64k (8-page) aligned. 673 */ 674 cpr_vaddr = vmem_xalloc(heap_arena, 675 mmu_ptob(NHMENTS), mmu_ptob(NHMENTS), 676 0, 0, NULL, NULL, VM_NOSLEEP); 677 return (cpr_vaddr); 678 } 679 680 /* 681 * create tmp locked tlb entries for a group of phys pages; 682 * 683 * i_cpr_mapin/i_cpr_mapout should always be called in pairs, 684 * otherwise would fill up a tlb with locked entries 685 */ 686 void 687 i_cpr_mapin(caddr_t vaddr, uint_t pages, pfn_t ppn) 688 { 689 tte_t tte; 690 extern pfn_t curthreadpfn; 691 extern int curthreadremapped; 692 693 curthreadremapped = (ppn <= curthreadpfn && curthreadpfn < ppn + pages); 694 695 for (; pages--; ppn++, vaddr += MMU_PAGESIZE) { 696 tte.tte_inthi = TTE_VALID_INT | TTE_PFN_INTHI(ppn); 697 tte.tte_intlo = TTE_PFN_INTLO(ppn) | TTE_LCK_INT | 698 TTE_CP_INT | TTE_PRIV_INT | TTE_HWWR_INT; 699 sfmmu_dtlb_ld_kva(vaddr, &tte); 700 } 701 } 702 703 void 704 i_cpr_mapout(caddr_t vaddr, uint_t pages) 705 { 706 extern int curthreadremapped; 707 708 if (curthreadremapped && vaddr <= (caddr_t)curthread && 709 (caddr_t)curthread < vaddr + pages * MMU_PAGESIZE) 710 curthreadremapped = 0; 711 712 for (; pages--; vaddr += MMU_PAGESIZE) 713 vtag_flushpage(vaddr, (uint64_t)ksfmmup); 714 } 715 716 /* 717 * We're done using the mapping area; release virtual space 718 */ 719 void 720 i_cpr_map_destroy(void) 721 { 722 vmem_free(heap_arena, cpr_vaddr, mmu_ptob(NHMENTS)); 723 cpr_vaddr = NULL; 724 } 725 726 /* ARGSUSED */ 727 void 728 i_cpr_handle_xc(int flag) 729 { 730 } 731 732 733 /* 734 * This function takes care of pages which are not in kas or need to be 735 * taken care of in a special way. For example, panicbuf pages are not 736 * in kas and their pages are allocated via prom_retain(). 737 */ 738 pgcnt_t 739 i_cpr_count_special_kpages(int mapflag, bitfunc_t bitfunc) 740 { 741 struct cpr_map_info *pri, *tail; 742 pgcnt_t pages, total = 0; 743 pfn_t pfn; 744 745 /* 746 * Save information about prom retained panicbuf pages 747 */ 748 if (bitfunc == cpr_setbit) { 749 pri = &cpr_prom_retain[CPR_PANICBUF]; 750 pri->virt = (cpr_ptr)panicbuf; 751 pri->phys = va_to_pa(panicbuf); 752 pri->size = sizeof (panicbuf); 753 } 754 755 /* 756 * Go through the prom_retain array to tag those pages. 757 */ 758 tail = &cpr_prom_retain[CPR_PROM_RETAIN_CNT]; 759 for (pri = cpr_prom_retain; pri < tail; pri++) { 760 pages = mmu_btopr(pri->size); 761 for (pfn = ADDR_TO_PN(pri->phys); pages--; pfn++) { 762 if (pf_is_memory(pfn)) { 763 if (bitfunc == cpr_setbit) { 764 if ((*bitfunc)(pfn, mapflag) == 0) 765 total++; 766 } else 767 total++; 768 } 769 } 770 } 771 772 return (total); 773 } 774 775 776 /* 777 * Free up memory-related resources here. We start by freeing buffers 778 * allocated during suspend initialization. Also, free up the mapping 779 * resources allocated in cpr_init(). 780 */ 781 void 782 i_cpr_free_memory_resources(void) 783 { 784 (void) i_cpr_prom_pages(CPR_PROM_FREE); 785 i_cpr_map_destroy(); 786 i_cpr_storage_free(); 787 } 788 789 790 /* 791 * Derived from cpr_write_statefile(). 792 * Save the sensitive pages to the storage area and do bookkeeping 793 * using the sensitive descriptors. Each descriptor will contain no more 794 * than CPR_MAXCONTIG amount of contiguous pages to match the max amount 795 * of pages that statefile gets written to disk at each write. 796 * XXX The CPR_MAXCONTIG can be changed to the size of the compression 797 * scratch area. 798 */ 799 static int 800 i_cpr_save_to_storage(void) 801 { 802 sensitive_size_saved = 0; 803 sensitive_pages_saved = 0; 804 sensitive_write_ptr = i_cpr_storage_data_base; 805 return (cpr_contig_pages(NULL, SAVE_TO_STORAGE)); 806 } 807 808 809 /* 810 * This routine allocates space to save the sensitive kernel pages, 811 * i.e. kernel data nucleus, kvalloc and kvseg segments. 812 * It's assumed that those segments are the only areas that can be 813 * contaminated by memory allocations during statefile dumping. 814 * The space allocated here contains: 815 * A list of descriptors describing the saved sensitive pages. 816 * The storage area for saving the compressed sensitive kernel pages. 817 * Since storage pages are allocated from segkmem, they need to be 818 * excluded when saving. 819 */ 820 int 821 i_cpr_save_sensitive_kpages(void) 822 { 823 static const char pages_fmt[] = "\n%s %s allocs\n" 824 " spages %ld, vpages %ld, diff %ld\n"; 825 int retry_cnt; 826 int error = 0; 827 pgcnt_t pages, spages, vpages; 828 caddr_t addr; 829 char *str; 830 831 /* 832 * Tag sensitive kpages. Allocate space for storage descriptors 833 * and storage data area based on the resulting bitmaps. 834 * Note: The storage space will be part of the sensitive 835 * segment, so we need to tag kpages here before the storage 836 * is actually allocated just so their space won't be accounted 837 * for. They will not be part of the statefile although those 838 * pages will be claimed by cprboot. 839 */ 840 cpr_clear_bitmaps(); 841 842 spages = i_cpr_count_sensitive_kpages(REGULAR_BITMAP, cpr_setbit); 843 vpages = cpr_count_volatile_pages(REGULAR_BITMAP, cpr_clrbit); 844 pages = spages - vpages; 845 846 str = "i_cpr_save_sensitive_kpages:"; 847 CPR_DEBUG(CPR_DEBUG7, pages_fmt, "before", str, spages, vpages, pages); 848 849 /* 850 * Allocate space to save the clean sensitive kpages 851 */ 852 for (retry_cnt = 0; retry_cnt < MAX_STORAGE_ALLOC_RETRY; retry_cnt++) { 853 /* 854 * Alloc on first pass or realloc if we are retrying because 855 * of insufficient storage for sensitive pages 856 */ 857 if (retry_cnt == 0 || error == ENOMEM) { 858 if (i_cpr_storage_data_base) { 859 kmem_free(i_cpr_storage_data_base, 860 mmu_ptob(i_cpr_storage_data_sz)); 861 i_cpr_storage_data_base = NULL; 862 i_cpr_storage_data_sz = 0; 863 } 864 addr = i_cpr_storage_data_alloc(pages, 865 &i_cpr_storage_data_sz, retry_cnt); 866 if (addr == NULL) { 867 CPR_DEBUG(CPR_DEBUG7, 868 "\n%s can't allocate data storage space!\n", 869 str); 870 return (ENOMEM); 871 } 872 i_cpr_storage_data_base = addr; 873 i_cpr_storage_data_end = 874 addr + mmu_ptob(i_cpr_storage_data_sz); 875 } 876 877 /* 878 * Allocate on first pass, only realloc if retry is because of 879 * insufficient descriptors, but reset contents on each pass 880 * (desc_alloc resets contents as well) 881 */ 882 if (retry_cnt == 0 || error == -1) { 883 error = i_cpr_storage_desc_alloc( 884 &i_cpr_storage_desc_base, &i_cpr_storage_desc_pgcnt, 885 &i_cpr_storage_desc_end, retry_cnt); 886 if (error != 0) 887 return (error); 888 } else { 889 i_cpr_storage_desc_init(i_cpr_storage_desc_base, 890 i_cpr_storage_desc_pgcnt, i_cpr_storage_desc_end); 891 } 892 893 /* 894 * We are ready to save the sensitive kpages to storage. 895 * We cannot trust what's tagged in the bitmaps anymore 896 * after storage allocations. Clear up the bitmaps and 897 * retag the sensitive kpages again. The storage pages 898 * should be untagged. 899 */ 900 cpr_clear_bitmaps(); 901 902 spages = 903 i_cpr_count_sensitive_kpages(REGULAR_BITMAP, cpr_setbit); 904 vpages = cpr_count_volatile_pages(REGULAR_BITMAP, cpr_clrbit); 905 906 CPR_DEBUG(CPR_DEBUG7, pages_fmt, "after ", str, 907 spages, vpages, spages - vpages); 908 909 /* 910 * Returns 0 on success, -1 if too few descriptors, and 911 * ENOMEM if not enough space to save sensitive pages 912 */ 913 CPR_DEBUG(CPR_DEBUG1, "compressing pages to storage...\n"); 914 error = i_cpr_save_to_storage(); 915 if (error == 0) { 916 /* Saving to storage succeeded */ 917 CPR_DEBUG(CPR_DEBUG1, "compressed %d pages\n", 918 sensitive_pages_saved); 919 break; 920 } else if (error == -1) 921 CPR_DEBUG(CPR_DEBUG1, "%s too few descriptors\n", str); 922 } 923 if (error == -1) 924 error = ENOMEM; 925 return (error); 926 } 927 928 929 /* 930 * Estimate how much memory we will need to save 931 * the sensitive pages with compression. 932 */ 933 static caddr_t 934 i_cpr_storage_data_alloc(pgcnt_t pages, pgcnt_t *alloc_pages, int retry_cnt) 935 { 936 pgcnt_t alloc_pcnt, last_pcnt; 937 caddr_t addr; 938 char *str; 939 940 str = "i_cpr_storage_data_alloc:"; 941 if (retry_cnt == 0) { 942 /* 943 * common compression ratio is about 3:1 944 * initial storage allocation is estimated at 40% 945 * to cover the majority of cases 946 */ 947 alloc_pcnt = INITIAL_ALLOC_PCNT; 948 *alloc_pages = (pages * alloc_pcnt) / INTEGRAL; 949 CPR_DEBUG(CPR_DEBUG7, "%s sensitive pages: %ld\n", str, pages); 950 CPR_DEBUG(CPR_DEBUG7, 951 "%s initial est pages: %ld, alloc %ld%%\n", 952 str, *alloc_pages, alloc_pcnt); 953 } else { 954 /* 955 * calculate the prior compression percentage (x100) 956 * from the last attempt to save sensitive pages 957 */ 958 ASSERT(sensitive_pages_saved != 0); 959 last_pcnt = (mmu_btopr(sensitive_size_saved) * INTEGRAL) / 960 sensitive_pages_saved; 961 CPR_DEBUG(CPR_DEBUG7, "%s last ratio %ld%%\n", str, last_pcnt); 962 963 /* 964 * new estimated storage size is based on 965 * the larger ratio + 5% for each retry: 966 * pages * (last + [5%, 10%]) 967 */ 968 alloc_pcnt = MAX(last_pcnt, INITIAL_ALLOC_PCNT) + 969 (retry_cnt * 5); 970 *alloc_pages = (pages * alloc_pcnt) / INTEGRAL; 971 CPR_DEBUG(CPR_DEBUG7, "%s Retry est pages: %ld, alloc %ld%%\n", 972 str, *alloc_pages, alloc_pcnt); 973 } 974 975 addr = kmem_alloc(mmu_ptob(*alloc_pages), KM_NOSLEEP); 976 CPR_DEBUG(CPR_DEBUG7, "%s alloc %ld pages\n", str, *alloc_pages); 977 return (addr); 978 } 979 980 981 void 982 i_cpr_storage_free(void) 983 { 984 /* Free descriptors */ 985 if (i_cpr_storage_desc_base) { 986 kmem_free(i_cpr_storage_desc_base, 987 mmu_ptob(i_cpr_storage_desc_pgcnt)); 988 i_cpr_storage_desc_base = NULL; 989 i_cpr_storage_desc_pgcnt = 0; 990 } 991 992 993 /* Data storage */ 994 if (i_cpr_storage_data_base) { 995 kmem_free(i_cpr_storage_data_base, 996 mmu_ptob(i_cpr_storage_data_sz)); 997 i_cpr_storage_data_base = NULL; 998 i_cpr_storage_data_sz = 0; 999 } 1000 } 1001 1002 1003 /* 1004 * This routine is derived from cpr_compress_and_write(). 1005 * 1. Do bookkeeping in the descriptor for the contiguous sensitive chunk. 1006 * 2. Compress and save the clean sensitive pages into the storage area. 1007 */ 1008 int 1009 i_cpr_compress_and_save(int chunks, pfn_t spfn, pgcnt_t pages) 1010 { 1011 extern char *cpr_compress_pages(cpd_t *, pgcnt_t, int); 1012 extern caddr_t i_cpr_storage_data_end; 1013 uint_t remaining, datalen; 1014 uint32_t test_usum; 1015 char *datap; 1016 csd_t *descp; 1017 cpd_t cpd; 1018 int error; 1019 1020 /* 1021 * Fill next empty storage descriptor 1022 */ 1023 descp = i_cpr_storage_desc_base + chunks - 1; 1024 if (descp >= i_cpr_storage_desc_end) { 1025 CPR_DEBUG(CPR_DEBUG1, "ran out of descriptors, base 0x%p, " 1026 "chunks %d, end 0x%p, descp 0x%p\n", 1027 i_cpr_storage_desc_base, chunks, 1028 i_cpr_storage_desc_end, descp); 1029 return (-1); 1030 } 1031 ASSERT(descp->csd_dirty_spfn == (uint_t)-1); 1032 i_cpr_storage_desc_last_used = descp; 1033 1034 descp->csd_dirty_spfn = spfn; 1035 descp->csd_dirty_npages = pages; 1036 1037 i_cpr_mapin(CPR->c_mapping_area, pages, spfn); 1038 1039 /* 1040 * try compressing pages and copy cpd fields 1041 * pfn is copied for debug use 1042 */ 1043 cpd.cpd_pfn = spfn; 1044 datap = cpr_compress_pages(&cpd, pages, C_COMPRESSING); 1045 datalen = cpd.cpd_length; 1046 descp->csd_clean_compressed = (cpd.cpd_flag & CPD_COMPRESS); 1047 #ifdef DEBUG 1048 descp->csd_usum = cpd.cpd_usum; 1049 descp->csd_csum = cpd.cpd_csum; 1050 #endif 1051 1052 error = 0; 1053 1054 /* 1055 * Save the raw or compressed data to the storage area pointed to by 1056 * sensitive_write_ptr. Make sure the storage space is big enough to 1057 * hold the result. Otherwise roll back to increase the storage space. 1058 */ 1059 descp->csd_clean_sva = (cpr_ptr)sensitive_write_ptr; 1060 descp->csd_clean_sz = datalen; 1061 if ((sensitive_write_ptr + datalen) < i_cpr_storage_data_end) { 1062 extern void cprbcopy(void *, void *, size_t); 1063 1064 cprbcopy(datap, sensitive_write_ptr, datalen); 1065 sensitive_size_saved += datalen; 1066 sensitive_pages_saved += descp->csd_dirty_npages; 1067 sensitive_write_ptr += datalen; 1068 } else { 1069 remaining = (i_cpr_storage_data_end - sensitive_write_ptr); 1070 CPR_DEBUG(CPR_DEBUG1, "i_cpr_compress_and_save: The storage " 1071 "space is too small!\ngot %d, want %d\n\n", 1072 remaining, (remaining + datalen)); 1073 #ifdef DEBUG 1074 /* 1075 * Check to see if the content of the sensitive pages that we 1076 * just copied have changed during this small time window. 1077 */ 1078 test_usum = checksum32(CPR->c_mapping_area, mmu_ptob(pages)); 1079 descp->csd_usum = cpd.cpd_usum; 1080 if (test_usum != descp->csd_usum) { 1081 CPR_DEBUG(CPR_DEBUG1, "\nWARNING: " 1082 "i_cpr_compress_and_save: " 1083 "Data in the range of pfn 0x%lx to pfn " 1084 "0x%lx has changed after they are saved " 1085 "into storage.", spfn, (spfn + pages - 1)); 1086 } 1087 #endif 1088 error = ENOMEM; 1089 } 1090 1091 i_cpr_mapout(CPR->c_mapping_area, pages); 1092 return (error); 1093 } 1094 1095 1096 /* 1097 * This routine is derived from cpr_count_kpages(). 1098 * It goes through kernel data nucleus and segkmem segments to select 1099 * pages in use and mark them in the corresponding bitmap. 1100 */ 1101 pgcnt_t 1102 i_cpr_count_sensitive_kpages(int mapflag, bitfunc_t bitfunc) 1103 { 1104 pgcnt_t kdata_cnt = 0, segkmem_cnt = 0; 1105 extern caddr_t e_moddata; 1106 extern struct seg kvalloc; 1107 extern struct seg kmem64; 1108 size_t size; 1109 1110 /* 1111 * Kernel data nucleus pages 1112 */ 1113 size = e_moddata - s_data; 1114 kdata_cnt += cpr_count_pages(s_data, size, 1115 mapflag, bitfunc, DBG_SHOWRANGE); 1116 1117 /* 1118 * kvseg and kvalloc pages 1119 */ 1120 segkmem_cnt += cpr_scan_kvseg(mapflag, bitfunc, &kvseg); 1121 segkmem_cnt += cpr_count_pages(kvalloc.s_base, kvalloc.s_size, 1122 mapflag, bitfunc, DBG_SHOWRANGE); 1123 1124 /* segment to support kernel memory usage above 32-bit space (4GB) */ 1125 if (kmem64.s_base) 1126 segkmem_cnt += cpr_count_pages(kmem64.s_base, kmem64.s_size, 1127 mapflag, bitfunc, DBG_SHOWRANGE); 1128 1129 CPR_DEBUG(CPR_DEBUG7, "\ni_cpr_count_sensitive_kpages:\n" 1130 "\tkdata_cnt %ld + segkmem_cnt %ld = %ld pages\n", 1131 kdata_cnt, segkmem_cnt, kdata_cnt + segkmem_cnt); 1132 1133 return (kdata_cnt + segkmem_cnt); 1134 } 1135 1136 1137 pgcnt_t 1138 i_cpr_count_storage_pages(int mapflag, bitfunc_t bitfunc) 1139 { 1140 pgcnt_t count = 0; 1141 1142 if (i_cpr_storage_desc_base) { 1143 count += cpr_count_pages((caddr_t)i_cpr_storage_desc_base, 1144 (size_t)mmu_ptob(i_cpr_storage_desc_pgcnt), 1145 mapflag, bitfunc, DBG_SHOWRANGE); 1146 } 1147 if (i_cpr_storage_data_base) { 1148 count += cpr_count_pages(i_cpr_storage_data_base, 1149 (size_t)mmu_ptob(i_cpr_storage_data_sz), 1150 mapflag, bitfunc, DBG_SHOWRANGE); 1151 } 1152 return (count); 1153 } 1154 1155 1156 /* 1157 * Derived from cpr_write_statefile(). 1158 * Allocate (or reallocate after exhausting the supply) descriptors for each 1159 * chunk of contiguous sensitive kpages. 1160 */ 1161 static int 1162 i_cpr_storage_desc_alloc(csd_t **basepp, pgcnt_t *pgsp, csd_t **endpp, 1163 int retry) 1164 { 1165 pgcnt_t npages; 1166 int chunks; 1167 csd_t *descp, *end; 1168 size_t len; 1169 char *str = "i_cpr_storage_desc_alloc:"; 1170 1171 /* 1172 * On initial allocation, add some extra to cover overhead caused 1173 * by the allocation for the storage area later. 1174 */ 1175 if (retry == 0) { 1176 chunks = cpr_contig_pages(NULL, STORAGE_DESC_ALLOC) + 1177 EXTRA_DESCS; 1178 npages = mmu_btopr(sizeof (**basepp) * (pgcnt_t)chunks); 1179 CPR_DEBUG(CPR_DEBUG7, "%s chunks %d, ", str, chunks); 1180 } else { 1181 CPR_DEBUG(CPR_DEBUG7, "%s retry %d: ", str, retry); 1182 npages = *pgsp + 1; 1183 } 1184 /* Free old descriptors, if any */ 1185 if (*basepp) 1186 kmem_free((caddr_t)*basepp, mmu_ptob(*pgsp)); 1187 1188 descp = *basepp = kmem_alloc(mmu_ptob(npages), KM_NOSLEEP); 1189 if (descp == NULL) { 1190 CPR_DEBUG(CPR_DEBUG7, "%s no space for descriptors!\n", str); 1191 return (ENOMEM); 1192 } 1193 1194 *pgsp = npages; 1195 len = mmu_ptob(npages); 1196 end = *endpp = descp + (len / (sizeof (**basepp))); 1197 CPR_DEBUG(CPR_DEBUG7, "npages 0x%lx, len 0x%lx, items 0x%lx\n\t*basepp " 1198 "%p, *endpp %p\n", npages, len, (len / (sizeof (**basepp))), 1199 *basepp, *endpp); 1200 i_cpr_storage_desc_init(descp, npages, end); 1201 return (0); 1202 } 1203 1204 static void 1205 i_cpr_storage_desc_init(csd_t *descp, pgcnt_t npages, csd_t *end) 1206 { 1207 size_t len = mmu_ptob(npages); 1208 1209 /* Initialize the descriptors to something impossible. */ 1210 bzero(descp, len); 1211 #ifdef DEBUG 1212 /* 1213 * This condition is tested by an ASSERT 1214 */ 1215 for (; descp < end; descp++) 1216 descp->csd_dirty_spfn = (uint_t)-1; 1217 #endif 1218 } 1219 1220 int 1221 i_cpr_dump_sensitive_kpages(vnode_t *vp) 1222 { 1223 int error = 0; 1224 uint_t spin_cnt = 0; 1225 csd_t *descp; 1226 1227 /* 1228 * These following two variables need to be reinitialized 1229 * for each cpr cycle. 1230 */ 1231 i_cpr_sensitive_bytes_dumped = 0; 1232 i_cpr_sensitive_pgs_dumped = 0; 1233 1234 if (i_cpr_storage_desc_base) { 1235 for (descp = i_cpr_storage_desc_base; 1236 descp <= i_cpr_storage_desc_last_used; descp++) { 1237 if (error = cpr_dump_sensitive(vp, descp)) 1238 return (error); 1239 spin_cnt++; 1240 if ((spin_cnt & 0x5F) == 1) 1241 cpr_spinning_bar(); 1242 } 1243 prom_printf(" \b"); 1244 } 1245 1246 CPR_DEBUG(CPR_DEBUG7, "\ni_cpr_dump_sensitive_kpages: dumped %ld\n", 1247 i_cpr_sensitive_pgs_dumped); 1248 return (0); 1249 } 1250 1251 1252 /* 1253 * 1. Fill the cpr page descriptor with the info of the dirty pages 1254 * and 1255 * write the descriptor out. It will be used at resume. 1256 * 2. Write the clean data in stead of the dirty data out. 1257 * Note: to save space, the clean data is already compressed. 1258 */ 1259 static int 1260 cpr_dump_sensitive(vnode_t *vp, csd_t *descp) 1261 { 1262 int error = 0; 1263 caddr_t datap; 1264 cpd_t cpd; /* cpr page descriptor */ 1265 pfn_t dirty_spfn; 1266 pgcnt_t dirty_npages; 1267 size_t clean_sz; 1268 caddr_t clean_sva; 1269 int clean_compressed; 1270 extern uchar_t cpr_pagecopy[]; 1271 1272 dirty_spfn = descp->csd_dirty_spfn; 1273 dirty_npages = descp->csd_dirty_npages; 1274 clean_sva = (caddr_t)descp->csd_clean_sva; 1275 clean_sz = descp->csd_clean_sz; 1276 clean_compressed = descp->csd_clean_compressed; 1277 1278 /* Fill cpr page descriptor. */ 1279 cpd.cpd_magic = (uint_t)CPR_PAGE_MAGIC; 1280 cpd.cpd_pfn = dirty_spfn; 1281 cpd.cpd_flag = 0; /* must init to zero */ 1282 cpd.cpd_pages = dirty_npages; 1283 1284 #ifdef DEBUG 1285 if ((cpd.cpd_usum = descp->csd_usum) != 0) 1286 cpd.cpd_flag |= CPD_USUM; 1287 if ((cpd.cpd_csum = descp->csd_csum) != 0) 1288 cpd.cpd_flag |= CPD_CSUM; 1289 #endif 1290 1291 STAT->cs_dumped_statefsz += mmu_ptob(dirty_npages); 1292 1293 /* 1294 * The sensitive kpages are usually saved with compression 1295 * unless compression could not reduce the size of the data. 1296 * If user choose not to have the statefile compressed, 1297 * we need to decompress the data back before dumping it to disk. 1298 */ 1299 if (CPR->c_flags & C_COMPRESSING) { 1300 cpd.cpd_length = clean_sz; 1301 datap = clean_sva; 1302 if (clean_compressed) 1303 cpd.cpd_flag |= CPD_COMPRESS; 1304 } else { 1305 if (clean_compressed) { 1306 cpd.cpd_length = decompress(clean_sva, cpr_pagecopy, 1307 clean_sz, mmu_ptob(dirty_npages)); 1308 datap = (caddr_t)cpr_pagecopy; 1309 ASSERT(cpd.cpd_length == mmu_ptob(dirty_npages)); 1310 } else { 1311 cpd.cpd_length = clean_sz; 1312 datap = clean_sva; 1313 } 1314 cpd.cpd_csum = 0; 1315 } 1316 1317 /* Write cpr page descriptor */ 1318 error = cpr_write(vp, (caddr_t)&cpd, sizeof (cpd)); 1319 if (error) { 1320 CPR_DEBUG(CPR_DEBUG7, "descp: %p\n", descp); 1321 #ifdef DEBUG 1322 debug_enter("cpr_dump_sensitive: cpr_write() page " 1323 "descriptor failed!\n"); 1324 #endif 1325 return (error); 1326 } 1327 1328 i_cpr_sensitive_bytes_dumped += sizeof (cpd_t); 1329 1330 /* Write page data */ 1331 error = cpr_write(vp, (caddr_t)datap, cpd.cpd_length); 1332 if (error) { 1333 CPR_DEBUG(CPR_DEBUG7, "error: %x\n", error); 1334 CPR_DEBUG(CPR_DEBUG7, "descp: %p\n", descp); 1335 CPR_DEBUG(CPR_DEBUG7, "cpr_write(%p, %p , %lx)\n", vp, datap, 1336 cpd.cpd_length); 1337 #ifdef DEBUG 1338 debug_enter("cpr_dump_sensitive: cpr_write() data failed!\n"); 1339 #endif 1340 return (error); 1341 } 1342 1343 i_cpr_sensitive_bytes_dumped += cpd.cpd_length; 1344 i_cpr_sensitive_pgs_dumped += dirty_npages; 1345 1346 return (error); 1347 } 1348 1349 1350 /* 1351 * Sanity check to make sure that we have dumped right amount 1352 * of pages from different sources to statefile. 1353 */ 1354 int 1355 i_cpr_check_pgs_dumped(uint_t pgs_expected, uint_t regular_pgs_dumped) 1356 { 1357 uint_t total_pgs_dumped; 1358 1359 total_pgs_dumped = regular_pgs_dumped + i_cpr_sensitive_pgs_dumped; 1360 1361 CPR_DEBUG(CPR_DEBUG7, "\ncheck_pgs: reg %d + sens %ld = %d, " 1362 "expect %d\n\n", regular_pgs_dumped, i_cpr_sensitive_pgs_dumped, 1363 total_pgs_dumped, pgs_expected); 1364 1365 if (pgs_expected == total_pgs_dumped) 1366 return (0); 1367 1368 return (EINVAL); 1369 } 1370 1371 1372 int 1373 i_cpr_reusefini(void) 1374 { 1375 struct vnode *vp; 1376 cdef_t *cdef; 1377 size_t size; 1378 char *bufp; 1379 int rc; 1380 1381 if (cpr_reusable_mode) 1382 cpr_reusable_mode = 0; 1383 1384 if (rc = cpr_open_deffile(FREAD|FWRITE, &vp)) { 1385 if (rc == EROFS) { 1386 cpr_err(CE_CONT, "uadmin A_FREEZE AD_REUSEFINI " 1387 "(uadmin %d %d)\nmust be done with / mounted " 1388 "writeable.\n", A_FREEZE, AD_REUSEFINI); 1389 } 1390 return (rc); 1391 } 1392 1393 cdef = kmem_alloc(sizeof (*cdef), KM_SLEEP); 1394 rc = cpr_rdwr(UIO_READ, vp, cdef, sizeof (*cdef)); 1395 1396 if (rc) { 1397 cpr_err(CE_WARN, "Failed reading %s, errno = %d", 1398 cpr_default_path, rc); 1399 } else if (cdef->mini.magic != CPR_DEFAULT_MAGIC) { 1400 cpr_err(CE_WARN, "bad magic number in %s, cannot restore " 1401 "prom values for %s", cpr_default_path, 1402 cpr_enumerate_promprops(&bufp, &size)); 1403 kmem_free(bufp, size); 1404 rc = EINVAL; 1405 } else { 1406 /* 1407 * clean up prom properties 1408 */ 1409 rc = cpr_update_nvram(cdef->props); 1410 if (rc == 0) { 1411 /* 1412 * invalidate the disk copy and turn off reusable 1413 */ 1414 cdef->mini.magic = 0; 1415 cdef->mini.reusable = 0; 1416 if (rc = cpr_rdwr(UIO_WRITE, vp, 1417 &cdef->mini, sizeof (cdef->mini))) { 1418 cpr_err(CE_WARN, "Failed writing %s, errno %d", 1419 cpr_default_path, rc); 1420 } 1421 } 1422 } 1423 1424 (void) VOP_CLOSE(vp, FREAD|FWRITE, 1, (offset_t)0, CRED(), NULL); 1425 VN_RELE(vp); 1426 kmem_free(cdef, sizeof (*cdef)); 1427 1428 return (rc); 1429 } 1430 1431 1432 int 1433 i_cpr_reuseinit(void) 1434 { 1435 int rc = 0; 1436 1437 if (rc = cpr_default_setup(1)) 1438 return (rc); 1439 1440 /* 1441 * We need to validate default file 1442 */ 1443 rc = cpr_validate_definfo(1); 1444 if (rc == 0) 1445 cpr_reusable_mode = 1; 1446 else if (rc == EROFS) { 1447 cpr_err(CE_NOTE, "reuseinit must be performed " 1448 "while / is mounted writeable"); 1449 } 1450 1451 (void) cpr_default_setup(0); 1452 1453 return (rc); 1454 } 1455 1456 1457 int 1458 i_cpr_check_cprinfo(void) 1459 { 1460 struct vnode *vp; 1461 cmini_t mini; 1462 int rc = 0; 1463 1464 if (rc = cpr_open_deffile(FREAD, &vp)) { 1465 if (rc == ENOENT) 1466 cpr_err(CE_NOTE, "cprinfo file does not " 1467 "exist. You must run 'uadmin %d %d' " 1468 "command while / is mounted writeable,\n" 1469 "then reboot and run 'uadmin %d %d' " 1470 "to create a reusable statefile", 1471 A_FREEZE, AD_REUSEINIT, A_FREEZE, AD_REUSABLE); 1472 return (rc); 1473 } 1474 1475 rc = cpr_rdwr(UIO_READ, vp, &mini, sizeof (mini)); 1476 (void) VOP_CLOSE(vp, FREAD, 1, (offset_t)0, CRED(), NULL); 1477 VN_RELE(vp); 1478 1479 if (rc) { 1480 cpr_err(CE_WARN, "Failed reading %s, errno = %d", 1481 cpr_default_path, rc); 1482 } else if (mini.magic != CPR_DEFAULT_MAGIC) { 1483 cpr_err(CE_CONT, "bad magic number in cprinfo file.\n" 1484 "You must run 'uadmin %d %d' while / is mounted " 1485 "writeable, then reboot and run 'uadmin %d %d' " 1486 "to create a reusable statefile\n", 1487 A_FREEZE, AD_REUSEINIT, A_FREEZE, AD_REUSABLE); 1488 rc = EINVAL; 1489 } 1490 1491 return (rc); 1492 } 1493 1494 1495 int 1496 i_cpr_reusable_supported(void) 1497 { 1498 return (1); 1499 } 1500 1501 1502 /* 1503 * find prom phys pages and alloc space for a tmp copy 1504 */ 1505 static int 1506 i_cpr_find_ppages(void) 1507 { 1508 extern struct vnode prom_ppages; 1509 struct page *pp; 1510 struct memlist *pmem; 1511 pgcnt_t npages, pcnt, scnt, vcnt; 1512 pfn_t ppn, plast, *dst; 1513 int mapflag; 1514 1515 cpr_clear_bitmaps(); 1516 mapflag = REGULAR_BITMAP; 1517 1518 /* 1519 * there should be a page_t for each phys page used by the kernel; 1520 * set a bit for each phys page not tracked by a page_t 1521 */ 1522 pcnt = 0; 1523 memlist_read_lock(); 1524 for (pmem = phys_install; pmem; pmem = pmem->next) { 1525 npages = mmu_btop(pmem->size); 1526 ppn = mmu_btop(pmem->address); 1527 for (plast = ppn + npages; ppn < plast; ppn++) { 1528 if (page_numtopp_nolock(ppn)) 1529 continue; 1530 (void) cpr_setbit(ppn, mapflag); 1531 pcnt++; 1532 } 1533 } 1534 memlist_read_unlock(); 1535 1536 /* 1537 * clear bits for phys pages in each segment 1538 */ 1539 scnt = cpr_count_seg_pages(mapflag, cpr_clrbit); 1540 1541 /* 1542 * set bits for phys pages referenced by the prom_ppages vnode; 1543 * these pages are mostly comprised of forthdebug words 1544 */ 1545 vcnt = 0; 1546 for (pp = prom_ppages.v_pages; pp; ) { 1547 if (cpr_setbit(pp->p_offset, mapflag) == 0) 1548 vcnt++; 1549 pp = pp->p_vpnext; 1550 if (pp == prom_ppages.v_pages) 1551 break; 1552 } 1553 1554 /* 1555 * total number of prom pages are: 1556 * (non-page_t pages - seg pages + vnode pages) 1557 */ 1558 ppage_count = pcnt - scnt + vcnt; 1559 CPR_DEBUG(CPR_DEBUG1, 1560 "find_ppages: pcnt %ld - scnt %ld + vcnt %ld = %ld\n", 1561 pcnt, scnt, vcnt, ppage_count); 1562 1563 /* 1564 * alloc array of pfn_t to store phys page list 1565 */ 1566 pphys_list_size = ppage_count * sizeof (pfn_t); 1567 pphys_list = kmem_alloc(pphys_list_size, KM_NOSLEEP); 1568 if (pphys_list == NULL) { 1569 cpr_err(CE_WARN, "cannot alloc pphys_list"); 1570 return (ENOMEM); 1571 } 1572 1573 /* 1574 * phys pages referenced in the bitmap should be 1575 * those used by the prom; scan bitmap and save 1576 * a list of prom phys page numbers 1577 */ 1578 dst = pphys_list; 1579 memlist_read_lock(); 1580 for (pmem = phys_install; pmem; pmem = pmem->next) { 1581 npages = mmu_btop(pmem->size); 1582 ppn = mmu_btop(pmem->address); 1583 for (plast = ppn + npages; ppn < plast; ppn++) { 1584 if (cpr_isset(ppn, mapflag)) { 1585 ASSERT(dst < (pphys_list + ppage_count)); 1586 *dst++ = ppn; 1587 } 1588 } 1589 } 1590 memlist_read_unlock(); 1591 1592 /* 1593 * allocate space to store prom pages 1594 */ 1595 ppage_buf = kmem_alloc(mmu_ptob(ppage_count), KM_NOSLEEP); 1596 if (ppage_buf == NULL) { 1597 kmem_free(pphys_list, pphys_list_size); 1598 pphys_list = NULL; 1599 cpr_err(CE_WARN, "cannot alloc ppage_buf"); 1600 return (ENOMEM); 1601 } 1602 1603 return (0); 1604 } 1605 1606 1607 /* 1608 * save prom pages to kmem pages 1609 */ 1610 static void 1611 i_cpr_save_ppages(void) 1612 { 1613 pfn_t *pphys, *plast; 1614 caddr_t dst; 1615 1616 /* 1617 * map in each prom page and copy to a kmem page 1618 */ 1619 dst = ppage_buf; 1620 plast = pphys_list + ppage_count; 1621 for (pphys = pphys_list; pphys < plast; pphys++) { 1622 i_cpr_mapin(cpr_vaddr, 1, *pphys); 1623 bcopy(cpr_vaddr, dst, MMU_PAGESIZE); 1624 i_cpr_mapout(cpr_vaddr, 1); 1625 dst += MMU_PAGESIZE; 1626 } 1627 1628 CPR_DEBUG(CPR_DEBUG1, "saved %ld prom pages\n", ppage_count); 1629 } 1630 1631 1632 /* 1633 * restore prom pages from kmem pages 1634 */ 1635 static void 1636 i_cpr_restore_ppages(void) 1637 { 1638 pfn_t *pphys, *plast; 1639 caddr_t src; 1640 1641 dcache_flushall(); 1642 1643 /* 1644 * map in each prom page and copy from a kmem page 1645 */ 1646 src = ppage_buf; 1647 plast = pphys_list + ppage_count; 1648 for (pphys = pphys_list; pphys < plast; pphys++) { 1649 i_cpr_mapin(cpr_vaddr, 1, *pphys); 1650 bcopy(src, cpr_vaddr, MMU_PAGESIZE); 1651 i_cpr_mapout(cpr_vaddr, 1); 1652 src += MMU_PAGESIZE; 1653 } 1654 1655 dcache_flushall(); 1656 1657 CPR_DEBUG(CPR_DEBUG1, "restored %ld prom pages\n", ppage_count); 1658 } 1659 1660 1661 /* 1662 * save/restore prom pages or free related allocs 1663 */ 1664 int 1665 i_cpr_prom_pages(int action) 1666 { 1667 int error; 1668 1669 if (action == CPR_PROM_SAVE) { 1670 if (ppage_buf == NULL) { 1671 ASSERT(pphys_list == NULL); 1672 if (error = i_cpr_find_ppages()) 1673 return (error); 1674 i_cpr_save_ppages(); 1675 } 1676 } else if (action == CPR_PROM_RESTORE) { 1677 i_cpr_restore_ppages(); 1678 } else if (action == CPR_PROM_FREE) { 1679 if (pphys_list) { 1680 ASSERT(pphys_list_size); 1681 kmem_free(pphys_list, pphys_list_size); 1682 pphys_list = NULL; 1683 pphys_list_size = 0; 1684 } 1685 if (ppage_buf) { 1686 ASSERT(ppage_count); 1687 kmem_free(ppage_buf, mmu_ptob(ppage_count)); 1688 CPR_DEBUG(CPR_DEBUG1, "freed %ld prom pages\n", 1689 ppage_count); 1690 ppage_buf = NULL; 1691 ppage_count = 0; 1692 } 1693 } 1694 return (0); 1695 } 1696 1697 1698 /* 1699 * record tlb data for the nucleus, bigktsb's, and the cpr module; 1700 * this data is later used by cprboot to install dtlb/itlb entries. 1701 * when we jump into the cpr module during the resume phase, those 1702 * mappings are needed until switching to the kernel trap table. 1703 * to make the dtte/itte info available during resume, we need 1704 * the info recorded prior to saving sensitive pages, otherwise 1705 * all the data would appear as NULLs. 1706 */ 1707 static void 1708 i_cpr_save_tlbinfo(void) 1709 { 1710 cti_t cti = {0}; 1711 1712 /* 1713 * during resume - shortly after jumping into the cpr module, 1714 * sfmmu_load_mmustate() will overwrite any dtlb entry at any 1715 * index used for TSBs; skip is set so that any saved tte will 1716 * target other tlb offsets and prevent being lost during 1717 * resume. now scan the dtlb and save locked entries, 1718 * then add entries for the tmp stack / data page and the 1719 * cpr thread structure. 1720 */ 1721 cti.dst = m_info.dtte; 1722 cti.tail = cti.dst + CPR_MAX_TLB; 1723 cti.reader = dtlb_rd_entry; 1724 cti.writer = NULL; 1725 cti.filter = i_cpr_lnb; 1726 cti.index = cpunodes[CPU->cpu_id].dtlb_size - 1; 1727 1728 if (utsb_dtlb_ttenum != -1) 1729 cti.skip = (1 << utsb_dtlb_ttenum); 1730 1731 if (utsb4m_dtlb_ttenum != -1) 1732 cti.skip |= (1 << utsb4m_dtlb_ttenum); 1733 1734 i_cpr_scan_tlb(&cti); 1735 i_cpr_make_tte(&cti, &i_cpr_data_page, datava); 1736 i_cpr_make_tte(&cti, curthread, datava); 1737 1738 /* 1739 * scan itlb and save locked entries; add an entry for 1740 * the first text page of the cpr module; cprboot will 1741 * jump to that page after restoring kernel pages. 1742 */ 1743 cti.dst = m_info.itte; 1744 cti.tail = cti.dst + CPR_MAX_TLB; 1745 cti.reader = itlb_rd_entry; 1746 cti.index = cpunodes[CPU->cpu_id].itlb_size - 1; 1747 cti.skip = 0; 1748 i_cpr_scan_tlb(&cti); 1749 i_cpr_make_tte(&cti, (void *)i_cpr_resume_setup, textva); 1750 } 1751 1752 1753 /* ARGSUSED */ 1754 int 1755 i_cpr_dump_setup(vnode_t *vp) 1756 { 1757 /* 1758 * zero out m_info and add info to dtte/itte arrays 1759 */ 1760 bzero(&m_info, sizeof (m_info)); 1761 i_cpr_save_tlbinfo(); 1762 return (0); 1763 } 1764 1765 1766 int 1767 i_cpr_is_supported(int sleeptype) 1768 { 1769 char es_prop[] = "energystar-v2"; 1770 pnode_t node; 1771 int last; 1772 extern int cpr_supported_override; 1773 extern int cpr_platform_enable; 1774 1775 if (sleeptype != CPR_TODISK) 1776 return (0); 1777 1778 /* 1779 * The next statement tests if a specific platform has turned off 1780 * cpr support. 1781 */ 1782 if (cpr_supported_override) 1783 return (0); 1784 1785 /* 1786 * Do not inspect energystar-v* property if a platform has 1787 * specifically turned on cpr support 1788 */ 1789 if (cpr_platform_enable) 1790 return (1); 1791 1792 node = prom_rootnode(); 1793 if (prom_getproplen(node, es_prop) != -1) 1794 return (1); 1795 last = strlen(es_prop) - 1; 1796 es_prop[last] = '3'; 1797 return (prom_getproplen(node, es_prop) != -1); 1798 } 1799 1800 1801 /* 1802 * the actual size of the statefile data isn't known until after all the 1803 * compressed pages are written; even the inode size doesn't reflect the 1804 * data size since there are usually many extra fs blocks. for recording 1805 * the actual data size, the first sector of the statefile is copied to 1806 * a tmp buf, and the copy is later updated and flushed to disk. 1807 */ 1808 int 1809 i_cpr_blockzero(char *base, char **bufpp, int *blkno, vnode_t *vp) 1810 { 1811 extern int cpr_flush_write(vnode_t *); 1812 static char cpr_sector[DEV_BSIZE]; 1813 cpr_ext bytes, *dst; 1814 1815 /* 1816 * this routine is called after cdd_t and csu_md_t are copied 1817 * to cpr_buf; mini-hack alert: the save/update method creates 1818 * a dependency on the combined struct size being >= one sector 1819 * or DEV_BSIZE; since introduction in Sol2.7, csu_md_t size is 1820 * over 1K bytes and will probably grow with any changes. 1821 * 1822 * copy when vp is NULL, flush when non-NULL 1823 */ 1824 if (vp == NULL) { 1825 ASSERT((*bufpp - base) >= DEV_BSIZE); 1826 bcopy(base, cpr_sector, sizeof (cpr_sector)); 1827 return (0); 1828 } else { 1829 bytes = dbtob(*blkno); 1830 dst = &((cdd_t *)cpr_sector)->cdd_filesize; 1831 bcopy(&bytes, dst, sizeof (bytes)); 1832 bcopy(cpr_sector, base, sizeof (cpr_sector)); 1833 *bufpp = base + sizeof (cpr_sector); 1834 *blkno = cpr_statefile_offset(); 1835 CPR_DEBUG(CPR_DEBUG1, "statefile data size: %ld\n\n", bytes); 1836 return (cpr_flush_write(vp)); 1837 } 1838 } 1839 1840 1841 /* 1842 * Allocate bitmaps according to the phys_install list. 1843 */ 1844 static int 1845 i_cpr_bitmap_setup(void) 1846 { 1847 struct memlist *pmem; 1848 cbd_t *dp, *tail; 1849 void *space; 1850 size_t size; 1851 1852 /* 1853 * The number of bitmap descriptors will be the count of 1854 * phys_install ranges plus 1 for a trailing NULL struct. 1855 */ 1856 cpr_nbitmaps = 1; 1857 for (pmem = phys_install; pmem; pmem = pmem->next) 1858 cpr_nbitmaps++; 1859 1860 if (cpr_nbitmaps > (CPR_MAX_BMDESC - 1)) { 1861 cpr_err(CE_WARN, "too many physical memory ranges %d, max %d", 1862 cpr_nbitmaps, CPR_MAX_BMDESC - 1); 1863 return (EFBIG); 1864 } 1865 1866 /* Alloc an array of bitmap descriptors. */ 1867 dp = kmem_zalloc(cpr_nbitmaps * sizeof (*dp), KM_NOSLEEP); 1868 if (dp == NULL) { 1869 cpr_nbitmaps = 0; 1870 return (ENOMEM); 1871 } 1872 tail = dp + cpr_nbitmaps; 1873 1874 CPR->c_bmda = dp; 1875 for (pmem = phys_install; pmem; pmem = pmem->next) { 1876 size = BITMAP_BYTES(pmem->size); 1877 space = kmem_zalloc(size * 2, KM_NOSLEEP); 1878 if (space == NULL) 1879 return (ENOMEM); 1880 ASSERT(dp < tail); 1881 dp->cbd_magic = CPR_BITMAP_MAGIC; 1882 dp->cbd_spfn = mmu_btop(pmem->address); 1883 dp->cbd_epfn = mmu_btop(pmem->address + pmem->size) - 1; 1884 dp->cbd_size = size; 1885 dp->cbd_reg_bitmap = (cpr_ptr)space; 1886 dp->cbd_vlt_bitmap = (cpr_ptr)((caddr_t)space + size); 1887 dp++; 1888 } 1889 1890 /* set magic for the last descriptor */ 1891 ASSERT(dp == (tail - 1)); 1892 dp->cbd_magic = CPR_BITMAP_MAGIC; 1893 1894 return (0); 1895 } 1896 1897 1898 void 1899 i_cpr_bitmap_cleanup(void) 1900 { 1901 cbd_t *dp; 1902 1903 if (CPR->c_bmda == NULL) 1904 return; 1905 for (dp = CPR->c_bmda; dp->cbd_size; dp++) 1906 kmem_free((void *)dp->cbd_reg_bitmap, dp->cbd_size * 2); 1907 kmem_free(CPR->c_bmda, cpr_nbitmaps * sizeof (*CPR->c_bmda)); 1908 CPR->c_bmda = NULL; 1909 cpr_nbitmaps = 0; 1910 } 1911 1912 1913 /* 1914 * A "regular" and "volatile" bitmap are created for each range of 1915 * physical memory. The volatile maps are used to count and track pages 1916 * susceptible to heap corruption - caused by drivers that allocate mem 1917 * during VOP_DUMP(); the regular maps are used for all the other non- 1918 * susceptible pages. Before writing the bitmaps to the statefile, 1919 * each bitmap pair gets merged to simplify handling within cprboot. 1920 */ 1921 int 1922 i_cpr_alloc_bitmaps(void) 1923 { 1924 int err; 1925 1926 memlist_read_lock(); 1927 err = i_cpr_bitmap_setup(); 1928 memlist_read_unlock(); 1929 if (err) 1930 i_cpr_bitmap_cleanup(); 1931 return (err); 1932 } 1933 1934 1935 1936 /* 1937 * Power down the system. 1938 */ 1939 int 1940 i_cpr_power_down(int sleeptype) 1941 { 1942 int is_defined = 0; 1943 char *wordexists = "p\" power-off\" find nip swap l! "; 1944 char *req = "power-off"; 1945 1946 ASSERT(sleeptype == CPR_TODISK); 1947 1948 /* 1949 * is_defined has value -1 when defined 1950 */ 1951 prom_interpret(wordexists, (uintptr_t)&is_defined, 0, 0, 0, 0); 1952 if (is_defined) { 1953 CPR_DEBUG(CPR_DEBUG1, "\ncpr: %s...\n", req); 1954 prom_interpret(req, 0, 0, 0, 0, 0); 1955 } 1956 /* 1957 * Only returns if failed 1958 */ 1959 return (EIO); 1960 } 1961 1962 void 1963 i_cpr_stop_other_cpus(void) 1964 { 1965 stop_other_cpus(); 1966 } 1967 1968 /* 1969 * Save context for the specified CPU 1970 */ 1971 /* ARGSUSED */ 1972 void * 1973 i_cpr_save_context(void *arg) 1974 { 1975 /* 1976 * Not yet 1977 */ 1978 ASSERT(0); 1979 return (NULL); 1980 } 1981 1982 void 1983 i_cpr_pre_resume_cpus(void) 1984 { 1985 /* 1986 * Not yet 1987 */ 1988 ASSERT(0); 1989 } 1990 1991 void 1992 i_cpr_post_resume_cpus(void) 1993 { 1994 /* 1995 * Not yet 1996 */ 1997 ASSERT(0); 1998 } 1999 2000 /* 2001 * nothing to do 2002 */ 2003 void 2004 i_cpr_alloc_cpus(void) 2005 { 2006 } 2007 2008 /* 2009 * nothing to do 2010 */ 2011 void 2012 i_cpr_free_cpus(void) 2013 { 2014 } 2015 2016 /* ARGSUSED */ 2017 void 2018 i_cpr_save_configuration(dev_info_t *dip) 2019 { 2020 /* 2021 * this is a no-op on sparc 2022 */ 2023 } 2024 2025 /* ARGSUSED */ 2026 void 2027 i_cpr_restore_configuration(dev_info_t *dip) 2028 { 2029 /* 2030 * this is a no-op on sparc 2031 */ 2032 } 2033