1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright 2008 Sun Microsystems, Inc. All rights reserved. 23 * Use is subject to license terms. 24 */ 25 26 #pragma ident "%Z%%M% %I% %E% SMI" 27 28 /* 29 * sunpm.c builds sunpm.o "power management framework" 30 * kernel-resident power management code. Implements power management 31 * policy 32 * Assumes: all backwards compat. device components wake up on & 33 * the pm_info pointer in dev_info is initially NULL 34 * 35 * PM - (device) Power Management 36 * 37 * Each device may have 0 or more components. If a device has no components, 38 * then it can't be power managed. Each component has 2 or more 39 * power states. 40 * 41 * "Backwards Compatible" (bc) devices: 42 * There are two different types of devices from the point of view of this 43 * code. The original type, left over from the original PM implementation on 44 * the voyager platform are known in this code as "backwards compatible" 45 * devices (PM_ISBC(dip) returns true). 46 * They are recognized by the pm code by the lack of a pm-components property 47 * and a call made by the driver to pm_create_components(9F). 48 * For these devices, component 0 is special, and represents the power state 49 * of the device. If component 0 is to be set to power level 0 (off), then 50 * the framework must first call into the driver's detach(9E) routine with 51 * DDI_PM_SUSPEND, to get the driver to save the hardware state of the device. 52 * After setting component 0 from 0 to a non-zero power level, a call must be 53 * made into the driver's attach(9E) routine with DDI_PM_RESUME. 54 * 55 * Currently, the only way to get a bc device power managed is via a set of 56 * ioctls (PM_DIRECT_PM, PM_SET_CURRENT_POWER) issued to /dev/pm. 57 * 58 * For non-bc devices, the driver describes the components by exporting a 59 * pm-components(9P) property that tells how many components there are, 60 * tells what each component's power state values are, and provides human 61 * readable strings (currently unused) for each component name and power state. 62 * Devices which export pm-components(9P) are automatically power managed 63 * whenever autopm is enabled (via PM_START_PM ioctl issued by pmconfig(1M) 64 * after parsing power.conf(4)). The exception to this rule is that power 65 * manageable CPU devices may be automatically managed independently of autopm 66 * by either enabling or disabling (via PM_START_CPUPM and PM_STOP_CPUPM 67 * ioctls) cpupm. If the CPU devices are not managed independently, then they 68 * are managed by autopm. In either case, for automatically power managed 69 * devices, all components are considered independent of each other, and it is 70 * up to the driver to decide when a transition requires saving or restoring 71 * hardware state. 72 * 73 * Each device component also has a threshold time associated with each power 74 * transition (see power.conf(4)), and a busy/idle state maintained by the 75 * driver calling pm_idle_component(9F) and pm_busy_component(9F). 76 * Components are created idle. 77 * 78 * The PM framework provides several functions: 79 * -implement PM policy as described in power.conf(4) 80 * Policy is set by pmconfig(1M) issuing pm ioctls based on power.conf(4). 81 * Policies consist of: 82 * -set threshold values (defaults if none provided by pmconfig) 83 * -set dependencies among devices 84 * -enable/disable autopm 85 * -enable/disable cpupm 86 * -turn down idle components based on thresholds (if autopm or cpupm is 87 * enabled) (aka scanning) 88 * -maintain power states based on dependencies among devices 89 * -upon request, or when the frame buffer powers off, attempt to turn off 90 * all components that are idle or become idle over the next (10 sec) 91 * period in an attempt to get down to an EnergyStar compliant state 92 * -prevent powering off of a device which exported the 93 * pm-no-involuntary-power-cycles property without active involvement of 94 * the device's driver (so no removing power when the device driver is 95 * not attached) 96 * -provide a mechanism for a device driver to request that a device's component 97 * be brought back to the power level necessary for the use of the device 98 * -allow a process to directly control the power levels of device components 99 * (via ioctls issued to /dev/pm--see usr/src/uts/common/io/pm.c) 100 * -ensure that the console frame buffer is powered up before being referenced 101 * via prom_printf() or other prom calls that might generate console output 102 * -maintain implicit dependencies (e.g. parent must be powered up if child is) 103 * -provide "backwards compatible" behavior for devices without pm-components 104 * property 105 * 106 * Scanning: 107 * Whenever autopm or cpupm is enabled, the framework attempts to bring each 108 * component of each managed device to its lowest power based on the threshold 109 * of idleness associated with each transition and the busy/idle state of the 110 * component. 111 * 112 * The actual work of this is done by pm_scan_dev(), which cycles through each 113 * component of a device, checking its idleness against its current threshold, 114 * and calling pm_set_power() as appropriate to change the power level. 115 * This function also indicates when it would next be profitable to scan the 116 * device again, and a new scan is scheduled after that time. 117 * 118 * Dependencies: 119 * It is possible to establish a dependency between the power states of two 120 * otherwise unrelated devices. This is currently done to ensure that the 121 * cdrom is always up whenever the console framebuffer is up, so that the user 122 * can insert a cdrom and see a popup as a result. 123 * 124 * The dependency terminology used in power.conf(4) is not easy to understand, 125 * so we've adopted a different terminology in the implementation. We write 126 * of a "keeps up" and a "kept up" device. A relationship can be established 127 * where one device keeps up another. That means that if the keepsup device 128 * has any component that is at a non-zero power level, all components of the 129 * "kept up" device must be brought to full power. This relationship is 130 * asynchronous. When the keeping device is powered up, a request is queued 131 * to a worker thread to bring up the kept device. The caller does not wait. 132 * Scan will not turn down a kept up device. 133 * 134 * Direct PM: 135 * A device may be directly power managed by a process. If a device is 136 * directly pm'd, then it will not be scanned, and dependencies will not be 137 * enforced. * If a directly pm'd device's driver requests a power change (via 138 * pm_raise_power(9F)), then the request is blocked and notification is sent 139 * to the controlling process, which must issue the requested power change for 140 * the driver to proceed. 141 * 142 */ 143 144 #include <sys/types.h> 145 #include <sys/errno.h> 146 #include <sys/callb.h> /* callback registration during CPR */ 147 #include <sys/conf.h> /* driver flags and functions */ 148 #include <sys/open.h> /* OTYP_CHR definition */ 149 #include <sys/stat.h> /* S_IFCHR definition */ 150 #include <sys/pathname.h> /* name -> dev_info xlation */ 151 #include <sys/ddi_impldefs.h> /* dev_info node fields */ 152 #include <sys/kmem.h> /* memory alloc stuff */ 153 #include <sys/debug.h> 154 #include <sys/archsystm.h> 155 #include <sys/pm.h> 156 #include <sys/ddi.h> 157 #include <sys/sunddi.h> 158 #include <sys/sunndi.h> 159 #include <sys/sunpm.h> 160 #include <sys/epm.h> 161 #include <sys/vfs.h> 162 #include <sys/mode.h> 163 #include <sys/mkdev.h> 164 #include <sys/promif.h> 165 #include <sys/consdev.h> 166 #include <sys/esunddi.h> 167 #include <sys/modctl.h> 168 #include <sys/fs/ufs_fs.h> 169 #include <sys/note.h> 170 #include <sys/taskq.h> 171 #include <sys/bootconf.h> 172 #include <sys/reboot.h> 173 #include <sys/spl.h> 174 #include <sys/disp.h> 175 #include <sys/sobject.h> 176 #include <sys/sunmdi.h> 177 #include <sys/systm.h> 178 #include <sys/cpuvar.h> 179 #include <sys/cyclic.h> 180 #include <sys/uadmin.h> 181 #include <sys/srn.h> 182 183 184 /* 185 * PM LOCKING 186 * The list of locks: 187 * Global pm mutex locks. 188 * 189 * pm_scan_lock: 190 * It protects the timeout id of the scan thread, and the value 191 * of autopm_enabled and cpupm. This lock is not held 192 * concurrently with any other PM locks. 193 * 194 * pm_clone_lock: Protects the clone list and count of poll events 195 * pending for the pm driver. 196 * Lock ordering: 197 * pm_clone_lock -> pm_pscc_interest_rwlock, 198 * pm_clone_lock -> pm_pscc_direct_rwlock. 199 * 200 * pm_rsvp_lock: 201 * Used to synchronize the data structures used for processes 202 * to rendezvous with state change information when doing 203 * direct PM. 204 * Lock ordering: 205 * pm_rsvp_lock -> pm_pscc_interest_rwlock, 206 * pm_rsvp_lock -> pm_pscc_direct_rwlock, 207 * pm_rsvp_lock -> pm_clone_lock. 208 * 209 * ppm_lock: protects the list of registered ppm drivers 210 * Lock ordering: 211 * ppm_lock -> ppm driver unit_lock 212 * 213 * pm_compcnt_lock: 214 * Protects count of components that are not at their lowest 215 * power level. 216 * Lock ordering: 217 * pm_compcnt_lock -> ppm_lock. 218 * 219 * pm_dep_thread_lock: 220 * Protects work list for pm_dep_thread. Not taken concurrently 221 * with any other pm lock. 222 * 223 * pm_remdrv_lock: 224 * Serializes the operation of removing noinvol data structure 225 * entries for a branch of the tree when a driver has been 226 * removed from the system (modctl_rem_major). 227 * Lock ordering: 228 * pm_remdrv_lock -> pm_noinvol_rwlock. 229 * 230 * pm_cfb_lock: (High level spin lock) 231 * Protects the count of how many components of the console 232 * frame buffer are off (so we know if we have to bring up the 233 * console as a result of a prom_printf, etc. 234 * No other locks are taken while holding this lock. 235 * 236 * pm_loan_lock: 237 * Protects the lock_loan list. List is used to record that one 238 * thread has acquired a power lock but has launched another thread 239 * to complete its processing. An entry in the list indicates that 240 * the worker thread can borrow the lock held by the other thread, 241 * which must block on the completion of the worker. Use is 242 * specific to module loading. 243 * No other locks are taken while holding this lock. 244 * 245 * Global PM rwlocks 246 * 247 * pm_thresh_rwlock: 248 * Protects the list of thresholds recorded for future use (when 249 * devices attach). 250 * Lock ordering: 251 * pm_thresh_rwlock -> devi_pm_lock 252 * 253 * pm_noinvol_rwlock: 254 * Protects list of detached nodes that had noinvol registered. 255 * No other PM locks are taken while holding pm_noinvol_rwlock. 256 * 257 * pm_pscc_direct_rwlock: 258 * Protects the list that maps devices being directly power 259 * managed to the processes that manage them. 260 * Lock ordering: 261 * pm_pscc_direct_rwlock -> psce_lock 262 * 263 * pm_pscc_interest_rwlock; 264 * Protects the list that maps state change events to processes 265 * that want to know about them. 266 * Lock ordering: 267 * pm_pscc_interest_rwlock -> psce_lock 268 * 269 * per-dip locks: 270 * 271 * Each node has these per-dip locks, which are only used if the device is 272 * a candidate for power management (e.g. has pm components) 273 * 274 * devi_pm_lock: 275 * Protects all power management state of the node except for 276 * power level, which is protected by ndi_devi_enter(). 277 * Encapsulated in macros PM_LOCK_DIP()/PM_UNLOCK_DIP(). 278 * Lock ordering: 279 * devi_pm_lock -> pm_rsvp_lock, 280 * devi_pm_lock -> pm_dep_thread_lock, 281 * devi_pm_lock -> pm_noinvol_rwlock, 282 * devi_pm_lock -> power lock 283 * 284 * power lock (ndi_devi_enter()): 285 * Since changing power level is possibly a slow operation (30 286 * seconds to spin up a disk drive), this is locked separately. 287 * Since a call into the driver to change the power level of one 288 * component may result in a call back into the framework to change 289 * the power level of another, this lock allows re-entrancy by 290 * the same thread (ndi_devi_enter is used for this because 291 * the USB framework uses ndi_devi_enter in its power entry point, 292 * and use of any other lock would produce a deadlock. 293 * 294 * devi_pm_busy_lock: 295 * This lock protects the integrity of the busy count. It is 296 * only taken by pm_busy_component() and pm_idle_component and 297 * some code that adjust the busy time after the timer gets set 298 * up or after a CPR operation. It is per-dip to keep from 299 * single-threading all the disk drivers on a system. 300 * It could be per component instead, but most devices have 301 * only one component. 302 * No other PM locks are taken while holding this lock. 303 * 304 */ 305 306 static int stdout_is_framebuffer; 307 static kmutex_t e_pm_power_lock; 308 static kmutex_t pm_loan_lock; 309 kmutex_t pm_scan_lock; 310 callb_id_t pm_cpr_cb_id; 311 callb_id_t pm_panic_cb_id; 312 callb_id_t pm_halt_cb_id; 313 int pm_comps_notlowest; /* no. of comps not at lowest power */ 314 int pm_powering_down; /* cpr is source of DDI_SUSPEND calls */ 315 316 clock_t pm_min_scan = PM_MIN_SCAN; 317 clock_t pm_id_ticks = 5; /* ticks to wait before scan during idle-down */ 318 319 static int pm_busop_set_power(dev_info_t *, 320 void *, pm_bus_power_op_t, void *, void *); 321 static int pm_busop_match_request(dev_info_t *, void *); 322 static int pm_all_to_normal_nexus(dev_info_t *, pm_canblock_t); 323 static void e_pm_set_max_power(dev_info_t *, int, int); 324 static int e_pm_get_max_power(dev_info_t *, int); 325 326 /* 327 * Dependency Processing is done thru a seperate thread. 328 */ 329 kmutex_t pm_dep_thread_lock; 330 kcondvar_t pm_dep_thread_cv; 331 pm_dep_wk_t *pm_dep_thread_workq = NULL; 332 pm_dep_wk_t *pm_dep_thread_tail = NULL; 333 334 /* 335 * Autopm must be turned on by a PM_START_PM ioctl, so we don't end up 336 * power managing things in single user mode that have been suppressed via 337 * power.conf entries. Protected by pm_scan_lock. 338 */ 339 int autopm_enabled; 340 341 /* 342 * cpupm is turned on and off, by the PM_START_CPUPM and PM_STOP_CPUPM ioctls, 343 * to define the power management behavior of CPU devices separate from 344 * autopm. Protected by pm_scan_lock. 345 */ 346 pm_cpupm_t cpupm = PM_CPUPM_NOTSET; 347 348 /* 349 * AutoS3 depends on autopm being enabled, and must be enabled by 350 * PM_START_AUTOS3 command. 351 */ 352 int autoS3_enabled; 353 354 #if !defined(__sparc) 355 /* 356 * on sparc these live in fillsysinfo.c 357 * 358 * If this variable is non-zero, cpr should return "not supported" when 359 * it is queried even though it would normally be supported on this platform. 360 */ 361 int cpr_supported_override; 362 363 /* 364 * Some platforms may need to support CPR even in the absence of 365 * having the correct platform id information. If this 366 * variable is non-zero, cpr should proceed even in the absence 367 * of otherwise being qualified. 368 */ 369 int cpr_platform_enable = 0; 370 371 #endif 372 373 /* 374 * pm_S3_enabled indicates that we believe the platform can support S3, 375 * which we get from pmconfig(1M) 376 */ 377 int pm_S3_enabled; 378 379 /* 380 * This flag is true while processes are stopped for a checkpoint/resume. 381 * Controlling processes of direct pm'd devices are not available to 382 * participate in power level changes, so we bypass them when this is set. 383 */ 384 static int pm_processes_stopped; 385 386 #ifdef DEBUG 387 388 /* 389 * see common/sys/epm.h for PMD_* values 390 */ 391 392 uint_t pm_debug = 0; 393 394 /* 395 * If pm_divertdebug is set, then no prom_printf calls will be made by 396 * PMD(), which will prevent debug output from bringing up the console 397 * frame buffer. Clearing this variable before setting pm_debug will result 398 * in PMD output going to the console. 399 * 400 * pm_divertdebug is incremented in pm_set_power() if dip == cfb_dip to avoid 401 * deadlocks and decremented at the end of pm_set_power() 402 */ 403 uint_t pm_divertdebug = 1; 404 volatile uint_t pm_debug_to_console = 0; 405 kmutex_t pm_debug_lock; /* protects pm_divertdebug */ 406 407 void prdeps(char *); 408 #endif 409 410 /* Globals */ 411 412 /* 413 * List of recorded thresholds and dependencies 414 */ 415 pm_thresh_rec_t *pm_thresh_head; 416 krwlock_t pm_thresh_rwlock; 417 418 pm_pdr_t *pm_dep_head; 419 static int pm_unresolved_deps = 0; 420 static int pm_prop_deps = 0; 421 422 /* 423 * List of devices that exported no-involuntary-power-cycles property 424 */ 425 pm_noinvol_t *pm_noinvol_head; 426 427 /* 428 * Locks used in noinvol processing 429 */ 430 krwlock_t pm_noinvol_rwlock; 431 kmutex_t pm_remdrv_lock; 432 433 int pm_default_idle_threshold = PM_DEFAULT_SYS_IDLENESS; 434 int pm_system_idle_threshold; 435 int pm_cpu_idle_threshold; 436 437 /* 438 * By default nexus has 0 threshold, and depends on its children to keep it up 439 */ 440 int pm_default_nexus_threshold = 0; 441 442 /* 443 * Data structures shared with common/io/pm.c 444 */ 445 kmutex_t pm_clone_lock; 446 kcondvar_t pm_clones_cv[PM_MAX_CLONE]; 447 uint_t pm_poll_cnt[PM_MAX_CLONE]; /* count of events for poll */ 448 unsigned char pm_interest[PM_MAX_CLONE]; 449 struct pollhead pm_pollhead; 450 451 /* 452 * Data structures shared with common/io/srn.c 453 */ 454 kmutex_t srn_clone_lock; /* protects srn_signal, srn_inuse */ 455 void (*srn_signal)(int type, int event); 456 int srn_inuse; /* stop srn detach */ 457 458 extern int hz; 459 extern char *platform_module_list[]; 460 461 /* 462 * Wrappers for use in ddi_walk_devs 463 */ 464 465 static int pm_set_dev_thr_walk(dev_info_t *, void *); 466 static int pm_restore_direct_lvl_walk(dev_info_t *, void *); 467 static int pm_save_direct_lvl_walk(dev_info_t *, void *); 468 static int pm_discard_dep_walk(dev_info_t *, void *); 469 #ifdef DEBUG 470 static int pm_desc_pwrchk_walk(dev_info_t *, void *); 471 #endif 472 473 /* 474 * Routines for managing noinvol devices 475 */ 476 int pm_noinvol_update(int, int, int, char *, dev_info_t *); 477 void pm_noinvol_update_node(dev_info_t *, 478 pm_bp_noinvol_t *req); 479 480 kmutex_t pm_rsvp_lock; 481 kmutex_t pm_compcnt_lock; 482 krwlock_t pm_pscc_direct_rwlock; 483 krwlock_t pm_pscc_interest_rwlock; 484 485 #define PSC_INTEREST 0 /* belongs to interest psc list */ 486 #define PSC_DIRECT 1 /* belongs to direct psc list */ 487 488 pscc_t *pm_pscc_interest; 489 pscc_t *pm_pscc_direct; 490 491 #define PM_MAJOR(dip) ddi_name_to_major(ddi_binding_name(dip)) 492 #define PM_IS_NEXUS(dip) NEXUS_DRV(devopsp[PM_MAJOR(dip)]) 493 #define POWERING_ON(old, new) ((old) == 0 && (new) != 0) 494 #define POWERING_OFF(old, new) ((old) != 0 && (new) == 0) 495 496 #define PM_INCR_NOTLOWEST(dip) { \ 497 mutex_enter(&pm_compcnt_lock); \ 498 if (!PM_IS_NEXUS(dip) || \ 499 (DEVI(dip)->devi_pm_flags & (PMC_DEV_THRESH|PMC_COMP_THRESH))) {\ 500 if (pm_comps_notlowest == 0) \ 501 pm_ppm_notify_all_lowest(dip, PM_NOT_ALL_LOWEST);\ 502 pm_comps_notlowest++; \ 503 PMD(PMD_LEVEL, ("%s: %s@%s(%s#%d) incr notlowest->%d\n",\ 504 pmf, PM_DEVICE(dip), pm_comps_notlowest)) \ 505 } \ 506 mutex_exit(&pm_compcnt_lock); \ 507 } 508 #define PM_DECR_NOTLOWEST(dip) { \ 509 mutex_enter(&pm_compcnt_lock); \ 510 if (!PM_IS_NEXUS(dip) || \ 511 (DEVI(dip)->devi_pm_flags & (PMC_DEV_THRESH|PMC_COMP_THRESH))) {\ 512 ASSERT(pm_comps_notlowest); \ 513 pm_comps_notlowest--; \ 514 PMD(PMD_LEVEL, ("%s: %s@%s(%s#%d) decr notlowest to " \ 515 "%d\n", pmf, PM_DEVICE(dip), pm_comps_notlowest))\ 516 if (pm_comps_notlowest == 0) \ 517 pm_ppm_notify_all_lowest(dip, PM_ALL_LOWEST); \ 518 } \ 519 mutex_exit(&pm_compcnt_lock); \ 520 } 521 522 /* 523 * console frame-buffer power-management is not enabled when 524 * debugging services are present. to override, set pm_cfb_override 525 * to non-zero. 526 */ 527 uint_t pm_cfb_comps_off = 0; /* PM_LEVEL_UNKNOWN is considered on */ 528 kmutex_t pm_cfb_lock; 529 int pm_cfb_enabled = 1; /* non-zero allows pm of console frame buffer */ 530 #ifdef DEBUG 531 int pm_cfb_override = 1; /* non-zero allows pm of cfb with debuggers */ 532 #else 533 int pm_cfb_override = 0; /* non-zero allows pm of cfb with debuggers */ 534 #endif 535 536 static dev_info_t *cfb_dip = 0; 537 static dev_info_t *cfb_dip_detaching = 0; 538 uint_t cfb_inuse = 0; 539 static ddi_softintr_t pm_soft_id; 540 static clock_t pm_soft_pending; 541 int pm_scans_disabled = 0; 542 543 /* 544 * A structure to record the fact that one thread has borrowed a lock held 545 * by another thread. The context requires that the lender block on the 546 * completion of the borrower. 547 */ 548 typedef struct lock_loan { 549 struct lock_loan *pmlk_next; 550 kthread_t *pmlk_borrower; 551 kthread_t *pmlk_lender; 552 dev_info_t *pmlk_dip; 553 } lock_loan_t; 554 static lock_loan_t lock_loan_head; /* list head is a dummy element */ 555 556 #ifdef DEBUG 557 #ifdef PMDDEBUG 558 #define PMD_FUNC(func, name) char *(func) = (name); 559 #else /* !PMDDEBUG */ 560 #define PMD_FUNC(func, name) 561 #endif /* PMDDEBUG */ 562 #else /* !DEBUG */ 563 #define PMD_FUNC(func, name) 564 #endif /* DEBUG */ 565 566 567 /* 568 * Must be called before first device (including pseudo) attach 569 */ 570 void 571 pm_init_locks(void) 572 { 573 mutex_init(&pm_scan_lock, NULL, MUTEX_DRIVER, NULL); 574 mutex_init(&pm_rsvp_lock, NULL, MUTEX_DRIVER, NULL); 575 mutex_init(&pm_compcnt_lock, NULL, MUTEX_DRIVER, NULL); 576 mutex_init(&pm_dep_thread_lock, NULL, MUTEX_DRIVER, NULL); 577 mutex_init(&pm_remdrv_lock, NULL, MUTEX_DRIVER, NULL); 578 mutex_init(&pm_loan_lock, NULL, MUTEX_DRIVER, NULL); 579 rw_init(&pm_thresh_rwlock, NULL, RW_DEFAULT, NULL); 580 rw_init(&pm_noinvol_rwlock, NULL, RW_DEFAULT, NULL); 581 cv_init(&pm_dep_thread_cv, NULL, CV_DEFAULT, NULL); 582 } 583 584 static boolean_t 585 pm_cpr_callb(void *arg, int code) 586 { 587 _NOTE(ARGUNUSED(arg)) 588 static int auto_save; 589 static pm_cpupm_t cpupm_save; 590 static int pm_reset_timestamps(dev_info_t *, void *); 591 592 switch (code) { 593 case CB_CODE_CPR_CHKPT: 594 /* 595 * Cancel scan or wait for scan in progress to finish 596 * Other threads may be trying to restart the scan, so we 597 * have to keep at it unil it sticks 598 */ 599 mutex_enter(&pm_scan_lock); 600 ASSERT(!pm_scans_disabled); 601 pm_scans_disabled = 1; 602 auto_save = autopm_enabled; 603 autopm_enabled = 0; 604 cpupm_save = cpupm; 605 cpupm = PM_CPUPM_NOTSET; 606 mutex_exit(&pm_scan_lock); 607 ddi_walk_devs(ddi_root_node(), pm_scan_stop_walk, NULL); 608 break; 609 610 case CB_CODE_CPR_RESUME: 611 ASSERT(!autopm_enabled); 612 ASSERT(cpupm == PM_CPUPM_NOTSET); 613 ASSERT(pm_scans_disabled); 614 pm_scans_disabled = 0; 615 /* 616 * Call pm_reset_timestamps to reset timestamps of each 617 * device to the time when the system is resumed so that their 618 * idleness can be re-calculated. That's to avoid devices from 619 * being powered down right after resume if the system was in 620 * suspended mode long enough. 621 */ 622 ddi_walk_devs(ddi_root_node(), pm_reset_timestamps, NULL); 623 624 autopm_enabled = auto_save; 625 cpupm = cpupm_save; 626 /* 627 * If there is any auto-pm device, get the scanning 628 * going. Otherwise don't bother. 629 */ 630 ddi_walk_devs(ddi_root_node(), pm_rescan_walk, NULL); 631 break; 632 } 633 return (B_TRUE); 634 } 635 636 /* 637 * This callback routine is called when there is a system panic. This function 638 * exists for prototype matching. 639 */ 640 static boolean_t 641 pm_panic_callb(void *arg, int code) 642 { 643 _NOTE(ARGUNUSED(arg, code)) 644 void pm_cfb_check_and_powerup(void); 645 PMD(PMD_CFB, ("pm_panic_callb\n")) 646 pm_cfb_check_and_powerup(); 647 return (B_TRUE); 648 } 649 650 static boolean_t 651 pm_halt_callb(void *arg, int code) 652 { 653 _NOTE(ARGUNUSED(arg, code)) 654 return (B_TRUE); 655 } 656 657 /* 658 * This needs to be called after the root and platform drivers are loaded 659 * and be single-threaded with respect to driver attach/detach 660 */ 661 void 662 pm_init(void) 663 { 664 PMD_FUNC(pmf, "pm_init") 665 char **mod; 666 extern pri_t minclsyspri; 667 static void pm_dep_thread(void); 668 669 pm_comps_notlowest = 0; 670 pm_system_idle_threshold = pm_default_idle_threshold; 671 pm_cpu_idle_threshold = 0; 672 673 pm_cpr_cb_id = callb_add(pm_cpr_callb, (void *)NULL, 674 CB_CL_CPR_PM, "pm_cpr"); 675 pm_panic_cb_id = callb_add(pm_panic_callb, (void *)NULL, 676 CB_CL_PANIC, "pm_panic"); 677 pm_halt_cb_id = callb_add(pm_halt_callb, (void *)NULL, 678 CB_CL_HALT, "pm_halt"); 679 680 /* 681 * Create a thread to do dependency processing. 682 */ 683 (void) thread_create(NULL, 0, (void (*)())pm_dep_thread, NULL, 0, &p0, 684 TS_RUN, minclsyspri); 685 686 /* 687 * loadrootmodules already loaded these ppm drivers, now get them 688 * attached so they can claim the root drivers as they attach 689 */ 690 for (mod = platform_module_list; *mod; mod++) { 691 if (i_ddi_attach_hw_nodes(*mod) != DDI_SUCCESS) { 692 cmn_err(CE_WARN, "!cannot load platform pm driver %s\n", 693 *mod); 694 } else { 695 PMD(PMD_DHR, ("%s: %s (%s)\n", pmf, *mod, 696 ddi_major_to_name(ddi_name_to_major(*mod)))) 697 } 698 } 699 } 700 701 /* 702 * pm_scan_init - create pm scan data structure. Called (if autopm or cpupm 703 * enabled) when device becomes power managed or after a failed detach and 704 * when autopm is started via PM_START_PM or PM_START_CPUPM ioctls, and after 705 * a CPR resume to get all the devices scanning again. 706 */ 707 void 708 pm_scan_init(dev_info_t *dip) 709 { 710 PMD_FUNC(pmf, "scan_init") 711 pm_scan_t *scanp; 712 713 ASSERT(!PM_ISBC(dip)); 714 715 PM_LOCK_DIP(dip); 716 scanp = PM_GET_PM_SCAN(dip); 717 if (!scanp) { 718 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d): create scan data\n", 719 pmf, PM_DEVICE(dip))) 720 scanp = kmem_zalloc(sizeof (pm_scan_t), KM_SLEEP); 721 DEVI(dip)->devi_pm_scan = scanp; 722 } else if (scanp->ps_scan_flags & PM_SCAN_STOP) { 723 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d): " 724 "clear PM_SCAN_STOP flag\n", pmf, PM_DEVICE(dip))) 725 scanp->ps_scan_flags &= ~PM_SCAN_STOP; 726 } 727 PM_UNLOCK_DIP(dip); 728 } 729 730 /* 731 * pm_scan_fini - remove pm scan data structure when stopping pm on the device 732 */ 733 void 734 pm_scan_fini(dev_info_t *dip) 735 { 736 PMD_FUNC(pmf, "scan_fini") 737 pm_scan_t *scanp; 738 739 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d)\n", pmf, PM_DEVICE(dip))) 740 ASSERT(!PM_ISBC(dip)); 741 PM_LOCK_DIP(dip); 742 scanp = PM_GET_PM_SCAN(dip); 743 if (!scanp) { 744 PM_UNLOCK_DIP(dip); 745 return; 746 } 747 748 ASSERT(!scanp->ps_scan_id && !(scanp->ps_scan_flags & 749 (PM_SCANNING | PM_SCAN_DISPATCHED | PM_SCAN_AGAIN))); 750 751 kmem_free(scanp, sizeof (pm_scan_t)); 752 DEVI(dip)->devi_pm_scan = NULL; 753 PM_UNLOCK_DIP(dip); 754 } 755 756 /* 757 * Given a pointer to a component struct, return the current power level 758 * (struct contains index unless it is a continuous level). 759 * Located here in hopes of getting both this and dev_is_needed into the 760 * cache together 761 */ 762 static int 763 cur_power(pm_component_t *cp) 764 { 765 if (cp->pmc_cur_pwr == PM_LEVEL_UNKNOWN) 766 return (cp->pmc_cur_pwr); 767 768 return (cp->pmc_comp.pmc_lvals[cp->pmc_cur_pwr]); 769 } 770 771 static char * 772 pm_decode_direction(int direction) 773 { 774 switch (direction) { 775 case PM_LEVEL_UPONLY: 776 return ("up"); 777 778 case PM_LEVEL_EXACT: 779 return ("exact"); 780 781 case PM_LEVEL_DOWNONLY: 782 return ("down"); 783 784 default: 785 return ("INVALID DIRECTION"); 786 } 787 } 788 789 char * 790 pm_decode_op(pm_bus_power_op_t op) 791 { 792 switch (op) { 793 case BUS_POWER_CHILD_PWRCHG: 794 return ("CHILD_PWRCHG"); 795 case BUS_POWER_NEXUS_PWRUP: 796 return ("NEXUS_PWRUP"); 797 case BUS_POWER_PRE_NOTIFICATION: 798 return ("PRE_NOTIFICATION"); 799 case BUS_POWER_POST_NOTIFICATION: 800 return ("POST_NOTIFICATION"); 801 case BUS_POWER_HAS_CHANGED: 802 return ("HAS_CHANGED"); 803 case BUS_POWER_NOINVOL: 804 return ("NOINVOL"); 805 default: 806 return ("UNKNOWN OP"); 807 } 808 } 809 810 /* 811 * Returns true if level is a possible (valid) power level for component 812 */ 813 int 814 e_pm_valid_power(dev_info_t *dip, int cmpt, int level) 815 { 816 PMD_FUNC(pmf, "e_pm_valid_power") 817 pm_component_t *cp = PM_CP(dip, cmpt); 818 int i; 819 int *ip = cp->pmc_comp.pmc_lvals; 820 int limit = cp->pmc_comp.pmc_numlevels; 821 822 if (level < 0) 823 return (0); 824 for (i = 0; i < limit; i++) { 825 if (level == *ip++) 826 return (1); 827 } 828 #ifdef DEBUG 829 if (pm_debug & PMD_FAIL) { 830 ip = cp->pmc_comp.pmc_lvals; 831 832 for (i = 0; i < limit; i++) 833 PMD(PMD_FAIL, ("%s: index=%d, level=%d\n", 834 pmf, i, *ip++)) 835 } 836 #endif 837 return (0); 838 } 839 840 /* 841 * Returns true if device is pm'd (after calling pm_start if need be) 842 */ 843 int 844 e_pm_valid_info(dev_info_t *dip, pm_info_t **infop) 845 { 846 pm_info_t *info; 847 static int pm_start(dev_info_t *dip); 848 849 /* 850 * Check if the device is power managed if not. 851 * To make the common case (device is power managed already) 852 * fast, we check without the lock. If device is not already 853 * power managed, then we take the lock and the long route through 854 * go get it managed. Devices never go unmanaged until they 855 * detach. 856 */ 857 info = PM_GET_PM_INFO(dip); 858 if (!info) { 859 if (!DEVI_IS_ATTACHING(dip)) { 860 return (0); 861 } 862 if (pm_start(dip) != DDI_SUCCESS) { 863 return (0); 864 } 865 info = PM_GET_PM_INFO(dip); 866 } 867 ASSERT(info); 868 if (infop != NULL) 869 *infop = info; 870 return (1); 871 } 872 873 int 874 e_pm_valid_comp(dev_info_t *dip, int cmpt, pm_component_t **cpp) 875 { 876 if (cmpt >= 0 && cmpt < PM_NUMCMPTS(dip)) { 877 if (cpp != NULL) 878 *cpp = PM_CP(dip, cmpt); 879 return (1); 880 } else { 881 return (0); 882 } 883 } 884 885 /* 886 * Internal guts of ddi_dev_is_needed and pm_raise/lower_power 887 */ 888 static int 889 dev_is_needed(dev_info_t *dip, int cmpt, int level, int direction) 890 { 891 PMD_FUNC(pmf, "din") 892 pm_component_t *cp; 893 char *pathbuf; 894 int result; 895 896 ASSERT(direction == PM_LEVEL_UPONLY || direction == PM_LEVEL_DOWNONLY); 897 if (!e_pm_valid_info(dip, NULL) || !e_pm_valid_comp(dip, cmpt, &cp) || 898 !e_pm_valid_power(dip, cmpt, level)) 899 return (DDI_FAILURE); 900 901 PMD(PMD_DIN, ("%s: %s@%s(%s#%d) cmpt=%d, dir=%s, new=%d, cur=%d\n", 902 pmf, PM_DEVICE(dip), cmpt, pm_decode_direction(direction), 903 level, cur_power(cp))) 904 905 if (pm_set_power(dip, cmpt, level, direction, 906 PM_CANBLOCK_BLOCK, 0, &result) != DDI_SUCCESS) { 907 if (direction == PM_LEVEL_UPONLY) { 908 pathbuf = kmem_alloc(MAXPATHLEN, KM_SLEEP); 909 (void) ddi_pathname(dip, pathbuf); 910 cmn_err(CE_WARN, "Device %s failed to power up.", 911 pathbuf); 912 kmem_free(pathbuf, MAXPATHLEN); 913 } 914 PMD(PMD_DIN | PMD_FAIL, ("%s: %s@%s(%s#%d) [%d] %s->%d failed, " 915 "errno %d\n", pmf, PM_DEVICE(dip), cmpt, 916 pm_decode_direction(direction), level, result)) 917 return (DDI_FAILURE); 918 } 919 920 PMD(PMD_RESCAN | PMD_DIN, ("%s: pm_rescan %s@%s(%s#%d)\n", pmf, 921 PM_DEVICE(dip))) 922 pm_rescan(dip); 923 return (DDI_SUCCESS); 924 } 925 926 /* 927 * We can get multiple pm_rescan() threads, if one of them discovers 928 * that no scan is running at the moment, it kicks it into action. 929 * Otherwise, it tells the current scanning thread to scan again when 930 * it is done by asserting the PM_SCAN_AGAIN flag. The PM_SCANNING and 931 * PM_SCAN_AGAIN flags are used to regulate scan, to make sure only one 932 * thread at a time runs the pm_scan_dev() code. 933 */ 934 void 935 pm_rescan(void *arg) 936 { 937 PMD_FUNC(pmf, "rescan") 938 dev_info_t *dip = (dev_info_t *)arg; 939 pm_info_t *info; 940 pm_scan_t *scanp; 941 timeout_id_t scanid; 942 943 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d)\n", pmf, PM_DEVICE(dip))) 944 PM_LOCK_DIP(dip); 945 info = PM_GET_PM_INFO(dip); 946 scanp = PM_GET_PM_SCAN(dip); 947 if (pm_scans_disabled || !PM_SCANABLE(dip) || !info || !scanp || 948 (scanp->ps_scan_flags & PM_SCAN_STOP)) { 949 PM_UNLOCK_DIP(dip); 950 return; 951 } 952 if (scanp->ps_scan_flags & PM_SCANNING) { 953 scanp->ps_scan_flags |= PM_SCAN_AGAIN; 954 PM_UNLOCK_DIP(dip); 955 return; 956 } else if (scanp->ps_scan_id) { 957 scanid = scanp->ps_scan_id; 958 scanp->ps_scan_id = 0; 959 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d): cancel timeout scanid %lx\n", 960 pmf, PM_DEVICE(dip), (ulong_t)scanid)) 961 PM_UNLOCK_DIP(dip); 962 (void) untimeout(scanid); 963 PM_LOCK_DIP(dip); 964 } 965 966 /* 967 * Dispatching pm_scan during attach time is risky due to the fact that 968 * attach might soon fail and dip dissolved, and panic may happen while 969 * attempting to stop scan. So schedule a pm_rescan instead. 970 * (Note that if either of the first two terms are true, taskq_dispatch 971 * will not be invoked). 972 * 973 * Multiple pm_scan dispatching is unecessary and costly to keep track 974 * of. The PM_SCAN_DISPATCHED flag is used between pm_rescan and pm_scan 975 * to regulate the dispatching. 976 * 977 * Scan is stopped before the device is detached (in pm_detaching()) 978 * but it may get re-started during the post_detach processing if the 979 * driver fails to detach. 980 */ 981 if (DEVI_IS_ATTACHING(dip) || 982 (scanp->ps_scan_flags & PM_SCAN_DISPATCHED) || 983 !taskq_dispatch(system_taskq, pm_scan, (void *)dip, TQ_NOSLEEP)) { 984 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d): attaching, pm_scan already " 985 "dispatched or dispatching failed\n", pmf, PM_DEVICE(dip))) 986 if (scanp->ps_scan_id) { 987 scanid = scanp->ps_scan_id; 988 scanp->ps_scan_id = 0; 989 PM_UNLOCK_DIP(dip); 990 (void) untimeout(scanid); 991 PM_LOCK_DIP(dip); 992 if (scanp->ps_scan_id) { 993 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d): a competing " 994 "thread scheduled pm_rescan, scanid %lx\n", 995 pmf, PM_DEVICE(dip), 996 (ulong_t)scanp->ps_scan_id)) 997 PM_UNLOCK_DIP(dip); 998 return; 999 } 1000 } 1001 scanp->ps_scan_id = timeout(pm_rescan, (void *)dip, 1002 (scanp->ps_idle_down ? pm_id_ticks : 1003 (pm_min_scan * hz))); 1004 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d): scheduled next pm_rescan, " 1005 "scanid %lx\n", pmf, PM_DEVICE(dip), 1006 (ulong_t)scanp->ps_scan_id)) 1007 } else { 1008 PMD(PMD_SCAN, ("%s: dispatched pm_scan for %s@%s(%s#%d)\n", 1009 pmf, PM_DEVICE(dip))) 1010 scanp->ps_scan_flags |= PM_SCAN_DISPATCHED; 1011 } 1012 PM_UNLOCK_DIP(dip); 1013 } 1014 1015 void 1016 pm_scan(void *arg) 1017 { 1018 PMD_FUNC(pmf, "scan") 1019 dev_info_t *dip = (dev_info_t *)arg; 1020 pm_scan_t *scanp; 1021 time_t nextscan; 1022 1023 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d)\n", pmf, PM_DEVICE(dip))) 1024 1025 PM_LOCK_DIP(dip); 1026 scanp = PM_GET_PM_SCAN(dip); 1027 ASSERT(scanp && PM_GET_PM_INFO(dip)); 1028 1029 if (pm_scans_disabled || !PM_SCANABLE(dip) || 1030 (scanp->ps_scan_flags & PM_SCAN_STOP)) { 1031 scanp->ps_scan_flags &= ~(PM_SCAN_AGAIN | PM_SCAN_DISPATCHED); 1032 PM_UNLOCK_DIP(dip); 1033 return; 1034 } 1035 1036 if (scanp->ps_idle_down) { 1037 /* 1038 * make sure we remember idledown was in affect until 1039 * we've completed the scan 1040 */ 1041 PMID_SET_SCANS(scanp->ps_idle_down) 1042 PMD(PMD_IDLEDOWN, ("%s: %s@%s(%s#%d): idledown starts " 1043 "(pmid %x)\n", pmf, PM_DEVICE(dip), scanp->ps_idle_down)) 1044 } 1045 1046 /* possible having two threads running pm_scan() */ 1047 if (scanp->ps_scan_flags & PM_SCANNING) { 1048 scanp->ps_scan_flags |= PM_SCAN_AGAIN; 1049 PMD(PMD_SCAN, ("%s: scanning, will scan %s@%s(%s#%d) again\n", 1050 pmf, PM_DEVICE(dip))) 1051 scanp->ps_scan_flags &= ~PM_SCAN_DISPATCHED; 1052 PM_UNLOCK_DIP(dip); 1053 return; 1054 } 1055 1056 scanp->ps_scan_flags |= PM_SCANNING; 1057 scanp->ps_scan_flags &= ~PM_SCAN_DISPATCHED; 1058 do { 1059 scanp->ps_scan_flags &= ~PM_SCAN_AGAIN; 1060 PM_UNLOCK_DIP(dip); 1061 nextscan = pm_scan_dev(dip); 1062 PM_LOCK_DIP(dip); 1063 } while (scanp->ps_scan_flags & PM_SCAN_AGAIN); 1064 1065 ASSERT(scanp->ps_scan_flags & PM_SCANNING); 1066 scanp->ps_scan_flags &= ~PM_SCANNING; 1067 1068 if (scanp->ps_idle_down) { 1069 scanp->ps_idle_down &= ~PMID_SCANS; 1070 PMD(PMD_IDLEDOWN, ("%s: %s@%s(%s#%d): idledown ends " 1071 "(pmid %x)\n", pmf, PM_DEVICE(dip), scanp->ps_idle_down)) 1072 } 1073 1074 /* schedule for next idle check */ 1075 if (nextscan != LONG_MAX) { 1076 if (nextscan > (LONG_MAX / hz)) 1077 nextscan = (LONG_MAX - 1) / hz; 1078 if (scanp->ps_scan_id) { 1079 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d): while scanning " 1080 "another rescan scheduled scanid(%lx)\n", pmf, 1081 PM_DEVICE(dip), (ulong_t)scanp->ps_scan_id)) 1082 PM_UNLOCK_DIP(dip); 1083 return; 1084 } else if (!(scanp->ps_scan_flags & PM_SCAN_STOP)) { 1085 scanp->ps_scan_id = timeout(pm_rescan, (void *)dip, 1086 (clock_t)(nextscan * hz)); 1087 PMD(PMD_SCAN, ("%s: nextscan for %s@%s(%s#%d) in " 1088 "%lx sec, scanid(%lx) \n", pmf, PM_DEVICE(dip), 1089 (ulong_t)nextscan, (ulong_t)scanp->ps_scan_id)) 1090 } 1091 } 1092 PM_UNLOCK_DIP(dip); 1093 } 1094 1095 void 1096 pm_get_timestamps(dev_info_t *dip, time_t *valuep) 1097 { 1098 int components = PM_NUMCMPTS(dip); 1099 int i; 1100 1101 ASSERT(components > 0); 1102 PM_LOCK_BUSY(dip); /* so we get a consistent view */ 1103 for (i = 0; i < components; i++) { 1104 valuep[i] = PM_CP(dip, i)->pmc_timestamp; 1105 } 1106 PM_UNLOCK_BUSY(dip); 1107 } 1108 1109 /* 1110 * Returns true if device needs to be kept up because it exported the 1111 * "no-involuntary-power-cycles" property or we're pretending it did (console 1112 * fb case) or it is an ancestor of such a device and has used up the "one 1113 * free cycle" allowed when all such leaf nodes have voluntarily powered down 1114 * upon detach 1115 */ 1116 int 1117 pm_noinvol(dev_info_t *dip) 1118 { 1119 PMD_FUNC(pmf, "noinvol") 1120 1121 /* 1122 * This doesn't change over the life of a driver, so no locking needed 1123 */ 1124 if (PM_IS_CFB(dip)) { 1125 PMD(PMD_NOINVOL | PMD_CFB, ("%s: inhibits CFB %s@%s(%s#%d)\n", 1126 pmf, PM_DEVICE(dip))) 1127 return (1); 1128 } 1129 /* 1130 * Not an issue if no such kids 1131 */ 1132 if (DEVI(dip)->devi_pm_noinvolpm == 0) { 1133 #ifdef DEBUG 1134 if (DEVI(dip)->devi_pm_volpmd != 0) { 1135 dev_info_t *pdip = dip; 1136 do { 1137 PMD(PMD_NOINVOL, ("%s: %s@%s(%s#%d) noinvol %d " 1138 "volpmd %d\n", pmf, PM_DEVICE(pdip), 1139 DEVI(pdip)->devi_pm_noinvolpm, 1140 DEVI(pdip)->devi_pm_volpmd)) 1141 pdip = ddi_get_parent(pdip); 1142 } while (pdip); 1143 } 1144 #endif 1145 ASSERT(DEVI(dip)->devi_pm_volpmd == 0); 1146 return (0); 1147 } 1148 1149 /* 1150 * Since we now maintain the counts correct at every node, we no longer 1151 * need to look up the tree. An ancestor cannot use up the free cycle 1152 * without the children getting their counts adjusted. 1153 */ 1154 1155 #ifdef DEBUG 1156 if (DEVI(dip)->devi_pm_noinvolpm != DEVI(dip)->devi_pm_volpmd) 1157 PMD(PMD_NOINVOL, ("%s: (%d != %d) inhibits %s@%s(%s#%d)\n", pmf, 1158 DEVI(dip)->devi_pm_noinvolpm, DEVI(dip)->devi_pm_volpmd, 1159 PM_DEVICE(dip))) 1160 #endif 1161 return (DEVI(dip)->devi_pm_noinvolpm != DEVI(dip)->devi_pm_volpmd); 1162 } 1163 1164 /* 1165 * This function performs the actual scanning of the device. 1166 * It attempts to power off the indicated device's components if they have 1167 * been idle and other restrictions are met. 1168 * pm_scan_dev calculates and returns when the next scan should happen for 1169 * this device. 1170 */ 1171 time_t 1172 pm_scan_dev(dev_info_t *dip) 1173 { 1174 PMD_FUNC(pmf, "scan_dev") 1175 pm_scan_t *scanp; 1176 time_t *timestamp, idletime, now, thresh; 1177 time_t timeleft = 0; 1178 #ifdef PMDDEBUG 1179 int curpwr; 1180 #endif 1181 int i, nxtpwr, pwrndx, unused; 1182 size_t size; 1183 pm_component_t *cp; 1184 dev_info_t *pdip = ddi_get_parent(dip); 1185 int circ; 1186 static int cur_threshold(dev_info_t *, int); 1187 static int pm_next_lower_power(pm_component_t *, int); 1188 1189 /* 1190 * skip attaching device 1191 */ 1192 if (DEVI_IS_ATTACHING(dip)) { 1193 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d) is attaching, timeleft(%lx)\n", 1194 pmf, PM_DEVICE(dip), pm_min_scan)) 1195 return (pm_min_scan); 1196 } 1197 1198 PM_LOCK_DIP(dip); 1199 scanp = PM_GET_PM_SCAN(dip); 1200 ASSERT(scanp && PM_GET_PM_INFO(dip)); 1201 1202 PMD(PMD_SCAN, ("%s: [BEGIN %s@%s(%s#%d)]\n", pmf, PM_DEVICE(dip))) 1203 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d): kuc is %d\n", pmf, PM_DEVICE(dip), 1204 PM_KUC(dip))) 1205 1206 /* no scan under the following conditions */ 1207 if (pm_scans_disabled || !PM_SCANABLE(dip) || 1208 (scanp->ps_scan_flags & PM_SCAN_STOP) || 1209 (PM_KUC(dip) != 0) || 1210 PM_ISDIRECT(dip) || pm_noinvol(dip)) { 1211 PM_UNLOCK_DIP(dip); 1212 PMD(PMD_SCAN, ("%s: [END, %s@%s(%s#%d)] no scan, " 1213 "scan_disabled(%d), apm_enabled(%d), cpupm(%d), " 1214 "kuc(%d), %s directpm, %s pm_noinvol\n", 1215 pmf, PM_DEVICE(dip), pm_scans_disabled, autopm_enabled, 1216 cpupm, PM_KUC(dip), 1217 PM_ISDIRECT(dip) ? "is" : "is not", 1218 pm_noinvol(dip) ? "is" : "is not")) 1219 return (LONG_MAX); 1220 } 1221 PM_UNLOCK_DIP(dip); 1222 1223 if (!ndi_devi_tryenter(pdip, &circ)) { 1224 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d) can't hold pdip", 1225 pmf, PM_DEVICE(pdip))) 1226 return ((time_t)1); 1227 } 1228 now = gethrestime_sec(); 1229 size = PM_NUMCMPTS(dip) * sizeof (time_t); 1230 timestamp = kmem_alloc(size, KM_SLEEP); 1231 pm_get_timestamps(dip, timestamp); 1232 1233 /* 1234 * Since we removed support for backwards compatible devices, 1235 * (see big comment at top of file) 1236 * it is no longer required to deal with component 0 last. 1237 */ 1238 for (i = 0; i < PM_NUMCMPTS(dip); i++) { 1239 /* 1240 * If already off (an optimization, perhaps) 1241 */ 1242 cp = PM_CP(dip, i); 1243 pwrndx = cp->pmc_cur_pwr; 1244 #ifdef PMDDEBUG 1245 curpwr = (pwrndx == PM_LEVEL_UNKNOWN) ? 1246 PM_LEVEL_UNKNOWN : 1247 cp->pmc_comp.pmc_lvals[pwrndx]; 1248 #endif 1249 1250 if (pwrndx == 0) { 1251 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d) comp %d off or " 1252 "lowest\n", pmf, PM_DEVICE(dip), i)) 1253 /* skip device if off or at its lowest */ 1254 continue; 1255 } 1256 1257 thresh = cur_threshold(dip, i); /* comp i threshold */ 1258 if ((timestamp[i] == 0) || (cp->pmc_busycount > 0)) { 1259 /* were busy or newly became busy by another thread */ 1260 if (timeleft == 0) 1261 timeleft = max(thresh, pm_min_scan); 1262 else 1263 timeleft = min( 1264 timeleft, max(thresh, pm_min_scan)); 1265 continue; 1266 } 1267 1268 idletime = now - timestamp[i]; /* idle time */ 1269 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d) comp %d idle time %lx\n", 1270 pmf, PM_DEVICE(dip), i, idletime)) 1271 if (idletime >= thresh || PM_IS_PID(dip)) { 1272 nxtpwr = pm_next_lower_power(cp, pwrndx); 1273 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d) comp %d, %d->%d\n", 1274 pmf, PM_DEVICE(dip), i, curpwr, nxtpwr)) 1275 if (pm_set_power(dip, i, nxtpwr, PM_LEVEL_DOWNONLY, 1276 PM_CANBLOCK_FAIL, 1, &unused) != DDI_SUCCESS && 1277 PM_CURPOWER(dip, i) != nxtpwr) { 1278 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d) comp %d, " 1279 "%d->%d Failed\n", pmf, PM_DEVICE(dip), 1280 i, curpwr, nxtpwr)) 1281 timeleft = pm_min_scan; 1282 continue; 1283 } else { 1284 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d) comp %d, " 1285 "%d->%d, GOOD curpwr %d\n", pmf, 1286 PM_DEVICE(dip), i, curpwr, nxtpwr, 1287 cur_power(cp))) 1288 1289 if (nxtpwr == 0) /* component went off */ 1290 continue; 1291 1292 /* 1293 * scan to next lower level 1294 */ 1295 if (timeleft == 0) 1296 timeleft = max( 1297 1, cur_threshold(dip, i)); 1298 else 1299 timeleft = min(timeleft, 1300 max(1, cur_threshold(dip, i))); 1301 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d) comp %d, " 1302 "timeleft(%lx)\n", pmf, PM_DEVICE(dip), 1303 i, timeleft)) 1304 } 1305 } else { /* comp not idle long enough */ 1306 if (timeleft == 0) 1307 timeleft = thresh - idletime; 1308 else 1309 timeleft = min(timeleft, (thresh - idletime)); 1310 PMD(PMD_SCAN, ("%s: %s@%s(%s#%d) comp %d, timeleft=" 1311 "%lx\n", pmf, PM_DEVICE(dip), i, timeleft)) 1312 } 1313 } 1314 ndi_devi_exit(pdip, circ); 1315 kmem_free(timestamp, size); 1316 PMD(PMD_SCAN, ("%s: [END %s@%s(%s#%d)] timeleft(%lx)\n", pmf, 1317 PM_DEVICE(dip), timeleft)) 1318 1319 /* 1320 * if components are already at lowest level, timeleft is left 0 1321 */ 1322 return ((timeleft == 0) ? LONG_MAX : timeleft); 1323 } 1324 1325 /* 1326 * pm_scan_stop - cancel scheduled pm_rescan, 1327 * wait for termination of dispatched pm_scan thread 1328 * and active pm_scan_dev thread. 1329 */ 1330 void 1331 pm_scan_stop(dev_info_t *dip) 1332 { 1333 PMD_FUNC(pmf, "scan_stop") 1334 pm_scan_t *scanp; 1335 timeout_id_t scanid; 1336 1337 PMD(PMD_SCAN, ("%s: [BEGIN %s@%s(%s#%d)]\n", pmf, PM_DEVICE(dip))) 1338 PM_LOCK_DIP(dip); 1339 scanp = PM_GET_PM_SCAN(dip); 1340 if (!scanp) { 1341 PMD(PMD_SCAN, ("%s: [END %s@%s(%s#%d)] scan not initialized\n", 1342 pmf, PM_DEVICE(dip))) 1343 PM_UNLOCK_DIP(dip); 1344 return; 1345 } 1346 scanp->ps_scan_flags |= PM_SCAN_STOP; 1347 1348 /* cancel scheduled scan taskq */ 1349 while (scanp->ps_scan_id) { 1350 scanid = scanp->ps_scan_id; 1351 scanp->ps_scan_id = 0; 1352 PM_UNLOCK_DIP(dip); 1353 (void) untimeout(scanid); 1354 PM_LOCK_DIP(dip); 1355 } 1356 1357 while (scanp->ps_scan_flags & (PM_SCANNING | PM_SCAN_DISPATCHED)) { 1358 PM_UNLOCK_DIP(dip); 1359 delay(1); 1360 PM_LOCK_DIP(dip); 1361 } 1362 PM_UNLOCK_DIP(dip); 1363 PMD(PMD_SCAN, ("%s: [END %s@%s(%s#%d)]\n", pmf, PM_DEVICE(dip))) 1364 } 1365 1366 int 1367 pm_scan_stop_walk(dev_info_t *dip, void *arg) 1368 { 1369 _NOTE(ARGUNUSED(arg)) 1370 1371 if (!PM_GET_PM_SCAN(dip)) 1372 return (DDI_WALK_CONTINUE); 1373 ASSERT(!PM_ISBC(dip)); 1374 pm_scan_stop(dip); 1375 return (DDI_WALK_CONTINUE); 1376 } 1377 1378 /* 1379 * Converts a power level value to its index 1380 */ 1381 static int 1382 power_val_to_index(pm_component_t *cp, int val) 1383 { 1384 int limit, i, *ip; 1385 1386 ASSERT(val != PM_LEVEL_UPONLY && val != PM_LEVEL_DOWNONLY && 1387 val != PM_LEVEL_EXACT); 1388 /* convert power value into index (i) */ 1389 limit = cp->pmc_comp.pmc_numlevels; 1390 ip = cp->pmc_comp.pmc_lvals; 1391 for (i = 0; i < limit; i++) 1392 if (val == *ip++) 1393 return (i); 1394 return (-1); 1395 } 1396 1397 /* 1398 * Converts a numeric power level to a printable string 1399 */ 1400 static char * 1401 power_val_to_string(pm_component_t *cp, int val) 1402 { 1403 int index; 1404 1405 if (val == PM_LEVEL_UPONLY) 1406 return ("<UPONLY>"); 1407 1408 if (val == PM_LEVEL_UNKNOWN || 1409 (index = power_val_to_index(cp, val)) == -1) 1410 return ("<LEVEL_UNKNOWN>"); 1411 1412 return (cp->pmc_comp.pmc_lnames[index]); 1413 } 1414 1415 /* 1416 * Return true if this node has been claimed by a ppm. 1417 */ 1418 static int 1419 pm_ppm_claimed(dev_info_t *dip) 1420 { 1421 return (PPM(dip) != NULL); 1422 } 1423 1424 /* 1425 * A node which was voluntarily power managed has just used up its "free cycle" 1426 * and need is volpmd field cleared, and the same done to all its descendents 1427 */ 1428 static void 1429 pm_clear_volpm_dip(dev_info_t *dip) 1430 { 1431 PMD_FUNC(pmf, "clear_volpm_dip") 1432 1433 if (dip == NULL) 1434 return; 1435 PMD(PMD_NOINVOL, ("%s: clear volpm from %s@%s(%s#%d)\n", pmf, 1436 PM_DEVICE(dip))) 1437 DEVI(dip)->devi_pm_volpmd = 0; 1438 for (dip = ddi_get_child(dip); dip; dip = ddi_get_next_sibling(dip)) { 1439 pm_clear_volpm_dip(dip); 1440 } 1441 } 1442 1443 /* 1444 * A node which was voluntarily power managed has used up the "free cycles" 1445 * for the subtree that it is the root of. Scan through the list of detached 1446 * nodes and adjust the counts of any that are descendents of the node. 1447 */ 1448 static void 1449 pm_clear_volpm_list(dev_info_t *dip) 1450 { 1451 PMD_FUNC(pmf, "clear_volpm_list") 1452 char *pathbuf; 1453 size_t len; 1454 pm_noinvol_t *ip; 1455 1456 pathbuf = kmem_alloc(MAXPATHLEN, KM_SLEEP); 1457 (void) ddi_pathname(dip, pathbuf); 1458 len = strlen(pathbuf); 1459 PMD(PMD_NOINVOL, ("%s: clear volpm list %s\n", pmf, pathbuf)) 1460 rw_enter(&pm_noinvol_rwlock, RW_WRITER); 1461 for (ip = pm_noinvol_head; ip; ip = ip->ni_next) { 1462 PMD(PMD_NOINVOL, ("%s: clear volpm: ni_path %s\n", pmf, 1463 ip->ni_path)) 1464 if (strncmp(pathbuf, ip->ni_path, len) == 0 && 1465 ip->ni_path[len] == '/') { 1466 PMD(PMD_NOINVOL, ("%s: clear volpm: %s\n", pmf, 1467 ip->ni_path)) 1468 ip->ni_volpmd = 0; 1469 ip->ni_wasvolpmd = 0; 1470 } 1471 } 1472 kmem_free(pathbuf, MAXPATHLEN); 1473 rw_exit(&pm_noinvol_rwlock); 1474 } 1475 1476 /* 1477 * Powers a device, suspending or resuming the driver if it is a backward 1478 * compatible device, calling into ppm to change power level. 1479 * Called with the component's power lock held. 1480 */ 1481 static int 1482 power_dev(dev_info_t *dip, int comp, int level, int old_level, 1483 pm_canblock_t canblock, pm_ppm_devlist_t **devlist) 1484 { 1485 PMD_FUNC(pmf, "power_dev") 1486 power_req_t power_req; 1487 int power_op_ret; /* DDI_SUCCESS or DDI_FAILURE */ 1488 int resume_needed = 0; 1489 int suspended = 0; 1490 int result; 1491 #ifdef PMDDEBUG 1492 struct pm_component *cp = PM_CP(dip, comp); 1493 #endif 1494 int bc = PM_ISBC(dip); 1495 int pm_all_components_off(dev_info_t *); 1496 int clearvolpmd = 0; 1497 char pathbuf[MAXNAMELEN]; 1498 #ifdef PMDDEBUG 1499 char *ppmname, *ppmaddr; 1500 #endif 1501 /* 1502 * If this is comp 0 of a backwards compat device and we are 1503 * going to take the power away, we need to detach it with 1504 * DDI_PM_SUSPEND command. 1505 */ 1506 if (bc && comp == 0 && POWERING_OFF(old_level, level)) { 1507 if (devi_detach(dip, DDI_PM_SUSPEND) != DDI_SUCCESS) { 1508 /* We could not suspend before turning cmpt zero off */ 1509 PMD(PMD_ERROR, ("%s: could not suspend %s@%s(%s#%d)\n", 1510 pmf, PM_DEVICE(dip))) 1511 return (DDI_FAILURE); 1512 } else { 1513 DEVI(dip)->devi_pm_flags |= PMC_SUSPENDED; 1514 suspended++; 1515 } 1516 } 1517 power_req.request_type = PMR_PPM_SET_POWER; 1518 power_req.req.ppm_set_power_req.who = dip; 1519 power_req.req.ppm_set_power_req.cmpt = comp; 1520 power_req.req.ppm_set_power_req.old_level = old_level; 1521 power_req.req.ppm_set_power_req.new_level = level; 1522 power_req.req.ppm_set_power_req.canblock = canblock; 1523 power_req.req.ppm_set_power_req.cookie = NULL; 1524 #ifdef PMDDEBUG 1525 if (pm_ppm_claimed(dip)) { 1526 ppmname = PM_NAME(PPM(dip)); 1527 ppmaddr = PM_ADDR(PPM(dip)); 1528 1529 } else { 1530 ppmname = "noppm"; 1531 ppmaddr = "0"; 1532 } 1533 PMD(PMD_PPM, ("%s: %s@%s(%s#%d):%s[%d] %s (%d) -> %s (%d) via %s@%s\n", 1534 pmf, PM_DEVICE(dip), cp->pmc_comp.pmc_name, comp, 1535 power_val_to_string(cp, old_level), old_level, 1536 power_val_to_string(cp, level), level, ppmname, ppmaddr)) 1537 #endif 1538 /* 1539 * If non-bc noinvolpm device is turning first comp on, or noinvolpm 1540 * bc device comp 0 is powering on, then we count it as a power cycle 1541 * against its voluntary count. 1542 */ 1543 if (DEVI(dip)->devi_pm_volpmd && 1544 (!bc && pm_all_components_off(dip) && level != 0) || 1545 (bc && comp == 0 && POWERING_ON(old_level, level))) 1546 clearvolpmd = 1; 1547 if ((power_op_ret = pm_ctlops(PPM(dip), dip, DDI_CTLOPS_POWER, 1548 &power_req, &result)) == DDI_SUCCESS) { 1549 /* 1550 * Now do involuntary pm accounting; If we've just cycled power 1551 * on a voluntarily pm'd node, and by inference on its entire 1552 * subtree, we need to set the subtree (including those nodes 1553 * already detached) volpmd counts to 0, and subtract out the 1554 * value of the current node's volpmd count from the ancestors 1555 */ 1556 if (clearvolpmd) { 1557 int volpmd = DEVI(dip)->devi_pm_volpmd; 1558 pm_clear_volpm_dip(dip); 1559 pm_clear_volpm_list(dip); 1560 if (volpmd) { 1561 (void) ddi_pathname(dip, pathbuf); 1562 (void) pm_noinvol_update(PM_BP_NOINVOL_POWER, 1563 volpmd, 0, pathbuf, dip); 1564 } 1565 } 1566 } else { 1567 PMD(PMD_FAIL, ("%s: can't set comp %d (%s) of %s@%s(%s#%d) " 1568 "to level %d (%s)\n", pmf, comp, cp->pmc_comp.pmc_name, 1569 PM_DEVICE(dip), level, power_val_to_string(cp, level))) 1570 } 1571 /* 1572 * If some other devices were also powered up (e.g. other cpus in 1573 * the same domain) return a pointer to that list 1574 */ 1575 if (devlist) { 1576 *devlist = (pm_ppm_devlist_t *) 1577 power_req.req.ppm_set_power_req.cookie; 1578 } 1579 /* 1580 * We will have to resume the device if the device is backwards compat 1581 * device and either of the following is true: 1582 * -This is comp 0 and we have successfully powered it up 1583 * -This is comp 0 and we have failed to power it down. Resume is 1584 * needed because we have suspended it above 1585 */ 1586 1587 if (bc && comp == 0) { 1588 ASSERT(PM_ISDIRECT(dip) || DEVI_IS_DETACHING(dip)); 1589 if (power_op_ret == DDI_SUCCESS) { 1590 if (POWERING_ON(old_level, level)) { 1591 /* 1592 * It must be either suspended or resumed 1593 * via pm_power_has_changed path 1594 */ 1595 ASSERT((DEVI(dip)->devi_pm_flags & 1596 PMC_SUSPENDED) || 1597 (PM_CP(dip, comp)->pmc_flags & 1598 PM_PHC_WHILE_SET_POWER)); 1599 1600 resume_needed = suspended; 1601 } 1602 } else { 1603 if (POWERING_OFF(old_level, level)) { 1604 /* 1605 * It must be either suspended or resumed 1606 * via pm_power_has_changed path 1607 */ 1608 ASSERT((DEVI(dip)->devi_pm_flags & 1609 PMC_SUSPENDED) || 1610 (PM_CP(dip, comp)->pmc_flags & 1611 PM_PHC_WHILE_SET_POWER)); 1612 1613 resume_needed = suspended; 1614 } 1615 } 1616 } 1617 if (resume_needed) { 1618 ASSERT(DEVI(dip)->devi_pm_flags & PMC_SUSPENDED); 1619 /* ppm is not interested in DDI_PM_RESUME */ 1620 if ((power_op_ret = devi_attach(dip, DDI_PM_RESUME)) == 1621 DDI_SUCCESS) { 1622 DEVI(dip)->devi_pm_flags &= ~PMC_SUSPENDED; 1623 } else 1624 cmn_err(CE_WARN, "!pm: Can't resume %s@%s(%s#%d)", 1625 PM_DEVICE(dip)); 1626 } 1627 return (power_op_ret); 1628 } 1629 1630 /* 1631 * Return true if we are the owner or a borrower of the devi lock. See 1632 * pm_lock_power_single() about borrowing the lock. 1633 */ 1634 static int 1635 pm_devi_lock_held(dev_info_t *dip) 1636 { 1637 lock_loan_t *cur; 1638 1639 if (DEVI_BUSY_OWNED(dip)) 1640 return (1); 1641 1642 /* return false if no locks borrowed */ 1643 if (lock_loan_head.pmlk_next == NULL) 1644 return (0); 1645 1646 mutex_enter(&pm_loan_lock); 1647 /* see if our thread is registered as a lock borrower. */ 1648 for (cur = lock_loan_head.pmlk_next; cur; cur = cur->pmlk_next) 1649 if (cur->pmlk_borrower == curthread) 1650 break; 1651 mutex_exit(&pm_loan_lock); 1652 1653 return (cur != NULL && cur->pmlk_lender == DEVI(dip)->devi_busy_thread); 1654 } 1655 1656 /* 1657 * pm_set_power: adjusts power level of device. Assumes device is power 1658 * manageable & component exists. 1659 * 1660 * Cases which require us to bring up devices we keep up ("wekeepups") for 1661 * backwards compatible devices: 1662 * component 0 is off and we're bringing it up from 0 1663 * bring up wekeepup first 1664 * and recursively when component 0 is off and we bring some other 1665 * component up from 0 1666 * For devices which are not backward compatible, our dependency notion is much 1667 * simpler. Unless all components are off, then wekeeps must be on. 1668 * We don't treat component 0 differently. 1669 * Canblock tells how to deal with a direct pm'd device. 1670 * Scan arg tells us if we were called from scan, in which case we don't need 1671 * to go back to the root node and walk down to change power. 1672 */ 1673 int 1674 pm_set_power(dev_info_t *dip, int comp, int level, int direction, 1675 pm_canblock_t canblock, int scan, int *retp) 1676 { 1677 PMD_FUNC(pmf, "set_power") 1678 char *pathbuf; 1679 pm_bp_child_pwrchg_t bpc; 1680 pm_sp_misc_t pspm; 1681 int ret = DDI_SUCCESS; 1682 int unused = DDI_SUCCESS; 1683 dev_info_t *pdip = ddi_get_parent(dip); 1684 1685 #ifdef DEBUG 1686 int diverted = 0; 1687 1688 /* 1689 * This prevents operations on the console from calling prom_printf and 1690 * either deadlocking or bringing up the console because of debug 1691 * output 1692 */ 1693 if (dip == cfb_dip) { 1694 diverted++; 1695 mutex_enter(&pm_debug_lock); 1696 pm_divertdebug++; 1697 mutex_exit(&pm_debug_lock); 1698 } 1699 #endif 1700 ASSERT(direction == PM_LEVEL_UPONLY || direction == PM_LEVEL_DOWNONLY || 1701 direction == PM_LEVEL_EXACT); 1702 PMD(PMD_SET, ("%s: %s@%s(%s#%d), comp=%d, dir=%s, new=%d\n", 1703 pmf, PM_DEVICE(dip), comp, pm_decode_direction(direction), level)) 1704 pathbuf = kmem_alloc(MAXPATHLEN, KM_SLEEP); 1705 (void) ddi_pathname(dip, pathbuf); 1706 bpc.bpc_dip = dip; 1707 bpc.bpc_path = pathbuf; 1708 bpc.bpc_comp = comp; 1709 bpc.bpc_olevel = PM_CURPOWER(dip, comp); 1710 bpc.bpc_nlevel = level; 1711 pspm.pspm_direction = direction; 1712 pspm.pspm_errnop = retp; 1713 pspm.pspm_canblock = canblock; 1714 pspm.pspm_scan = scan; 1715 bpc.bpc_private = &pspm; 1716 1717 /* 1718 * If a config operation is being done (we've locked the parent) or 1719 * we already hold the power lock (we've locked the node) 1720 * then we can operate directly on the node because we have already 1721 * brought up all the ancestors, otherwise, we have to go back to the 1722 * top of the tree. 1723 */ 1724 if (pm_devi_lock_held(pdip) || pm_devi_lock_held(dip)) 1725 ret = pm_busop_set_power(dip, NULL, BUS_POWER_CHILD_PWRCHG, 1726 (void *)&bpc, (void *)&unused); 1727 else 1728 ret = pm_busop_bus_power(ddi_root_node(), NULL, 1729 BUS_POWER_CHILD_PWRCHG, (void *)&bpc, (void *)&unused); 1730 #ifdef DEBUG 1731 if (ret != DDI_SUCCESS || *retp != DDI_SUCCESS) { 1732 PMD(PMD_ERROR, ("%s: %s@%s(%s#%d) can't change power, ret=%d, " 1733 "errno=%d\n", pmf, PM_DEVICE(dip), ret, *retp)) 1734 } 1735 if (diverted) { 1736 mutex_enter(&pm_debug_lock); 1737 pm_divertdebug--; 1738 mutex_exit(&pm_debug_lock); 1739 } 1740 #endif 1741 kmem_free(pathbuf, MAXPATHLEN); 1742 return (ret); 1743 } 1744 1745 /* 1746 * If holddip is set, then if a dip is found we return with the node held. 1747 * 1748 * This code uses the same locking scheme as e_ddi_hold_devi_by_path 1749 * (resolve_pathname), but it does not drive attach. 1750 */ 1751 dev_info_t * 1752 pm_name_to_dip(char *pathname, int holddip) 1753 { 1754 struct pathname pn; 1755 char *component; 1756 dev_info_t *parent, *child; 1757 int circ; 1758 1759 if ((pathname == NULL) || (*pathname != '/')) 1760 return (NULL); 1761 1762 /* setup pathname and allocate component */ 1763 if (pn_get(pathname, UIO_SYSSPACE, &pn)) 1764 return (NULL); 1765 component = kmem_alloc(MAXNAMELEN, KM_SLEEP); 1766 1767 /* start at top, process '/' component */ 1768 parent = child = ddi_root_node(); 1769 ndi_hold_devi(parent); 1770 pn_skipslash(&pn); 1771 ASSERT(i_ddi_devi_attached(parent)); 1772 1773 /* process components of pathname */ 1774 while (pn_pathleft(&pn)) { 1775 (void) pn_getcomponent(&pn, component); 1776 1777 /* enter parent and search for component child */ 1778 ndi_devi_enter(parent, &circ); 1779 child = ndi_devi_findchild(parent, component); 1780 if ((child == NULL) || !i_ddi_devi_attached(child)) { 1781 child = NULL; 1782 ndi_devi_exit(parent, circ); 1783 ndi_rele_devi(parent); 1784 goto out; 1785 } 1786 1787 /* attached child found, hold child and release parent */ 1788 ndi_hold_devi(child); 1789 ndi_devi_exit(parent, circ); 1790 ndi_rele_devi(parent); 1791 1792 /* child becomes parent, and process next component */ 1793 parent = child; 1794 pn_skipslash(&pn); 1795 1796 /* loop with active ndi_devi_hold of child->parent */ 1797 } 1798 1799 out: 1800 pn_free(&pn); 1801 kmem_free(component, MAXNAMELEN); 1802 1803 /* if we are not asked to return with hold, drop current hold */ 1804 if (child && !holddip) 1805 ndi_rele_devi(child); 1806 return (child); 1807 } 1808 1809 /* 1810 * Search for a dependency and mark it unsatisfied 1811 */ 1812 static void 1813 pm_unsatisfy(char *keeper, char *kept) 1814 { 1815 PMD_FUNC(pmf, "unsatisfy") 1816 pm_pdr_t *dp; 1817 1818 PMD(PMD_KEEPS, ("%s: keeper=%s, kept=%s\n", pmf, keeper, kept)) 1819 for (dp = pm_dep_head; dp; dp = dp->pdr_next) { 1820 if (!dp->pdr_isprop) { 1821 if (strcmp(dp->pdr_keeper, keeper) == 0 && 1822 (dp->pdr_kept_count > 0) && 1823 strcmp(dp->pdr_kept_paths[0], kept) == 0) { 1824 if (dp->pdr_satisfied) { 1825 dp->pdr_satisfied = 0; 1826 pm_unresolved_deps++; 1827 PMD(PMD_KEEPS, ("%s: clear satisfied, " 1828 "pm_unresolved_deps now %d\n", pmf, 1829 pm_unresolved_deps)) 1830 } 1831 } 1832 } 1833 } 1834 } 1835 1836 /* 1837 * Device dip is being un power managed, it keeps up count other devices. 1838 * We need to release any hold we have on the kept devices, and also 1839 * mark the dependency no longer satisfied. 1840 */ 1841 static void 1842 pm_unkeeps(int count, char *keeper, char **keptpaths, int pwr) 1843 { 1844 PMD_FUNC(pmf, "unkeeps") 1845 int i, j; 1846 dev_info_t *kept; 1847 dev_info_t *dip; 1848 struct pm_component *cp; 1849 int keeper_on = 0, circ; 1850 1851 PMD(PMD_KEEPS, ("%s: count=%d, keeper=%s, keptpaths=%p\n", pmf, count, 1852 keeper, (void *)keptpaths)) 1853 /* 1854 * Try to grab keeper. Keeper may have gone away by now, 1855 * in this case, used the passed in value pwr 1856 */ 1857 dip = pm_name_to_dip(keeper, 1); 1858 for (i = 0; i < count; i++) { 1859 /* Release power hold */ 1860 kept = pm_name_to_dip(keptpaths[i], 1); 1861 if (kept) { 1862 PMD(PMD_KEEPS, ("%s: %s@%s(%s#%d)[%d]\n", pmf, 1863 PM_DEVICE(kept), i)) 1864 /* 1865 * We need to check if we skipped a bringup here 1866 * because we could have failed the bringup 1867 * (ie DIRECT PM device) and have 1868 * not increment the count. 1869 */ 1870 if ((dip != NULL) && (PM_GET_PM_INFO(dip) != NULL)) { 1871 keeper_on = 0; 1872 PM_LOCK_POWER(dip, &circ); 1873 for (j = 0; j < PM_NUMCMPTS(dip); j++) { 1874 cp = &DEVI(dip)->devi_pm_components[j]; 1875 if (cur_power(cp)) { 1876 keeper_on++; 1877 break; 1878 } 1879 } 1880 if (keeper_on && (PM_SKBU(kept) == 0)) { 1881 pm_rele_power(kept); 1882 DEVI(kept)->devi_pm_flags 1883 &= ~PMC_SKIP_BRINGUP; 1884 } 1885 PM_UNLOCK_POWER(dip, circ); 1886 } else if (pwr) { 1887 if (PM_SKBU(kept) == 0) { 1888 pm_rele_power(kept); 1889 DEVI(kept)->devi_pm_flags 1890 &= ~PMC_SKIP_BRINGUP; 1891 } 1892 } 1893 ddi_release_devi(kept); 1894 } 1895 /* 1896 * mark this dependency not satisfied 1897 */ 1898 pm_unsatisfy(keeper, keptpaths[i]); 1899 } 1900 if (dip) 1901 ddi_release_devi(dip); 1902 } 1903 1904 /* 1905 * Device kept is being un power managed, it is kept up by keeper. 1906 * We need to mark the dependency no longer satisfied. 1907 */ 1908 static void 1909 pm_unkepts(char *kept, char *keeper) 1910 { 1911 PMD_FUNC(pmf, "unkepts") 1912 PMD(PMD_KEEPS, ("%s: kept=%s, keeper=%s\n", pmf, kept, keeper)) 1913 ASSERT(keeper != NULL); 1914 /* 1915 * mark this dependency not satisfied 1916 */ 1917 pm_unsatisfy(keeper, kept); 1918 } 1919 1920 /* 1921 * Removes dependency information and hold on the kepts, if the path is a 1922 * path of a keeper. 1923 */ 1924 static void 1925 pm_free_keeper(char *path, int pwr) 1926 { 1927 pm_pdr_t *dp; 1928 int i; 1929 size_t length; 1930 1931 for (dp = pm_dep_head; dp; dp = dp->pdr_next) { 1932 if (strcmp(dp->pdr_keeper, path) != 0) 1933 continue; 1934 /* 1935 * Remove all our kept holds and the dependency records, 1936 * then free up the kept lists. 1937 */ 1938 pm_unkeeps(dp->pdr_kept_count, path, dp->pdr_kept_paths, pwr); 1939 if (dp->pdr_kept_count) { 1940 for (i = 0; i < dp->pdr_kept_count; i++) { 1941 length = strlen(dp->pdr_kept_paths[i]); 1942 kmem_free(dp->pdr_kept_paths[i], length + 1); 1943 } 1944 kmem_free(dp->pdr_kept_paths, 1945 dp->pdr_kept_count * sizeof (char **)); 1946 dp->pdr_kept_paths = NULL; 1947 dp->pdr_kept_count = 0; 1948 } 1949 } 1950 } 1951 1952 /* 1953 * Removes the device represented by path from the list of kepts, if the 1954 * path is a path of a kept 1955 */ 1956 static void 1957 pm_free_kept(char *path) 1958 { 1959 pm_pdr_t *dp; 1960 int i; 1961 int j, count; 1962 size_t length; 1963 char **paths; 1964 1965 for (dp = pm_dep_head; dp; dp = dp->pdr_next) { 1966 if (dp->pdr_kept_count == 0) 1967 continue; 1968 count = dp->pdr_kept_count; 1969 /* Remove this device from the kept path lists */ 1970 for (i = 0; i < count; i++) { 1971 if (strcmp(dp->pdr_kept_paths[i], path) == 0) { 1972 pm_unkepts(path, dp->pdr_keeper); 1973 length = strlen(dp->pdr_kept_paths[i]) + 1; 1974 kmem_free(dp->pdr_kept_paths[i], length); 1975 dp->pdr_kept_paths[i] = NULL; 1976 dp->pdr_kept_count--; 1977 } 1978 } 1979 /* Compact the kept paths array */ 1980 if (dp->pdr_kept_count) { 1981 length = dp->pdr_kept_count * sizeof (char **); 1982 paths = kmem_zalloc(length, KM_SLEEP); 1983 j = 0; 1984 for (i = 0; i < count; i++) { 1985 if (dp->pdr_kept_paths[i] != NULL) { 1986 paths[j] = dp->pdr_kept_paths[i]; 1987 j++; 1988 } 1989 } 1990 ASSERT(j == dp->pdr_kept_count); 1991 } 1992 /* Now free the old array and point to the new one */ 1993 kmem_free(dp->pdr_kept_paths, count * sizeof (char **)); 1994 if (dp->pdr_kept_count) 1995 dp->pdr_kept_paths = paths; 1996 else 1997 dp->pdr_kept_paths = NULL; 1998 } 1999 } 2000 2001 /* 2002 * Free the dependency information for a device. 2003 */ 2004 void 2005 pm_free_keeps(char *path, int pwr) 2006 { 2007 PMD_FUNC(pmf, "free_keeps") 2008 2009 #ifdef DEBUG 2010 int doprdeps = 0; 2011 void prdeps(char *); 2012 2013 PMD(PMD_KEEPS, ("%s: %s\n", pmf, path)) 2014 if (pm_debug & PMD_KEEPS) { 2015 doprdeps = 1; 2016 prdeps("pm_free_keeps before"); 2017 } 2018 #endif 2019 /* 2020 * First assume we are a keeper and remove all our kepts. 2021 */ 2022 pm_free_keeper(path, pwr); 2023 /* 2024 * Now assume we a kept device, and remove all our records. 2025 */ 2026 pm_free_kept(path); 2027 #ifdef DEBUG 2028 if (doprdeps) { 2029 prdeps("pm_free_keeps after"); 2030 } 2031 #endif 2032 } 2033 2034 static int 2035 pm_is_kept(char *path) 2036 { 2037 pm_pdr_t *dp; 2038 int i; 2039 2040 for (dp = pm_dep_head; dp; dp = dp->pdr_next) { 2041 if (dp->pdr_kept_count == 0) 2042 continue; 2043 for (i = 0; i < dp->pdr_kept_count; i++) { 2044 if (strcmp(dp->pdr_kept_paths[i], path) == 0) 2045 return (1); 2046 } 2047 } 2048 return (0); 2049 } 2050 2051 static void 2052 e_pm_hold_rele_power(dev_info_t *dip, int cnt) 2053 { 2054 PMD_FUNC(pmf, "hold_rele_power") 2055 int circ; 2056 2057 if ((dip == NULL) || 2058 (PM_GET_PM_INFO(dip) == NULL) || PM_ISBC(dip)) 2059 return; 2060 2061 PM_LOCK_POWER(dip, &circ); 2062 ASSERT(cnt >= 0 && PM_KUC(dip) >= 0 || cnt < 0 && PM_KUC(dip) > 0); 2063 PMD(PMD_KIDSUP, ("%s: kidsupcnt for %s@%s(%s#%d) %d->%d\n", pmf, 2064 PM_DEVICE(dip), PM_KUC(dip), (PM_KUC(dip) + cnt))) 2065 2066 PM_KUC(dip) += cnt; 2067 2068 ASSERT(PM_KUC(dip) >= 0); 2069 PM_UNLOCK_POWER(dip, circ); 2070 2071 if (cnt < 0 && PM_KUC(dip) == 0) 2072 pm_rescan(dip); 2073 } 2074 2075 #define MAX_PPM_HANDLERS 4 2076 2077 kmutex_t ppm_lock; /* in case we ever do multi-threaded startup */ 2078 2079 struct ppm_callbacks { 2080 int (*ppmc_func)(dev_info_t *); 2081 dev_info_t *ppmc_dip; 2082 } ppm_callbacks[MAX_PPM_HANDLERS + 1]; 2083 2084 2085 /* 2086 * This routine calls into all the registered ppms to notify them 2087 * that either all components of power-managed devices are at their 2088 * lowest levels or no longer all are at their lowest levels. 2089 */ 2090 static void 2091 pm_ppm_notify_all_lowest(dev_info_t *dip, int mode) 2092 { 2093 struct ppm_callbacks *ppmcp; 2094 power_req_t power_req; 2095 int result = 0; 2096 2097 power_req.request_type = PMR_PPM_ALL_LOWEST; 2098 power_req.req.ppm_all_lowest_req.mode = mode; 2099 mutex_enter(&ppm_lock); 2100 for (ppmcp = ppm_callbacks; ppmcp->ppmc_func; ppmcp++) 2101 (void) pm_ctlops((dev_info_t *)ppmcp->ppmc_dip, dip, 2102 DDI_CTLOPS_POWER, &power_req, &result); 2103 mutex_exit(&ppm_lock); 2104 if (mode == PM_ALL_LOWEST) { 2105 if (autoS3_enabled) { 2106 PMD(PMD_SX, ("pm_ppm_notify_all_lowest triggering " 2107 "autos3\n")) 2108 mutex_enter(&srn_clone_lock); 2109 if (srn_signal) { 2110 srn_inuse++; 2111 PMD(PMD_SX, ("(*srn_signal)(AUTOSX, 3)\n")) 2112 (*srn_signal)(SRN_TYPE_AUTOSX, 3); 2113 srn_inuse--; 2114 } else { 2115 PMD(PMD_SX, ("srn_signal NULL\n")) 2116 } 2117 mutex_exit(&srn_clone_lock); 2118 } else { 2119 PMD(PMD_SX, ("pm_ppm_notify_all_lowest autos3 " 2120 "disabled\n")); 2121 } 2122 } 2123 } 2124 2125 static void 2126 pm_set_pm_info(dev_info_t *dip, void *value) 2127 { 2128 DEVI(dip)->devi_pm_info = value; 2129 } 2130 2131 pm_rsvp_t *pm_blocked_list; 2132 2133 /* 2134 * Look up an entry in the blocked list by dip and component 2135 */ 2136 static pm_rsvp_t * 2137 pm_rsvp_lookup(dev_info_t *dip, int comp) 2138 { 2139 pm_rsvp_t *p; 2140 ASSERT(MUTEX_HELD(&pm_rsvp_lock)); 2141 for (p = pm_blocked_list; p; p = p->pr_next) 2142 if (p->pr_dip == dip && p->pr_comp == comp) { 2143 return (p); 2144 } 2145 return (NULL); 2146 } 2147 2148 /* 2149 * Called when a device which is direct power managed (or the parent or 2150 * dependent of such a device) changes power, or when a pm clone is closed 2151 * that was direct power managing a device. This call results in pm_blocked() 2152 * (below) returning. 2153 */ 2154 void 2155 pm_proceed(dev_info_t *dip, int cmd, int comp, int newlevel) 2156 { 2157 PMD_FUNC(pmf, "proceed") 2158 pm_rsvp_t *found = NULL; 2159 pm_rsvp_t *p; 2160 2161 mutex_enter(&pm_rsvp_lock); 2162 switch (cmd) { 2163 /* 2164 * we're giving up control, let any pending op continue 2165 */ 2166 case PMP_RELEASE: 2167 for (p = pm_blocked_list; p; p = p->pr_next) { 2168 if (dip == p->pr_dip) { 2169 p->pr_retval = PMP_RELEASE; 2170 PMD(PMD_DPM, ("%s: RELEASE %s@%s(%s#%d)\n", 2171 pmf, PM_DEVICE(dip))) 2172 cv_signal(&p->pr_cv); 2173 } 2174 } 2175 break; 2176 2177 /* 2178 * process has done PM_SET_CURRENT_POWER; let a matching request 2179 * succeed and a non-matching request for the same device fail 2180 */ 2181 case PMP_SETPOWER: 2182 found = pm_rsvp_lookup(dip, comp); 2183 if (!found) /* if driver not waiting */ 2184 break; 2185 /* 2186 * This cannot be pm_lower_power, since that can only happen 2187 * during detach or probe 2188 */ 2189 if (found->pr_newlevel <= newlevel) { 2190 found->pr_retval = PMP_SUCCEED; 2191 PMD(PMD_DPM, ("%s: SUCCEED %s@%s(%s#%d)\n", pmf, 2192 PM_DEVICE(dip))) 2193 } else { 2194 found->pr_retval = PMP_FAIL; 2195 PMD(PMD_DPM, ("%s: FAIL %s@%s(%s#%d)\n", pmf, 2196 PM_DEVICE(dip))) 2197 } 2198 cv_signal(&found->pr_cv); 2199 break; 2200 2201 default: 2202 panic("pm_proceed unknown cmd %d", cmd); 2203 } 2204 mutex_exit(&pm_rsvp_lock); 2205 } 2206 2207 /* 2208 * This routine dispatches new work to the dependency thread. Caller must 2209 * be prepared to block for memory if necessary. 2210 */ 2211 void 2212 pm_dispatch_to_dep_thread(int cmd, char *keeper, char *kept, int wait, 2213 int *res, int cached_pwr) 2214 { 2215 pm_dep_wk_t *new_work; 2216 2217 new_work = kmem_zalloc(sizeof (pm_dep_wk_t), KM_SLEEP); 2218 new_work->pdw_type = cmd; 2219 new_work->pdw_wait = wait; 2220 new_work->pdw_done = 0; 2221 new_work->pdw_ret = 0; 2222 new_work->pdw_pwr = cached_pwr; 2223 cv_init(&new_work->pdw_cv, NULL, CV_DEFAULT, NULL); 2224 if (keeper != NULL) { 2225 new_work->pdw_keeper = kmem_zalloc(strlen(keeper) + 1, 2226 KM_SLEEP); 2227 (void) strcpy(new_work->pdw_keeper, keeper); 2228 } 2229 if (kept != NULL) { 2230 new_work->pdw_kept = kmem_zalloc(strlen(kept) + 1, KM_SLEEP); 2231 (void) strcpy(new_work->pdw_kept, kept); 2232 } 2233 mutex_enter(&pm_dep_thread_lock); 2234 if (pm_dep_thread_workq == NULL) { 2235 pm_dep_thread_workq = new_work; 2236 pm_dep_thread_tail = new_work; 2237 new_work->pdw_next = NULL; 2238 } else { 2239 pm_dep_thread_tail->pdw_next = new_work; 2240 pm_dep_thread_tail = new_work; 2241 new_work->pdw_next = NULL; 2242 } 2243 cv_signal(&pm_dep_thread_cv); 2244 /* If caller asked for it, wait till it is done. */ 2245 if (wait) { 2246 while (!new_work->pdw_done) 2247 cv_wait(&new_work->pdw_cv, &pm_dep_thread_lock); 2248 /* 2249 * Pass return status, if any, back. 2250 */ 2251 if (res != NULL) 2252 *res = new_work->pdw_ret; 2253 /* 2254 * If we asked to wait, it is our job to free the request 2255 * structure. 2256 */ 2257 if (new_work->pdw_keeper) 2258 kmem_free(new_work->pdw_keeper, 2259 strlen(new_work->pdw_keeper) + 1); 2260 if (new_work->pdw_kept) 2261 kmem_free(new_work->pdw_kept, 2262 strlen(new_work->pdw_kept) + 1); 2263 kmem_free(new_work, sizeof (pm_dep_wk_t)); 2264 } 2265 mutex_exit(&pm_dep_thread_lock); 2266 } 2267 2268 /* 2269 * Release the pm resource for this device. 2270 */ 2271 void 2272 pm_rem_info(dev_info_t *dip) 2273 { 2274 PMD_FUNC(pmf, "rem_info") 2275 int i, count = 0; 2276 pm_info_t *info = PM_GET_PM_INFO(dip); 2277 dev_info_t *pdip = ddi_get_parent(dip); 2278 char *pathbuf; 2279 int work_type = PM_DEP_WK_DETACH; 2280 2281 ASSERT(info); 2282 2283 ASSERT(!PM_IAM_LOCKING_DIP(dip)); 2284 if (PM_ISDIRECT(dip)) { 2285 info->pmi_dev_pm_state &= ~PM_DIRECT; 2286 ASSERT(info->pmi_clone); 2287 info->pmi_clone = 0; 2288 pm_proceed(dip, PMP_RELEASE, -1, -1); 2289 } 2290 ASSERT(!PM_GET_PM_SCAN(dip)); 2291 2292 /* 2293 * Now adjust parent's kidsupcnt. BC nodes we check only comp 0, 2294 * Others we check all components. BC node that has already 2295 * called pm_destroy_components() has zero component count. 2296 * Parents that get notification are not adjusted because their 2297 * kidsupcnt is always 0 (or 1 during configuration). 2298 */ 2299 PMD(PMD_KEEPS, ("%s: %s@%s(%s#%d) has %d components\n", pmf, 2300 PM_DEVICE(dip), PM_NUMCMPTS(dip))) 2301 2302 /* node is detached, so we can examine power without locking */ 2303 if (PM_ISBC(dip)) { 2304 count = (PM_CURPOWER(dip, 0) != 0); 2305 } else { 2306 for (i = 0; i < PM_NUMCMPTS(dip); i++) 2307 count += (PM_CURPOWER(dip, i) != 0); 2308 } 2309 2310 if (PM_NUMCMPTS(dip) && pdip && !PM_WANTS_NOTIFICATION(pdip)) 2311 e_pm_hold_rele_power(pdip, -count); 2312 2313 /* Schedule a request to clean up dependency records */ 2314 pathbuf = kmem_zalloc(MAXPATHLEN, KM_SLEEP); 2315 (void) ddi_pathname(dip, pathbuf); 2316 pm_dispatch_to_dep_thread(work_type, pathbuf, pathbuf, 2317 PM_DEP_NOWAIT, NULL, (count > 0)); 2318 kmem_free(pathbuf, MAXPATHLEN); 2319 2320 /* 2321 * Adjust the pm_comps_notlowest count since this device is 2322 * not being power-managed anymore. 2323 */ 2324 for (i = 0; i < PM_NUMCMPTS(dip); i++) { 2325 if (PM_CURPOWER(dip, i) != 0) 2326 PM_DECR_NOTLOWEST(dip); 2327 } 2328 /* 2329 * Once we clear the info pointer, it looks like it is not power 2330 * managed to everybody else. 2331 */ 2332 pm_set_pm_info(dip, NULL); 2333 kmem_free(info, sizeof (pm_info_t)); 2334 } 2335 2336 int 2337 pm_get_norm_pwrs(dev_info_t *dip, int **valuep, size_t *length) 2338 { 2339 int components = PM_NUMCMPTS(dip); 2340 int *bufp; 2341 size_t size; 2342 int i; 2343 2344 if (components <= 0) { 2345 cmn_err(CE_NOTE, "!pm: %s@%s(%s#%d) has no components, " 2346 "can't get normal power values\n", PM_DEVICE(dip)); 2347 return (DDI_FAILURE); 2348 } else { 2349 size = components * sizeof (int); 2350 bufp = kmem_alloc(size, KM_SLEEP); 2351 for (i = 0; i < components; i++) { 2352 bufp[i] = pm_get_normal_power(dip, i); 2353 } 2354 } 2355 *length = size; 2356 *valuep = bufp; 2357 return (DDI_SUCCESS); 2358 } 2359 2360 static int 2361 pm_reset_timestamps(dev_info_t *dip, void *arg) 2362 { 2363 _NOTE(ARGUNUSED(arg)) 2364 2365 int components; 2366 int i; 2367 2368 if (!PM_GET_PM_INFO(dip)) 2369 return (DDI_WALK_CONTINUE); 2370 components = PM_NUMCMPTS(dip); 2371 ASSERT(components > 0); 2372 PM_LOCK_BUSY(dip); 2373 for (i = 0; i < components; i++) { 2374 struct pm_component *cp; 2375 /* 2376 * If the component was not marked as busy, 2377 * reset its timestamp to now. 2378 */ 2379 cp = PM_CP(dip, i); 2380 if (cp->pmc_timestamp) 2381 cp->pmc_timestamp = gethrestime_sec(); 2382 } 2383 PM_UNLOCK_BUSY(dip); 2384 return (DDI_WALK_CONTINUE); 2385 } 2386 2387 /* 2388 * Convert a power level to an index into the levels array (or 2389 * just PM_LEVEL_UNKNOWN in that special case). 2390 */ 2391 static int 2392 pm_level_to_index(dev_info_t *dip, pm_component_t *cp, int level) 2393 { 2394 PMD_FUNC(pmf, "level_to_index") 2395 int i; 2396 int limit = cp->pmc_comp.pmc_numlevels; 2397 int *ip = cp->pmc_comp.pmc_lvals; 2398 2399 if (level == PM_LEVEL_UNKNOWN) 2400 return (level); 2401 2402 for (i = 0; i < limit; i++) { 2403 if (level == *ip++) { 2404 PMD(PMD_LEVEL, ("%s: %s@%s(%s#%d)[%d] to %x\n", 2405 pmf, PM_DEVICE(dip), 2406 (int)(cp - DEVI(dip)->devi_pm_components), level)) 2407 return (i); 2408 } 2409 } 2410 panic("pm_level_to_index: level %d not found for device " 2411 "%s@%s(%s#%d)", level, PM_DEVICE(dip)); 2412 /*NOTREACHED*/ 2413 } 2414 2415 /* 2416 * Internal function to set current power level 2417 */ 2418 static void 2419 e_pm_set_cur_pwr(dev_info_t *dip, pm_component_t *cp, int level) 2420 { 2421 PMD_FUNC(pmf, "set_cur_pwr") 2422 int curpwr = (cp->pmc_flags & PM_PHC_WHILE_SET_POWER ? 2423 cp->pmc_phc_pwr : cp->pmc_cur_pwr); 2424 2425 /* 2426 * Nothing to adjust if current & new levels are the same. 2427 */ 2428 if (curpwr != PM_LEVEL_UNKNOWN && 2429 level == cp->pmc_comp.pmc_lvals[curpwr]) 2430 return; 2431 2432 /* 2433 * Keep the count for comps doing transition to/from lowest 2434 * level. 2435 */ 2436 if (curpwr == 0) { 2437 PM_INCR_NOTLOWEST(dip); 2438 } else if (level == cp->pmc_comp.pmc_lvals[0]) { 2439 PM_DECR_NOTLOWEST(dip); 2440 } 2441 cp->pmc_phc_pwr = PM_LEVEL_UNKNOWN; 2442 cp->pmc_cur_pwr = pm_level_to_index(dip, cp, level); 2443 } 2444 2445 /* 2446 * This is the default method of setting the power of a device if no ppm 2447 * driver has claimed it. 2448 */ 2449 int 2450 pm_power(dev_info_t *dip, int comp, int level) 2451 { 2452 PMD_FUNC(pmf, "power") 2453 struct dev_ops *ops; 2454 int (*fn)(dev_info_t *, int, int); 2455 struct pm_component *cp = PM_CP(dip, comp); 2456 int retval; 2457 pm_info_t *info = PM_GET_PM_INFO(dip); 2458 static int pm_phc_impl(dev_info_t *, int, int, int); 2459 2460 PMD(PMD_KIDSUP, ("%s: %s@%s(%s#%d), comp=%d, level=%d\n", pmf, 2461 PM_DEVICE(dip), comp, level)) 2462 if (!(ops = ddi_get_driver(dip))) { 2463 PMD(PMD_FAIL, ("%s: %s@%s(%s#%d) has no ops\n", pmf, 2464 PM_DEVICE(dip))) 2465 return (DDI_FAILURE); 2466 } 2467 if ((ops->devo_rev < 2) || !(fn = ops->devo_power)) { 2468 PMD(PMD_FAIL, ("%s: %s%s\n", pmf, 2469 (ops->devo_rev < 2 ? " wrong devo_rev" : ""), 2470 (!fn ? " devo_power NULL" : ""))) 2471 return (DDI_FAILURE); 2472 } 2473 cp->pmc_flags |= PM_POWER_OP; 2474 retval = (*fn)(dip, comp, level); 2475 cp->pmc_flags &= ~PM_POWER_OP; 2476 if (retval == DDI_SUCCESS) { 2477 e_pm_set_cur_pwr(dip, PM_CP(dip, comp), level); 2478 return (DDI_SUCCESS); 2479 } 2480 2481 /* 2482 * If pm_power_has_changed() detected a deadlock with pm_power() it 2483 * updated only the power level of the component. If our attempt to 2484 * set the device new to a power level above has failed we sync the 2485 * total power state via phc code now. 2486 */ 2487 if (cp->pmc_flags & PM_PHC_WHILE_SET_POWER) { 2488 int phc_lvl = 2489 cp->pmc_comp.pmc_lvals[cp->pmc_cur_pwr]; 2490 2491 ASSERT(info); 2492 (void) pm_phc_impl(dip, comp, phc_lvl, 0); 2493 PMD(PMD_PHC, ("%s: phc %s@%s(%s#%d) comp=%d level=%d\n", 2494 pmf, PM_DEVICE(dip), comp, phc_lvl)) 2495 } 2496 2497 PMD(PMD_FAIL, ("%s: can't set comp=%d (%s) of %s@%s(%s#%d) to " 2498 "level=%d (%s)\n", pmf, comp, cp->pmc_comp.pmc_name, PM_DEVICE(dip), 2499 level, power_val_to_string(cp, level))); 2500 return (DDI_FAILURE); 2501 } 2502 2503 int 2504 pm_unmanage(dev_info_t *dip) 2505 { 2506 PMD_FUNC(pmf, "unmanage") 2507 power_req_t power_req; 2508 int result, retval = 0; 2509 2510 ASSERT(!PM_IAM_LOCKING_DIP(dip)); 2511 PMD(PMD_REMDEV | PMD_KIDSUP, ("%s: %s@%s(%s#%d)\n", pmf, 2512 PM_DEVICE(dip))) 2513 power_req.request_type = PMR_PPM_UNMANAGE; 2514 power_req.req.ppm_config_req.who = dip; 2515 if (pm_ppm_claimed(dip)) 2516 retval = pm_ctlops(PPM(dip), dip, DDI_CTLOPS_POWER, 2517 &power_req, &result); 2518 #ifdef DEBUG 2519 else 2520 retval = pm_ctlops(PPM(dip), dip, DDI_CTLOPS_POWER, 2521 &power_req, &result); 2522 #endif 2523 ASSERT(retval == DDI_SUCCESS); 2524 pm_rem_info(dip); 2525 return (retval); 2526 } 2527 2528 int 2529 pm_raise_power(dev_info_t *dip, int comp, int level) 2530 { 2531 if (level < 0) 2532 return (DDI_FAILURE); 2533 if (!e_pm_valid_info(dip, NULL) || !e_pm_valid_comp(dip, comp, NULL) || 2534 !e_pm_valid_power(dip, comp, level)) 2535 return (DDI_FAILURE); 2536 2537 return (dev_is_needed(dip, comp, level, PM_LEVEL_UPONLY)); 2538 } 2539 2540 int 2541 pm_lower_power(dev_info_t *dip, int comp, int level) 2542 { 2543 PMD_FUNC(pmf, "pm_lower_power") 2544 2545 if (!e_pm_valid_info(dip, NULL) || !e_pm_valid_comp(dip, comp, NULL) || 2546 !e_pm_valid_power(dip, comp, level)) { 2547 PMD(PMD_FAIL, ("%s: validation checks failed for %s@%s(%s#%d) " 2548 "comp=%d level=%d\n", pmf, PM_DEVICE(dip), comp, level)) 2549 return (DDI_FAILURE); 2550 } 2551 2552 if (!DEVI_IS_DETACHING(dip)) { 2553 PMD(PMD_FAIL, ("%s: %s@%s(%s#%d) not detaching\n", 2554 pmf, PM_DEVICE(dip))) 2555 return (DDI_FAILURE); 2556 } 2557 2558 /* 2559 * If we don't care about saving power, or we're treating this node 2560 * specially, then this is a no-op 2561 */ 2562 if (!PM_SCANABLE(dip) || pm_noinvol(dip)) { 2563 PMD(PMD_FAIL, ("%s: %s@%s(%s#%d) %s%s%s%s\n", 2564 pmf, PM_DEVICE(dip), 2565 !autopm_enabled ? "!autopm_enabled " : "", 2566 !PM_CPUPM_ENABLED ? "!cpupm_enabled " : "", 2567 PM_CPUPM_DISABLED ? "cpupm_disabled " : "", 2568 pm_noinvol(dip) ? "pm_noinvol()" : "")) 2569 return (DDI_SUCCESS); 2570 } 2571 2572 if (dev_is_needed(dip, comp, level, PM_LEVEL_DOWNONLY) != DDI_SUCCESS) { 2573 PMD(PMD_FAIL, ("%s: %s@%s(%s#%d) dev_is_needed failed\n", pmf, 2574 PM_DEVICE(dip))) 2575 return (DDI_FAILURE); 2576 } 2577 return (DDI_SUCCESS); 2578 } 2579 2580 /* 2581 * Find the entries struct for a given dip in the blocked list, return it locked 2582 */ 2583 static psce_t * 2584 pm_psc_dip_to_direct(dev_info_t *dip, pscc_t **psccp) 2585 { 2586 pscc_t *p; 2587 psce_t *psce; 2588 2589 rw_enter(&pm_pscc_direct_rwlock, RW_READER); 2590 for (p = pm_pscc_direct; p; p = p->pscc_next) { 2591 if (p->pscc_dip == dip) { 2592 *psccp = p; 2593 psce = p->pscc_entries; 2594 mutex_enter(&psce->psce_lock); 2595 ASSERT(psce); 2596 rw_exit(&pm_pscc_direct_rwlock); 2597 return (psce); 2598 } 2599 } 2600 rw_exit(&pm_pscc_direct_rwlock); 2601 panic("sunpm: no entry for dip %p in direct list", (void *)dip); 2602 /*NOTREACHED*/ 2603 } 2604 2605 /* 2606 * Write an entry indicating a power level change (to be passed to a process 2607 * later) in the given psce. 2608 * If we were called in the path that brings up the console fb in the 2609 * case of entering the prom, we don't want to sleep. If the alloc fails, then 2610 * we create a record that has a size of -1, a physaddr of NULL, and that 2611 * has the overflow flag set. 2612 */ 2613 static int 2614 psc_entry(ushort_t event, psce_t *psce, dev_info_t *dip, int comp, int new, 2615 int old, int which, pm_canblock_t canblock) 2616 { 2617 char buf[MAXNAMELEN]; 2618 pm_state_change_t *p; 2619 size_t size; 2620 caddr_t physpath = NULL; 2621 int overrun = 0; 2622 2623 ASSERT(MUTEX_HELD(&psce->psce_lock)); 2624 (void) ddi_pathname(dip, buf); 2625 size = strlen(buf) + 1; 2626 p = psce->psce_in; 2627 if (canblock == PM_CANBLOCK_BYPASS) { 2628 physpath = kmem_alloc(size, KM_NOSLEEP); 2629 if (physpath == NULL) { 2630 /* 2631 * mark current entry as overrun 2632 */ 2633 p->flags |= PSC_EVENT_LOST; 2634 size = (size_t)-1; 2635 } 2636 } else 2637 physpath = kmem_alloc(size, KM_SLEEP); 2638 if (p->size) { /* overflow; mark the next entry */ 2639 if (p->size != (size_t)-1) 2640 kmem_free(p->physpath, p->size); 2641 ASSERT(psce->psce_out == p); 2642 if (p == psce->psce_last) { 2643 psce->psce_first->flags |= PSC_EVENT_LOST; 2644 psce->psce_out = psce->psce_first; 2645 } else { 2646 (p + 1)->flags |= PSC_EVENT_LOST; 2647 psce->psce_out = (p + 1); 2648 } 2649 overrun++; 2650 } else if (physpath == NULL) { /* alloc failed, mark this entry */ 2651 p->flags |= PSC_EVENT_LOST; 2652 p->size = 0; 2653 p->physpath = NULL; 2654 } 2655 if (which == PSC_INTEREST) { 2656 mutex_enter(&pm_compcnt_lock); 2657 if (pm_comps_notlowest == 0) 2658 p->flags |= PSC_ALL_LOWEST; 2659 else 2660 p->flags &= ~PSC_ALL_LOWEST; 2661 mutex_exit(&pm_compcnt_lock); 2662 } 2663 p->event = event; 2664 p->timestamp = gethrestime_sec(); 2665 p->component = comp; 2666 p->old_level = old; 2667 p->new_level = new; 2668 p->physpath = physpath; 2669 p->size = size; 2670 if (physpath != NULL) 2671 (void) strcpy(p->physpath, buf); 2672 if (p == psce->psce_last) 2673 psce->psce_in = psce->psce_first; 2674 else 2675 psce->psce_in = ++p; 2676 mutex_exit(&psce->psce_lock); 2677 return (overrun); 2678 } 2679 2680 /* 2681 * Find the next entry on the interest list. We keep a pointer to the item we 2682 * last returned in the user's cooke. Returns a locked entries struct. 2683 */ 2684 static psce_t * 2685 psc_interest(void **cookie, pscc_t **psccp) 2686 { 2687 pscc_t *pscc; 2688 pscc_t **cookiep = (pscc_t **)cookie; 2689 2690 if (*cookiep == NULL) 2691 pscc = pm_pscc_interest; 2692 else 2693 pscc = (*cookiep)->pscc_next; 2694 if (pscc) { 2695 *cookiep = pscc; 2696 *psccp = pscc; 2697 mutex_enter(&pscc->pscc_entries->psce_lock); 2698 return (pscc->pscc_entries); 2699 } else { 2700 return (NULL); 2701 } 2702 } 2703 2704 /* 2705 * Create an entry for a process to pick up indicating a power level change. 2706 */ 2707 static void 2708 pm_enqueue_notify(ushort_t cmd, dev_info_t *dip, int comp, 2709 int newlevel, int oldlevel, pm_canblock_t canblock) 2710 { 2711 PMD_FUNC(pmf, "enqueue_notify") 2712 pscc_t *pscc; 2713 psce_t *psce; 2714 void *cookie = NULL; 2715 int overrun; 2716 2717 ASSERT(MUTEX_HELD(&pm_rsvp_lock)); 2718 switch (cmd) { 2719 case PSC_PENDING_CHANGE: /* only for controlling process */ 2720 PMD(PMD_DPM, ("%s: PENDING %s@%s(%s#%d), comp %d, %d -> %d\n", 2721 pmf, PM_DEVICE(dip), comp, oldlevel, newlevel)) 2722 psce = pm_psc_dip_to_direct(dip, &pscc); 2723 ASSERT(psce); 2724 PMD(PMD_IOCTL, ("%s: PENDING: %s@%s(%s#%d) pm_poll_cnt[%d] " 2725 "%d\n", pmf, PM_DEVICE(dip), pscc->pscc_clone, 2726 pm_poll_cnt[pscc->pscc_clone])) 2727 overrun = psc_entry(cmd, psce, dip, comp, newlevel, oldlevel, 2728 PSC_DIRECT, canblock); 2729 PMD(PMD_DPM, ("%s: sig %d\n", pmf, pscc->pscc_clone)) 2730 mutex_enter(&pm_clone_lock); 2731 if (!overrun) 2732 pm_poll_cnt[pscc->pscc_clone]++; 2733 cv_signal(&pm_clones_cv[pscc->pscc_clone]); 2734 pollwakeup(&pm_pollhead, (POLLRDNORM | POLLIN)); 2735 mutex_exit(&pm_clone_lock); 2736 break; 2737 case PSC_HAS_CHANGED: 2738 PMD(PMD_DPM, ("%s: HAS %s@%s(%s#%d), comp %d, %d -> %d\n", 2739 pmf, PM_DEVICE(dip), comp, oldlevel, newlevel)) 2740 if (PM_ISDIRECT(dip) && canblock != PM_CANBLOCK_BYPASS) { 2741 psce = pm_psc_dip_to_direct(dip, &pscc); 2742 PMD(PMD_IOCTL, ("%s: HAS: %s@%s(%s#%d) pm_poll_cnt[%d] " 2743 "%d\n", pmf, PM_DEVICE(dip), pscc->pscc_clone, 2744 pm_poll_cnt[pscc->pscc_clone])) 2745 overrun = psc_entry(cmd, psce, dip, comp, newlevel, 2746 oldlevel, PSC_DIRECT, canblock); 2747 PMD(PMD_DPM, ("%s: sig %d\n", pmf, pscc->pscc_clone)) 2748 mutex_enter(&pm_clone_lock); 2749 if (!overrun) 2750 pm_poll_cnt[pscc->pscc_clone]++; 2751 cv_signal(&pm_clones_cv[pscc->pscc_clone]); 2752 pollwakeup(&pm_pollhead, (POLLRDNORM | POLLIN)); 2753 mutex_exit(&pm_clone_lock); 2754 } 2755 mutex_enter(&pm_clone_lock); 2756 rw_enter(&pm_pscc_interest_rwlock, RW_READER); 2757 while ((psce = psc_interest(&cookie, &pscc)) != NULL) { 2758 (void) psc_entry(cmd, psce, dip, comp, newlevel, 2759 oldlevel, PSC_INTEREST, canblock); 2760 cv_signal(&pm_clones_cv[pscc->pscc_clone]); 2761 } 2762 rw_exit(&pm_pscc_interest_rwlock); 2763 mutex_exit(&pm_clone_lock); 2764 break; 2765 #ifdef DEBUG 2766 default: 2767 ASSERT(0); 2768 #endif 2769 } 2770 } 2771 2772 static void 2773 pm_enqueue_notify_others(pm_ppm_devlist_t **listp, pm_canblock_t canblock) 2774 { 2775 if (listp) { 2776 pm_ppm_devlist_t *p, *next = NULL; 2777 2778 for (p = *listp; p; p = next) { 2779 next = p->ppd_next; 2780 pm_enqueue_notify(PSC_HAS_CHANGED, p->ppd_who, 2781 p->ppd_cmpt, p->ppd_new_level, p->ppd_old_level, 2782 canblock); 2783 kmem_free(p, sizeof (pm_ppm_devlist_t)); 2784 } 2785 *listp = NULL; 2786 } 2787 } 2788 2789 /* 2790 * Try to get the power locks of the parent node and target (child) 2791 * node. Return true if successful (with both locks held) or false 2792 * (with no locks held). 2793 */ 2794 static int 2795 pm_try_parent_child_locks(dev_info_t *pdip, 2796 dev_info_t *dip, int *pcircp, int *circp) 2797 { 2798 if (ndi_devi_tryenter(pdip, pcircp)) 2799 if (PM_TRY_LOCK_POWER(dip, circp)) { 2800 return (1); 2801 } else { 2802 ndi_devi_exit(pdip, *pcircp); 2803 } 2804 return (0); 2805 } 2806 2807 /* 2808 * Determine if the power lock owner is blocked by current thread. 2809 * returns : 2810 * 1 - If the thread owning the effective power lock (the first lock on 2811 * which a thread blocks when it does PM_LOCK_POWER) is blocked by 2812 * a mutex held by the current thread. 2813 * 2814 * 0 - otherwise 2815 * 2816 * Note : This function is called by pm_power_has_changed to determine whether 2817 * it is executing in parallel with pm_set_power. 2818 */ 2819 static int 2820 pm_blocked_by_us(dev_info_t *dip) 2821 { 2822 power_req_t power_req; 2823 kthread_t *owner; 2824 int result; 2825 kmutex_t *mp; 2826 dev_info_t *ppm = (dev_info_t *)DEVI(dip)->devi_pm_ppm; 2827 2828 power_req.request_type = PMR_PPM_POWER_LOCK_OWNER; 2829 power_req.req.ppm_power_lock_owner_req.who = dip; 2830 if (pm_ctlops(ppm, dip, DDI_CTLOPS_POWER, &power_req, &result) != 2831 DDI_SUCCESS) { 2832 /* 2833 * It is assumed that if the device is claimed by ppm, ppm 2834 * will always implement this request type and it'll always 2835 * return success. We panic here, if it fails. 2836 */ 2837 panic("pm: Can't determine power lock owner of %s@%s(%s#%d)\n", 2838 PM_DEVICE(dip)); 2839 /*NOTREACHED*/ 2840 } 2841 2842 if ((owner = power_req.req.ppm_power_lock_owner_req.owner) != NULL && 2843 owner->t_state == TS_SLEEP && 2844 owner->t_sobj_ops && 2845 SOBJ_TYPE(owner->t_sobj_ops) == SOBJ_MUTEX && 2846 (mp = (kmutex_t *)owner->t_wchan) && 2847 mutex_owner(mp) == curthread) 2848 return (1); 2849 2850 return (0); 2851 } 2852 2853 /* 2854 * Notify parent which wants to hear about a child's power changes. 2855 */ 2856 static void 2857 pm_notify_parent(dev_info_t *dip, 2858 dev_info_t *pdip, int comp, int old_level, int level) 2859 { 2860 pm_bp_has_changed_t bphc; 2861 pm_sp_misc_t pspm; 2862 char *pathbuf = kmem_alloc(MAXPATHLEN, KM_SLEEP); 2863 int result = DDI_SUCCESS; 2864 2865 bphc.bphc_dip = dip; 2866 bphc.bphc_path = ddi_pathname(dip, pathbuf); 2867 bphc.bphc_comp = comp; 2868 bphc.bphc_olevel = old_level; 2869 bphc.bphc_nlevel = level; 2870 pspm.pspm_canblock = PM_CANBLOCK_BLOCK; 2871 pspm.pspm_scan = 0; 2872 bphc.bphc_private = &pspm; 2873 (void) (*PM_BUS_POWER_FUNC(pdip))(pdip, NULL, 2874 BUS_POWER_HAS_CHANGED, (void *)&bphc, (void *)&result); 2875 kmem_free(pathbuf, MAXPATHLEN); 2876 } 2877 2878 /* 2879 * Check if we need to resume a BC device, and make the attach call as required. 2880 */ 2881 static int 2882 pm_check_and_resume(dev_info_t *dip, int comp, int old_level, int level) 2883 { 2884 int ret = DDI_SUCCESS; 2885 2886 if (PM_ISBC(dip) && comp == 0 && old_level == 0 && level != 0) { 2887 ASSERT(DEVI(dip)->devi_pm_flags & PMC_SUSPENDED); 2888 /* ppm is not interested in DDI_PM_RESUME */ 2889 if ((ret = devi_attach(dip, DDI_PM_RESUME)) != DDI_SUCCESS) 2890 /* XXX Should we mark it resumed, */ 2891 /* even though it failed? */ 2892 cmn_err(CE_WARN, "!pm: Can't resume %s@%s", 2893 PM_NAME(dip), PM_ADDR(dip)); 2894 DEVI(dip)->devi_pm_flags &= ~PMC_SUSPENDED; 2895 } 2896 2897 return (ret); 2898 } 2899 2900 /* 2901 * Tests outside the lock to see if we should bother to enqueue an entry 2902 * for any watching process. If yes, then caller will take the lock and 2903 * do the full protocol 2904 */ 2905 static int 2906 pm_watchers() 2907 { 2908 if (pm_processes_stopped) 2909 return (0); 2910 return (pm_pscc_direct || pm_pscc_interest); 2911 } 2912 2913 /* 2914 * A driver is reporting that the power of one of its device's components 2915 * has changed. Update the power state accordingly. 2916 */ 2917 int 2918 pm_power_has_changed(dev_info_t *dip, int comp, int level) 2919 { 2920 PMD_FUNC(pmf, "pm_power_has_changed") 2921 int ret; 2922 dev_info_t *pdip = ddi_get_parent(dip); 2923 struct pm_component *cp; 2924 int blocked, circ, pcirc, old_level; 2925 static int pm_phc_impl(dev_info_t *, int, int, int); 2926 2927 if (level < 0) { 2928 PMD(PMD_FAIL, ("%s: %s@%s(%s#%d): bad level=%d\n", pmf, 2929 PM_DEVICE(dip), level)) 2930 return (DDI_FAILURE); 2931 } 2932 2933 PMD(PMD_KIDSUP | PMD_DEP, ("%s: %s@%s(%s#%d), comp=%d, level=%d\n", pmf, 2934 PM_DEVICE(dip), comp, level)) 2935 2936 if (!e_pm_valid_info(dip, NULL) || !e_pm_valid_comp(dip, comp, &cp) || 2937 !e_pm_valid_power(dip, comp, level)) 2938 return (DDI_FAILURE); 2939 2940 /* 2941 * A driver thread calling pm_power_has_changed and another thread 2942 * calling pm_set_power can deadlock. The problem is not resolvable 2943 * by changing lock order, so we use pm_blocked_by_us() to detect 2944 * this specific deadlock. If we can't get the lock immediately 2945 * and we are deadlocked, just update the component's level, do 2946 * notifications, and return. We intend to update the total power 2947 * state later (if the other thread fails to set power to the 2948 * desired level). If we were called because of a power change on a 2949 * component that isn't involved in a set_power op, update all state 2950 * immediately. 2951 */ 2952 cp = PM_CP(dip, comp); 2953 while (!pm_try_parent_child_locks(pdip, dip, &pcirc, &circ)) { 2954 if (((blocked = pm_blocked_by_us(dip)) != 0) && 2955 (cp->pmc_flags & PM_POWER_OP)) { 2956 if (pm_watchers()) { 2957 mutex_enter(&pm_rsvp_lock); 2958 pm_enqueue_notify(PSC_HAS_CHANGED, dip, comp, 2959 level, cur_power(cp), PM_CANBLOCK_BLOCK); 2960 mutex_exit(&pm_rsvp_lock); 2961 } 2962 if (pdip && PM_WANTS_NOTIFICATION(pdip)) 2963 pm_notify_parent(dip, 2964 pdip, comp, cur_power(cp), level); 2965 (void) pm_check_and_resume(dip, 2966 comp, cur_power(cp), level); 2967 2968 /* 2969 * Stash the old power index, update curpwr, and flag 2970 * that the total power state needs to be synched. 2971 */ 2972 cp->pmc_flags |= PM_PHC_WHILE_SET_POWER; 2973 /* 2974 * Several pm_power_has_changed calls could arrive 2975 * while the set power path remains blocked. Keep the 2976 * oldest old power and the newest new power of any 2977 * sequence of phc calls which arrive during deadlock. 2978 */ 2979 if (cp->pmc_phc_pwr == PM_LEVEL_UNKNOWN) 2980 cp->pmc_phc_pwr = cp->pmc_cur_pwr; 2981 cp->pmc_cur_pwr = 2982 pm_level_to_index(dip, cp, level); 2983 PMD(PMD_PHC, ("%s: deadlock for %s@%s(%s#%d), comp=%d, " 2984 "level=%d\n", pmf, PM_DEVICE(dip), comp, level)) 2985 return (DDI_SUCCESS); 2986 } else 2987 if (blocked) { /* blocked, but different cmpt? */ 2988 if (!ndi_devi_tryenter(pdip, &pcirc)) { 2989 cmn_err(CE_NOTE, 2990 "!pm: parent kuc not updated due " 2991 "to possible deadlock.\n"); 2992 return (pm_phc_impl(dip, 2993 comp, level, 1)); 2994 } 2995 old_level = cur_power(cp); 2996 if (pdip && !PM_WANTS_NOTIFICATION(pdip) && 2997 (!PM_ISBC(dip) || comp == 0) && 2998 POWERING_ON(old_level, level)) 2999 pm_hold_power(pdip); 3000 ret = pm_phc_impl(dip, comp, level, 1); 3001 if (pdip && !PM_WANTS_NOTIFICATION(pdip)) { 3002 if ((!PM_ISBC(dip) || 3003 comp == 0) && level == 0 && 3004 old_level != PM_LEVEL_UNKNOWN) 3005 pm_rele_power(pdip); 3006 } 3007 ndi_devi_exit(pdip, pcirc); 3008 /* child lock not held: deadlock */ 3009 return (ret); 3010 } 3011 delay(1); 3012 PMD(PMD_PHC, ("%s: try lock again\n", pmf)) 3013 } 3014 3015 /* non-deadlock case */ 3016 old_level = cur_power(cp); 3017 if (pdip && !PM_WANTS_NOTIFICATION(pdip) && 3018 (!PM_ISBC(dip) || comp == 0) && POWERING_ON(old_level, level)) 3019 pm_hold_power(pdip); 3020 ret = pm_phc_impl(dip, comp, level, 1); 3021 if (pdip && !PM_WANTS_NOTIFICATION(pdip)) { 3022 if ((!PM_ISBC(dip) || comp == 0) && level == 0 && 3023 old_level != PM_LEVEL_UNKNOWN) 3024 pm_rele_power(pdip); 3025 } 3026 PM_UNLOCK_POWER(dip, circ); 3027 ndi_devi_exit(pdip, pcirc); 3028 return (ret); 3029 } 3030 3031 /* 3032 * Account for power changes to a component of the the console frame buffer. 3033 * If lowering power from full (or "unkown", which is treatd as full) 3034 * we will increment the "components off" count of the fb device. 3035 * Subsequent lowering of the same component doesn't affect the count. If 3036 * raising a component back to full power, we will decrement the count. 3037 * 3038 * Return: the increment value for pm_cfb_comps_off (-1, 0, or 1) 3039 */ 3040 static int 3041 calc_cfb_comps_incr(dev_info_t *dip, int cmpt, int old, int new) 3042 { 3043 struct pm_component *cp = PM_CP(dip, cmpt); 3044 int on = (old == PM_LEVEL_UNKNOWN || old == cp->pmc_norm_pwr); 3045 int want_normal = (new == cp->pmc_norm_pwr); 3046 int incr = 0; 3047 3048 if (on && !want_normal) 3049 incr = 1; 3050 else if (!on && want_normal) 3051 incr = -1; 3052 return (incr); 3053 } 3054 3055 /* 3056 * Adjust the count of console frame buffer components < full power. 3057 */ 3058 static void 3059 update_comps_off(int incr, dev_info_t *dip) 3060 { 3061 mutex_enter(&pm_cfb_lock); 3062 pm_cfb_comps_off += incr; 3063 ASSERT(pm_cfb_comps_off <= PM_NUMCMPTS(dip)); 3064 mutex_exit(&pm_cfb_lock); 3065 } 3066 3067 /* 3068 * Update the power state in the framework (via the ppm). The 'notify' 3069 * argument tells whether to notify watchers. Power lock is already held. 3070 */ 3071 static int 3072 pm_phc_impl(dev_info_t *dip, int comp, int level, int notify) 3073 { 3074 PMD_FUNC(pmf, "phc_impl") 3075 power_req_t power_req; 3076 int i, dodeps = 0; 3077 dev_info_t *pdip = ddi_get_parent(dip); 3078 int result; 3079 int old_level; 3080 struct pm_component *cp; 3081 int incr = 0; 3082 dev_info_t *ppm = (dev_info_t *)DEVI(dip)->devi_pm_ppm; 3083 int work_type = 0; 3084 char *pathbuf; 3085 3086 /* Must use "official" power level for this test. */ 3087 cp = PM_CP(dip, comp); 3088 old_level = (cp->pmc_flags & PM_PHC_WHILE_SET_POWER ? 3089 cp->pmc_phc_pwr : cp->pmc_cur_pwr); 3090 if (old_level != PM_LEVEL_UNKNOWN) 3091 old_level = cp->pmc_comp.pmc_lvals[old_level]; 3092 3093 if (level == old_level) { 3094 PMD(PMD_SET, ("%s: %s@%s(%s#%d), comp=%d is already at " 3095 "level=%d\n", pmf, PM_DEVICE(dip), comp, level)) 3096 return (DDI_SUCCESS); 3097 } 3098 3099 /* 3100 * Tell ppm about this. 3101 */ 3102 power_req.request_type = PMR_PPM_POWER_CHANGE_NOTIFY; 3103 power_req.req.ppm_notify_level_req.who = dip; 3104 power_req.req.ppm_notify_level_req.cmpt = comp; 3105 power_req.req.ppm_notify_level_req.new_level = level; 3106 power_req.req.ppm_notify_level_req.old_level = old_level; 3107 if (pm_ctlops(ppm, dip, DDI_CTLOPS_POWER, &power_req, 3108 &result) == DDI_FAILURE) { 3109 PMD(PMD_FAIL, ("%s: pm_ctlops %s@%s(%s#%d) to %d failed\n", 3110 pmf, PM_DEVICE(dip), level)) 3111 return (DDI_FAILURE); 3112 } 3113 3114 if (PM_IS_CFB(dip)) { 3115 incr = calc_cfb_comps_incr(dip, comp, old_level, level); 3116 3117 if (incr) { 3118 update_comps_off(incr, dip); 3119 PMD(PMD_CFB, ("%s: %s@%s(%s#%d) comp=%d %d->%d " 3120 "cfb_comps_off->%d\n", pmf, PM_DEVICE(dip), 3121 comp, old_level, level, pm_cfb_comps_off)) 3122 } 3123 } 3124 e_pm_set_cur_pwr(dip, PM_CP(dip, comp), level); 3125 result = DDI_SUCCESS; 3126 3127 if (notify) { 3128 if (pdip && PM_WANTS_NOTIFICATION(pdip)) 3129 pm_notify_parent(dip, pdip, comp, old_level, level); 3130 (void) pm_check_and_resume(dip, comp, old_level, level); 3131 } 3132 3133 /* 3134 * Decrement the dependency kidsup count if we turn a device 3135 * off. 3136 */ 3137 if (POWERING_OFF(old_level, level)) { 3138 dodeps = 1; 3139 for (i = 0; i < PM_NUMCMPTS(dip); i++) { 3140 cp = PM_CP(dip, i); 3141 if (cur_power(cp)) { 3142 dodeps = 0; 3143 break; 3144 } 3145 } 3146 if (dodeps) 3147 work_type = PM_DEP_WK_POWER_OFF; 3148 } 3149 3150 /* 3151 * Increment if we turn it on. Check to see 3152 * if other comps are already on, if so, 3153 * dont increment. 3154 */ 3155 if (POWERING_ON(old_level, level)) { 3156 dodeps = 1; 3157 for (i = 0; i < PM_NUMCMPTS(dip); i++) { 3158 cp = PM_CP(dip, i); 3159 if (comp == i) 3160 continue; 3161 /* -1 also treated as 0 in this case */ 3162 if (cur_power(cp) > 0) { 3163 dodeps = 0; 3164 break; 3165 } 3166 } 3167 if (dodeps) 3168 work_type = PM_DEP_WK_POWER_ON; 3169 } 3170 3171 if (dodeps) { 3172 pathbuf = kmem_zalloc(MAXPATHLEN, KM_SLEEP); 3173 (void) ddi_pathname(dip, pathbuf); 3174 pm_dispatch_to_dep_thread(work_type, pathbuf, NULL, 3175 PM_DEP_NOWAIT, NULL, 0); 3176 kmem_free(pathbuf, MAXPATHLEN); 3177 } 3178 3179 if (notify && (level != old_level) && pm_watchers()) { 3180 mutex_enter(&pm_rsvp_lock); 3181 pm_enqueue_notify(PSC_HAS_CHANGED, dip, comp, level, old_level, 3182 PM_CANBLOCK_BLOCK); 3183 mutex_exit(&pm_rsvp_lock); 3184 } 3185 3186 PMD(PMD_RESCAN, ("%s: %s@%s(%s#%d): pm_rescan\n", pmf, PM_DEVICE(dip))) 3187 pm_rescan(dip); 3188 return (DDI_SUCCESS); 3189 } 3190 3191 /* 3192 * This function is called at startup time to notify pm of the existence 3193 * of any platform power managers for this platform. As a result of 3194 * this registration, each function provided will be called each time 3195 * a device node is attached, until one returns true, and it must claim the 3196 * device node (by returning non-zero) if it wants to be involved in the 3197 * node's power management. If it does claim the node, then it will 3198 * subsequently be notified of attach and detach events. 3199 * 3200 */ 3201 3202 int 3203 pm_register_ppm(int (*func)(dev_info_t *), dev_info_t *dip) 3204 { 3205 PMD_FUNC(pmf, "register_ppm") 3206 struct ppm_callbacks *ppmcp; 3207 pm_component_t *cp; 3208 int i, pwr, result, circ; 3209 power_req_t power_req; 3210 struct ppm_notify_level_req *p = &power_req.req.ppm_notify_level_req; 3211 void pm_ppm_claim(dev_info_t *); 3212 3213 mutex_enter(&ppm_lock); 3214 ppmcp = ppm_callbacks; 3215 for (i = 0; i < MAX_PPM_HANDLERS; i++, ppmcp++) { 3216 if (ppmcp->ppmc_func == NULL) { 3217 ppmcp->ppmc_func = func; 3218 ppmcp->ppmc_dip = dip; 3219 break; 3220 } 3221 } 3222 mutex_exit(&ppm_lock); 3223 3224 if (i >= MAX_PPM_HANDLERS) 3225 return (DDI_FAILURE); 3226 while ((dip = ddi_get_parent(dip)) != NULL) { 3227 if (dip != ddi_root_node() && PM_GET_PM_INFO(dip) == NULL) 3228 continue; 3229 pm_ppm_claim(dip); 3230 /* don't bother with the not power-manageable nodes */ 3231 if (pm_ppm_claimed(dip) && PM_GET_PM_INFO(dip)) { 3232 /* 3233 * Tell ppm about this. 3234 */ 3235 power_req.request_type = PMR_PPM_POWER_CHANGE_NOTIFY; 3236 p->old_level = PM_LEVEL_UNKNOWN; 3237 p->who = dip; 3238 PM_LOCK_POWER(dip, &circ); 3239 for (i = 0; i < PM_NUMCMPTS(dip); i++) { 3240 cp = PM_CP(dip, i); 3241 pwr = cp->pmc_cur_pwr; 3242 if (pwr != PM_LEVEL_UNKNOWN) { 3243 p->cmpt = i; 3244 p->new_level = cur_power(cp); 3245 p->old_level = PM_LEVEL_UNKNOWN; 3246 if (pm_ctlops(PPM(dip), dip, 3247 DDI_CTLOPS_POWER, &power_req, 3248 &result) == DDI_FAILURE) { 3249 PMD(PMD_FAIL, ("%s: pc " 3250 "%s@%s(%s#%d) to %d " 3251 "fails\n", pmf, 3252 PM_DEVICE(dip), pwr)) 3253 } 3254 } 3255 } 3256 PM_UNLOCK_POWER(dip, circ); 3257 } 3258 } 3259 return (DDI_SUCCESS); 3260 } 3261 3262 /* 3263 * Call the ppm's that have registered and adjust the devinfo struct as 3264 * appropriate. First one to claim it gets it. The sets of devices claimed 3265 * by each ppm are assumed to be disjoint. 3266 */ 3267 void 3268 pm_ppm_claim(dev_info_t *dip) 3269 { 3270 struct ppm_callbacks *ppmcp; 3271 3272 if (PPM(dip)) { 3273 return; 3274 } 3275 mutex_enter(&ppm_lock); 3276 for (ppmcp = ppm_callbacks; ppmcp->ppmc_func; ppmcp++) { 3277 if ((*ppmcp->ppmc_func)(dip)) { 3278 DEVI(dip)->devi_pm_ppm = 3279 (struct dev_info *)ppmcp->ppmc_dip; 3280 mutex_exit(&ppm_lock); 3281 return; 3282 } 3283 } 3284 mutex_exit(&ppm_lock); 3285 } 3286 3287 /* 3288 * Node is being detached so stop autopm until we see if it succeeds, in which 3289 * case pm_stop will be called. For backwards compatible devices we bring the 3290 * device up to full power on the assumption the detach will succeed. 3291 */ 3292 void 3293 pm_detaching(dev_info_t *dip) 3294 { 3295 PMD_FUNC(pmf, "detaching") 3296 pm_info_t *info = PM_GET_PM_INFO(dip); 3297 int iscons; 3298 3299 PMD(PMD_REMDEV, ("%s: %s@%s(%s#%d), %d comps\n", pmf, PM_DEVICE(dip), 3300 PM_NUMCMPTS(dip))) 3301 if (info == NULL) 3302 return; 3303 ASSERT(DEVI_IS_DETACHING(dip)); 3304 PM_LOCK_DIP(dip); 3305 info->pmi_dev_pm_state |= PM_DETACHING; 3306 PM_UNLOCK_DIP(dip); 3307 if (!PM_ISBC(dip)) 3308 pm_scan_stop(dip); 3309 3310 /* 3311 * console and old-style devices get brought up when detaching. 3312 */ 3313 iscons = PM_IS_CFB(dip); 3314 if (iscons || PM_ISBC(dip)) { 3315 (void) pm_all_to_normal(dip, PM_CANBLOCK_BYPASS); 3316 if (iscons) { 3317 mutex_enter(&pm_cfb_lock); 3318 while (cfb_inuse) { 3319 mutex_exit(&pm_cfb_lock); 3320 PMD(PMD_CFB, ("%s: delay; cfb_inuse\n", pmf)) 3321 delay(1); 3322 mutex_enter(&pm_cfb_lock); 3323 } 3324 ASSERT(cfb_dip_detaching == NULL); 3325 ASSERT(cfb_dip); 3326 cfb_dip_detaching = cfb_dip; /* case detach fails */ 3327 cfb_dip = NULL; 3328 mutex_exit(&pm_cfb_lock); 3329 } 3330 } 3331 } 3332 3333 /* 3334 * Node failed to detach. If it used to be autopm'd, make it so again. 3335 */ 3336 void 3337 pm_detach_failed(dev_info_t *dip) 3338 { 3339 PMD_FUNC(pmf, "detach_failed") 3340 pm_info_t *info = PM_GET_PM_INFO(dip); 3341 int pm_all_at_normal(dev_info_t *); 3342 3343 if (info == NULL) 3344 return; 3345 ASSERT(DEVI_IS_DETACHING(dip)); 3346 if (info->pmi_dev_pm_state & PM_DETACHING) { 3347 info->pmi_dev_pm_state &= ~PM_DETACHING; 3348 if (info->pmi_dev_pm_state & PM_ALLNORM_DEFERRED) { 3349 /* Make sure the operation is still needed */ 3350 if (!pm_all_at_normal(dip)) { 3351 if (pm_all_to_normal(dip, 3352 PM_CANBLOCK_FAIL) != DDI_SUCCESS) { 3353 PMD(PMD_ERROR, ("%s: could not bring " 3354 "%s@%s(%s#%d) to normal\n", pmf, 3355 PM_DEVICE(dip))) 3356 } 3357 } 3358 info->pmi_dev_pm_state &= ~PM_ALLNORM_DEFERRED; 3359 } 3360 } 3361 if (!PM_ISBC(dip)) { 3362 mutex_enter(&pm_scan_lock); 3363 if (PM_SCANABLE(dip)) 3364 pm_scan_init(dip); 3365 mutex_exit(&pm_scan_lock); 3366 pm_rescan(dip); 3367 } 3368 } 3369 3370 /* generic Backwards Compatible component */ 3371 static char *bc_names[] = {"off", "on"}; 3372 3373 static pm_comp_t bc_comp = {"unknown", 2, NULL, NULL, &bc_names[0]}; 3374 3375 static void 3376 e_pm_default_levels(dev_info_t *dip, pm_component_t *cp, int norm) 3377 { 3378 pm_comp_t *pmc; 3379 pmc = &cp->pmc_comp; 3380 pmc->pmc_numlevels = 2; 3381 pmc->pmc_lvals[0] = 0; 3382 pmc->pmc_lvals[1] = norm; 3383 e_pm_set_cur_pwr(dip, cp, norm); 3384 } 3385 3386 static void 3387 e_pm_default_components(dev_info_t *dip, int cmpts) 3388 { 3389 int i; 3390 pm_component_t *p = DEVI(dip)->devi_pm_components; 3391 3392 p = DEVI(dip)->devi_pm_components; 3393 for (i = 0; i < cmpts; i++, p++) { 3394 p->pmc_comp = bc_comp; /* struct assignment */ 3395 p->pmc_comp.pmc_lvals = kmem_zalloc(2 * sizeof (int), 3396 KM_SLEEP); 3397 p->pmc_comp.pmc_thresh = kmem_alloc(2 * sizeof (int), 3398 KM_SLEEP); 3399 p->pmc_comp.pmc_numlevels = 2; 3400 p->pmc_comp.pmc_thresh[0] = INT_MAX; 3401 p->pmc_comp.pmc_thresh[1] = INT_MAX; 3402 } 3403 } 3404 3405 /* 3406 * Called from functions that require components to exist already to allow 3407 * for their creation by parsing the pm-components property. 3408 * Device will not be power managed as a result of this call 3409 * No locking needed because we're single threaded by the ndi_devi_enter 3410 * done while attaching, and the device isn't visible until after it has 3411 * attached 3412 */ 3413 int 3414 pm_premanage(dev_info_t *dip, int style) 3415 { 3416 PMD_FUNC(pmf, "premanage") 3417 pm_comp_t *pcp, *compp; 3418 int cmpts, i, norm, error; 3419 pm_component_t *p = DEVI(dip)->devi_pm_components; 3420 pm_comp_t *pm_autoconfig(dev_info_t *, int *); 3421 3422 ASSERT(!PM_IAM_LOCKING_DIP(dip)); 3423 /* 3424 * If this dip has already been processed, don't mess with it 3425 */ 3426 if (DEVI(dip)->devi_pm_flags & PMC_COMPONENTS_DONE) 3427 return (DDI_SUCCESS); 3428 if (DEVI(dip)->devi_pm_flags & PMC_COMPONENTS_FAILED) { 3429 return (DDI_FAILURE); 3430 } 3431 /* 3432 * Look up pm-components property and create components accordingly 3433 * If that fails, fall back to backwards compatibility 3434 */ 3435 if ((compp = pm_autoconfig(dip, &error)) == NULL) { 3436 /* 3437 * If error is set, the property existed but was not well formed 3438 */ 3439 if (error || (style == PM_STYLE_NEW)) { 3440 DEVI(dip)->devi_pm_flags |= PMC_COMPONENTS_FAILED; 3441 return (DDI_FAILURE); 3442 } 3443 /* 3444 * If they don't have the pm-components property, then we 3445 * want the old "no pm until PM_SET_DEVICE_THRESHOLDS ioctl" 3446 * behavior driver must have called pm_create_components, and 3447 * we need to flesh out dummy components 3448 */ 3449 if ((cmpts = PM_NUMCMPTS(dip)) == 0) { 3450 /* 3451 * Not really failure, but we don't want the 3452 * caller to treat it as success 3453 */ 3454 return (DDI_FAILURE); 3455 } 3456 DEVI(dip)->devi_pm_flags |= PMC_BC; 3457 e_pm_default_components(dip, cmpts); 3458 for (i = 0; i < cmpts; i++) { 3459 /* 3460 * if normal power not set yet, we don't really know 3461 * what *ANY* of the power values are. If normal 3462 * power is set, then we assume for this backwards 3463 * compatible case that the values are 0, normal power. 3464 */ 3465 norm = pm_get_normal_power(dip, i); 3466 if (norm == (uint_t)-1) { 3467 PMD(PMD_ERROR, ("%s: %s@%s(%s#%d)[%d]\n", pmf, 3468 PM_DEVICE(dip), i)) 3469 return (DDI_FAILURE); 3470 } 3471 /* 3472 * Components of BC devices start at their normal power, 3473 * so count them to be not at their lowest power. 3474 */ 3475 PM_INCR_NOTLOWEST(dip); 3476 e_pm_default_levels(dip, PM_CP(dip, i), norm); 3477 } 3478 } else { 3479 /* 3480 * e_pm_create_components was called from pm_autoconfig(), it 3481 * creates components with no descriptions (or known levels) 3482 */ 3483 cmpts = PM_NUMCMPTS(dip); 3484 ASSERT(cmpts != 0); 3485 pcp = compp; 3486 p = DEVI(dip)->devi_pm_components; 3487 for (i = 0; i < cmpts; i++, p++) { 3488 p->pmc_comp = *pcp++; /* struct assignment */ 3489 ASSERT(PM_CP(dip, i)->pmc_cur_pwr == 0); 3490 e_pm_set_cur_pwr(dip, PM_CP(dip, i), PM_LEVEL_UNKNOWN); 3491 } 3492 if (DEVI(dip)->devi_pm_flags & PMC_CPU_THRESH) 3493 pm_set_device_threshold(dip, pm_cpu_idle_threshold, 3494 PMC_CPU_THRESH); 3495 else 3496 pm_set_device_threshold(dip, pm_system_idle_threshold, 3497 PMC_DEF_THRESH); 3498 kmem_free(compp, cmpts * sizeof (pm_comp_t)); 3499 } 3500 return (DDI_SUCCESS); 3501 } 3502 3503 /* 3504 * Called from during or after the device's attach to let us know it is ready 3505 * to play autopm. Look up the pm model and manage the device accordingly. 3506 * Returns system call errno value. 3507 * If DDI_ATTACH and DDI_DETACH were in same namespace, this would be 3508 * a little cleaner 3509 * 3510 * Called with dip lock held, return with dip lock unheld. 3511 */ 3512 3513 int 3514 e_pm_manage(dev_info_t *dip, int style) 3515 { 3516 PMD_FUNC(pmf, "e_manage") 3517 pm_info_t *info; 3518 dev_info_t *pdip = ddi_get_parent(dip); 3519 int pm_thresh_specd(dev_info_t *); 3520 int count; 3521 char *pathbuf; 3522 3523 if (pm_premanage(dip, style) != DDI_SUCCESS) { 3524 return (DDI_FAILURE); 3525 } 3526 PMD(PMD_KIDSUP, ("%s: %s@%s(%s#%d)\n", pmf, PM_DEVICE(dip))) 3527 ASSERT(PM_GET_PM_INFO(dip) == NULL); 3528 info = kmem_zalloc(sizeof (pm_info_t), KM_SLEEP); 3529 3530 /* 3531 * Now set up parent's kidsupcnt. BC nodes are assumed to start 3532 * out at their normal power, so they are "up", others start out 3533 * unknown, which is effectively "up". Parent which want notification 3534 * get kidsupcnt of 0 always. 3535 */ 3536 count = (PM_ISBC(dip)) ? 1 : PM_NUMCMPTS(dip); 3537 if (count && pdip && !PM_WANTS_NOTIFICATION(pdip)) 3538 e_pm_hold_rele_power(pdip, count); 3539 3540 pm_set_pm_info(dip, info); 3541 /* 3542 * Apply any recorded thresholds 3543 */ 3544 (void) pm_thresh_specd(dip); 3545 3546 /* 3547 * Do dependency processing. 3548 */ 3549 pathbuf = kmem_zalloc(MAXPATHLEN, KM_SLEEP); 3550 (void) ddi_pathname(dip, pathbuf); 3551 pm_dispatch_to_dep_thread(PM_DEP_WK_ATTACH, pathbuf, pathbuf, 3552 PM_DEP_NOWAIT, NULL, 0); 3553 kmem_free(pathbuf, MAXPATHLEN); 3554 3555 if (!PM_ISBC(dip)) { 3556 mutex_enter(&pm_scan_lock); 3557 if (PM_SCANABLE(dip)) { 3558 pm_scan_init(dip); 3559 mutex_exit(&pm_scan_lock); 3560 pm_rescan(dip); 3561 } else { 3562 mutex_exit(&pm_scan_lock); 3563 } 3564 } 3565 return (0); 3566 } 3567 3568 /* 3569 * This is the obsolete exported interface for a driver to find out its 3570 * "normal" (max) power. 3571 * We only get components destroyed while no power management is 3572 * going on (and the device is detached), so we don't need a mutex here 3573 */ 3574 int 3575 pm_get_normal_power(dev_info_t *dip, int comp) 3576 { 3577 3578 if (comp >= 0 && comp < PM_NUMCMPTS(dip)) { 3579 return (PM_CP(dip, comp)->pmc_norm_pwr); 3580 } 3581 return (DDI_FAILURE); 3582 } 3583 3584 /* 3585 * Fetches the current power level. Return DDI_SUCCESS or DDI_FAILURE. 3586 */ 3587 int 3588 pm_get_current_power(dev_info_t *dip, int comp, int *levelp) 3589 { 3590 if (comp >= 0 && comp < PM_NUMCMPTS(dip)) { 3591 *levelp = PM_CURPOWER(dip, comp); 3592 return (DDI_SUCCESS); 3593 } 3594 return (DDI_FAILURE); 3595 } 3596 3597 /* 3598 * Returns current threshold of indicated component 3599 */ 3600 static int 3601 cur_threshold(dev_info_t *dip, int comp) 3602 { 3603 pm_component_t *cp = PM_CP(dip, comp); 3604 int pwr; 3605 3606 if (PM_ISBC(dip)) { 3607 /* 3608 * backwards compatible nodes only have one threshold 3609 */ 3610 return (cp->pmc_comp.pmc_thresh[1]); 3611 } 3612 pwr = cp->pmc_cur_pwr; 3613 if (pwr == PM_LEVEL_UNKNOWN) { 3614 int thresh; 3615 if (DEVI(dip)->devi_pm_flags & PMC_NEXDEF_THRESH) 3616 thresh = pm_default_nexus_threshold; 3617 else if (DEVI(dip)->devi_pm_flags & PMC_CPU_THRESH) 3618 thresh = pm_cpu_idle_threshold; 3619 else 3620 thresh = pm_system_idle_threshold; 3621 return (thresh); 3622 } 3623 ASSERT(cp->pmc_comp.pmc_thresh); 3624 return (cp->pmc_comp.pmc_thresh[pwr]); 3625 } 3626 3627 /* 3628 * Compute next lower component power level given power index. 3629 */ 3630 static int 3631 pm_next_lower_power(pm_component_t *cp, int pwrndx) 3632 { 3633 int nxt_pwr; 3634 3635 if (pwrndx == PM_LEVEL_UNKNOWN) { 3636 nxt_pwr = cp->pmc_comp.pmc_lvals[0]; 3637 } else { 3638 pwrndx--; 3639 ASSERT(pwrndx >= 0); 3640 nxt_pwr = cp->pmc_comp.pmc_lvals[pwrndx]; 3641 } 3642 return (nxt_pwr); 3643 } 3644 3645 /* 3646 * Update the maxpower (normal) power of a component. Note that the 3647 * component's power level is only changed if it's current power level 3648 * is higher than the new max power. 3649 */ 3650 int 3651 pm_update_maxpower(dev_info_t *dip, int comp, int level) 3652 { 3653 PMD_FUNC(pmf, "update_maxpower") 3654 int old; 3655 int result; 3656 3657 if (!e_pm_valid_info(dip, NULL) || !e_pm_valid_comp(dip, comp, NULL) || 3658 !e_pm_valid_power(dip, comp, level)) { 3659 PMD(PMD_FAIL, ("%s: validation checks failed for %s@%s(%s#%d) " 3660 "comp=%d level=%d\n", pmf, PM_DEVICE(dip), comp, level)) 3661 return (DDI_FAILURE); 3662 } 3663 old = e_pm_get_max_power(dip, comp); 3664 e_pm_set_max_power(dip, comp, level); 3665 3666 if (pm_set_power(dip, comp, level, PM_LEVEL_DOWNONLY, 3667 PM_CANBLOCK_BLOCK, 0, &result) != DDI_SUCCESS) { 3668 e_pm_set_max_power(dip, comp, old); 3669 PMD(PMD_FAIL, ("%s: %s@%s(%s#%d) pm_set_power failed\n", pmf, 3670 PM_DEVICE(dip))) 3671 return (DDI_FAILURE); 3672 } 3673 return (DDI_SUCCESS); 3674 } 3675 3676 /* 3677 * Bring all components of device to normal power 3678 */ 3679 int 3680 pm_all_to_normal(dev_info_t *dip, pm_canblock_t canblock) 3681 { 3682 PMD_FUNC(pmf, "all_to_normal") 3683 int *normal; 3684 int i, ncomps, result; 3685 size_t size; 3686 int changefailed = 0; 3687 3688 PMD(PMD_ALLNORM, ("%s: %s@%s(%s#%d)\n", pmf, PM_DEVICE(dip))) 3689 ASSERT(PM_GET_PM_INFO(dip)); 3690 if (pm_get_norm_pwrs(dip, &normal, &size) != DDI_SUCCESS) { 3691 PMD(PMD_ALLNORM, ("%s: can't get norm pwrs for " 3692 "%s@%s(%s#%d)\n", pmf, PM_DEVICE(dip))) 3693 return (DDI_FAILURE); 3694 } 3695 ncomps = PM_NUMCMPTS(dip); 3696 for (i = 0; i < ncomps; i++) { 3697 if (pm_set_power(dip, i, normal[i], 3698 PM_LEVEL_UPONLY, canblock, 0, &result) != DDI_SUCCESS) { 3699 changefailed++; 3700 PMD(PMD_ALLNORM | PMD_FAIL, ("%s: failed to set " 3701 "%s@%s(%s#%d)[%d] to %d, errno %d\n", pmf, 3702 PM_DEVICE(dip), i, normal[i], result)) 3703 } 3704 } 3705 kmem_free(normal, size); 3706 if (changefailed) { 3707 PMD(PMD_FAIL, ("%s: failed to set %d comps %s@%s(%s#%d) " 3708 "to full power\n", pmf, changefailed, PM_DEVICE(dip))) 3709 return (DDI_FAILURE); 3710 } 3711 return (DDI_SUCCESS); 3712 } 3713 3714 /* 3715 * Returns true if all components of device are at normal power 3716 */ 3717 int 3718 pm_all_at_normal(dev_info_t *dip) 3719 { 3720 PMD_FUNC(pmf, "all_at_normal") 3721 int *normal; 3722 int i; 3723 size_t size; 3724 3725 PMD(PMD_ALLNORM, ("%s: %s@%s(%s#%d)\n", pmf, PM_DEVICE(dip))) 3726 if (pm_get_norm_pwrs(dip, &normal, &size) != DDI_SUCCESS) { 3727 PMD(PMD_ALLNORM, ("%s: can't get normal power\n", pmf)) 3728 return (DDI_FAILURE); 3729 } 3730 for (i = 0; i < PM_NUMCMPTS(dip); i++) { 3731 int current = PM_CURPOWER(dip, i); 3732 if (normal[i] > current) { 3733 PMD(PMD_ALLNORM, ("%s: %s@%s(%s#%d) comp=%d, " 3734 "norm=%d, cur=%d\n", pmf, PM_DEVICE(dip), i, 3735 normal[i], current)) 3736 break; 3737 } 3738 } 3739 kmem_free(normal, size); 3740 if (i != PM_NUMCMPTS(dip)) { 3741 return (0); 3742 } 3743 return (1); 3744 } 3745 3746 static void 3747 bring_wekeeps_up(char *keeper) 3748 { 3749 PMD_FUNC(pmf, "bring_wekeeps_up") 3750 int i; 3751 pm_pdr_t *dp; 3752 pm_info_t *wku_info; 3753 char *kept_path; 3754 dev_info_t *kept; 3755 static void bring_pmdep_up(dev_info_t *, int); 3756 3757 if (panicstr) { 3758 return; 3759 } 3760 /* 3761 * We process the request even if the keeper detaches because 3762 * detach processing expects this to increment kidsupcnt of kept. 3763 */ 3764 PMD(PMD_BRING, ("%s: keeper= %s\n", pmf, keeper)) 3765 for (dp = pm_dep_head; dp; dp = dp->pdr_next) { 3766 if (strcmp(dp->pdr_keeper, keeper) != 0) 3767 continue; 3768 for (i = 0; i < dp->pdr_kept_count; i++) { 3769 kept_path = dp->pdr_kept_paths[i]; 3770 if (kept_path == NULL) 3771 continue; 3772 ASSERT(kept_path[0] != '\0'); 3773 if ((kept = pm_name_to_dip(kept_path, 1)) == NULL) 3774 continue; 3775 wku_info = PM_GET_PM_INFO(kept); 3776 if (wku_info == NULL) { 3777 if (kept) 3778 ddi_release_devi(kept); 3779 continue; 3780 } 3781 /* 3782 * Don't mess with it if it is being detached, it isn't 3783 * safe to call its power entry point 3784 */ 3785 if (wku_info->pmi_dev_pm_state & PM_DETACHING) { 3786 if (kept) 3787 ddi_release_devi(kept); 3788 continue; 3789 } 3790 bring_pmdep_up(kept, 1); 3791 ddi_release_devi(kept); 3792 } 3793 } 3794 } 3795 3796 /* 3797 * Bring up the 'kept' device passed as argument 3798 */ 3799 static void 3800 bring_pmdep_up(dev_info_t *kept_dip, int hold) 3801 { 3802 PMD_FUNC(pmf, "bring_pmdep_up") 3803 int is_all_at_normal = 0; 3804 3805 /* 3806 * If the kept device has been unmanaged, do nothing. 3807 */ 3808 if (!PM_GET_PM_INFO(kept_dip)) 3809 return; 3810 3811 /* Just ignore DIRECT PM device till they are released. */ 3812 if (!pm_processes_stopped && PM_ISDIRECT(kept_dip) && 3813 !(is_all_at_normal = pm_all_at_normal(kept_dip))) { 3814 PMD(PMD_BRING, ("%s: can't bring up PM_DIRECT %s@%s(%s#%d) " 3815 "controlling process did something else\n", pmf, 3816 PM_DEVICE(kept_dip))) 3817 DEVI(kept_dip)->devi_pm_flags |= PMC_SKIP_BRINGUP; 3818 return; 3819 } 3820 /* if we got here the keeper had a transition from OFF->ON */ 3821 if (hold) 3822 pm_hold_power(kept_dip); 3823 3824 if (!is_all_at_normal) 3825 (void) pm_all_to_normal(kept_dip, PM_CANBLOCK_FAIL); 3826 } 3827 3828 /* 3829 * A bunch of stuff that belongs only to the next routine (or two) 3830 */ 3831 3832 static const char namestr[] = "NAME="; 3833 static const int nameln = sizeof (namestr) - 1; 3834 static const char pmcompstr[] = "pm-components"; 3835 3836 struct pm_comp_pkg { 3837 pm_comp_t *comp; 3838 struct pm_comp_pkg *next; 3839 }; 3840 3841 #define isdigit(ch) ((ch) >= '0' && (ch) <= '9') 3842 3843 #define isxdigit(ch) (isdigit(ch) || ((ch) >= 'a' && (ch) <= 'f') || \ 3844 ((ch) >= 'A' && (ch) <= 'F')) 3845 3846 /* 3847 * Rather than duplicate this code ... 3848 * (this code excerpted from the function that follows it) 3849 */ 3850 #define FINISH_COMP { \ 3851 ASSERT(compp); \ 3852 compp->pmc_lnames_sz = size; \ 3853 tp = compp->pmc_lname_buf = kmem_alloc(size, KM_SLEEP); \ 3854 compp->pmc_numlevels = level; \ 3855 compp->pmc_lnames = kmem_alloc(level * sizeof (char *), KM_SLEEP); \ 3856 compp->pmc_lvals = kmem_alloc(level * sizeof (int), KM_SLEEP); \ 3857 compp->pmc_thresh = kmem_alloc(level * sizeof (int), KM_SLEEP); \ 3858 /* copy string out of prop array into buffer */ \ 3859 for (j = 0; j < level; j++) { \ 3860 compp->pmc_thresh[j] = INT_MAX; /* only [0] sticks */ \ 3861 compp->pmc_lvals[j] = lvals[j]; \ 3862 (void) strcpy(tp, lnames[j]); \ 3863 compp->pmc_lnames[j] = tp; \ 3864 tp += lszs[j]; \ 3865 } \ 3866 ASSERT(tp > compp->pmc_lname_buf && tp <= \ 3867 compp->pmc_lname_buf + compp->pmc_lnames_sz); \ 3868 } 3869 3870 /* 3871 * Create (empty) component data structures. 3872 */ 3873 static void 3874 e_pm_create_components(dev_info_t *dip, int num_components) 3875 { 3876 struct pm_component *compp, *ocompp; 3877 int i, size = 0; 3878 3879 ASSERT(!PM_IAM_LOCKING_DIP(dip)); 3880 ASSERT(!DEVI(dip)->devi_pm_components); 3881 ASSERT(!(DEVI(dip)->devi_pm_flags & PMC_COMPONENTS_DONE)); 3882 size = sizeof (struct pm_component) * num_components; 3883 3884 compp = kmem_zalloc(size, KM_SLEEP); 3885 ocompp = compp; 3886 DEVI(dip)->devi_pm_comp_size = size; 3887 DEVI(dip)->devi_pm_num_components = num_components; 3888 PM_LOCK_BUSY(dip); 3889 for (i = 0; i < num_components; i++) { 3890 compp->pmc_timestamp = gethrestime_sec(); 3891 compp->pmc_norm_pwr = (uint_t)-1; 3892 compp++; 3893 } 3894 PM_UNLOCK_BUSY(dip); 3895 DEVI(dip)->devi_pm_components = ocompp; 3896 DEVI(dip)->devi_pm_flags |= PMC_COMPONENTS_DONE; 3897 } 3898 3899 /* 3900 * Parse hex or decimal value from char string 3901 */ 3902 static char * 3903 pm_parsenum(char *cp, int *valp) 3904 { 3905 int ch, offset; 3906 char numbuf[256]; 3907 char *np = numbuf; 3908 int value = 0; 3909 3910 ch = *cp++; 3911 if (isdigit(ch)) { 3912 if (ch == '0') { 3913 if ((ch = *cp++) == 'x' || ch == 'X') { 3914 ch = *cp++; 3915 while (isxdigit(ch)) { 3916 *np++ = (char)ch; 3917 ch = *cp++; 3918 } 3919 *np = 0; 3920 cp--; 3921 goto hexval; 3922 } else { 3923 goto digit; 3924 } 3925 } else { 3926 digit: 3927 while (isdigit(ch)) { 3928 *np++ = (char)ch; 3929 ch = *cp++; 3930 } 3931 *np = 0; 3932 cp--; 3933 goto decval; 3934 } 3935 } else 3936 return (NULL); 3937 3938 hexval: 3939 for (np = numbuf; *np; np++) { 3940 if (*np >= 'a' && *np <= 'f') 3941 offset = 'a' - 10; 3942 else if (*np >= 'A' && *np <= 'F') 3943 offset = 'A' - 10; 3944 else if (*np >= '0' && *np <= '9') 3945 offset = '0'; 3946 value *= 16; 3947 value += *np - offset; 3948 } 3949 *valp = value; 3950 return (cp); 3951 3952 decval: 3953 offset = '0'; 3954 for (np = numbuf; *np; np++) { 3955 value *= 10; 3956 value += *np - offset; 3957 } 3958 *valp = value; 3959 return (cp); 3960 } 3961 3962 /* 3963 * Set max (previously documented as "normal") power. 3964 */ 3965 static void 3966 e_pm_set_max_power(dev_info_t *dip, int component_number, int level) 3967 { 3968 PM_CP(dip, component_number)->pmc_norm_pwr = level; 3969 } 3970 3971 /* 3972 * Get max (previously documented as "normal") power. 3973 */ 3974 static int 3975 e_pm_get_max_power(dev_info_t *dip, int component_number) 3976 { 3977 return (PM_CP(dip, component_number)->pmc_norm_pwr); 3978 } 3979 3980 /* 3981 * Internal routine for destroying components 3982 * It is called even when there might not be any, so it must be forgiving. 3983 */ 3984 static void 3985 e_pm_destroy_components(dev_info_t *dip) 3986 { 3987 int i; 3988 struct pm_component *cp; 3989 3990 ASSERT(!PM_IAM_LOCKING_DIP(dip)); 3991 if (PM_NUMCMPTS(dip) == 0) 3992 return; 3993 cp = DEVI(dip)->devi_pm_components; 3994 ASSERT(cp); 3995 for (i = 0; i < PM_NUMCMPTS(dip); i++, cp++) { 3996 int nlevels = cp->pmc_comp.pmc_numlevels; 3997 kmem_free(cp->pmc_comp.pmc_lvals, nlevels * sizeof (int)); 3998 kmem_free(cp->pmc_comp.pmc_thresh, nlevels * sizeof (int)); 3999 /* 4000 * For BC nodes, the rest is static in bc_comp, so skip it 4001 */ 4002 if (PM_ISBC(dip)) 4003 continue; 4004 kmem_free(cp->pmc_comp.pmc_name, cp->pmc_comp.pmc_name_sz); 4005 kmem_free(cp->pmc_comp.pmc_lnames, nlevels * sizeof (char *)); 4006 kmem_free(cp->pmc_comp.pmc_lname_buf, 4007 cp->pmc_comp.pmc_lnames_sz); 4008 } 4009 kmem_free(DEVI(dip)->devi_pm_components, DEVI(dip)->devi_pm_comp_size); 4010 DEVI(dip)->devi_pm_components = NULL; 4011 DEVI(dip)->devi_pm_num_components = 0; 4012 DEVI(dip)->devi_pm_flags &= 4013 ~(PMC_COMPONENTS_DONE | PMC_COMPONENTS_FAILED); 4014 } 4015 4016 /* 4017 * Read the pm-components property (if there is one) and use it to set up 4018 * components. Returns a pointer to an array of component structures if 4019 * pm-components found and successfully parsed, else returns NULL. 4020 * Sets error return *errp to true to indicate a failure (as opposed to no 4021 * property being present). 4022 */ 4023 pm_comp_t * 4024 pm_autoconfig(dev_info_t *dip, int *errp) 4025 { 4026 PMD_FUNC(pmf, "autoconfig") 4027 uint_t nelems; 4028 char **pp; 4029 pm_comp_t *compp = NULL; 4030 int i, j, level, components = 0; 4031 size_t size = 0; 4032 struct pm_comp_pkg *p, *ptail; 4033 struct pm_comp_pkg *phead = NULL; 4034 int *lvals = NULL; 4035 int *lszs = NULL; 4036 int *np = NULL; 4037 int npi = 0; 4038 char **lnames = NULL; 4039 char *cp, *tp; 4040 pm_comp_t *ret = NULL; 4041 4042 ASSERT(!PM_IAM_LOCKING_DIP(dip)); 4043 *errp = 0; /* assume success */ 4044 if (ddi_prop_lookup_string_array(DDI_DEV_T_ANY, dip, DDI_PROP_DONTPASS, 4045 (char *)pmcompstr, &pp, &nelems) != DDI_PROP_SUCCESS) { 4046 return (NULL); 4047 } 4048 4049 if (nelems < 3) { /* need at least one name and two levels */ 4050 goto errout; 4051 } 4052 4053 /* 4054 * pm_create_components is no longer allowed 4055 */ 4056 if (PM_NUMCMPTS(dip) != 0) { 4057 PMD(PMD_ERROR, ("%s: %s@%s(%s#%d) has %d comps\n", 4058 pmf, PM_DEVICE(dip), PM_NUMCMPTS(dip))) 4059 goto errout; 4060 } 4061 4062 lvals = kmem_alloc(nelems * sizeof (int), KM_SLEEP); 4063 lszs = kmem_alloc(nelems * sizeof (int), KM_SLEEP); 4064 lnames = kmem_alloc(nelems * sizeof (char *), KM_SLEEP); 4065 np = kmem_alloc(nelems * sizeof (int), KM_SLEEP); 4066 4067 level = 0; 4068 phead = NULL; 4069 for (i = 0; i < nelems; i++) { 4070 cp = pp[i]; 4071 if (!isdigit(*cp)) { /* must be name */ 4072 if (strncmp(cp, namestr, nameln) != 0) { 4073 goto errout; 4074 } 4075 if (i != 0) { 4076 if (level == 0) { /* no level spec'd */ 4077 PMD(PMD_ERROR, ("%s: no level spec'd\n", 4078 pmf)) 4079 goto errout; 4080 } 4081 np[npi++] = lvals[level - 1]; 4082 /* finish up previous component levels */ 4083 FINISH_COMP; 4084 } 4085 cp += nameln; 4086 if (!*cp) { 4087 PMD(PMD_ERROR, ("%s: nsa\n", pmf)) 4088 goto errout; 4089 } 4090 p = kmem_zalloc(sizeof (*phead), KM_SLEEP); 4091 if (phead == NULL) { 4092 phead = ptail = p; 4093 } else { 4094 ptail->next = p; 4095 ptail = p; 4096 } 4097 compp = p->comp = kmem_zalloc(sizeof (pm_comp_t), 4098 KM_SLEEP); 4099 compp->pmc_name_sz = strlen(cp) + 1; 4100 compp->pmc_name = kmem_zalloc(compp->pmc_name_sz, 4101 KM_SLEEP); 4102 (void) strncpy(compp->pmc_name, cp, compp->pmc_name_sz); 4103 components++; 4104 level = 0; 4105 } else { /* better be power level <num>=<name> */ 4106 #ifdef DEBUG 4107 tp = cp; 4108 #endif 4109 if (i == 0 || 4110 (cp = pm_parsenum(cp, &lvals[level])) == NULL) { 4111 PMD(PMD_ERROR, ("%s: parsenum(%s)\n", pmf, tp)) 4112 goto errout; 4113 } 4114 #ifdef DEBUG 4115 tp = cp; 4116 #endif 4117 if (*cp++ != '=' || !*cp) { 4118 PMD(PMD_ERROR, ("%s: ex =, got %s\n", pmf, tp)) 4119 goto errout; 4120 } 4121 4122 lszs[level] = strlen(cp) + 1; 4123 size += lszs[level]; 4124 lnames[level] = cp; /* points into prop string */ 4125 level++; 4126 } 4127 } 4128 np[npi++] = lvals[level - 1]; 4129 if (level == 0) { /* ended with a name */ 4130 PMD(PMD_ERROR, ("%s: ewn\n", pmf)) 4131 goto errout; 4132 } 4133 FINISH_COMP; 4134 4135 4136 /* 4137 * Now we have a list of components--we have to return instead an 4138 * array of them, but we can just copy the top level and leave 4139 * the rest as is 4140 */ 4141 (void) e_pm_create_components(dip, components); 4142 for (i = 0; i < components; i++) 4143 e_pm_set_max_power(dip, i, np[i]); 4144 4145 ret = kmem_zalloc(components * sizeof (pm_comp_t), KM_SLEEP); 4146 for (i = 0, p = phead; i < components; i++) { 4147 ASSERT(p); 4148 /* 4149 * Now sanity-check values: levels must be monotonically 4150 * increasing 4151 */ 4152 if (p->comp->pmc_numlevels < 2) { 4153 PMD(PMD_ERROR, ("%s: comp %s of %s@%s(%s#%d) only %d " 4154 "levels\n", pmf, 4155 p->comp->pmc_name, PM_DEVICE(dip), 4156 p->comp->pmc_numlevels)) 4157 goto errout; 4158 } 4159 for (j = 0; j < p->comp->pmc_numlevels; j++) { 4160 if ((p->comp->pmc_lvals[j] < 0) || ((j > 0) && 4161 (p->comp->pmc_lvals[j] <= 4162 p->comp->pmc_lvals[j - 1]))) { 4163 PMD(PMD_ERROR, ("%s: comp %s of %s@%s(%s#%d) " 4164 "not mono. incr, %d follows %d\n", pmf, 4165 p->comp->pmc_name, PM_DEVICE(dip), 4166 p->comp->pmc_lvals[j], 4167 p->comp->pmc_lvals[j - 1])) 4168 goto errout; 4169 } 4170 } 4171 ret[i] = *p->comp; /* struct assignment */ 4172 for (j = 0; j < i; j++) { 4173 /* 4174 * Test for unique component names 4175 */ 4176 if (strcmp(ret[j].pmc_name, ret[i].pmc_name) == 0) { 4177 PMD(PMD_ERROR, ("%s: %s of %s@%s(%s#%d) not " 4178 "unique\n", pmf, ret[j].pmc_name, 4179 PM_DEVICE(dip))) 4180 goto errout; 4181 } 4182 } 4183 ptail = p; 4184 p = p->next; 4185 phead = p; /* errout depends on phead making sense */ 4186 kmem_free(ptail->comp, sizeof (*ptail->comp)); 4187 kmem_free(ptail, sizeof (*ptail)); 4188 } 4189 out: 4190 ddi_prop_free(pp); 4191 if (lvals) 4192 kmem_free(lvals, nelems * sizeof (int)); 4193 if (lszs) 4194 kmem_free(lszs, nelems * sizeof (int)); 4195 if (lnames) 4196 kmem_free(lnames, nelems * sizeof (char *)); 4197 if (np) 4198 kmem_free(np, nelems * sizeof (int)); 4199 return (ret); 4200 4201 errout: 4202 e_pm_destroy_components(dip); 4203 *errp = 1; /* signal failure */ 4204 cmn_err(CE_CONT, "!pm: %s property ", pmcompstr); 4205 for (i = 0; i < nelems - 1; i++) 4206 cmn_err(CE_CONT, "!'%s', ", pp[i]); 4207 if (nelems != 0) 4208 cmn_err(CE_CONT, "!'%s'", pp[nelems - 1]); 4209 cmn_err(CE_CONT, "! for %s@%s(%s#%d) is ill-formed.\n", PM_DEVICE(dip)); 4210 for (p = phead; p; ) { 4211 pm_comp_t *pp; 4212 int n; 4213 4214 ptail = p; 4215 /* 4216 * Free component data structures 4217 */ 4218 pp = p->comp; 4219 n = pp->pmc_numlevels; 4220 if (pp->pmc_name_sz) { 4221 kmem_free(pp->pmc_name, pp->pmc_name_sz); 4222 } 4223 if (pp->pmc_lnames_sz) { 4224 kmem_free(pp->pmc_lname_buf, pp->pmc_lnames_sz); 4225 } 4226 if (pp->pmc_lnames) { 4227 kmem_free(pp->pmc_lnames, n * (sizeof (char *))); 4228 } 4229 if (pp->pmc_thresh) { 4230 kmem_free(pp->pmc_thresh, n * (sizeof (int))); 4231 } 4232 if (pp->pmc_lvals) { 4233 kmem_free(pp->pmc_lvals, n * (sizeof (int))); 4234 } 4235 p = ptail->next; 4236 kmem_free(ptail, sizeof (*ptail)); 4237 } 4238 if (ret != NULL) 4239 kmem_free(ret, components * sizeof (pm_comp_t)); 4240 ret = NULL; 4241 goto out; 4242 } 4243 4244 /* 4245 * Set threshold values for a devices components by dividing the target 4246 * threshold (base) by the number of transitions and assign each transition 4247 * that threshold. This will get the entire device down in the target time if 4248 * all components are idle and even if there are dependencies among components. 4249 * 4250 * Devices may well get powered all the way down before the target time, but 4251 * at least the EPA will be happy. 4252 */ 4253 void 4254 pm_set_device_threshold(dev_info_t *dip, int base, int flag) 4255 { 4256 PMD_FUNC(pmf, "set_device_threshold") 4257 int target_threshold = (base * 95) / 100; 4258 int level, comp; /* loop counters */ 4259 int transitions = 0; 4260 int ncomp = PM_NUMCMPTS(dip); 4261 int thresh; 4262 int remainder; 4263 pm_comp_t *pmc; 4264 int i, circ; 4265 4266 ASSERT(!PM_IAM_LOCKING_DIP(dip)); 4267 PM_LOCK_DIP(dip); 4268 /* 4269 * First we handle the easy one. If we're setting the default 4270 * threshold for a node with children, then we set it to the 4271 * default nexus threshold (currently 0) and mark it as default 4272 * nexus threshold instead 4273 */ 4274 if (PM_IS_NEXUS(dip)) { 4275 if (flag == PMC_DEF_THRESH) { 4276 PMD(PMD_THRESH, ("%s: [%s@%s(%s#%d) NEXDEF]\n", pmf, 4277 PM_DEVICE(dip))) 4278 thresh = pm_default_nexus_threshold; 4279 for (comp = 0; comp < ncomp; comp++) { 4280 pmc = &PM_CP(dip, comp)->pmc_comp; 4281 for (level = 1; level < pmc->pmc_numlevels; 4282 level++) { 4283 pmc->pmc_thresh[level] = thresh; 4284 } 4285 } 4286 DEVI(dip)->devi_pm_dev_thresh = 4287 pm_default_nexus_threshold; 4288 /* 4289 * If the nexus node is being reconfigured back to 4290 * the default threshold, adjust the notlowest count. 4291 */ 4292 if (DEVI(dip)->devi_pm_flags & 4293 (PMC_DEV_THRESH|PMC_COMP_THRESH)) { 4294 PM_LOCK_POWER(dip, &circ); 4295 for (i = 0; i < PM_NUMCMPTS(dip); i++) { 4296 if (PM_CURPOWER(dip, i) == 0) 4297 continue; 4298 mutex_enter(&pm_compcnt_lock); 4299 ASSERT(pm_comps_notlowest); 4300 pm_comps_notlowest--; 4301 PMD(PMD_LEVEL, ("%s: %s@%s(%s#%d) decr " 4302 "notlowest to %d\n", pmf, 4303 PM_DEVICE(dip), pm_comps_notlowest)) 4304 if (pm_comps_notlowest == 0) 4305 pm_ppm_notify_all_lowest(dip, 4306 PM_ALL_LOWEST); 4307 mutex_exit(&pm_compcnt_lock); 4308 } 4309 PM_UNLOCK_POWER(dip, circ); 4310 } 4311 DEVI(dip)->devi_pm_flags &= PMC_THRESH_NONE; 4312 DEVI(dip)->devi_pm_flags |= PMC_NEXDEF_THRESH; 4313 PM_UNLOCK_DIP(dip); 4314 return; 4315 } else if (DEVI(dip)->devi_pm_flags & PMC_NEXDEF_THRESH) { 4316 /* 4317 * If the nexus node is being configured for a 4318 * non-default threshold, include that node in 4319 * the notlowest accounting. 4320 */ 4321 PM_LOCK_POWER(dip, &circ); 4322 for (i = 0; i < PM_NUMCMPTS(dip); i++) { 4323 if (PM_CURPOWER(dip, i) == 0) 4324 continue; 4325 mutex_enter(&pm_compcnt_lock); 4326 if (pm_comps_notlowest == 0) 4327 pm_ppm_notify_all_lowest(dip, 4328 PM_NOT_ALL_LOWEST); 4329 pm_comps_notlowest++; 4330 PMD(PMD_LEVEL, ("%s: %s@%s(%s#%d) incr " 4331 "notlowest to %d\n", pmf, 4332 PM_DEVICE(dip), pm_comps_notlowest)) 4333 mutex_exit(&pm_compcnt_lock); 4334 } 4335 PM_UNLOCK_POWER(dip, circ); 4336 } 4337 } 4338 /* 4339 * Compute the total number of transitions for all components 4340 * of the device. Distribute the threshold evenly over them 4341 */ 4342 for (comp = 0; comp < ncomp; comp++) { 4343 pmc = &PM_CP(dip, comp)->pmc_comp; 4344 ASSERT(pmc->pmc_numlevels > 1); 4345 transitions += pmc->pmc_numlevels - 1; 4346 } 4347 ASSERT(transitions); 4348 thresh = target_threshold / transitions; 4349 4350 for (comp = 0; comp < ncomp; comp++) { 4351 pmc = &PM_CP(dip, comp)->pmc_comp; 4352 for (level = 1; level < pmc->pmc_numlevels; level++) { 4353 pmc->pmc_thresh[level] = thresh; 4354 } 4355 } 4356 4357 #ifdef DEBUG 4358 for (comp = 0; comp < ncomp; comp++) { 4359 pmc = &PM_CP(dip, comp)->pmc_comp; 4360 for (level = 1; level < pmc->pmc_numlevels; level++) { 4361 PMD(PMD_THRESH, ("%s: thresh before %s@%s(%s#%d) " 4362 "comp=%d, level=%d, %d\n", pmf, PM_DEVICE(dip), 4363 comp, level, pmc->pmc_thresh[level])) 4364 } 4365 } 4366 #endif 4367 /* 4368 * Distribute any remainder till they are all gone 4369 */ 4370 remainder = target_threshold - thresh * transitions; 4371 level = 1; 4372 #ifdef DEBUG 4373 PMD(PMD_THRESH, ("%s: remainder=%d target_threshold=%d thresh=%d " 4374 "trans=%d\n", pmf, remainder, target_threshold, thresh, 4375 transitions)) 4376 #endif 4377 while (remainder > 0) { 4378 comp = 0; 4379 while (remainder && (comp < ncomp)) { 4380 pmc = &PM_CP(dip, comp)->pmc_comp; 4381 if (level < pmc->pmc_numlevels) { 4382 pmc->pmc_thresh[level] += 1; 4383 remainder--; 4384 } 4385 comp++; 4386 } 4387 level++; 4388 } 4389 #ifdef DEBUG 4390 for (comp = 0; comp < ncomp; comp++) { 4391 pmc = &PM_CP(dip, comp)->pmc_comp; 4392 for (level = 1; level < pmc->pmc_numlevels; level++) { 4393 PMD(PMD_THRESH, ("%s: thresh after %s@%s(%s#%d) " 4394 "comp=%d level=%d, %d\n", pmf, PM_DEVICE(dip), 4395 comp, level, pmc->pmc_thresh[level])) 4396 } 4397 } 4398 #endif 4399 ASSERT(PM_IAM_LOCKING_DIP(dip)); 4400 DEVI(dip)->devi_pm_dev_thresh = base; 4401 DEVI(dip)->devi_pm_flags &= PMC_THRESH_NONE; 4402 DEVI(dip)->devi_pm_flags |= flag; 4403 PM_UNLOCK_DIP(dip); 4404 } 4405 4406 /* 4407 * Called when there is no old-style platform power management driver 4408 */ 4409 static int 4410 ddi_no_platform_power(power_req_t *req) 4411 { 4412 _NOTE(ARGUNUSED(req)) 4413 return (DDI_FAILURE); 4414 } 4415 4416 /* 4417 * This function calls the entry point supplied by the platform-specific 4418 * pm driver to bring the device component 'pm_cmpt' to power level 'pm_level'. 4419 * The use of global for getting the function name from platform-specific 4420 * pm driver is not ideal, but it is simple and efficient. 4421 * The previous property lookup was being done in the idle loop on swift 4422 * systems without pmc chips and hurt deskbench performance as well as 4423 * violating scheduler locking rules 4424 */ 4425 int (*pm_platform_power)(power_req_t *) = ddi_no_platform_power; 4426 4427 /* 4428 * Old obsolete interface for a device to request a power change (but only 4429 * an increase in power) 4430 */ 4431 int 4432 ddi_dev_is_needed(dev_info_t *dip, int cmpt, int level) 4433 { 4434 return (pm_raise_power(dip, cmpt, level)); 4435 } 4436 4437 /* 4438 * The old obsolete interface to platform power management. Only used by 4439 * Gypsy platform and APM on X86. 4440 */ 4441 int 4442 ddi_power(dev_info_t *dip, int pm_cmpt, int pm_level) 4443 { 4444 power_req_t request; 4445 4446 request.request_type = PMR_SET_POWER; 4447 request.req.set_power_req.who = dip; 4448 request.req.set_power_req.cmpt = pm_cmpt; 4449 request.req.set_power_req.level = pm_level; 4450 return (ddi_ctlops(dip, dip, DDI_CTLOPS_POWER, &request, NULL)); 4451 } 4452 4453 /* 4454 * A driver can invoke this from its detach routine when DDI_SUSPEND is 4455 * passed. Returns true if subsequent processing could result in power being 4456 * removed from the device. The arg is not currently used because it is 4457 * implicit in the operation of cpr/DR. 4458 */ 4459 int 4460 ddi_removing_power(dev_info_t *dip) 4461 { 4462 _NOTE(ARGUNUSED(dip)) 4463 return (pm_powering_down); 4464 } 4465 4466 /* 4467 * Returns true if a device indicates that its parent handles suspend/resume 4468 * processing for it. 4469 */ 4470 int 4471 e_ddi_parental_suspend_resume(dev_info_t *dip) 4472 { 4473 return (DEVI(dip)->devi_pm_flags & PMC_PARENTAL_SR); 4474 } 4475 4476 /* 4477 * Called for devices which indicate that their parent does suspend/resume 4478 * handling for them 4479 */ 4480 int 4481 e_ddi_suspend(dev_info_t *dip, ddi_detach_cmd_t cmd) 4482 { 4483 power_req_t request; 4484 request.request_type = PMR_SUSPEND; 4485 request.req.suspend_req.who = dip; 4486 request.req.suspend_req.cmd = cmd; 4487 return (ddi_ctlops(dip, dip, DDI_CTLOPS_POWER, &request, NULL)); 4488 } 4489 4490 /* 4491 * Called for devices which indicate that their parent does suspend/resume 4492 * handling for them 4493 */ 4494 int 4495 e_ddi_resume(dev_info_t *dip, ddi_attach_cmd_t cmd) 4496 { 4497 power_req_t request; 4498 request.request_type = PMR_RESUME; 4499 request.req.resume_req.who = dip; 4500 request.req.resume_req.cmd = cmd; 4501 return (ddi_ctlops(dip, dip, DDI_CTLOPS_POWER, &request, NULL)); 4502 } 4503 4504 /* 4505 * Old obsolete exported interface for drivers to create components. 4506 * This is now handled by exporting the pm-components property. 4507 */ 4508 int 4509 pm_create_components(dev_info_t *dip, int num_components) 4510 { 4511 PMD_FUNC(pmf, "pm_create_components") 4512 4513 if (num_components < 1) 4514 return (DDI_FAILURE); 4515 4516 if (!DEVI_IS_ATTACHING(dip)) { 4517 return (DDI_FAILURE); 4518 } 4519 4520 /* don't need to lock dip because attach is single threaded */ 4521 if (DEVI(dip)->devi_pm_components) { 4522 PMD(PMD_ERROR, ("%s: %s@%s(%s#%d) already has %d\n", pmf, 4523 PM_DEVICE(dip), PM_NUMCMPTS(dip))) 4524 return (DDI_FAILURE); 4525 } 4526 e_pm_create_components(dip, num_components); 4527 DEVI(dip)->devi_pm_flags |= PMC_BC; 4528 e_pm_default_components(dip, num_components); 4529 return (DDI_SUCCESS); 4530 } 4531 4532 /* 4533 * Obsolete interface previously called by drivers to destroy their components 4534 * at detach time. This is now done automatically. However, we need to keep 4535 * this for the old drivers. 4536 */ 4537 void 4538 pm_destroy_components(dev_info_t *dip) 4539 { 4540 PMD_FUNC(pmf, "pm_destroy_components") 4541 dev_info_t *pdip = ddi_get_parent(dip); 4542 4543 PMD(PMD_REMDEV | PMD_KIDSUP, ("%s: %s@%s(%s#%d)\n", pmf, 4544 PM_DEVICE(dip))) 4545 ASSERT(DEVI_IS_DETACHING(dip)); 4546 #ifdef DEBUG 4547 if (!PM_ISBC(dip)) 4548 cmn_err(CE_WARN, "!driver exporting pm-components property " 4549 "(%s@%s) calls pm_destroy_components", PM_NAME(dip), 4550 PM_ADDR(dip)); 4551 #endif 4552 /* 4553 * We ignore this unless this is an old-style driver, except for 4554 * printing the message above 4555 */ 4556 if (PM_NUMCMPTS(dip) == 0 || !PM_ISBC(dip)) { 4557 PMD(PMD_REMDEV, ("%s: ignore %s@%s(%s#%d)\n", pmf, 4558 PM_DEVICE(dip))) 4559 return; 4560 } 4561 ASSERT(PM_GET_PM_INFO(dip)); 4562 4563 /* 4564 * pm_unmanage will clear info pointer later, after dealing with 4565 * dependencies 4566 */ 4567 ASSERT(!PM_GET_PM_SCAN(dip)); /* better be gone already */ 4568 /* 4569 * Now adjust parent's kidsupcnt. We check only comp 0. 4570 * Parents that get notification are not adjusted because their 4571 * kidsupcnt is always 0 (or 1 during probe and attach). 4572 */ 4573 if ((PM_CURPOWER(dip, 0) != 0) && pdip && !PM_WANTS_NOTIFICATION(pdip)) 4574 pm_rele_power(pdip); 4575 #ifdef DEBUG 4576 else { 4577 PMD(PMD_KIDSUP, ("%s: kuc stays %s@%s(%s#%d) comps gone\n", 4578 pmf, PM_DEVICE(dip))) 4579 } 4580 #endif 4581 e_pm_destroy_components(dip); 4582 /* 4583 * Forget we ever knew anything about the components of this device 4584 */ 4585 DEVI(dip)->devi_pm_flags &= 4586 ~(PMC_BC | PMC_COMPONENTS_DONE | PMC_COMPONENTS_FAILED); 4587 } 4588 4589 /* 4590 * Exported interface for a driver to set a component busy. 4591 */ 4592 int 4593 pm_busy_component(dev_info_t *dip, int cmpt) 4594 { 4595 struct pm_component *cp; 4596 4597 ASSERT(dip != NULL); 4598 if (!e_pm_valid_info(dip, NULL) || !e_pm_valid_comp(dip, cmpt, &cp)) 4599 return (DDI_FAILURE); 4600 PM_LOCK_BUSY(dip); 4601 cp->pmc_busycount++; 4602 cp->pmc_timestamp = 0; 4603 PM_UNLOCK_BUSY(dip); 4604 return (DDI_SUCCESS); 4605 } 4606 4607 /* 4608 * Exported interface for a driver to set a component idle. 4609 */ 4610 int 4611 pm_idle_component(dev_info_t *dip, int cmpt) 4612 { 4613 PMD_FUNC(pmf, "pm_idle_component") 4614 struct pm_component *cp; 4615 pm_scan_t *scanp = PM_GET_PM_SCAN(dip); 4616 4617 if (!e_pm_valid_info(dip, NULL) || !e_pm_valid_comp(dip, cmpt, &cp)) 4618 return (DDI_FAILURE); 4619 4620 PM_LOCK_BUSY(dip); 4621 if (cp->pmc_busycount) { 4622 if (--(cp->pmc_busycount) == 0) 4623 cp->pmc_timestamp = gethrestime_sec(); 4624 } else { 4625 cp->pmc_timestamp = gethrestime_sec(); 4626 } 4627 4628 PM_UNLOCK_BUSY(dip); 4629 4630 /* 4631 * if device becomes idle during idle down period, try scan it down 4632 */ 4633 if (scanp && PM_IS_PID(dip)) { 4634 PMD(PMD_IDLEDOWN, ("%s: %s@%s(%s#%d) idle.\n", pmf, 4635 PM_DEVICE(dip))) 4636 pm_rescan(dip); 4637 return (DDI_SUCCESS); 4638 } 4639 4640 /* 4641 * handle scan not running with nexus threshold == 0 4642 */ 4643 4644 if (PM_IS_NEXUS(dip) && (cp->pmc_busycount == 0)) { 4645 pm_rescan(dip); 4646 } 4647 4648 return (DDI_SUCCESS); 4649 } 4650 4651 /* 4652 * This is the old obsolete interface called by drivers to set their normal 4653 * power. Thus we can't fix its behavior or return a value. 4654 * This functionality is replaced by the pm-component property. 4655 * We'll only get components destroyed while no power management is 4656 * going on (and the device is detached), so we don't need a mutex here 4657 */ 4658 void 4659 pm_set_normal_power(dev_info_t *dip, int comp, int level) 4660 { 4661 PMD_FUNC(pmf, "set_normal_power") 4662 #ifdef DEBUG 4663 if (!PM_ISBC(dip)) 4664 cmn_err(CE_WARN, "!call to pm_set_normal_power() by %s@%s " 4665 "(driver exporting pm-components property) ignored", 4666 PM_NAME(dip), PM_ADDR(dip)); 4667 #endif 4668 if (PM_ISBC(dip)) { 4669 PMD(PMD_NORM, ("%s: %s@%s(%s#%d) set normal power comp=%d, " 4670 "level=%d\n", pmf, PM_DEVICE(dip), comp, level)) 4671 e_pm_set_max_power(dip, comp, level); 4672 e_pm_default_levels(dip, PM_CP(dip, comp), level); 4673 } 4674 } 4675 4676 /* 4677 * Called on a successfully detached driver to free pm resources 4678 */ 4679 static void 4680 pm_stop(dev_info_t *dip) 4681 { 4682 PMD_FUNC(pmf, "stop") 4683 dev_info_t *pdip = ddi_get_parent(dip); 4684 4685 ASSERT(!PM_IAM_LOCKING_DIP(dip)); 4686 /* stopping scan, destroy scan data structure */ 4687 if (!PM_ISBC(dip)) { 4688 pm_scan_stop(dip); 4689 pm_scan_fini(dip); 4690 } 4691 4692 if (PM_GET_PM_INFO(dip) != NULL) { 4693 if (pm_unmanage(dip) == DDI_SUCCESS) { 4694 /* 4695 * Old style driver may have called 4696 * pm_destroy_components already, but just in case ... 4697 */ 4698 e_pm_destroy_components(dip); 4699 } else { 4700 PMD(PMD_FAIL, ("%s: can't pm_unmanage %s@%s(%s#%d)\n", 4701 pmf, PM_DEVICE(dip))) 4702 } 4703 } else { 4704 if (PM_NUMCMPTS(dip)) 4705 e_pm_destroy_components(dip); 4706 else { 4707 if (DEVI(dip)->devi_pm_flags & PMC_NOPMKID) { 4708 DEVI(dip)->devi_pm_flags &= ~PMC_NOPMKID; 4709 if (pdip && !PM_WANTS_NOTIFICATION(pdip)) { 4710 pm_rele_power(pdip); 4711 } else if (pdip && 4712 MDI_VHCI(pdip) && MDI_CLIENT(dip)) { 4713 (void) mdi_power(pdip, 4714 MDI_PM_RELE_POWER, 4715 (void *)dip, NULL, 0); 4716 } 4717 } 4718 } 4719 } 4720 } 4721 4722 /* 4723 * The node is the subject of a reparse pm props ioctl. Throw away the old 4724 * info and start over. 4725 */ 4726 int 4727 e_new_pm_props(dev_info_t *dip) 4728 { 4729 if (PM_GET_PM_INFO(dip) != NULL) { 4730 pm_stop(dip); 4731 4732 if (e_pm_manage(dip, PM_STYLE_NEW) != DDI_SUCCESS) { 4733 return (DDI_FAILURE); 4734 } 4735 } 4736 e_pm_props(dip); 4737 return (DDI_SUCCESS); 4738 } 4739 4740 /* 4741 * Device has been attached, so process its pm properties 4742 */ 4743 void 4744 e_pm_props(dev_info_t *dip) 4745 { 4746 char *pp; 4747 int len; 4748 int flags = 0; 4749 int propflag = DDI_PROP_DONTPASS|DDI_PROP_CANSLEEP; 4750 4751 /* 4752 * It doesn't matter if we do this more than once, we should always 4753 * get the same answers, and if not, then the last one in is the 4754 * best one. 4755 */ 4756 if (ddi_getlongprop(DDI_DEV_T_ANY, dip, propflag, "pm-hardware-state", 4757 (caddr_t)&pp, &len) == DDI_PROP_SUCCESS) { 4758 if (strcmp(pp, "needs-suspend-resume") == 0) { 4759 flags = PMC_NEEDS_SR; 4760 } else if (strcmp(pp, "no-suspend-resume") == 0) { 4761 flags = PMC_NO_SR; 4762 } else if (strcmp(pp, "parental-suspend-resume") == 0) { 4763 flags = PMC_PARENTAL_SR; 4764 } else { 4765 cmn_err(CE_NOTE, "!device %s@%s has unrecognized " 4766 "%s property value '%s'", PM_NAME(dip), 4767 PM_ADDR(dip), "pm-hardware-state", pp); 4768 } 4769 kmem_free(pp, len); 4770 } 4771 /* 4772 * This next segment (PMC_WANTS_NOTIFY) is in 4773 * support of nexus drivers which will want to be involved in 4774 * (or at least notified of) their child node's power level transitions. 4775 * "pm-want-child-notification?" is defined by the parent. 4776 */ 4777 if (ddi_prop_exists(DDI_DEV_T_ANY, dip, propflag, 4778 "pm-want-child-notification?") && PM_HAS_BUS_POWER(dip)) 4779 flags |= PMC_WANTS_NOTIFY; 4780 ASSERT(PM_HAS_BUS_POWER(dip) || !ddi_prop_exists(DDI_DEV_T_ANY, 4781 dip, propflag, "pm-want-child-notification?")); 4782 if (ddi_prop_exists(DDI_DEV_T_ANY, dip, propflag, 4783 "no-involuntary-power-cycles")) 4784 flags |= PMC_NO_INVOL; 4785 /* 4786 * Is the device a CPU device? 4787 */ 4788 if (ddi_getlongprop(DDI_DEV_T_ANY, dip, propflag, "pm-class", 4789 (caddr_t)&pp, &len) == DDI_PROP_SUCCESS) { 4790 if (strcmp(pp, "CPU") == 0) { 4791 flags |= PMC_CPU_DEVICE; 4792 } else { 4793 cmn_err(CE_NOTE, "!device %s@%s has unrecognized " 4794 "%s property value '%s'", PM_NAME(dip), 4795 PM_ADDR(dip), "pm-class", pp); 4796 } 4797 kmem_free(pp, len); 4798 } 4799 /* devfs single threads us */ 4800 DEVI(dip)->devi_pm_flags |= flags; 4801 } 4802 4803 /* 4804 * This is the DDI_CTLOPS_POWER handler that is used when there is no ppm 4805 * driver which has claimed a node. 4806 * Sets old_power in arg struct. 4807 */ 4808 static int 4809 pm_default_ctlops(dev_info_t *dip, dev_info_t *rdip, 4810 ddi_ctl_enum_t ctlop, void *arg, void *result) 4811 { 4812 _NOTE(ARGUNUSED(dip)) 4813 PMD_FUNC(pmf, "ctlops") 4814 power_req_t *reqp = (power_req_t *)arg; 4815 int retval; 4816 dev_info_t *target_dip; 4817 int new_level, old_level, cmpt; 4818 #ifdef PMDDEBUG 4819 char *format; 4820 #endif 4821 4822 /* 4823 * The interface for doing the actual power level changes is now 4824 * through the DDI_CTLOPS_POWER bus_ctl, so that we can plug in 4825 * different platform-specific power control drivers. 4826 * 4827 * This driver implements the "default" version of this interface. 4828 * If no ppm driver has been installed then this interface is called 4829 * instead. 4830 */ 4831 ASSERT(dip == NULL); 4832 switch (ctlop) { 4833 case DDI_CTLOPS_POWER: 4834 switch (reqp->request_type) { 4835 case PMR_PPM_SET_POWER: 4836 { 4837 target_dip = reqp->req.ppm_set_power_req.who; 4838 ASSERT(target_dip == rdip); 4839 new_level = reqp->req.ppm_set_power_req.new_level; 4840 cmpt = reqp->req.ppm_set_power_req.cmpt; 4841 /* pass back old power for the PM_LEVEL_UNKNOWN case */ 4842 old_level = PM_CURPOWER(target_dip, cmpt); 4843 reqp->req.ppm_set_power_req.old_level = old_level; 4844 retval = pm_power(target_dip, cmpt, new_level); 4845 PMD(PMD_PPM, ("%s: PPM_SET_POWER %s@%s(%s#%d)[%d] %d->" 4846 "%d %s\n", pmf, PM_DEVICE(target_dip), cmpt, 4847 old_level, new_level, (retval == DDI_SUCCESS ? 4848 "chd" : "no chg"))) 4849 return (retval); 4850 } 4851 4852 case PMR_PPM_PRE_DETACH: 4853 case PMR_PPM_POST_DETACH: 4854 case PMR_PPM_PRE_ATTACH: 4855 case PMR_PPM_POST_ATTACH: 4856 case PMR_PPM_PRE_PROBE: 4857 case PMR_PPM_POST_PROBE: 4858 case PMR_PPM_PRE_RESUME: 4859 case PMR_PPM_INIT_CHILD: 4860 case PMR_PPM_UNINIT_CHILD: 4861 #ifdef PMDDEBUG 4862 switch (reqp->request_type) { 4863 case PMR_PPM_PRE_DETACH: 4864 format = "%s: PMR_PPM_PRE_DETACH " 4865 "%s@%s(%s#%d)\n"; 4866 break; 4867 case PMR_PPM_POST_DETACH: 4868 format = "%s: PMR_PPM_POST_DETACH " 4869 "%s@%s(%s#%d) rets %d\n"; 4870 break; 4871 case PMR_PPM_PRE_ATTACH: 4872 format = "%s: PMR_PPM_PRE_ATTACH " 4873 "%s@%s(%s#%d)\n"; 4874 break; 4875 case PMR_PPM_POST_ATTACH: 4876 format = "%s: PMR_PPM_POST_ATTACH " 4877 "%s@%s(%s#%d) rets %d\n"; 4878 break; 4879 case PMR_PPM_PRE_PROBE: 4880 format = "%s: PMR_PPM_PRE_PROBE " 4881 "%s@%s(%s#%d)\n"; 4882 break; 4883 case PMR_PPM_POST_PROBE: 4884 format = "%s: PMR_PPM_POST_PROBE " 4885 "%s@%s(%s#%d) rets %d\n"; 4886 break; 4887 case PMR_PPM_PRE_RESUME: 4888 format = "%s: PMR_PPM_PRE_RESUME " 4889 "%s@%s(%s#%d) rets %d\n"; 4890 break; 4891 case PMR_PPM_INIT_CHILD: 4892 format = "%s: PMR_PPM_INIT_CHILD " 4893 "%s@%s(%s#%d)\n"; 4894 break; 4895 case PMR_PPM_UNINIT_CHILD: 4896 format = "%s: PMR_PPM_UNINIT_CHILD " 4897 "%s@%s(%s#%d)\n"; 4898 break; 4899 default: 4900 break; 4901 } 4902 PMD(PMD_PPM, (format, pmf, PM_DEVICE(rdip), 4903 reqp->req.ppm_config_req.result)) 4904 #endif 4905 return (DDI_SUCCESS); 4906 4907 case PMR_PPM_POWER_CHANGE_NOTIFY: 4908 /* 4909 * Nothing for us to do 4910 */ 4911 ASSERT(reqp->req.ppm_notify_level_req.who == rdip); 4912 PMD(PMD_PPM, ("%s: PMR_PPM_POWER_CHANGE_NOTIFY " 4913 "%s@%s(%s#%d)[%d] %d->%d\n", pmf, 4914 PM_DEVICE(reqp->req.ppm_notify_level_req.who), 4915 reqp->req.ppm_notify_level_req.cmpt, 4916 PM_CURPOWER(reqp->req.ppm_notify_level_req.who, 4917 reqp->req.ppm_notify_level_req.cmpt), 4918 reqp->req.ppm_notify_level_req.new_level)) 4919 return (DDI_SUCCESS); 4920 4921 case PMR_PPM_UNMANAGE: 4922 PMD(PMD_PPM, ("%s: PMR_PPM_UNMANAGE %s@%s(%s#%d)\n", 4923 pmf, PM_DEVICE(rdip))) 4924 return (DDI_SUCCESS); 4925 4926 case PMR_PPM_LOCK_POWER: 4927 pm_lock_power_single(reqp->req.ppm_lock_power_req.who, 4928 reqp->req.ppm_lock_power_req.circp); 4929 return (DDI_SUCCESS); 4930 4931 case PMR_PPM_UNLOCK_POWER: 4932 pm_unlock_power_single( 4933 reqp->req.ppm_unlock_power_req.who, 4934 reqp->req.ppm_unlock_power_req.circ); 4935 return (DDI_SUCCESS); 4936 4937 case PMR_PPM_TRY_LOCK_POWER: 4938 *(int *)result = pm_try_locking_power_single( 4939 reqp->req.ppm_lock_power_req.who, 4940 reqp->req.ppm_lock_power_req.circp); 4941 return (DDI_SUCCESS); 4942 4943 case PMR_PPM_POWER_LOCK_OWNER: 4944 target_dip = reqp->req.ppm_power_lock_owner_req.who; 4945 ASSERT(target_dip == rdip); 4946 reqp->req.ppm_power_lock_owner_req.owner = 4947 DEVI(rdip)->devi_busy_thread; 4948 return (DDI_SUCCESS); 4949 default: 4950 PMD(PMD_ERROR, ("%s: default!\n", pmf)) 4951 return (DDI_FAILURE); 4952 } 4953 4954 default: 4955 PMD(PMD_ERROR, ("%s: unknown\n", pmf)) 4956 return (DDI_FAILURE); 4957 } 4958 } 4959 4960 /* 4961 * We overload the bus_ctl ops here--perhaps we ought to have a distinct 4962 * power_ops struct for this functionality instead? 4963 * However, we only ever do this on a ppm driver. 4964 */ 4965 int 4966 pm_ctlops(dev_info_t *d, dev_info_t *r, ddi_ctl_enum_t op, void *a, void *v) 4967 { 4968 int (*fp)(); 4969 4970 /* if no ppm handler, call the default routine */ 4971 if (d == NULL) { 4972 return (pm_default_ctlops(d, r, op, a, v)); 4973 } 4974 if (!d || !r) 4975 return (DDI_FAILURE); 4976 ASSERT(DEVI(d)->devi_ops && DEVI(d)->devi_ops->devo_bus_ops && 4977 DEVI(d)->devi_ops->devo_bus_ops->bus_ctl); 4978 4979 fp = DEVI(d)->devi_ops->devo_bus_ops->bus_ctl; 4980 return ((*fp)(d, r, op, a, v)); 4981 } 4982 4983 /* 4984 * Called on a node when attach completes or the driver makes its first pm 4985 * call (whichever comes first). 4986 * In the attach case, device may not be power manageable at all. 4987 * Don't need to lock the dip because we're single threaded by the devfs code 4988 */ 4989 static int 4990 pm_start(dev_info_t *dip) 4991 { 4992 PMD_FUNC(pmf, "start") 4993 int ret; 4994 dev_info_t *pdip = ddi_get_parent(dip); 4995 int e_pm_manage(dev_info_t *, int); 4996 void pm_noinvol_specd(dev_info_t *dip); 4997 4998 e_pm_props(dip); 4999 pm_noinvol_specd(dip); 5000 /* 5001 * If this dip has already been processed, don't mess with it 5002 * (but decrement the speculative count we did above, as whatever 5003 * code put it under pm already will have dealt with it) 5004 */ 5005 if (PM_GET_PM_INFO(dip)) { 5006 PMD(PMD_KIDSUP, ("%s: pm already done for %s@%s(%s#%d)\n", 5007 pmf, PM_DEVICE(dip))) 5008 return (0); 5009 } 5010 ret = e_pm_manage(dip, PM_STYLE_UNKNOWN); 5011 5012 if (PM_GET_PM_INFO(dip) == NULL) { 5013 /* 5014 * keep the kidsupcount increment as is 5015 */ 5016 DEVI(dip)->devi_pm_flags |= PMC_NOPMKID; 5017 if (pdip && !PM_WANTS_NOTIFICATION(pdip)) { 5018 pm_hold_power(pdip); 5019 } else if (pdip && MDI_VHCI(pdip) && MDI_CLIENT(dip)) { 5020 (void) mdi_power(pdip, MDI_PM_HOLD_POWER, 5021 (void *)dip, NULL, 0); 5022 } 5023 5024 PMD(PMD_KIDSUP, ("%s: pm of %s@%s(%s#%d) failed, parent " 5025 "left up\n", pmf, PM_DEVICE(dip))) 5026 } 5027 5028 return (ret); 5029 } 5030 5031 /* 5032 * Keep a list of recorded thresholds. For now we just keep a list and 5033 * search it linearly. We don't expect too many entries. Can always hash it 5034 * later if we need to. 5035 */ 5036 void 5037 pm_record_thresh(pm_thresh_rec_t *rp) 5038 { 5039 pm_thresh_rec_t *pptr, *ptr; 5040 5041 ASSERT(*rp->ptr_physpath); 5042 rw_enter(&pm_thresh_rwlock, RW_WRITER); 5043 for (pptr = NULL, ptr = pm_thresh_head; 5044 ptr; pptr = ptr, ptr = ptr->ptr_next) { 5045 if (strcmp(rp->ptr_physpath, ptr->ptr_physpath) == 0) { 5046 /* replace this one */ 5047 rp->ptr_next = ptr->ptr_next; 5048 if (pptr) { 5049 pptr->ptr_next = rp; 5050 } else { 5051 pm_thresh_head = rp; 5052 } 5053 rw_exit(&pm_thresh_rwlock); 5054 kmem_free(ptr, ptr->ptr_size); 5055 return; 5056 } 5057 continue; 5058 } 5059 /* 5060 * There was not a match in the list, insert this one in front 5061 */ 5062 if (pm_thresh_head) { 5063 rp->ptr_next = pm_thresh_head; 5064 pm_thresh_head = rp; 5065 } else { 5066 rp->ptr_next = NULL; 5067 pm_thresh_head = rp; 5068 } 5069 rw_exit(&pm_thresh_rwlock); 5070 } 5071 5072 /* 5073 * Create a new dependency record and hang a new dependency entry off of it 5074 */ 5075 pm_pdr_t * 5076 newpdr(char *kept, char *keeps, int isprop) 5077 { 5078 size_t size = strlen(kept) + strlen(keeps) + 2 + sizeof (pm_pdr_t); 5079 pm_pdr_t *p = kmem_zalloc(size, KM_SLEEP); 5080 p->pdr_size = size; 5081 p->pdr_isprop = isprop; 5082 p->pdr_kept_paths = NULL; 5083 p->pdr_kept_count = 0; 5084 p->pdr_kept = (char *)((intptr_t)p + sizeof (pm_pdr_t)); 5085 (void) strcpy(p->pdr_kept, kept); 5086 p->pdr_keeper = (char *)((intptr_t)p->pdr_kept + strlen(kept) + 1); 5087 (void) strcpy(p->pdr_keeper, keeps); 5088 ASSERT((intptr_t)p->pdr_keeper + strlen(p->pdr_keeper) + 1 <= 5089 (intptr_t)p + size); 5090 ASSERT((intptr_t)p->pdr_kept + strlen(p->pdr_kept) + 1 <= 5091 (intptr_t)p + size); 5092 return (p); 5093 } 5094 5095 /* 5096 * Keep a list of recorded dependencies. We only keep the 5097 * keeper -> kept list for simplification. At this point We do not 5098 * care about whether the devices are attached or not yet, 5099 * this would be done in pm_keeper() and pm_kept(). 5100 * If a PM_RESET_PM happens, then we tear down and forget the dependencies, 5101 * and it is up to the user to issue the ioctl again if they want it 5102 * (e.g. pmconfig) 5103 * Returns true if dependency already exists in the list. 5104 */ 5105 int 5106 pm_record_keeper(char *kept, char *keeper, int isprop) 5107 { 5108 PMD_FUNC(pmf, "record_keeper") 5109 pm_pdr_t *npdr, *ppdr, *pdr; 5110 5111 PMD(PMD_KEEPS, ("%s: %s, %s\n", pmf, kept, keeper)) 5112 ASSERT(kept && keeper); 5113 #ifdef DEBUG 5114 if (pm_debug & PMD_KEEPS) 5115 prdeps("pm_record_keeper entry"); 5116 #endif 5117 for (ppdr = NULL, pdr = pm_dep_head; pdr; 5118 ppdr = pdr, pdr = pdr->pdr_next) { 5119 PMD(PMD_KEEPS, ("%s: check %s, %s\n", pmf, pdr->pdr_kept, 5120 pdr->pdr_keeper)) 5121 if (strcmp(kept, pdr->pdr_kept) == 0 && 5122 strcmp(keeper, pdr->pdr_keeper) == 0) { 5123 PMD(PMD_KEEPS, ("%s: match\n", pmf)) 5124 return (1); 5125 } 5126 } 5127 /* 5128 * We did not find any match, so we have to make an entry 5129 */ 5130 npdr = newpdr(kept, keeper, isprop); 5131 if (ppdr) { 5132 ASSERT(ppdr->pdr_next == NULL); 5133 ppdr->pdr_next = npdr; 5134 } else { 5135 ASSERT(pm_dep_head == NULL); 5136 pm_dep_head = npdr; 5137 } 5138 #ifdef DEBUG 5139 if (pm_debug & PMD_KEEPS) 5140 prdeps("pm_record_keeper after new record"); 5141 #endif 5142 if (!isprop) 5143 pm_unresolved_deps++; 5144 else 5145 pm_prop_deps++; 5146 return (0); 5147 } 5148 5149 /* 5150 * Look up this device in the set of devices we've seen ioctls for 5151 * to see if we are holding a threshold spec for it. If so, make it so. 5152 * At ioctl time, we were given the physical path of the device. 5153 */ 5154 int 5155 pm_thresh_specd(dev_info_t *dip) 5156 { 5157 void pm_apply_recorded_thresh(dev_info_t *, pm_thresh_rec_t *); 5158 char *path = 0; 5159 char pathbuf[MAXNAMELEN]; 5160 pm_thresh_rec_t *rp; 5161 5162 path = ddi_pathname(dip, pathbuf); 5163 5164 rw_enter(&pm_thresh_rwlock, RW_READER); 5165 for (rp = pm_thresh_head; rp; rp = rp->ptr_next) { 5166 if (strcmp(rp->ptr_physpath, path) != 0) 5167 continue; 5168 pm_apply_recorded_thresh(dip, rp); 5169 rw_exit(&pm_thresh_rwlock); 5170 return (1); 5171 } 5172 rw_exit(&pm_thresh_rwlock); 5173 return (0); 5174 } 5175 5176 static int 5177 pm_set_keeping(dev_info_t *keeper, dev_info_t *kept) 5178 { 5179 PMD_FUNC(pmf, "set_keeping") 5180 pm_info_t *kept_info; 5181 int j, up = 0, circ; 5182 void prdeps(char *); 5183 5184 PMD(PMD_KEEPS, ("%s: keeper=%s@%s(%s#%d), kept=%s@%s(%s#%d)\n", pmf, 5185 PM_DEVICE(keeper), PM_DEVICE(kept))) 5186 #ifdef DEBUG 5187 if (pm_debug & PMD_KEEPS) 5188 prdeps("Before PAD\n"); 5189 #endif 5190 ASSERT(keeper != kept); 5191 if (PM_GET_PM_INFO(keeper) == NULL) { 5192 cmn_err(CE_CONT, "!device %s@%s(%s#%d) keeps up device " 5193 "%s@%s(%s#%d), but the latter is not power managed", 5194 PM_DEVICE(keeper), PM_DEVICE(kept)); 5195 PMD((PMD_FAIL | PMD_KEEPS), ("%s: keeper %s@%s(%s#%d) is not" 5196 "power managed\n", pmf, PM_DEVICE(keeper))) 5197 return (0); 5198 } 5199 kept_info = PM_GET_PM_INFO(kept); 5200 ASSERT(kept_info); 5201 PM_LOCK_POWER(keeper, &circ); 5202 for (j = 0; j < PM_NUMCMPTS(keeper); j++) { 5203 if (PM_CURPOWER(keeper, j)) { 5204 up++; 5205 break; 5206 } 5207 } 5208 if (up) { 5209 /* Bringup and maintain a hold on the kept */ 5210 PMD(PMD_KEEPS, ("%s: place a hold on kept %s@%s(%s#%d)\n", pmf, 5211 PM_DEVICE(kept))) 5212 bring_pmdep_up(kept, 1); 5213 } 5214 PM_UNLOCK_POWER(keeper, circ); 5215 #ifdef DEBUG 5216 if (pm_debug & PMD_KEEPS) 5217 prdeps("After PAD\n"); 5218 #endif 5219 return (1); 5220 } 5221 5222 /* 5223 * Should this device keep up another device? 5224 * Look up this device in the set of devices we've seen ioctls for 5225 * to see if we are holding a dependency spec for it. If so, make it so. 5226 * Because we require the kept device to be attached already in order to 5227 * make the list entry (and hold it), we only need to look for keepers. 5228 * At ioctl time, we were given the physical path of the device. 5229 */ 5230 int 5231 pm_keeper(char *keeper) 5232 { 5233 PMD_FUNC(pmf, "keeper") 5234 int pm_apply_recorded_dep(dev_info_t *, pm_pdr_t *); 5235 dev_info_t *dip; 5236 pm_pdr_t *dp; 5237 dev_info_t *kept = NULL; 5238 int ret = 0; 5239 int i; 5240 5241 if (!pm_unresolved_deps && !pm_prop_deps) 5242 return (0); 5243 ASSERT(keeper != NULL); 5244 dip = pm_name_to_dip(keeper, 1); 5245 if (dip == NULL) 5246 return (0); 5247 PMD(PMD_KEEPS, ("%s: keeper=%s\n", pmf, keeper)) 5248 for (dp = pm_dep_head; dp; dp = dp->pdr_next) { 5249 if (!dp->pdr_isprop) { 5250 if (!pm_unresolved_deps) 5251 continue; 5252 PMD(PMD_KEEPS, ("%s: keeper %s\n", pmf, dp->pdr_keeper)) 5253 if (dp->pdr_satisfied) { 5254 PMD(PMD_KEEPS, ("%s: satisfied\n", pmf)) 5255 continue; 5256 } 5257 if (strcmp(dp->pdr_keeper, keeper) == 0) { 5258 ret += pm_apply_recorded_dep(dip, dp); 5259 } 5260 } else { 5261 if (strcmp(dp->pdr_keeper, keeper) != 0) 5262 continue; 5263 for (i = 0; i < dp->pdr_kept_count; i++) { 5264 if (dp->pdr_kept_paths[i] == NULL) 5265 continue; 5266 kept = pm_name_to_dip(dp->pdr_kept_paths[i], 1); 5267 if (kept == NULL) 5268 continue; 5269 ASSERT(ddi_prop_exists(DDI_DEV_T_ANY, kept, 5270 DDI_PROP_DONTPASS, dp->pdr_kept)); 5271 PMD(PMD_KEEPS, ("%s: keeper=%s@%s(%s#%d), " 5272 "kept=%s@%s(%s#%d) keptcnt=%d\n", 5273 pmf, PM_DEVICE(dip), PM_DEVICE(kept), 5274 dp->pdr_kept_count)) 5275 if (kept != dip) { 5276 ret += pm_set_keeping(dip, kept); 5277 } 5278 ddi_release_devi(kept); 5279 } 5280 5281 } 5282 } 5283 ddi_release_devi(dip); 5284 return (ret); 5285 } 5286 5287 /* 5288 * Should this device be kept up by another device? 5289 * Look up all dependency recorded from PM_ADD_DEPENDENT and 5290 * PM_ADD_DEPENDENT_PROPERTY ioctls. Record down on the keeper's 5291 * kept device lists. 5292 */ 5293 static int 5294 pm_kept(char *keptp) 5295 { 5296 PMD_FUNC(pmf, "kept") 5297 pm_pdr_t *dp; 5298 int found = 0; 5299 int ret = 0; 5300 dev_info_t *keeper; 5301 dev_info_t *kept; 5302 size_t length; 5303 int i; 5304 char **paths; 5305 char *path; 5306 5307 ASSERT(keptp != NULL); 5308 kept = pm_name_to_dip(keptp, 1); 5309 if (kept == NULL) 5310 return (0); 5311 PMD(PMD_KEEPS, ("%s: %s@%s(%s#%d)\n", pmf, PM_DEVICE(kept))) 5312 for (dp = pm_dep_head; dp; dp = dp->pdr_next) { 5313 if (dp->pdr_isprop) { 5314 PMD(PMD_KEEPS, ("%s: property %s\n", pmf, dp->pdr_kept)) 5315 if (ddi_prop_exists(DDI_DEV_T_ANY, kept, 5316 DDI_PROP_DONTPASS, dp->pdr_kept)) { 5317 /* 5318 * Dont allow self dependency. 5319 */ 5320 if (strcmp(dp->pdr_keeper, keptp) == 0) 5321 continue; 5322 keeper = pm_name_to_dip(dp->pdr_keeper, 1); 5323 if (keeper == NULL) 5324 continue; 5325 PMD(PMD_KEEPS, ("%s: adding to kepts path list " 5326 "%p\n", pmf, (void *)kept)) 5327 #ifdef DEBUG 5328 if (pm_debug & PMD_DEP) 5329 prdeps("Before Adding from pm_kept\n"); 5330 #endif 5331 /* 5332 * Add ourselves to the dip list. 5333 */ 5334 if (dp->pdr_kept_count == 0) { 5335 length = strlen(keptp) + 1; 5336 path = 5337 kmem_alloc(length, KM_SLEEP); 5338 paths = kmem_alloc(sizeof (char **), 5339 KM_SLEEP); 5340 (void) strcpy(path, keptp); 5341 paths[0] = path; 5342 dp->pdr_kept_paths = paths; 5343 dp->pdr_kept_count++; 5344 } else { 5345 /* Check to see if already on list */ 5346 for (i = 0; i < dp->pdr_kept_count; 5347 i++) { 5348 if (strcmp(keptp, 5349 dp->pdr_kept_paths[i]) 5350 == 0) { 5351 found++; 5352 break; 5353 } 5354 } 5355 if (found) { 5356 ddi_release_devi(keeper); 5357 continue; 5358 } 5359 length = dp->pdr_kept_count * 5360 sizeof (char **); 5361 paths = kmem_alloc( 5362 length + sizeof (char **), 5363 KM_SLEEP); 5364 if (dp->pdr_kept_count) { 5365 bcopy(dp->pdr_kept_paths, 5366 paths, length); 5367 kmem_free(dp->pdr_kept_paths, 5368 length); 5369 } 5370 dp->pdr_kept_paths = paths; 5371 length = strlen(keptp) + 1; 5372 path = 5373 kmem_alloc(length, KM_SLEEP); 5374 (void) strcpy(path, keptp); 5375 dp->pdr_kept_paths[i] = path; 5376 dp->pdr_kept_count++; 5377 } 5378 #ifdef DEBUG 5379 if (pm_debug & PMD_DEP) 5380 prdeps("After from pm_kept\n"); 5381 #endif 5382 if (keeper) { 5383 ret += pm_set_keeping(keeper, kept); 5384 ddi_release_devi(keeper); 5385 } 5386 } 5387 } else { 5388 /* 5389 * pm_keeper would be called later to do 5390 * the actual pm_set_keeping. 5391 */ 5392 PMD(PMD_KEEPS, ("%s: adding to kepts path list %p\n", 5393 pmf, (void *)kept)) 5394 #ifdef DEBUG 5395 if (pm_debug & PMD_DEP) 5396 prdeps("Before Adding from pm_kept\n"); 5397 #endif 5398 if (strcmp(keptp, dp->pdr_kept) == 0) { 5399 if (dp->pdr_kept_paths == NULL) { 5400 length = strlen(keptp) + 1; 5401 path = 5402 kmem_alloc(length, KM_SLEEP); 5403 paths = kmem_alloc(sizeof (char **), 5404 KM_SLEEP); 5405 (void) strcpy(path, keptp); 5406 paths[0] = path; 5407 dp->pdr_kept_paths = paths; 5408 dp->pdr_kept_count++; 5409 } 5410 } 5411 #ifdef DEBUG 5412 if (pm_debug & PMD_DEP) 5413 prdeps("After from pm_kept\n"); 5414 #endif 5415 } 5416 } 5417 ddi_release_devi(kept); 5418 return (ret); 5419 } 5420 5421 /* 5422 * Apply a recorded dependency. dp specifies the dependency, and 5423 * keeper is already known to be the device that keeps up the other (kept) one. 5424 * We have to the whole tree for the "kept" device, then apply 5425 * the dependency (which may already be applied). 5426 */ 5427 int 5428 pm_apply_recorded_dep(dev_info_t *keeper, pm_pdr_t *dp) 5429 { 5430 PMD_FUNC(pmf, "apply_recorded_dep") 5431 dev_info_t *kept = NULL; 5432 int ret = 0; 5433 char *keptp = NULL; 5434 5435 /* 5436 * Device to Device dependency can only be 1 to 1. 5437 */ 5438 if (dp->pdr_kept_paths == NULL) 5439 return (0); 5440 keptp = dp->pdr_kept_paths[0]; 5441 if (keptp == NULL) 5442 return (0); 5443 ASSERT(*keptp != '\0'); 5444 kept = pm_name_to_dip(keptp, 1); 5445 if (kept == NULL) 5446 return (0); 5447 if (kept) { 5448 PMD(PMD_KEEPS, ("%s: keeper=%s, kept=%s\n", pmf, 5449 dp->pdr_keeper, keptp)) 5450 if (pm_set_keeping(keeper, kept)) { 5451 ASSERT(dp->pdr_satisfied == 0); 5452 dp->pdr_satisfied = 1; 5453 ASSERT(pm_unresolved_deps); 5454 pm_unresolved_deps--; 5455 ret++; 5456 } 5457 } 5458 ddi_release_devi(kept); 5459 5460 return (ret); 5461 } 5462 5463 /* 5464 * Called from common/io/pm.c 5465 */ 5466 int 5467 pm_cur_power(pm_component_t *cp) 5468 { 5469 return (cur_power(cp)); 5470 } 5471 5472 /* 5473 * External interface to sanity-check a power level. 5474 */ 5475 int 5476 pm_valid_power(dev_info_t *dip, int comp, int level) 5477 { 5478 PMD_FUNC(pmf, "valid_power") 5479 5480 if (comp >= 0 && comp < PM_NUMCMPTS(dip) && level >= 0) 5481 return (e_pm_valid_power(dip, comp, level)); 5482 else { 5483 PMD(PMD_FAIL, ("%s: comp=%d, ncomp=%d, level=%d\n", 5484 pmf, comp, PM_NUMCMPTS(dip), level)) 5485 return (0); 5486 } 5487 } 5488 5489 /* 5490 * Called when a device that is direct power managed needs to change state. 5491 * This routine arranges to block the request until the process managing 5492 * the device makes the change (or some other incompatible change) or 5493 * the process closes /dev/pm. 5494 */ 5495 static int 5496 pm_block(dev_info_t *dip, int comp, int newpower, int oldpower) 5497 { 5498 pm_rsvp_t *new = kmem_zalloc(sizeof (*new), KM_SLEEP); 5499 int ret = 0; 5500 void pm_dequeue_blocked(pm_rsvp_t *); 5501 void pm_enqueue_blocked(pm_rsvp_t *); 5502 5503 ASSERT(!pm_processes_stopped); 5504 ASSERT(PM_IAM_LOCKING_DIP(dip)); 5505 new->pr_dip = dip; 5506 new->pr_comp = comp; 5507 new->pr_newlevel = newpower; 5508 new->pr_oldlevel = oldpower; 5509 cv_init(&new->pr_cv, NULL, CV_DEFAULT, NULL); 5510 mutex_enter(&pm_rsvp_lock); 5511 pm_enqueue_blocked(new); 5512 pm_enqueue_notify(PSC_PENDING_CHANGE, dip, comp, newpower, oldpower, 5513 PM_CANBLOCK_BLOCK); 5514 PM_UNLOCK_DIP(dip); 5515 /* 5516 * truss may make the cv_wait_sig return prematurely 5517 */ 5518 while (ret == 0) { 5519 /* 5520 * Normally there will be no user context involved, but if 5521 * there is (e.g. we are here via an ioctl call to a driver) 5522 * then we should allow the process to abort the request, 5523 * or we get an unkillable process if the same thread does 5524 * PM_DIRECT_PM and pm_raise_power 5525 */ 5526 if (cv_wait_sig(&new->pr_cv, &pm_rsvp_lock) == 0) { 5527 ret = PMP_FAIL; 5528 } else { 5529 ret = new->pr_retval; 5530 } 5531 } 5532 pm_dequeue_blocked(new); 5533 mutex_exit(&pm_rsvp_lock); 5534 cv_destroy(&new->pr_cv); 5535 kmem_free(new, sizeof (*new)); 5536 return (ret); 5537 } 5538 5539 /* 5540 * Returns true if the process is interested in power level changes (has issued 5541 * PM_GET_STATE_CHANGE ioctl). 5542 */ 5543 int 5544 pm_interest_registered(int clone) 5545 { 5546 ASSERT(clone >= 0 && clone < PM_MAX_CLONE - 1); 5547 return (pm_interest[clone]); 5548 } 5549 5550 /* 5551 * Process with clone has just done PM_DIRECT_PM on dip, or has asked to 5552 * watch all state transitions (dip == NULL). Set up data 5553 * structs to communicate with process about state changes. 5554 */ 5555 void 5556 pm_register_watcher(int clone, dev_info_t *dip) 5557 { 5558 pscc_t *p; 5559 psce_t *psce; 5560 static void pm_enqueue_pscc(pscc_t *, pscc_t **); 5561 5562 /* 5563 * We definitely need a control struct, then we have to search to see 5564 * there is already an entries struct (in the dip != NULL case). 5565 */ 5566 pscc_t *pscc = kmem_zalloc(sizeof (*pscc), KM_SLEEP); 5567 pscc->pscc_clone = clone; 5568 pscc->pscc_dip = dip; 5569 5570 if (dip) { 5571 int found = 0; 5572 rw_enter(&pm_pscc_direct_rwlock, RW_WRITER); 5573 for (p = pm_pscc_direct; p; p = p->pscc_next) { 5574 /* 5575 * Already an entry for this clone, so just use it 5576 * for the new one (for the case where a single 5577 * process is watching multiple devices) 5578 */ 5579 if (p->pscc_clone == clone) { 5580 pscc->pscc_entries = p->pscc_entries; 5581 pscc->pscc_entries->psce_references++; 5582 found++; 5583 break; 5584 } 5585 } 5586 if (!found) { /* create a new one */ 5587 psce = kmem_zalloc(sizeof (psce_t), KM_SLEEP); 5588 mutex_init(&psce->psce_lock, NULL, MUTEX_DEFAULT, NULL); 5589 psce->psce_first = 5590 kmem_zalloc(sizeof (pm_state_change_t) * PSCCOUNT, 5591 KM_SLEEP); 5592 psce->psce_in = psce->psce_out = psce->psce_first; 5593 psce->psce_last = &psce->psce_first[PSCCOUNT - 1]; 5594 psce->psce_references = 1; 5595 pscc->pscc_entries = psce; 5596 } 5597 pm_enqueue_pscc(pscc, &pm_pscc_direct); 5598 rw_exit(&pm_pscc_direct_rwlock); 5599 } else { 5600 ASSERT(!pm_interest_registered(clone)); 5601 rw_enter(&pm_pscc_interest_rwlock, RW_WRITER); 5602 #ifdef DEBUG 5603 for (p = pm_pscc_interest; p; p = p->pscc_next) { 5604 /* 5605 * Should not be an entry for this clone! 5606 */ 5607 ASSERT(p->pscc_clone != clone); 5608 } 5609 #endif 5610 psce = kmem_zalloc(sizeof (psce_t), KM_SLEEP); 5611 psce->psce_first = kmem_zalloc(sizeof (pm_state_change_t) * 5612 PSCCOUNT, KM_SLEEP); 5613 psce->psce_in = psce->psce_out = psce->psce_first; 5614 psce->psce_last = &psce->psce_first[PSCCOUNT - 1]; 5615 psce->psce_references = 1; 5616 pscc->pscc_entries = psce; 5617 pm_enqueue_pscc(pscc, &pm_pscc_interest); 5618 pm_interest[clone] = 1; 5619 rw_exit(&pm_pscc_interest_rwlock); 5620 } 5621 } 5622 5623 /* 5624 * Remove the given entry from the blocked list 5625 */ 5626 void 5627 pm_dequeue_blocked(pm_rsvp_t *p) 5628 { 5629 ASSERT(MUTEX_HELD(&pm_rsvp_lock)); 5630 if (pm_blocked_list == p) { 5631 ASSERT(p->pr_prev == NULL); 5632 if (p->pr_next != NULL) 5633 p->pr_next->pr_prev = NULL; 5634 pm_blocked_list = p->pr_next; 5635 } else { 5636 ASSERT(p->pr_prev != NULL); 5637 p->pr_prev->pr_next = p->pr_next; 5638 if (p->pr_next != NULL) 5639 p->pr_next->pr_prev = p->pr_prev; 5640 } 5641 } 5642 5643 /* 5644 * Remove the given control struct from the given list 5645 */ 5646 static void 5647 pm_dequeue_pscc(pscc_t *p, pscc_t **list) 5648 { 5649 if (*list == p) { 5650 ASSERT(p->pscc_prev == NULL); 5651 if (p->pscc_next != NULL) 5652 p->pscc_next->pscc_prev = NULL; 5653 *list = p->pscc_next; 5654 } else { 5655 ASSERT(p->pscc_prev != NULL); 5656 p->pscc_prev->pscc_next = p->pscc_next; 5657 if (p->pscc_next != NULL) 5658 p->pscc_next->pscc_prev = p->pscc_prev; 5659 } 5660 } 5661 5662 /* 5663 * Stick the control struct specified on the front of the list 5664 */ 5665 static void 5666 pm_enqueue_pscc(pscc_t *p, pscc_t **list) 5667 { 5668 pscc_t *h; /* entry at head of list */ 5669 if ((h = *list) == NULL) { 5670 *list = p; 5671 ASSERT(p->pscc_next == NULL); 5672 ASSERT(p->pscc_prev == NULL); 5673 } else { 5674 p->pscc_next = h; 5675 ASSERT(h->pscc_prev == NULL); 5676 h->pscc_prev = p; 5677 ASSERT(p->pscc_prev == NULL); 5678 *list = p; 5679 } 5680 } 5681 5682 /* 5683 * If dip is NULL, process is closing "clone" clean up all its registrations. 5684 * Otherwise only clean up those for dip because process is just giving up 5685 * control of a direct device. 5686 */ 5687 void 5688 pm_deregister_watcher(int clone, dev_info_t *dip) 5689 { 5690 pscc_t *p, *pn; 5691 psce_t *psce; 5692 int found = 0; 5693 5694 if (dip == NULL) { 5695 rw_enter(&pm_pscc_interest_rwlock, RW_WRITER); 5696 for (p = pm_pscc_interest; p; p = pn) { 5697 pn = p->pscc_next; 5698 if (p->pscc_clone == clone) { 5699 pm_dequeue_pscc(p, &pm_pscc_interest); 5700 psce = p->pscc_entries; 5701 ASSERT(psce->psce_references == 1); 5702 mutex_destroy(&psce->psce_lock); 5703 kmem_free(psce->psce_first, 5704 sizeof (pm_state_change_t) * PSCCOUNT); 5705 kmem_free(psce, sizeof (*psce)); 5706 kmem_free(p, sizeof (*p)); 5707 } 5708 } 5709 pm_interest[clone] = 0; 5710 rw_exit(&pm_pscc_interest_rwlock); 5711 } 5712 found = 0; 5713 rw_enter(&pm_pscc_direct_rwlock, RW_WRITER); 5714 for (p = pm_pscc_direct; p; p = pn) { 5715 pn = p->pscc_next; 5716 if ((dip && p->pscc_dip == dip) || 5717 (dip == NULL && clone == p->pscc_clone)) { 5718 ASSERT(clone == p->pscc_clone); 5719 found++; 5720 /* 5721 * Remove from control list 5722 */ 5723 pm_dequeue_pscc(p, &pm_pscc_direct); 5724 /* 5725 * If we're the last reference, free the 5726 * entries struct. 5727 */ 5728 psce = p->pscc_entries; 5729 ASSERT(psce); 5730 if (psce->psce_references == 1) { 5731 kmem_free(psce->psce_first, 5732 PSCCOUNT * sizeof (pm_state_change_t)); 5733 kmem_free(psce, sizeof (*psce)); 5734 } else { 5735 psce->psce_references--; 5736 } 5737 kmem_free(p, sizeof (*p)); 5738 } 5739 } 5740 ASSERT(dip == NULL || found); 5741 rw_exit(&pm_pscc_direct_rwlock); 5742 } 5743 5744 /* 5745 * Search the indicated list for an entry that matches clone, and return a 5746 * pointer to it. To be interesting, the entry must have something ready to 5747 * be passed up to the controlling process. 5748 * The returned entry will be locked upon return from this call. 5749 */ 5750 static psce_t * 5751 pm_psc_find_clone(int clone, pscc_t **list, krwlock_t *lock) 5752 { 5753 pscc_t *p; 5754 psce_t *psce; 5755 rw_enter(lock, RW_READER); 5756 for (p = *list; p; p = p->pscc_next) { 5757 if (clone == p->pscc_clone) { 5758 psce = p->pscc_entries; 5759 mutex_enter(&psce->psce_lock); 5760 if (psce->psce_out->size) { 5761 rw_exit(lock); 5762 return (psce); 5763 } else { 5764 mutex_exit(&psce->psce_lock); 5765 } 5766 } 5767 } 5768 rw_exit(lock); 5769 return (NULL); 5770 } 5771 5772 /* 5773 * Find an entry for a particular clone in the direct list. 5774 */ 5775 psce_t * 5776 pm_psc_clone_to_direct(int clone) 5777 { 5778 static psce_t *pm_psc_find_clone(int, pscc_t **, krwlock_t *); 5779 return (pm_psc_find_clone(clone, &pm_pscc_direct, 5780 &pm_pscc_direct_rwlock)); 5781 } 5782 5783 /* 5784 * Find an entry for a particular clone in the interest list. 5785 */ 5786 psce_t * 5787 pm_psc_clone_to_interest(int clone) 5788 { 5789 static psce_t *pm_psc_find_clone(int, pscc_t **, krwlock_t *); 5790 return (pm_psc_find_clone(clone, &pm_pscc_interest, 5791 &pm_pscc_interest_rwlock)); 5792 } 5793 5794 /* 5795 * Put the given entry at the head of the blocked list 5796 */ 5797 void 5798 pm_enqueue_blocked(pm_rsvp_t *p) 5799 { 5800 ASSERT(MUTEX_HELD(&pm_rsvp_lock)); 5801 ASSERT(p->pr_next == NULL); 5802 ASSERT(p->pr_prev == NULL); 5803 if (pm_blocked_list != NULL) { 5804 p->pr_next = pm_blocked_list; 5805 ASSERT(pm_blocked_list->pr_prev == NULL); 5806 pm_blocked_list->pr_prev = p; 5807 pm_blocked_list = p; 5808 } else { 5809 pm_blocked_list = p; 5810 } 5811 } 5812 5813 /* 5814 * Sets every power managed device back to its default threshold 5815 */ 5816 void 5817 pm_all_to_default_thresholds(void) 5818 { 5819 ddi_walk_devs(ddi_root_node(), pm_set_dev_thr_walk, 5820 (void *) &pm_system_idle_threshold); 5821 } 5822 5823 static int 5824 pm_set_dev_thr_walk(dev_info_t *dip, void *arg) 5825 { 5826 int thr = (int)(*(int *)arg); 5827 5828 if (!PM_GET_PM_INFO(dip)) 5829 return (DDI_WALK_CONTINUE); 5830 pm_set_device_threshold(dip, thr, PMC_DEF_THRESH); 5831 return (DDI_WALK_CONTINUE); 5832 } 5833 5834 /* 5835 * Returns the current threshold value (in seconds) for the indicated component 5836 */ 5837 int 5838 pm_current_threshold(dev_info_t *dip, int comp, int *threshp) 5839 { 5840 if (comp < 0 || comp >= PM_NUMCMPTS(dip)) { 5841 return (DDI_FAILURE); 5842 } else { 5843 *threshp = cur_threshold(dip, comp); 5844 return (DDI_SUCCESS); 5845 } 5846 } 5847 5848 /* 5849 * To be called when changing the power level of a component of a device. 5850 * On some platforms, changing power on one device may require that power 5851 * be changed on other, related devices in the same transaction. Thus, we 5852 * always pass this request to the platform power manager so that all the 5853 * affected devices will be locked. 5854 */ 5855 void 5856 pm_lock_power(dev_info_t *dip, int *circp) 5857 { 5858 power_req_t power_req; 5859 int result; 5860 5861 power_req.request_type = PMR_PPM_LOCK_POWER; 5862 power_req.req.ppm_lock_power_req.who = dip; 5863 power_req.req.ppm_lock_power_req.circp = circp; 5864 (void) pm_ctlops(PPM(dip), dip, DDI_CTLOPS_POWER, &power_req, &result); 5865 } 5866 5867 /* 5868 * Release the lock (or locks) acquired to change the power of a device. 5869 * See comments for pm_lock_power. 5870 */ 5871 void 5872 pm_unlock_power(dev_info_t *dip, int circ) 5873 { 5874 power_req_t power_req; 5875 int result; 5876 5877 power_req.request_type = PMR_PPM_UNLOCK_POWER; 5878 power_req.req.ppm_unlock_power_req.who = dip; 5879 power_req.req.ppm_unlock_power_req.circ = circ; 5880 (void) pm_ctlops(PPM(dip), dip, DDI_CTLOPS_POWER, &power_req, &result); 5881 } 5882 5883 5884 /* 5885 * Attempt (without blocking) to acquire the lock(s) needed to change the 5886 * power of a component of a device. See comments for pm_lock_power. 5887 * 5888 * Return: 1 if lock(s) acquired, 0 if not. 5889 */ 5890 int 5891 pm_try_locking_power(dev_info_t *dip, int *circp) 5892 { 5893 power_req_t power_req; 5894 int result; 5895 5896 power_req.request_type = PMR_PPM_TRY_LOCK_POWER; 5897 power_req.req.ppm_lock_power_req.who = dip; 5898 power_req.req.ppm_lock_power_req.circp = circp; 5899 (void) pm_ctlops(PPM(dip), dip, DDI_CTLOPS_POWER, &power_req, &result); 5900 return (result); 5901 } 5902 5903 5904 /* 5905 * Lock power state of a device. 5906 * 5907 * The implementation handles a special case where another thread may have 5908 * acquired the lock and created/launched this thread to do the work. If 5909 * the lock cannot be acquired immediately, we check to see if this thread 5910 * is registered as a borrower of the lock. If so, we may proceed without 5911 * the lock. This assumes that the lending thread blocks on the completion 5912 * of this thread. 5913 * 5914 * Note 1: for use by ppm only. 5915 * 5916 * Note 2: On failing to get the lock immediately, we search lock_loan list 5917 * for curthread (as borrower of the lock). On a hit, we check that the 5918 * lending thread already owns the lock we want. It is safe to compare 5919 * devi_busy_thread and thread id of the lender because in the == case (the 5920 * only one we care about) we know that the owner is blocked. Similarly, 5921 * If we find that curthread isn't registered as a lock borrower, it is safe 5922 * to use the blocking call (ndi_devi_enter) because we know that if we 5923 * weren't already listed as a borrower (upstream on the call stack) we won't 5924 * become one. 5925 */ 5926 void 5927 pm_lock_power_single(dev_info_t *dip, int *circp) 5928 { 5929 lock_loan_t *cur; 5930 5931 /* if the lock is available, we are done. */ 5932 if (ndi_devi_tryenter(dip, circp)) 5933 return; 5934 5935 mutex_enter(&pm_loan_lock); 5936 /* see if our thread is registered as a lock borrower. */ 5937 for (cur = lock_loan_head.pmlk_next; cur; cur = cur->pmlk_next) 5938 if (cur->pmlk_borrower == curthread) 5939 break; 5940 mutex_exit(&pm_loan_lock); 5941 5942 /* if this thread not already registered, it is safe to block */ 5943 if (cur == NULL) 5944 ndi_devi_enter(dip, circp); 5945 else { 5946 /* registered: does lender own the lock we want? */ 5947 if (cur->pmlk_lender == DEVI(dip)->devi_busy_thread) { 5948 ASSERT(cur->pmlk_dip == NULL || cur->pmlk_dip == dip); 5949 cur->pmlk_dip = dip; 5950 } else /* no: just block for it */ 5951 ndi_devi_enter(dip, circp); 5952 5953 } 5954 } 5955 5956 /* 5957 * Drop the lock on the device's power state. See comment for 5958 * pm_lock_power_single() for special implementation considerations. 5959 * 5960 * Note: for use by ppm only. 5961 */ 5962 void 5963 pm_unlock_power_single(dev_info_t *dip, int circ) 5964 { 5965 lock_loan_t *cur; 5966 5967 /* optimization: mutex not needed to check empty list */ 5968 if (lock_loan_head.pmlk_next == NULL) { 5969 ndi_devi_exit(dip, circ); 5970 return; 5971 } 5972 5973 mutex_enter(&pm_loan_lock); 5974 /* see if our thread is registered as a lock borrower. */ 5975 for (cur = lock_loan_head.pmlk_next; cur; cur = cur->pmlk_next) 5976 if (cur->pmlk_borrower == curthread) 5977 break; 5978 mutex_exit(&pm_loan_lock); 5979 5980 if (cur == NULL || cur->pmlk_dip != dip) 5981 /* we acquired the lock directly, so return it */ 5982 ndi_devi_exit(dip, circ); 5983 } 5984 5985 /* 5986 * Try to take the lock for changing the power level of a component. 5987 * 5988 * Note: for use by ppm only. 5989 */ 5990 int 5991 pm_try_locking_power_single(dev_info_t *dip, int *circp) 5992 { 5993 return (ndi_devi_tryenter(dip, circp)); 5994 } 5995 5996 #ifdef DEBUG 5997 /* 5998 * The following are used only to print out data structures for debugging 5999 */ 6000 void 6001 prdeps(char *msg) 6002 { 6003 6004 pm_pdr_t *rp; 6005 int i; 6006 6007 pm_log("pm_dep_head %s %p\n", msg, (void *)pm_dep_head); 6008 for (rp = pm_dep_head; rp; rp = rp->pdr_next) { 6009 pm_log("%p: %s keeper %s, kept %s, kept count %d, next %p\n", 6010 (void *)rp, (rp->pdr_isprop ? "property" : "device"), 6011 rp->pdr_keeper, rp->pdr_kept, rp->pdr_kept_count, 6012 (void *)rp->pdr_next); 6013 if (rp->pdr_kept_count != 0) { 6014 pm_log("kept list = "); 6015 i = 0; 6016 while (i < rp->pdr_kept_count) { 6017 pm_log("%s ", rp->pdr_kept_paths[i]); 6018 i++; 6019 } 6020 pm_log("\n"); 6021 } 6022 } 6023 } 6024 6025 void 6026 pr_noinvol(char *hdr) 6027 { 6028 pm_noinvol_t *ip; 6029 6030 pm_log("%s\n", hdr); 6031 rw_enter(&pm_noinvol_rwlock, RW_READER); 6032 for (ip = pm_noinvol_head; ip; ip = ip->ni_next) 6033 pm_log("\tmaj %d, flags %x, noinvolpm %d %s\n", 6034 ip->ni_major, ip->ni_flags, ip->ni_noinvolpm, ip->ni_path); 6035 rw_exit(&pm_noinvol_rwlock); 6036 } 6037 #endif 6038 6039 /* 6040 * Attempt to apply the thresholds indicated by rp to the node specified by 6041 * dip. 6042 */ 6043 void 6044 pm_apply_recorded_thresh(dev_info_t *dip, pm_thresh_rec_t *rp) 6045 { 6046 PMD_FUNC(pmf, "apply_recorded_thresh") 6047 int i, j; 6048 int comps = PM_NUMCMPTS(dip); 6049 struct pm_component *cp; 6050 pm_pte_t *ep; 6051 int pm_valid_thresh(dev_info_t *, pm_thresh_rec_t *); 6052 6053 PMD(PMD_THRESH, ("%s: part: %s@%s(%s#%d), rp %p, %s\n", pmf, 6054 PM_DEVICE(dip), (void *)rp, rp->ptr_physpath)) 6055 PM_LOCK_DIP(dip); 6056 if (!PM_GET_PM_INFO(dip) || PM_ISBC(dip) || !pm_valid_thresh(dip, rp)) { 6057 PMD(PMD_FAIL, ("%s: part: %s@%s(%s#%d) PM_GET_PM_INFO %p\n", 6058 pmf, PM_DEVICE(dip), (void*)PM_GET_PM_INFO(dip))) 6059 PMD(PMD_FAIL, ("%s: part: %s@%s(%s#%d) PM_ISBC %d\n", 6060 pmf, PM_DEVICE(dip), PM_ISBC(dip))) 6061 PMD(PMD_FAIL, ("%s: part: %s@%s(%s#%d) pm_valid_thresh %d\n", 6062 pmf, PM_DEVICE(dip), pm_valid_thresh(dip, rp))) 6063 PM_UNLOCK_DIP(dip); 6064 return; 6065 } 6066 6067 ep = rp->ptr_entries; 6068 /* 6069 * Here we do the special case of a device threshold 6070 */ 6071 if (rp->ptr_numcomps == 0) { /* PM_SET_DEVICE_THRESHOLD product */ 6072 ASSERT(ep && ep->pte_numthresh == 1); 6073 PMD(PMD_THRESH, ("%s: set dev thr %s@%s(%s#%d) to 0x%x\n", 6074 pmf, PM_DEVICE(dip), ep->pte_thresh[0])) 6075 PM_UNLOCK_DIP(dip); 6076 pm_set_device_threshold(dip, ep->pte_thresh[0], PMC_DEV_THRESH); 6077 if (PM_SCANABLE(dip)) 6078 pm_rescan(dip); 6079 return; 6080 } 6081 for (i = 0; i < comps; i++) { 6082 cp = PM_CP(dip, i); 6083 for (j = 0; j < ep->pte_numthresh; j++) { 6084 PMD(PMD_THRESH, ("%s: set thr %d for %s@%s(%s#%d)[%d] " 6085 "to %x\n", pmf, j, PM_DEVICE(dip), 6086 i, ep->pte_thresh[j])) 6087 cp->pmc_comp.pmc_thresh[j + 1] = ep->pte_thresh[j]; 6088 } 6089 ep++; 6090 } 6091 DEVI(dip)->devi_pm_flags &= PMC_THRESH_NONE; 6092 DEVI(dip)->devi_pm_flags |= PMC_COMP_THRESH; 6093 PM_UNLOCK_DIP(dip); 6094 6095 if (PM_SCANABLE(dip)) 6096 pm_rescan(dip); 6097 } 6098 6099 /* 6100 * Returns true if the threshold specified by rp could be applied to dip 6101 * (that is, the number of components and transitions are the same) 6102 */ 6103 int 6104 pm_valid_thresh(dev_info_t *dip, pm_thresh_rec_t *rp) 6105 { 6106 PMD_FUNC(pmf, "valid_thresh") 6107 int comps, i; 6108 pm_component_t *cp; 6109 pm_pte_t *ep; 6110 6111 if (!PM_GET_PM_INFO(dip) || PM_ISBC(dip)) { 6112 PMD(PMD_ERROR, ("%s: %s: no pm_info or BC\n", pmf, 6113 rp->ptr_physpath)) 6114 return (0); 6115 } 6116 /* 6117 * Special case: we represent the PM_SET_DEVICE_THRESHOLD case by 6118 * an entry with numcomps == 0, (since we don't know how many 6119 * components there are in advance). This is always a valid 6120 * spec. 6121 */ 6122 if (rp->ptr_numcomps == 0) { 6123 ASSERT(rp->ptr_entries && rp->ptr_entries->pte_numthresh == 1); 6124 return (1); 6125 } 6126 if (rp->ptr_numcomps != (comps = PM_NUMCMPTS(dip))) { 6127 PMD(PMD_ERROR, ("%s: comp # mm (dip %d cmd %d) for %s\n", 6128 pmf, PM_NUMCMPTS(dip), rp->ptr_numcomps, rp->ptr_physpath)) 6129 return (0); 6130 } 6131 ep = rp->ptr_entries; 6132 for (i = 0; i < comps; i++) { 6133 cp = PM_CP(dip, i); 6134 if ((ep + i)->pte_numthresh != 6135 cp->pmc_comp.pmc_numlevels - 1) { 6136 PMD(PMD_ERROR, ("%s: %s[%d]: thresh=%d, record=%d\n", 6137 pmf, rp->ptr_physpath, i, 6138 cp->pmc_comp.pmc_numlevels - 1, 6139 (ep + i)->pte_numthresh)) 6140 return (0); 6141 } 6142 } 6143 return (1); 6144 } 6145 6146 /* 6147 * Remove any recorded threshold for device physpath 6148 * We know there will be at most one. 6149 */ 6150 void 6151 pm_unrecord_threshold(char *physpath) 6152 { 6153 pm_thresh_rec_t *pptr, *ptr; 6154 6155 rw_enter(&pm_thresh_rwlock, RW_WRITER); 6156 for (pptr = NULL, ptr = pm_thresh_head; ptr; ptr = ptr->ptr_next) { 6157 if (strcmp(physpath, ptr->ptr_physpath) == 0) { 6158 if (pptr) { 6159 pptr->ptr_next = ptr->ptr_next; 6160 } else { 6161 ASSERT(pm_thresh_head == ptr); 6162 pm_thresh_head = ptr->ptr_next; 6163 } 6164 kmem_free(ptr, ptr->ptr_size); 6165 break; 6166 } 6167 pptr = ptr; 6168 } 6169 rw_exit(&pm_thresh_rwlock); 6170 } 6171 6172 /* 6173 * Discard all recorded thresholds. We are returning to the default pm state. 6174 */ 6175 void 6176 pm_discard_thresholds(void) 6177 { 6178 pm_thresh_rec_t *rp; 6179 rw_enter(&pm_thresh_rwlock, RW_WRITER); 6180 while (pm_thresh_head) { 6181 rp = pm_thresh_head; 6182 pm_thresh_head = rp->ptr_next; 6183 kmem_free(rp, rp->ptr_size); 6184 } 6185 rw_exit(&pm_thresh_rwlock); 6186 } 6187 6188 /* 6189 * Discard all recorded dependencies. We are returning to the default pm state. 6190 */ 6191 void 6192 pm_discard_dependencies(void) 6193 { 6194 pm_pdr_t *rp; 6195 int i; 6196 size_t length; 6197 6198 #ifdef DEBUG 6199 if (pm_debug & PMD_DEP) 6200 prdeps("Before discard\n"); 6201 #endif 6202 ddi_walk_devs(ddi_root_node(), pm_discard_dep_walk, NULL); 6203 6204 #ifdef DEBUG 6205 if (pm_debug & PMD_DEP) 6206 prdeps("After discard\n"); 6207 #endif 6208 while (pm_dep_head) { 6209 rp = pm_dep_head; 6210 if (!rp->pdr_isprop) { 6211 ASSERT(rp->pdr_satisfied == 0); 6212 ASSERT(pm_unresolved_deps); 6213 pm_unresolved_deps--; 6214 } else { 6215 ASSERT(pm_prop_deps); 6216 pm_prop_deps--; 6217 } 6218 pm_dep_head = rp->pdr_next; 6219 if (rp->pdr_kept_count) { 6220 for (i = 0; i < rp->pdr_kept_count; i++) { 6221 length = strlen(rp->pdr_kept_paths[i]) + 1; 6222 kmem_free(rp->pdr_kept_paths[i], length); 6223 } 6224 kmem_free(rp->pdr_kept_paths, 6225 rp->pdr_kept_count * sizeof (char **)); 6226 } 6227 kmem_free(rp, rp->pdr_size); 6228 } 6229 } 6230 6231 6232 static int 6233 pm_discard_dep_walk(dev_info_t *dip, void *arg) 6234 { 6235 _NOTE(ARGUNUSED(arg)) 6236 char *pathbuf; 6237 6238 if (PM_GET_PM_INFO(dip) == NULL) 6239 return (DDI_WALK_CONTINUE); 6240 pathbuf = kmem_alloc(MAXPATHLEN, KM_SLEEP); 6241 (void) ddi_pathname(dip, pathbuf); 6242 pm_free_keeper(pathbuf, 0); 6243 kmem_free(pathbuf, MAXPATHLEN); 6244 return (DDI_WALK_CONTINUE); 6245 } 6246 6247 static int 6248 pm_kept_walk(dev_info_t *dip, void *arg) 6249 { 6250 _NOTE(ARGUNUSED(arg)) 6251 char *pathbuf; 6252 6253 pathbuf = kmem_alloc(MAXPATHLEN, KM_SLEEP); 6254 (void) ddi_pathname(dip, pathbuf); 6255 (void) pm_kept(pathbuf); 6256 kmem_free(pathbuf, MAXPATHLEN); 6257 6258 return (DDI_WALK_CONTINUE); 6259 } 6260 6261 static int 6262 pm_keeper_walk(dev_info_t *dip, void *arg) 6263 { 6264 _NOTE(ARGUNUSED(arg)) 6265 char *pathbuf; 6266 6267 pathbuf = kmem_alloc(MAXPATHLEN, KM_SLEEP); 6268 (void) ddi_pathname(dip, pathbuf); 6269 (void) pm_keeper(pathbuf); 6270 kmem_free(pathbuf, MAXPATHLEN); 6271 6272 return (DDI_WALK_CONTINUE); 6273 } 6274 6275 static char * 6276 pdw_type_decode(int type) 6277 { 6278 switch (type) { 6279 case PM_DEP_WK_POWER_ON: 6280 return ("power on"); 6281 case PM_DEP_WK_POWER_OFF: 6282 return ("power off"); 6283 case PM_DEP_WK_DETACH: 6284 return ("detach"); 6285 case PM_DEP_WK_REMOVE_DEP: 6286 return ("remove dep"); 6287 case PM_DEP_WK_BRINGUP_SELF: 6288 return ("bringup self"); 6289 case PM_DEP_WK_RECORD_KEEPER: 6290 return ("add dependent"); 6291 case PM_DEP_WK_RECORD_KEEPER_PROP: 6292 return ("add dependent property"); 6293 case PM_DEP_WK_KEPT: 6294 return ("kept"); 6295 case PM_DEP_WK_KEEPER: 6296 return ("keeper"); 6297 case PM_DEP_WK_ATTACH: 6298 return ("attach"); 6299 case PM_DEP_WK_CHECK_KEPT: 6300 return ("check kept"); 6301 case PM_DEP_WK_CPR_SUSPEND: 6302 return ("suspend"); 6303 case PM_DEP_WK_CPR_RESUME: 6304 return ("resume"); 6305 default: 6306 return ("unknown"); 6307 } 6308 6309 } 6310 6311 static void 6312 pm_rele_dep(char *keeper) 6313 { 6314 PMD_FUNC(pmf, "rele_dep") 6315 pm_pdr_t *dp; 6316 char *kept_path = NULL; 6317 dev_info_t *kept = NULL; 6318 int count = 0; 6319 6320 for (dp = pm_dep_head; dp; dp = dp->pdr_next) { 6321 if (strcmp(dp->pdr_keeper, keeper) != 0) 6322 continue; 6323 for (count = 0; count < dp->pdr_kept_count; count++) { 6324 kept_path = dp->pdr_kept_paths[count]; 6325 if (kept_path == NULL) 6326 continue; 6327 kept = pm_name_to_dip(kept_path, 1); 6328 if (kept) { 6329 PMD(PMD_KEEPS, ("%s: release kept=%s@%s(%s#%d) " 6330 "of keeper=%s\n", pmf, PM_DEVICE(kept), 6331 keeper)) 6332 ASSERT(DEVI(kept)->devi_pm_kidsupcnt > 0); 6333 pm_rele_power(kept); 6334 ddi_release_devi(kept); 6335 } 6336 } 6337 } 6338 } 6339 6340 /* 6341 * Called when we are just released from direct PM. Bring ourself up 6342 * if our keeper is up since dependency is not honored while a kept 6343 * device is under direct PM. 6344 */ 6345 static void 6346 pm_bring_self_up(char *keptpath) 6347 { 6348 PMD_FUNC(pmf, "bring_self_up") 6349 dev_info_t *kept; 6350 dev_info_t *keeper; 6351 pm_pdr_t *dp; 6352 int i, j; 6353 int up = 0, circ; 6354 6355 kept = pm_name_to_dip(keptpath, 1); 6356 if (kept == NULL) 6357 return; 6358 PMD(PMD_KEEPS, ("%s: kept=%s@%s(%s#%d)\n", pmf, PM_DEVICE(kept))) 6359 for (dp = pm_dep_head; dp; dp = dp->pdr_next) { 6360 if (dp->pdr_kept_count == 0) 6361 continue; 6362 for (i = 0; i < dp->pdr_kept_count; i++) { 6363 if (strcmp(dp->pdr_kept_paths[i], keptpath) != 0) 6364 continue; 6365 keeper = pm_name_to_dip(dp->pdr_keeper, 1); 6366 if (keeper) { 6367 PMD(PMD_KEEPS, ("%s: keeper=%s@%s(%s#%d)\n", 6368 pmf, PM_DEVICE(keeper))) 6369 PM_LOCK_POWER(keeper, &circ); 6370 for (j = 0; j < PM_NUMCMPTS(keeper); 6371 j++) { 6372 if (PM_CURPOWER(keeper, j)) { 6373 PMD(PMD_KEEPS, ("%s: comp=" 6374 "%d is up\n", pmf, j)) 6375 up++; 6376 } 6377 } 6378 if (up) { 6379 if (PM_SKBU(kept)) 6380 DEVI(kept)->devi_pm_flags &= 6381 ~PMC_SKIP_BRINGUP; 6382 bring_pmdep_up(kept, 1); 6383 } 6384 PM_UNLOCK_POWER(keeper, circ); 6385 ddi_release_devi(keeper); 6386 } 6387 } 6388 } 6389 ddi_release_devi(kept); 6390 } 6391 6392 static void 6393 pm_process_dep_request(pm_dep_wk_t *work) 6394 { 6395 PMD_FUNC(pmf, "dep_req") 6396 int ret; 6397 6398 PMD(PMD_DEP, ("%s: work=%s\n", pmf, 6399 pdw_type_decode(work->pdw_type))) 6400 PMD(PMD_DEP, ("%s: keeper=%s, kept=%s\n", pmf, 6401 (work->pdw_keeper ? work->pdw_keeper : "NULL"), 6402 (work->pdw_kept ? work->pdw_kept : "NULL"))) 6403 6404 switch (work->pdw_type) { 6405 case PM_DEP_WK_POWER_ON: 6406 /* Bring up the kept devices and put a hold on them */ 6407 bring_wekeeps_up(work->pdw_keeper); 6408 break; 6409 case PM_DEP_WK_POWER_OFF: 6410 /* Release the kept devices */ 6411 pm_rele_dep(work->pdw_keeper); 6412 break; 6413 case PM_DEP_WK_DETACH: 6414 pm_free_keeps(work->pdw_keeper, work->pdw_pwr); 6415 break; 6416 case PM_DEP_WK_REMOVE_DEP: 6417 pm_discard_dependencies(); 6418 break; 6419 case PM_DEP_WK_BRINGUP_SELF: 6420 /* 6421 * We deferred satisfying our dependency till now, so satisfy 6422 * it again and bring ourselves up. 6423 */ 6424 pm_bring_self_up(work->pdw_kept); 6425 break; 6426 case PM_DEP_WK_RECORD_KEEPER: 6427 (void) pm_record_keeper(work->pdw_kept, work->pdw_keeper, 0); 6428 ddi_walk_devs(ddi_root_node(), pm_kept_walk, NULL); 6429 ddi_walk_devs(ddi_root_node(), pm_keeper_walk, NULL); 6430 break; 6431 case PM_DEP_WK_RECORD_KEEPER_PROP: 6432 (void) pm_record_keeper(work->pdw_kept, work->pdw_keeper, 1); 6433 ddi_walk_devs(ddi_root_node(), pm_keeper_walk, NULL); 6434 ddi_walk_devs(ddi_root_node(), pm_kept_walk, NULL); 6435 break; 6436 case PM_DEP_WK_KEPT: 6437 ret = pm_kept(work->pdw_kept); 6438 PMD(PMD_DEP, ("%s: PM_DEP_WK_KEPT: pm_kept returns %d\n", pmf, 6439 ret)) 6440 break; 6441 case PM_DEP_WK_KEEPER: 6442 ret = pm_keeper(work->pdw_keeper); 6443 PMD(PMD_DEP, ("%s: PM_DEP_WK_KEEPER: pm_keeper returns %d\n", 6444 pmf, ret)) 6445 break; 6446 case PM_DEP_WK_ATTACH: 6447 ret = pm_keeper(work->pdw_keeper); 6448 PMD(PMD_DEP, ("%s: PM_DEP_WK_ATTACH: pm_keeper returns %d\n", 6449 pmf, ret)) 6450 ret = pm_kept(work->pdw_kept); 6451 PMD(PMD_DEP, ("%s: PM_DEP_WK_ATTACH: pm_kept returns %d\n", 6452 pmf, ret)) 6453 break; 6454 case PM_DEP_WK_CHECK_KEPT: 6455 ret = pm_is_kept(work->pdw_kept); 6456 PMD(PMD_DEP, ("%s: PM_DEP_WK_CHECK_KEPT: kept=%s, ret=%d\n", 6457 pmf, work->pdw_kept, ret)) 6458 break; 6459 case PM_DEP_WK_CPR_SUSPEND: 6460 pm_discard_dependencies(); 6461 break; 6462 case PM_DEP_WK_CPR_RESUME: 6463 ddi_walk_devs(ddi_root_node(), pm_kept_walk, NULL); 6464 ddi_walk_devs(ddi_root_node(), pm_keeper_walk, NULL); 6465 break; 6466 default: 6467 ASSERT(0); 6468 break; 6469 } 6470 /* 6471 * Free the work structure if the requester is not waiting 6472 * Otherwise it is the requester's responsiblity to free it. 6473 */ 6474 if (!work->pdw_wait) { 6475 if (work->pdw_keeper) 6476 kmem_free(work->pdw_keeper, 6477 strlen(work->pdw_keeper) + 1); 6478 if (work->pdw_kept) 6479 kmem_free(work->pdw_kept, strlen(work->pdw_kept) + 1); 6480 kmem_free(work, sizeof (pm_dep_wk_t)); 6481 } else { 6482 /* 6483 * Notify requester if it is waiting for it. 6484 */ 6485 work->pdw_ret = ret; 6486 work->pdw_done = 1; 6487 cv_signal(&work->pdw_cv); 6488 } 6489 } 6490 6491 /* 6492 * Process PM dependency requests. 6493 */ 6494 static void 6495 pm_dep_thread(void) 6496 { 6497 pm_dep_wk_t *work; 6498 callb_cpr_t cprinfo; 6499 6500 CALLB_CPR_INIT(&cprinfo, &pm_dep_thread_lock, callb_generic_cpr, 6501 "pm_dep_thread"); 6502 for (;;) { 6503 mutex_enter(&pm_dep_thread_lock); 6504 if (pm_dep_thread_workq == NULL) { 6505 CALLB_CPR_SAFE_BEGIN(&cprinfo); 6506 cv_wait(&pm_dep_thread_cv, &pm_dep_thread_lock); 6507 CALLB_CPR_SAFE_END(&cprinfo, &pm_dep_thread_lock); 6508 } 6509 work = pm_dep_thread_workq; 6510 pm_dep_thread_workq = work->pdw_next; 6511 if (pm_dep_thread_tail == work) 6512 pm_dep_thread_tail = work->pdw_next; 6513 mutex_exit(&pm_dep_thread_lock); 6514 pm_process_dep_request(work); 6515 6516 } 6517 /*NOTREACHED*/ 6518 } 6519 6520 /* 6521 * Set the power level of the indicated device to unknown (if it is not a 6522 * backwards compatible device), as it has just been resumed, and it won't 6523 * know if the power was removed or not. Adjust parent's kidsupcnt if necessary. 6524 */ 6525 void 6526 pm_forget_power_level(dev_info_t *dip) 6527 { 6528 dev_info_t *pdip = ddi_get_parent(dip); 6529 int i, count = 0; 6530 6531 if (!PM_ISBC(dip)) { 6532 for (i = 0; i < PM_NUMCMPTS(dip); i++) 6533 count += (PM_CURPOWER(dip, i) == 0); 6534 6535 if (count && pdip && !PM_WANTS_NOTIFICATION(pdip)) 6536 e_pm_hold_rele_power(pdip, count); 6537 6538 /* 6539 * Count this as a power cycle if we care 6540 */ 6541 if (DEVI(dip)->devi_pm_volpmd && 6542 PM_CP(dip, 0)->pmc_cur_pwr == 0) 6543 DEVI(dip)->devi_pm_volpmd = 0; 6544 for (i = 0; i < PM_NUMCMPTS(dip); i++) 6545 e_pm_set_cur_pwr(dip, PM_CP(dip, i), PM_LEVEL_UNKNOWN); 6546 } 6547 } 6548 6549 /* 6550 * This function advises the caller whether it should make a power-off 6551 * transition at this time or not. If the transition is not advised 6552 * at this time, the time that the next power-off transition can 6553 * be made from now is returned through "intervalp" pointer. 6554 * This function returns: 6555 * 6556 * 1 power-off advised 6557 * 0 power-off not advised, intervalp will point to seconds from 6558 * now that a power-off is advised. If it is passed the number 6559 * of years that policy specifies the device should last, 6560 * a large number is returned as the time interval. 6561 * -1 error 6562 */ 6563 int 6564 pm_trans_check(struct pm_trans_data *datap, time_t *intervalp) 6565 { 6566 PMD_FUNC(pmf, "pm_trans_check") 6567 char dbuf[DC_SCSI_MFR_LEN]; 6568 struct pm_scsi_cycles *scp; 6569 int service_years, service_weeks, full_years; 6570 time_t now, service_seconds, tdiff; 6571 time_t within_year, when_allowed; 6572 char *ptr; 6573 int lower_bound_cycles, upper_bound_cycles, cycles_allowed; 6574 int cycles_diff, cycles_over; 6575 6576 if (datap == NULL) { 6577 PMD(PMD_TCHECK, ("%s: NULL data pointer!\n", pmf)) 6578 return (-1); 6579 } 6580 6581 if (datap->format == DC_SCSI_FORMAT) { 6582 /* 6583 * Power cycles of the scsi drives are distributed 6584 * over 5 years with the following percentage ratio: 6585 * 6586 * 30%, 25%, 20%, 15%, and 10% 6587 * 6588 * The power cycle quota for each year is distributed 6589 * linearly through out the year. The equation for 6590 * determining the expected cycles is: 6591 * 6592 * e = a * (n / y) 6593 * 6594 * e = expected cycles 6595 * a = allocated cycles for this year 6596 * n = number of seconds since beginning of this year 6597 * y = number of seconds in a year 6598 * 6599 * Note that beginning of the year starts the day that 6600 * the drive has been put on service. 6601 * 6602 * If the drive has passed its expected cycles, we 6603 * can determine when it can start to power cycle 6604 * again to keep it on track to meet the 5-year 6605 * life expectancy. The equation for determining 6606 * when to power cycle is: 6607 * 6608 * w = y * (c / a) 6609 * 6610 * w = when it can power cycle again 6611 * y = number of seconds in a year 6612 * c = current number of cycles 6613 * a = allocated cycles for the year 6614 * 6615 */ 6616 char pcnt[DC_SCSI_NPY] = { 30, 55, 75, 90, 100 }; 6617 6618 scp = &datap->un.scsi_cycles; 6619 PMD(PMD_TCHECK, ("%s: format=%d, lifemax=%d, ncycles=%d, " 6620 "svc_date=%s, svc_flag=%d\n", pmf, datap->format, 6621 scp->lifemax, scp->ncycles, scp->svc_date, scp->flag)) 6622 if (scp->ncycles < 0 || scp->flag != 0) { 6623 PMD(PMD_TCHECK, ("%s: ncycles < 0 || flag != 0\n", pmf)) 6624 return (-1); 6625 } 6626 6627 if (scp->ncycles > scp->lifemax) { 6628 *intervalp = (LONG_MAX / hz); 6629 return (0); 6630 } 6631 6632 /* 6633 * convert service date to time_t 6634 */ 6635 bcopy(scp->svc_date, dbuf, DC_SCSI_YEAR_LEN); 6636 dbuf[DC_SCSI_YEAR_LEN] = '\0'; 6637 ptr = dbuf; 6638 service_years = stoi(&ptr) - EPOCH_YEAR; 6639 bcopy(&scp->svc_date[DC_SCSI_YEAR_LEN], dbuf, 6640 DC_SCSI_WEEK_LEN); 6641 dbuf[DC_SCSI_WEEK_LEN] = '\0'; 6642 6643 /* 6644 * scsi standard does not specify WW data, 6645 * could be (00-51) or (01-52) 6646 */ 6647 ptr = dbuf; 6648 service_weeks = stoi(&ptr); 6649 if (service_years < 0 || 6650 service_weeks < 0 || service_weeks > 52) { 6651 PMD(PMD_TCHECK, ("%s: service year %d and week %d\n", 6652 pmf, service_years, service_weeks)) 6653 return (-1); 6654 } 6655 6656 /* 6657 * calculate service date in seconds-since-epoch, 6658 * adding one day for each leap-year. 6659 * 6660 * (years-since-epoch + 2) fixes integer truncation, 6661 * example: (8) leap-years during [1972, 2000] 6662 * (2000 - 1970) = 30; and (30 + 2) / 4 = 8; 6663 */ 6664 service_seconds = (service_years * DC_SPY) + 6665 (service_weeks * DC_SPW) + 6666 (((service_years + 2) / 4) * DC_SPD); 6667 6668 now = gethrestime_sec(); 6669 /* 6670 * since the granularity of 'svc_date' is day not second, 6671 * 'now' should be rounded up to full day. 6672 */ 6673 now = ((now + DC_SPD -1) / DC_SPD) * DC_SPD; 6674 if (service_seconds > now) { 6675 PMD(PMD_TCHECK, ("%s: service date (%ld) later " 6676 "than now (%ld)!\n", pmf, service_seconds, now)) 6677 return (-1); 6678 } 6679 6680 tdiff = now - service_seconds; 6681 PMD(PMD_TCHECK, ("%s: age is %ld sec\n", pmf, tdiff)) 6682 6683 /* 6684 * NOTE - Leap years are not considered in the calculations 6685 * below. 6686 */ 6687 full_years = (tdiff / DC_SPY); 6688 if ((full_years >= DC_SCSI_NPY) && 6689 (scp->ncycles <= scp->lifemax)) 6690 return (1); 6691 6692 /* 6693 * Determine what is the normal cycle usage for the 6694 * device at the beginning and the end of this year. 6695 */ 6696 lower_bound_cycles = (!full_years) ? 0 : 6697 ((scp->lifemax * pcnt[full_years - 1]) / 100); 6698 upper_bound_cycles = (scp->lifemax * pcnt[full_years]) / 100; 6699 6700 if (scp->ncycles <= lower_bound_cycles) 6701 return (1); 6702 6703 /* 6704 * The linear slope that determines how many cycles 6705 * are allowed this year is number of seconds 6706 * passed this year over total number of seconds in a year. 6707 */ 6708 cycles_diff = (upper_bound_cycles - lower_bound_cycles); 6709 within_year = (tdiff % DC_SPY); 6710 cycles_allowed = lower_bound_cycles + 6711 (((uint64_t)cycles_diff * (uint64_t)within_year) / DC_SPY); 6712 PMD(PMD_TCHECK, ("%s: lived %d yrs and %ld secs\n", pmf, 6713 full_years, within_year)) 6714 PMD(PMD_TCHECK, ("%s: # of cycles allowed %d\n", pmf, 6715 cycles_allowed)) 6716 6717 if (scp->ncycles <= cycles_allowed) 6718 return (1); 6719 6720 /* 6721 * The transition is not advised now but we can 6722 * determine when the next transition can be made. 6723 * 6724 * Depending on how many cycles the device has been 6725 * over-used, we may need to skip years with 6726 * different percentage quota in order to determine 6727 * when the next transition can be made. 6728 */ 6729 cycles_over = (scp->ncycles - lower_bound_cycles); 6730 while (cycles_over > cycles_diff) { 6731 full_years++; 6732 if (full_years >= DC_SCSI_NPY) { 6733 *intervalp = (LONG_MAX / hz); 6734 return (0); 6735 } 6736 cycles_over -= cycles_diff; 6737 lower_bound_cycles = upper_bound_cycles; 6738 upper_bound_cycles = 6739 (scp->lifemax * pcnt[full_years]) / 100; 6740 cycles_diff = (upper_bound_cycles - lower_bound_cycles); 6741 } 6742 6743 /* 6744 * The linear slope that determines when the next transition 6745 * can be made is the relative position of used cycles within a 6746 * year over total number of cycles within that year. 6747 */ 6748 when_allowed = service_seconds + (full_years * DC_SPY) + 6749 (((uint64_t)DC_SPY * (uint64_t)cycles_over) / cycles_diff); 6750 *intervalp = (when_allowed - now); 6751 if (*intervalp > (LONG_MAX / hz)) 6752 *intervalp = (LONG_MAX / hz); 6753 PMD(PMD_TCHECK, ("%s: no cycle is allowed in %ld secs\n", pmf, 6754 *intervalp)) 6755 return (0); 6756 } 6757 6758 PMD(PMD_TCHECK, ("%s: unknown format!\n", pmf)) 6759 return (-1); 6760 } 6761 6762 /* 6763 * Nexus drivers call into pm framework to indicate which child driver is about 6764 * to be installed. In some platforms, ppm may need to configure the hardware 6765 * for successful installation of a driver. 6766 */ 6767 int 6768 pm_init_child(dev_info_t *dip) 6769 { 6770 power_req_t power_req; 6771 6772 ASSERT(ddi_binding_name(dip)); 6773 ASSERT(ddi_get_name_addr(dip)); 6774 pm_ppm_claim(dip); 6775 if (pm_ppm_claimed(dip)) { /* if ppm driver claims the node */ 6776 power_req.request_type = PMR_PPM_INIT_CHILD; 6777 power_req.req.ppm_config_req.who = dip; 6778 ASSERT(PPM(dip) != NULL); 6779 return (pm_ctlops(PPM(dip), dip, DDI_CTLOPS_POWER, &power_req, 6780 NULL)); 6781 } else { 6782 #ifdef DEBUG 6783 /* pass it to the default handler so we can debug things */ 6784 power_req.request_type = PMR_PPM_INIT_CHILD; 6785 power_req.req.ppm_config_req.who = dip; 6786 (void) pm_ctlops(NULL, dip, 6787 DDI_CTLOPS_POWER, &power_req, NULL); 6788 #endif 6789 } 6790 return (DDI_SUCCESS); 6791 } 6792 6793 /* 6794 * Bring parent of a node that is about to be probed up to full power, and 6795 * arrange for it to stay up until pm_post_probe() or pm_post_attach() decide 6796 * it is time to let it go down again 6797 */ 6798 void 6799 pm_pre_probe(dev_info_t *dip, pm_ppm_cookie_t *cp) 6800 { 6801 int result; 6802 power_req_t power_req; 6803 6804 bzero(cp, sizeof (*cp)); 6805 cp->ppc_dip = dip; 6806 6807 pm_ppm_claim(dip); 6808 if (pm_ppm_claimed(dip)) { /* if ppm driver claims the node */ 6809 power_req.request_type = PMR_PPM_PRE_PROBE; 6810 power_req.req.ppm_config_req.who = dip; 6811 ASSERT(PPM(dip) != NULL); 6812 (void) pm_ctlops(PPM(dip), dip, 6813 DDI_CTLOPS_POWER, &power_req, &result); 6814 cp->ppc_ppm = PPM(dip); 6815 } else { 6816 #ifdef DEBUG 6817 /* pass it to the default handler so we can debug things */ 6818 power_req.request_type = PMR_PPM_PRE_PROBE; 6819 power_req.req.ppm_config_req.who = dip; 6820 (void) pm_ctlops(NULL, dip, 6821 DDI_CTLOPS_POWER, &power_req, &result); 6822 #endif 6823 cp->ppc_ppm = NULL; 6824 } 6825 } 6826 6827 int 6828 pm_pre_config(dev_info_t *dip, char *devnm) 6829 { 6830 PMD_FUNC(pmf, "pre_config") 6831 int ret; 6832 6833 if (MDI_VHCI(dip)) { 6834 PMD(PMD_SET, ("%s: %s@%s(%s#%d)\n", pmf, PM_DEVICE(dip))) 6835 ret = mdi_power(dip, MDI_PM_PRE_CONFIG, NULL, devnm, 0); 6836 return (ret == MDI_SUCCESS ? DDI_SUCCESS : DDI_FAILURE); 6837 } else if (!PM_GET_PM_INFO(dip)) 6838 return (DDI_SUCCESS); 6839 6840 PMD(PMD_SET, ("%s: %s@%s(%s#%d)\n", pmf, PM_DEVICE(dip))) 6841 pm_hold_power(dip); 6842 ret = pm_all_to_normal(dip, PM_CANBLOCK_BLOCK); 6843 if (ret != DDI_SUCCESS) 6844 pm_rele_power(dip); 6845 return (ret); 6846 } 6847 6848 /* 6849 * This routine is called by devfs during its walk to unconfigue a node. 6850 * If the call is due to auto mod_unloads and the dip is not at its 6851 * full power, we return DDI_FAILURE to terminate the walk, otherwise 6852 * return DDI_SUCCESS. 6853 */ 6854 int 6855 pm_pre_unconfig(dev_info_t *dip, int flags, int *held, char *devnm) 6856 { 6857 PMD_FUNC(pmf, "pre_unconfig") 6858 int ret; 6859 6860 if (MDI_VHCI(dip)) { 6861 PMD(PMD_SET, ("%s: %s@%s(%s#%d), flags=%x\n", pmf, 6862 PM_DEVICE(dip), flags)) 6863 ret = mdi_power(dip, MDI_PM_PRE_UNCONFIG, held, devnm, flags); 6864 return (ret == MDI_SUCCESS ? DDI_SUCCESS : DDI_FAILURE); 6865 } else if (!PM_GET_PM_INFO(dip)) 6866 return (DDI_SUCCESS); 6867 6868 PMD(PMD_SET, ("%s: %s@%s(%s#%d), flags=%x\n", pmf, PM_DEVICE(dip), 6869 flags)) 6870 *held = 0; 6871 6872 /* 6873 * If the dip is a leaf node, don't power it up. 6874 */ 6875 if (!ddi_get_child(dip)) 6876 return (DDI_SUCCESS); 6877 6878 /* 6879 * Do not power up the node if it is called due to auto-modunload. 6880 */ 6881 if ((flags & NDI_AUTODETACH) && !pm_all_at_normal(dip)) 6882 return (DDI_FAILURE); 6883 6884 pm_hold_power(dip); 6885 *held = 1; 6886 ret = pm_all_to_normal(dip, PM_CANBLOCK_BLOCK); 6887 if (ret != DDI_SUCCESS) { 6888 pm_rele_power(dip); 6889 *held = 0; 6890 } 6891 return (ret); 6892 } 6893 6894 /* 6895 * Notify ppm of attach action. Parent is already held at full power by 6896 * probe action. 6897 */ 6898 void 6899 pm_pre_attach(dev_info_t *dip, pm_ppm_cookie_t *cp, ddi_attach_cmd_t cmd) 6900 { 6901 static char *me = "pm_pre_attach"; 6902 power_req_t power_req; 6903 int result; 6904 6905 /* 6906 * Initialize and fill in the PPM cookie 6907 */ 6908 bzero(cp, sizeof (*cp)); 6909 cp->ppc_cmd = (int)cmd; 6910 cp->ppc_ppm = PPM(dip); 6911 cp->ppc_dip = dip; 6912 6913 /* 6914 * DDI_ATTACH and DDI_RESUME cmds need to call platform specific 6915 * Power Management stuff. DDI_RESUME also has to purge it's 6916 * powerlevel information. 6917 */ 6918 switch (cmd) { 6919 case DDI_ATTACH: 6920 if (cp->ppc_ppm) { /* if ppm driver claims the node */ 6921 power_req.request_type = PMR_PPM_PRE_ATTACH; 6922 power_req.req.ppm_config_req.who = dip; 6923 ASSERT(PPM(dip)); 6924 (void) pm_ctlops(cp->ppc_ppm, dip, DDI_CTLOPS_POWER, 6925 &power_req, &result); 6926 } 6927 #ifdef DEBUG 6928 else { 6929 power_req.request_type = PMR_PPM_PRE_ATTACH; 6930 power_req.req.ppm_config_req.who = dip; 6931 (void) pm_ctlops(NULL, dip, 6932 DDI_CTLOPS_POWER, &power_req, &result); 6933 } 6934 #endif 6935 break; 6936 case DDI_RESUME: 6937 pm_forget_power_level(dip); 6938 6939 if (cp->ppc_ppm) { /* if ppm driver claims the node */ 6940 power_req.request_type = PMR_PPM_PRE_RESUME; 6941 power_req.req.resume_req.who = cp->ppc_dip; 6942 power_req.req.resume_req.cmd = 6943 (ddi_attach_cmd_t)cp->ppc_cmd; 6944 ASSERT(PPM(cp->ppc_dip) == cp->ppc_ppm); 6945 (void) pm_ctlops(cp->ppc_ppm, cp->ppc_dip, 6946 DDI_CTLOPS_POWER, &power_req, &result); 6947 } 6948 #ifdef DEBUG 6949 else { 6950 power_req.request_type = PMR_PPM_PRE_RESUME; 6951 power_req.req.resume_req.who = cp->ppc_dip; 6952 power_req.req.resume_req.cmd = 6953 (ddi_attach_cmd_t)cp->ppc_cmd; 6954 (void) pm_ctlops(NULL, cp->ppc_dip, 6955 DDI_CTLOPS_POWER, &power_req, &result); 6956 } 6957 #endif 6958 break; 6959 6960 case DDI_PM_RESUME: 6961 break; 6962 6963 default: 6964 panic(me); 6965 } 6966 } 6967 6968 /* 6969 * Nexus drivers call into pm framework to indicate which child driver is 6970 * being uninstalled. In some platforms, ppm may need to reconfigure the 6971 * hardware since the device driver is no longer installed. 6972 */ 6973 int 6974 pm_uninit_child(dev_info_t *dip) 6975 { 6976 power_req_t power_req; 6977 6978 ASSERT(ddi_binding_name(dip)); 6979 ASSERT(ddi_get_name_addr(dip)); 6980 pm_ppm_claim(dip); 6981 if (pm_ppm_claimed(dip)) { /* if ppm driver claims the node */ 6982 power_req.request_type = PMR_PPM_UNINIT_CHILD; 6983 power_req.req.ppm_config_req.who = dip; 6984 ASSERT(PPM(dip)); 6985 return (pm_ctlops(PPM(dip), dip, DDI_CTLOPS_POWER, &power_req, 6986 NULL)); 6987 } else { 6988 #ifdef DEBUG 6989 /* pass it to the default handler so we can debug things */ 6990 power_req.request_type = PMR_PPM_UNINIT_CHILD; 6991 power_req.req.ppm_config_req.who = dip; 6992 (void) pm_ctlops(NULL, dip, DDI_CTLOPS_POWER, &power_req, NULL); 6993 #endif 6994 } 6995 return (DDI_SUCCESS); 6996 } 6997 /* 6998 * Decrement kidsupcnt so scan can turn the parent back off if it is idle 6999 * Also notify ppm of result of probe if there is a ppm that cares 7000 */ 7001 void 7002 pm_post_probe(pm_ppm_cookie_t *cp, int ret, int probe_failed) 7003 { 7004 _NOTE(ARGUNUSED(probe_failed)) 7005 int result; 7006 power_req_t power_req; 7007 7008 if (cp->ppc_ppm) { /* if ppm driver claims the node */ 7009 power_req.request_type = PMR_PPM_POST_PROBE; 7010 power_req.req.ppm_config_req.who = cp->ppc_dip; 7011 power_req.req.ppm_config_req.result = ret; 7012 ASSERT(PPM(cp->ppc_dip) == cp->ppc_ppm); 7013 (void) pm_ctlops(cp->ppc_ppm, cp->ppc_dip, DDI_CTLOPS_POWER, 7014 &power_req, &result); 7015 } 7016 #ifdef DEBUG 7017 else { 7018 power_req.request_type = PMR_PPM_POST_PROBE; 7019 power_req.req.ppm_config_req.who = cp->ppc_dip; 7020 power_req.req.ppm_config_req.result = ret; 7021 (void) pm_ctlops(NULL, cp->ppc_dip, DDI_CTLOPS_POWER, 7022 &power_req, &result); 7023 } 7024 #endif 7025 } 7026 7027 void 7028 pm_post_config(dev_info_t *dip, char *devnm) 7029 { 7030 PMD_FUNC(pmf, "post_config") 7031 7032 if (MDI_VHCI(dip)) { 7033 PMD(PMD_SET, ("%s: %s@%s(%s#%d)\n", pmf, PM_DEVICE(dip))) 7034 (void) mdi_power(dip, MDI_PM_POST_CONFIG, NULL, devnm, 0); 7035 return; 7036 } else if (!PM_GET_PM_INFO(dip)) 7037 return; 7038 7039 PMD(PMD_SET, ("%s: %s@%s(%s#%d)\n", pmf, PM_DEVICE(dip))) 7040 pm_rele_power(dip); 7041 } 7042 7043 void 7044 pm_post_unconfig(dev_info_t *dip, int held, char *devnm) 7045 { 7046 PMD_FUNC(pmf, "post_unconfig") 7047 7048 if (MDI_VHCI(dip)) { 7049 PMD(PMD_SET, ("%s: %s@%s(%s#%d), held = %d\n", pmf, 7050 PM_DEVICE(dip), held)) 7051 (void) mdi_power(dip, MDI_PM_POST_UNCONFIG, &held, devnm, 0); 7052 return; 7053 } else if (!PM_GET_PM_INFO(dip)) 7054 return; 7055 7056 PMD(PMD_SET, ("%s: %s@%s(%s#%d), held = %d\n", pmf, PM_DEVICE(dip), 7057 held)) 7058 if (!held) 7059 return; 7060 /* 7061 * We have held power in pre_unconfig, release it here. 7062 */ 7063 pm_rele_power(dip); 7064 } 7065 7066 /* 7067 * Notify ppm of result of attach if there is a ppm that cares 7068 */ 7069 void 7070 pm_post_attach(pm_ppm_cookie_t *cp, int ret) 7071 { 7072 int result; 7073 power_req_t power_req; 7074 dev_info_t *dip; 7075 7076 if (cp->ppc_cmd != DDI_ATTACH) 7077 return; 7078 7079 dip = cp->ppc_dip; 7080 7081 if (ret == DDI_SUCCESS) { 7082 /* 7083 * Attach succeeded, so proceed to doing post-attach pm tasks 7084 */ 7085 if (PM_GET_PM_INFO(dip) == NULL) 7086 (void) pm_start(dip); 7087 } else { 7088 /* 7089 * Attach may have got pm started before failing 7090 */ 7091 pm_stop(dip); 7092 } 7093 7094 if (cp->ppc_ppm) { /* if ppm driver claims the node */ 7095 power_req.request_type = PMR_PPM_POST_ATTACH; 7096 power_req.req.ppm_config_req.who = cp->ppc_dip; 7097 power_req.req.ppm_config_req.result = ret; 7098 ASSERT(PPM(cp->ppc_dip) == cp->ppc_ppm); 7099 (void) pm_ctlops(cp->ppc_ppm, cp->ppc_dip, 7100 DDI_CTLOPS_POWER, &power_req, &result); 7101 } 7102 #ifdef DEBUG 7103 else { 7104 power_req.request_type = PMR_PPM_POST_ATTACH; 7105 power_req.req.ppm_config_req.who = cp->ppc_dip; 7106 power_req.req.ppm_config_req.result = ret; 7107 (void) pm_ctlops(NULL, cp->ppc_dip, 7108 DDI_CTLOPS_POWER, &power_req, &result); 7109 } 7110 #endif 7111 } 7112 7113 /* 7114 * Notify ppm of attach action. Parent is already held at full power by 7115 * probe action. 7116 */ 7117 void 7118 pm_pre_detach(dev_info_t *dip, ddi_detach_cmd_t cmd, pm_ppm_cookie_t *cp) 7119 { 7120 int result; 7121 power_req_t power_req; 7122 7123 bzero(cp, sizeof (*cp)); 7124 cp->ppc_dip = dip; 7125 cp->ppc_cmd = (int)cmd; 7126 7127 switch (cmd) { 7128 case DDI_DETACH: 7129 pm_detaching(dip); /* suspend pm while detaching */ 7130 if (pm_ppm_claimed(dip)) { /* if ppm driver claims node */ 7131 power_req.request_type = PMR_PPM_PRE_DETACH; 7132 power_req.req.ppm_config_req.who = dip; 7133 ASSERT(PPM(dip)); 7134 (void) pm_ctlops(PPM(dip), dip, DDI_CTLOPS_POWER, 7135 &power_req, &result); 7136 cp->ppc_ppm = PPM(dip); 7137 } else { 7138 #ifdef DEBUG 7139 /* pass to the default handler so we can debug things */ 7140 power_req.request_type = PMR_PPM_PRE_DETACH; 7141 power_req.req.ppm_config_req.who = dip; 7142 (void) pm_ctlops(NULL, dip, 7143 DDI_CTLOPS_POWER, &power_req, &result); 7144 #endif 7145 cp->ppc_ppm = NULL; 7146 } 7147 break; 7148 7149 default: 7150 break; 7151 } 7152 } 7153 7154 /* 7155 * Dip is either a leaf node that exported "no-involuntary-power-cycles" prop., 7156 * (if devi_pm_noinvol count is 0) or an ancestor of such a node. We need to 7157 * make an entry to record the details, which includes certain flag settings. 7158 */ 7159 static void 7160 pm_record_invol_path(char *path, int flags, int noinvolpm, int volpmd, 7161 int wasvolpmd, major_t major) 7162 { 7163 PMD_FUNC(pmf, "record_invol_path") 7164 major_t pm_path_to_major(char *); 7165 size_t plen; 7166 pm_noinvol_t *ip, *np, *pp; 7167 pp = NULL; 7168 7169 plen = strlen(path) + 1; 7170 np = kmem_zalloc(sizeof (*np), KM_SLEEP); 7171 np->ni_size = plen; 7172 np->ni_path = kmem_alloc(plen, KM_SLEEP); 7173 np->ni_noinvolpm = noinvolpm; 7174 np->ni_volpmd = volpmd; 7175 np->ni_wasvolpmd = wasvolpmd; 7176 np->ni_flags = flags; 7177 (void) strcpy(np->ni_path, path); 7178 /* 7179 * If we haven't actually seen the node attached, it is hard to figure 7180 * out its major. If we could hold the node by path, we would be much 7181 * happier here. 7182 */ 7183 if (major == DDI_MAJOR_T_NONE) { 7184 np->ni_major = pm_path_to_major(path); 7185 } else { 7186 np->ni_major = major; 7187 } 7188 rw_enter(&pm_noinvol_rwlock, RW_WRITER); 7189 for (ip = pm_noinvol_head; ip; pp = ip, ip = ip->ni_next) { 7190 int comp = strcmp(path, ip->ni_path); 7191 if (comp < 0) { 7192 PMD(PMD_NOINVOL, ("%s: %s insert before %s\n", 7193 pmf, path, ip->ni_path)) 7194 /* insert before current entry */ 7195 np->ni_next = ip; 7196 if (pp) { 7197 pp->ni_next = np; 7198 } else { 7199 pm_noinvol_head = np; 7200 } 7201 rw_exit(&pm_noinvol_rwlock); 7202 #ifdef DEBUG 7203 if (pm_debug & PMD_NOINVOL) 7204 pr_noinvol("record_invol_path exit0"); 7205 #endif 7206 return; 7207 } else if (comp == 0) { 7208 panic("%s already in pm_noinvol list", path); 7209 } 7210 } 7211 /* 7212 * If we did not find an entry in the list that this should go before, 7213 * then it must go at the end 7214 */ 7215 if (pp) { 7216 PMD(PMD_NOINVOL, ("%s: %s append after %s\n", pmf, path, 7217 pp->ni_path)) 7218 ASSERT(pp->ni_next == 0); 7219 pp->ni_next = np; 7220 } else { 7221 PMD(PMD_NOINVOL, ("%s: %s added to end-of-list\n", pmf, path)) 7222 ASSERT(!pm_noinvol_head); 7223 pm_noinvol_head = np; 7224 } 7225 rw_exit(&pm_noinvol_rwlock); 7226 #ifdef DEBUG 7227 if (pm_debug & PMD_NOINVOL) 7228 pr_noinvol("record_invol_path exit"); 7229 #endif 7230 } 7231 7232 void 7233 pm_record_invol(dev_info_t *dip) 7234 { 7235 char *pathbuf; 7236 int pm_all_components_off(dev_info_t *); 7237 int volpmd = (PM_NUMCMPTS(dip) > 0) && pm_all_components_off(dip); 7238 7239 pathbuf = kmem_alloc(MAXPATHLEN, KM_SLEEP); 7240 (void) ddi_pathname(dip, pathbuf); 7241 7242 pm_record_invol_path(pathbuf, (DEVI(dip)->devi_pm_flags & 7243 (PMC_NO_INVOL | PMC_CONSOLE_FB)), DEVI(dip)->devi_pm_noinvolpm, 7244 DEVI(dip)->devi_pm_volpmd, volpmd, PM_MAJOR(dip)); 7245 7246 /* 7247 * If this child's detach will be holding up its ancestors, then we 7248 * allow for an exception to that if all children of this type have 7249 * gone down voluntarily. 7250 * Now walk down the tree incrementing devi_pm_noinvolpm 7251 */ 7252 (void) pm_noinvol_update(PM_BP_NOINVOL_DETACH, 0, volpmd, pathbuf, 7253 dip); 7254 kmem_free(pathbuf, MAXPATHLEN); 7255 } 7256 7257 void 7258 pm_post_detach(pm_ppm_cookie_t *cp, int ret) 7259 { 7260 dev_info_t *dip = cp->ppc_dip; 7261 int result; 7262 power_req_t power_req; 7263 7264 switch (cp->ppc_cmd) { 7265 case DDI_DETACH: 7266 if (cp->ppc_ppm) { /* if ppm driver claims the node */ 7267 power_req.request_type = PMR_PPM_POST_DETACH; 7268 power_req.req.ppm_config_req.who = cp->ppc_dip; 7269 power_req.req.ppm_config_req.result = ret; 7270 ASSERT(PPM(cp->ppc_dip) == cp->ppc_ppm); 7271 (void) pm_ctlops(cp->ppc_ppm, cp->ppc_dip, 7272 DDI_CTLOPS_POWER, &power_req, &result); 7273 } 7274 #ifdef DEBUG 7275 else { 7276 power_req.request_type = PMR_PPM_POST_DETACH; 7277 power_req.req.ppm_config_req.who = cp->ppc_dip; 7278 power_req.req.ppm_config_req.result = ret; 7279 (void) pm_ctlops(NULL, cp->ppc_dip, 7280 DDI_CTLOPS_POWER, &power_req, &result); 7281 } 7282 #endif 7283 if (ret == DDI_SUCCESS) { 7284 /* 7285 * For hotplug detach we assume it is *really* gone 7286 */ 7287 if (cp->ppc_cmd == DDI_DETACH && 7288 ((DEVI(dip)->devi_pm_flags & 7289 (PMC_NO_INVOL | PMC_CONSOLE_FB)) || 7290 DEVI(dip)->devi_pm_noinvolpm)) 7291 pm_record_invol(dip); 7292 DEVI(dip)->devi_pm_flags &= 7293 ~(PMC_NO_INVOL | PMC_NOINVOL_DONE); 7294 7295 /* 7296 * If console fb is detaching, then we don't need to 7297 * worry any more about it going off (pm_detaching has 7298 * brought up all components) 7299 */ 7300 if (PM_IS_CFB(dip)) { 7301 mutex_enter(&pm_cfb_lock); 7302 ASSERT(cfb_dip_detaching); 7303 ASSERT(cfb_dip == NULL); 7304 ASSERT(pm_cfb_comps_off == 0); 7305 cfb_dip_detaching = NULL; 7306 mutex_exit(&pm_cfb_lock); 7307 } 7308 pm_stop(dip); /* make it permanent */ 7309 } else { 7310 if (PM_IS_CFB(dip)) { 7311 mutex_enter(&pm_cfb_lock); 7312 ASSERT(cfb_dip_detaching); 7313 ASSERT(cfb_dip == NULL); 7314 ASSERT(pm_cfb_comps_off == 0); 7315 cfb_dip = cfb_dip_detaching; 7316 cfb_dip_detaching = NULL; 7317 mutex_exit(&pm_cfb_lock); 7318 } 7319 pm_detach_failed(dip); /* resume power management */ 7320 } 7321 break; 7322 case DDI_PM_SUSPEND: 7323 break; 7324 case DDI_SUSPEND: 7325 break; /* legal, but nothing to do */ 7326 default: 7327 #ifdef DEBUG 7328 panic("pm_post_detach: unrecognized cmd %d for detach", 7329 cp->ppc_cmd); 7330 /*NOTREACHED*/ 7331 #else 7332 break; 7333 #endif 7334 } 7335 } 7336 7337 /* 7338 * Called after vfs_mountroot has got the clock started to fix up timestamps 7339 * that were set when root bush drivers attached. hresttime was 0 then, so the 7340 * devices look busy but have a 0 busycnt 7341 */ 7342 int 7343 pm_adjust_timestamps(dev_info_t *dip, void *arg) 7344 { 7345 _NOTE(ARGUNUSED(arg)) 7346 7347 pm_info_t *info = PM_GET_PM_INFO(dip); 7348 struct pm_component *cp; 7349 int i; 7350 7351 if (!info) 7352 return (DDI_WALK_CONTINUE); 7353 PM_LOCK_BUSY(dip); 7354 for (i = 0; i < PM_NUMCMPTS(dip); i++) { 7355 cp = PM_CP(dip, i); 7356 if (cp->pmc_timestamp == 0 && cp->pmc_busycount == 0) 7357 cp->pmc_timestamp = gethrestime_sec(); 7358 } 7359 PM_UNLOCK_BUSY(dip); 7360 return (DDI_WALK_CONTINUE); 7361 } 7362 7363 /* 7364 * Called at attach time to see if the device being attached has a record in 7365 * the no involuntary power cycles list. If so, we do some bookkeeping on the 7366 * parents and set a flag in the dip 7367 */ 7368 void 7369 pm_noinvol_specd(dev_info_t *dip) 7370 { 7371 PMD_FUNC(pmf, "noinvol_specd") 7372 char *pathbuf; 7373 pm_noinvol_t *ip, *pp = NULL; 7374 int wasvolpmd; 7375 int found = 0; 7376 7377 if (DEVI(dip)->devi_pm_flags & PMC_NOINVOL_DONE) 7378 return; 7379 DEVI(dip)->devi_pm_flags |= PMC_NOINVOL_DONE; 7380 pathbuf = kmem_alloc(MAXPATHLEN, KM_SLEEP); 7381 (void) ddi_pathname(dip, pathbuf); 7382 7383 PM_LOCK_DIP(dip); 7384 DEVI(dip)->devi_pm_volpmd = 0; 7385 DEVI(dip)->devi_pm_noinvolpm = 0; 7386 rw_enter(&pm_noinvol_rwlock, RW_READER); 7387 for (ip = pm_noinvol_head; ip; pp = ip, ip = ip->ni_next) { 7388 PMD(PMD_NOINVOL, ("%s: comparing '%s' to '%s'\n", 7389 pmf, pathbuf, ip->ni_path)) 7390 if (strcmp(pathbuf, ip->ni_path) == 0) { 7391 found++; 7392 break; 7393 } 7394 } 7395 rw_exit(&pm_noinvol_rwlock); 7396 if (!found) { 7397 PM_UNLOCK_DIP(dip); 7398 kmem_free(pathbuf, MAXPATHLEN); 7399 return; 7400 } 7401 rw_enter(&pm_noinvol_rwlock, RW_WRITER); 7402 pp = NULL; 7403 for (ip = pm_noinvol_head; ip; pp = ip, ip = ip->ni_next) { 7404 PMD(PMD_NOINVOL, ("%s: comparing '%s' to '%s'\n", 7405 pmf, pathbuf, ip->ni_path)) 7406 if (strcmp(pathbuf, ip->ni_path) == 0) { 7407 ip->ni_flags &= ~PMC_DRIVER_REMOVED; 7408 DEVI(dip)->devi_pm_flags |= ip->ni_flags; 7409 /* 7410 * Handle special case of console fb 7411 */ 7412 if (PM_IS_CFB(dip)) { 7413 mutex_enter(&pm_cfb_lock); 7414 cfb_dip = dip; 7415 PMD(PMD_CFB, ("%s: %s@%s(%s#%d) setting " 7416 "cfb_dip\n", pmf, PM_DEVICE(dip))) 7417 mutex_exit(&pm_cfb_lock); 7418 } 7419 DEVI(dip)->devi_pm_noinvolpm = ip->ni_noinvolpm; 7420 ASSERT((DEVI(dip)->devi_pm_flags & 7421 (PMC_NO_INVOL | PMC_CONSOLE_FB)) || 7422 DEVI(dip)->devi_pm_noinvolpm); 7423 DEVI(dip)->devi_pm_volpmd = ip->ni_volpmd; 7424 PMD(PMD_NOINVOL, ("%s: noinvol=%d, volpmd=%d, " 7425 "wasvolpmd=%d, flags=%x, path=%s\n", pmf, 7426 ip->ni_noinvolpm, ip->ni_volpmd, 7427 ip->ni_wasvolpmd, ip->ni_flags, ip->ni_path)) 7428 /* 7429 * free the entry in hopes the list will now be empty 7430 * and we won't have to search it any more until the 7431 * device detaches 7432 */ 7433 if (pp) { 7434 PMD(PMD_NOINVOL, ("%s: free %s, prev %s\n", 7435 pmf, ip->ni_path, pp->ni_path)) 7436 pp->ni_next = ip->ni_next; 7437 } else { 7438 PMD(PMD_NOINVOL, ("%s: free %s head\n", 7439 pmf, ip->ni_path)) 7440 ASSERT(pm_noinvol_head == ip); 7441 pm_noinvol_head = ip->ni_next; 7442 } 7443 PM_UNLOCK_DIP(dip); 7444 wasvolpmd = ip->ni_wasvolpmd; 7445 rw_exit(&pm_noinvol_rwlock); 7446 kmem_free(ip->ni_path, ip->ni_size); 7447 kmem_free(ip, sizeof (*ip)); 7448 /* 7449 * Now walk up the tree decrementing devi_pm_noinvolpm 7450 * (and volpmd if appropriate) 7451 */ 7452 (void) pm_noinvol_update(PM_BP_NOINVOL_ATTACH, 0, 7453 wasvolpmd, pathbuf, dip); 7454 #ifdef DEBUG 7455 if (pm_debug & PMD_NOINVOL) 7456 pr_noinvol("noinvol_specd exit"); 7457 #endif 7458 kmem_free(pathbuf, MAXPATHLEN); 7459 return; 7460 } 7461 } 7462 kmem_free(pathbuf, MAXPATHLEN); 7463 rw_exit(&pm_noinvol_rwlock); 7464 PM_UNLOCK_DIP(dip); 7465 } 7466 7467 int 7468 pm_all_components_off(dev_info_t *dip) 7469 { 7470 int i; 7471 pm_component_t *cp; 7472 7473 for (i = 0; i < PM_NUMCMPTS(dip); i++) { 7474 cp = PM_CP(dip, i); 7475 if (cp->pmc_cur_pwr == PM_LEVEL_UNKNOWN || 7476 cp->pmc_comp.pmc_lvals[cp->pmc_cur_pwr]) 7477 return (0); 7478 } 7479 return (1); /* all off */ 7480 } 7481 7482 /* 7483 * Make sure that all "no involuntary power cycles" devices are attached. 7484 * Called before doing a cpr suspend to make sure the driver has a say about 7485 * the power cycle 7486 */ 7487 int 7488 pm_reattach_noinvol(void) 7489 { 7490 PMD_FUNC(pmf, "reattach_noinvol") 7491 pm_noinvol_t *ip; 7492 char *path; 7493 dev_info_t *dip; 7494 7495 /* 7496 * Prevent the modunload thread from unloading any modules until we 7497 * have completely stopped all kernel threads. 7498 */ 7499 modunload_disable(); 7500 for (ip = pm_noinvol_head; ip; ip = ip->ni_next) { 7501 /* 7502 * Forget we'v ever seen any entry 7503 */ 7504 ip->ni_persistent = 0; 7505 } 7506 restart: 7507 rw_enter(&pm_noinvol_rwlock, RW_READER); 7508 for (ip = pm_noinvol_head; ip; ip = ip->ni_next) { 7509 #ifdef PMDDEBUG 7510 major_t maj; 7511 maj = ip->ni_major; 7512 #endif 7513 path = ip->ni_path; 7514 if (path != NULL && !(ip->ni_flags & PMC_DRIVER_REMOVED)) { 7515 if (ip->ni_persistent) { 7516 /* 7517 * If we weren't able to make this entry 7518 * go away, then we give up, as 7519 * holding/attaching the driver ought to have 7520 * resulted in this entry being deleted 7521 */ 7522 PMD(PMD_NOINVOL, ("%s: can't reattach %s " 7523 "(%s|%d)\n", pmf, ip->ni_path, 7524 ddi_major_to_name(maj), (int)maj)) 7525 cmn_err(CE_WARN, "cpr: unable to reattach %s ", 7526 ip->ni_path); 7527 modunload_enable(); 7528 rw_exit(&pm_noinvol_rwlock); 7529 return (0); 7530 } 7531 ip->ni_persistent++; 7532 rw_exit(&pm_noinvol_rwlock); 7533 PMD(PMD_NOINVOL, ("%s: holding %s\n", pmf, path)) 7534 dip = e_ddi_hold_devi_by_path(path, 0); 7535 if (dip == NULL) { 7536 PMD(PMD_NOINVOL, ("%s: can't hold (%s|%d)\n", 7537 pmf, path, (int)maj)) 7538 cmn_err(CE_WARN, "cpr: unable to hold %s " 7539 "driver", path); 7540 modunload_enable(); 7541 return (0); 7542 } else { 7543 PMD(PMD_DHR, ("%s: release %s\n", pmf, path)) 7544 /* 7545 * Since the modunload thread is stopped, we 7546 * don't have to keep the driver held, which 7547 * saves a ton of bookkeeping 7548 */ 7549 ddi_release_devi(dip); 7550 goto restart; 7551 } 7552 } else { 7553 PMD(PMD_NOINVOL, ("%s: skip %s; unknown major\n", 7554 pmf, ip->ni_path)) 7555 continue; 7556 } 7557 } 7558 rw_exit(&pm_noinvol_rwlock); 7559 return (1); 7560 } 7561 7562 void 7563 pm_reattach_noinvol_fini(void) 7564 { 7565 modunload_enable(); 7566 } 7567 7568 /* 7569 * Display pm support code 7570 */ 7571 7572 7573 /* 7574 * console frame-buffer power-mgmt gets enabled when debugging 7575 * services are not present or console fbpm override is set 7576 */ 7577 void 7578 pm_cfb_setup(const char *stdout_path) 7579 { 7580 PMD_FUNC(pmf, "cfb_setup") 7581 extern int obpdebug; 7582 char *devname; 7583 dev_info_t *dip; 7584 int devname_len; 7585 extern dev_info_t *fbdip; 7586 7587 /* 7588 * By virtue of this function being called (from consconfig), 7589 * we know stdout is a framebuffer. 7590 */ 7591 stdout_is_framebuffer = 1; 7592 7593 if (obpdebug || (boothowto & RB_DEBUG)) { 7594 if (pm_cfb_override == 0) { 7595 /* 7596 * Console is frame buffer, but we want to suppress 7597 * pm on it because of debugging setup 7598 */ 7599 pm_cfb_enabled = 0; 7600 cmn_err(CE_NOTE, "Kernel debugger present: disabling " 7601 "console power management."); 7602 /* 7603 * however, we still need to know which is the console 7604 * fb in order to suppress pm on it 7605 */ 7606 } else { 7607 cmn_err(CE_WARN, "Kernel debugger present: see " 7608 "kmdb(1M) for interaction with power management."); 7609 } 7610 } 7611 #ifdef DEBUG 7612 /* 7613 * IF console is fb and is power managed, don't do prom_printfs from 7614 * pm debug macro 7615 */ 7616 if (pm_cfb_enabled && !pm_debug_to_console) { 7617 if (pm_debug) 7618 prom_printf("pm debug output will be to log only\n"); 7619 pm_divertdebug++; 7620 } 7621 #endif 7622 devname = i_ddi_strdup((char *)stdout_path, KM_SLEEP); 7623 devname_len = strlen(devname) + 1; 7624 PMD(PMD_CFB, ("%s: stripped %s\n", pmf, devname)) 7625 /* if the driver is attached */ 7626 if ((dip = fbdip) != NULL) { 7627 PMD(PMD_CFB, ("%s: attached: %s@%s(%s#%d)\n", pmf, 7628 PM_DEVICE(dip))) 7629 /* 7630 * We set up here as if the driver were power manageable in case 7631 * we get a later attach of a pm'able driver (which would result 7632 * in a panic later) 7633 */ 7634 cfb_dip = dip; 7635 DEVI(dip)->devi_pm_flags |= (PMC_CONSOLE_FB | PMC_NO_INVOL); 7636 PMD(PMD_CFB, ("%s: cfb_dip -> %s@%s(%s#%d)\n", pmf, 7637 PM_DEVICE(dip))) 7638 #ifdef DEBUG 7639 if (!(PM_GET_PM_INFO(dip) != NULL && PM_NUMCMPTS(dip))) { 7640 PMD(PMD_CFB, ("%s: %s@%s(%s#%d) not power-managed\n", 7641 pmf, PM_DEVICE(dip))) 7642 } 7643 #endif 7644 } else { 7645 char *ep; 7646 PMD(PMD_CFB, ("%s: pntd %s failed\n", pmf, devname)) 7647 pm_record_invol_path(devname, 7648 (PMC_CONSOLE_FB | PMC_NO_INVOL), 1, 0, 0, 7649 DDI_MAJOR_T_NONE); 7650 for (ep = strrchr(devname, '/'); ep != devname; 7651 ep = strrchr(devname, '/')) { 7652 PMD(PMD_CFB, ("%s: devname %s\n", pmf, devname)) 7653 *ep = '\0'; 7654 dip = pm_name_to_dip(devname, 0); 7655 if (dip != NULL) { 7656 /* 7657 * Walk up the tree incrementing 7658 * devi_pm_noinvolpm 7659 */ 7660 (void) pm_noinvol_update(PM_BP_NOINVOL_CFB, 7661 0, 0, devname, dip); 7662 break; 7663 } else { 7664 pm_record_invol_path(devname, 7665 PMC_NO_INVOL, 1, 0, 0, DDI_MAJOR_T_NONE); 7666 } 7667 } 7668 } 7669 kmem_free(devname, devname_len); 7670 } 7671 7672 void 7673 pm_cfb_rele(void) 7674 { 7675 mutex_enter(&pm_cfb_lock); 7676 /* 7677 * this call isn't using the console any more, it is ok to take it 7678 * down if the count goes to 0 7679 */ 7680 cfb_inuse--; 7681 mutex_exit(&pm_cfb_lock); 7682 } 7683 7684 /* 7685 * software interrupt handler for fbpm; this function exists because we can't 7686 * bring up the frame buffer power from above lock level. So if we need to, 7687 * we instead schedule a softint that runs this routine and takes us into 7688 * debug_enter (a bit delayed from the original request, but avoiding a panic). 7689 */ 7690 static uint_t 7691 pm_cfb_softint(caddr_t int_handler_arg) 7692 { 7693 _NOTE(ARGUNUSED(int_handler_arg)) 7694 int rval = DDI_INTR_UNCLAIMED; 7695 7696 mutex_enter(&pm_cfb_lock); 7697 if (pm_soft_pending) { 7698 mutex_exit(&pm_cfb_lock); 7699 debug_enter((char *)NULL); 7700 /* acquired in debug_enter before calling pm_cfb_trigger */ 7701 pm_cfb_rele(); 7702 mutex_enter(&pm_cfb_lock); 7703 pm_soft_pending = 0; 7704 mutex_exit(&pm_cfb_lock); 7705 rval = DDI_INTR_CLAIMED; 7706 } else 7707 mutex_exit(&pm_cfb_lock); 7708 7709 return (rval); 7710 } 7711 7712 void 7713 pm_cfb_setup_intr(void) 7714 { 7715 PMD_FUNC(pmf, "cfb_setup_intr") 7716 extern void prom_set_outfuncs(void (*)(void), void (*)(void)); 7717 void pm_cfb_check_and_powerup(void); 7718 7719 mutex_init(&pm_cfb_lock, NULL, MUTEX_SPIN, (void *)ipltospl(SPL8)); 7720 #ifdef PMDDEBUG 7721 mutex_init(&pm_debug_lock, NULL, MUTEX_SPIN, (void *)ipltospl(SPL8)); 7722 #endif 7723 7724 if (!stdout_is_framebuffer) { 7725 PMD(PMD_CFB, ("%s: console not fb\n", pmf)) 7726 return; 7727 } 7728 7729 /* 7730 * setup software interrupt handler 7731 */ 7732 if (ddi_add_softintr(ddi_root_node(), DDI_SOFTINT_HIGH, &pm_soft_id, 7733 NULL, NULL, pm_cfb_softint, NULL) != DDI_SUCCESS) 7734 panic("pm: unable to register soft intr."); 7735 7736 prom_set_outfuncs(pm_cfb_check_and_powerup, pm_cfb_rele); 7737 } 7738 7739 /* 7740 * Checks to see if it is safe to write to the console wrt power management 7741 * (i.e. if the console is a framebuffer, then it must be at full power) 7742 * returns 1 when power is off (power-up is needed) 7743 * returns 0 when power is on (power-up not needed) 7744 */ 7745 int 7746 pm_cfb_check_and_hold(void) 7747 { 7748 /* 7749 * cfb_dip is set iff console is a power manageable frame buffer 7750 * device 7751 */ 7752 extern int modrootloaded; 7753 7754 mutex_enter(&pm_cfb_lock); 7755 cfb_inuse++; 7756 ASSERT(cfb_inuse); /* wrap? */ 7757 if (modrootloaded && cfb_dip) { 7758 /* 7759 * don't power down the frame buffer, the prom is using it 7760 */ 7761 if (pm_cfb_comps_off) { 7762 mutex_exit(&pm_cfb_lock); 7763 return (1); 7764 } 7765 } 7766 mutex_exit(&pm_cfb_lock); 7767 return (0); 7768 } 7769 7770 /* 7771 * turn on cfb power (which is known to be off). 7772 * Must be called below lock level! 7773 */ 7774 void 7775 pm_cfb_powerup(void) 7776 { 7777 pm_info_t *info; 7778 int norm; 7779 int ccount, ci; 7780 int unused; 7781 #ifdef DEBUG 7782 /* 7783 * Can't reenter prom_prekern, so suppress pm debug messages 7784 * (still go to circular buffer). 7785 */ 7786 mutex_enter(&pm_debug_lock); 7787 pm_divertdebug++; 7788 mutex_exit(&pm_debug_lock); 7789 #endif 7790 info = PM_GET_PM_INFO(cfb_dip); 7791 ASSERT(info); 7792 7793 ccount = PM_NUMCMPTS(cfb_dip); 7794 for (ci = 0; ci < ccount; ci++) { 7795 norm = pm_get_normal_power(cfb_dip, ci); 7796 (void) pm_set_power(cfb_dip, ci, norm, PM_LEVEL_UPONLY, 7797 PM_CANBLOCK_BYPASS, 0, &unused); 7798 } 7799 #ifdef DEBUG 7800 mutex_enter(&pm_debug_lock); 7801 pm_divertdebug--; 7802 mutex_exit(&pm_debug_lock); 7803 #endif 7804 } 7805 7806 /* 7807 * Check if the console framebuffer is powered up. If not power it up. 7808 * Note: Calling pm_cfb_check_and_hold has put a hold on the power state which 7809 * must be released by calling pm_cfb_rele when the console fb operation 7810 * is completed. 7811 */ 7812 void 7813 pm_cfb_check_and_powerup(void) 7814 { 7815 if (pm_cfb_check_and_hold()) 7816 pm_cfb_powerup(); 7817 } 7818 7819 /* 7820 * Trigger a low level interrupt to power up console frame buffer. 7821 */ 7822 void 7823 pm_cfb_trigger(void) 7824 { 7825 if (cfb_dip == NULL) 7826 return; 7827 7828 mutex_enter(&pm_cfb_lock); 7829 /* 7830 * If machine appears to be hung, pulling the keyboard connector of 7831 * the console will cause a high level interrupt and go to debug_enter. 7832 * But, if the fb is powered down, this routine will be called to bring 7833 * it up (by generating a softint to do the work). If soft interrupts 7834 * are not running, and the keyboard connector is pulled again, the 7835 * following code detects this condition and calls panic which allows 7836 * the fb to be brought up from high level. 7837 * 7838 * If two nearly simultaneous calls to debug_enter occur (both from 7839 * high level) the code described above will cause a panic. 7840 */ 7841 if (lbolt <= pm_soft_pending) { 7842 panicstr = "pm_cfb_trigger: lbolt not advancing"; 7843 panic(panicstr); /* does a power up at any intr level */ 7844 /* NOTREACHED */ 7845 } 7846 pm_soft_pending = lbolt; 7847 mutex_exit(&pm_cfb_lock); 7848 ddi_trigger_softintr(pm_soft_id); 7849 } 7850 7851 major_t 7852 pm_path_to_major(char *path) 7853 { 7854 PMD_FUNC(pmf, "path_to_major") 7855 char *np, *ap, *bp; 7856 major_t ret; 7857 size_t len; 7858 static major_t i_path_to_major(char *, char *); 7859 7860 PMD(PMD_NOINVOL, ("%s: %s\n", pmf, path)) 7861 7862 np = strrchr(path, '/'); 7863 if (np != NULL) 7864 np++; 7865 else 7866 np = path; 7867 len = strlen(np) + 1; 7868 bp = kmem_alloc(len, KM_SLEEP); 7869 (void) strcpy(bp, np); 7870 if ((ap = strchr(bp, '@')) != NULL) { 7871 *ap = '\0'; 7872 } 7873 PMD(PMD_NOINVOL, ("%s: %d\n", pmf, ddi_name_to_major(np))) 7874 ret = i_path_to_major(path, np); 7875 kmem_free(bp, len); 7876 return (ret); 7877 } 7878 7879 #ifdef DEBUG 7880 #ifndef sparc 7881 clock_t pt_sleep = 1; 7882 #endif 7883 7884 char *pm_msgp; 7885 char *pm_bufend; 7886 char *pm_msgbuf = NULL; 7887 int pm_logpages = 0x100; 7888 #include <sys/sunldi.h> 7889 #include <sys/uio.h> 7890 clock_t pm_log_sleep = 1000; 7891 int pm_extra_cr = 1; 7892 volatile int pm_tty = 1; 7893 7894 #define PMLOGPGS pm_logpages 7895 7896 #if defined(__x86) 7897 void pm_printf(char *s); 7898 #endif 7899 7900 /*PRINTFLIKE1*/ 7901 void 7902 pm_log(const char *fmt, ...) 7903 { 7904 va_list adx; 7905 size_t size; 7906 7907 mutex_enter(&pm_debug_lock); 7908 if (pm_msgbuf == NULL) { 7909 pm_msgbuf = kmem_zalloc(mmu_ptob(PMLOGPGS), KM_SLEEP); 7910 pm_bufend = pm_msgbuf + mmu_ptob(PMLOGPGS) - 1; 7911 pm_msgp = pm_msgbuf; 7912 } 7913 va_start(adx, fmt); 7914 size = vsnprintf(NULL, 0, fmt, adx) + 1; 7915 va_end(adx); 7916 va_start(adx, fmt); 7917 if (size > (pm_bufend - pm_msgp)) { /* wraps */ 7918 bzero(pm_msgp, pm_bufend - pm_msgp); 7919 (void) vsnprintf(pm_msgbuf, size, fmt, adx); 7920 if (!pm_divertdebug) 7921 prom_printf("%s", pm_msgp); 7922 #if defined(__x86) 7923 if (pm_tty) { 7924 pm_printf(pm_msgp); 7925 if (pm_extra_cr) 7926 pm_printf("\r"); 7927 } 7928 #endif 7929 pm_msgp = pm_msgbuf + size; 7930 } else { 7931 (void) vsnprintf(pm_msgp, size, fmt, adx); 7932 #if defined(__x86) 7933 if (pm_tty) { 7934 pm_printf(pm_msgp); 7935 if (pm_extra_cr) 7936 pm_printf("\r"); 7937 } 7938 #endif 7939 if (!pm_divertdebug) 7940 prom_printf("%s", pm_msgp); 7941 pm_msgp += size; 7942 } 7943 va_end(adx); 7944 mutex_exit(&pm_debug_lock); 7945 drv_usecwait((clock_t)pm_log_sleep); 7946 } 7947 #endif /* DEBUG */ 7948 7949 /* 7950 * We want to save the state of any directly pm'd devices over the suspend/ 7951 * resume process so that we can put them back the way the controlling 7952 * process left them. 7953 */ 7954 void 7955 pm_save_direct_levels(void) 7956 { 7957 pm_processes_stopped = 1; 7958 ddi_walk_devs(ddi_root_node(), pm_save_direct_lvl_walk, 0); 7959 } 7960 7961 static int 7962 pm_save_direct_lvl_walk(dev_info_t *dip, void *arg) 7963 { 7964 _NOTE(ARGUNUSED(arg)) 7965 int i; 7966 int *ip; 7967 pm_info_t *info = PM_GET_PM_INFO(dip); 7968 7969 if (!info) 7970 return (DDI_WALK_CONTINUE); 7971 7972 if (PM_ISDIRECT(dip) && !PM_ISBC(dip)) { 7973 if (PM_NUMCMPTS(dip) > 2) { 7974 info->pmi_lp = kmem_alloc(PM_NUMCMPTS(dip) * 7975 sizeof (int), KM_SLEEP); 7976 ip = info->pmi_lp; 7977 } else { 7978 ip = info->pmi_levels; 7979 } 7980 /* autopm and processes are stopped, ok not to lock power */ 7981 for (i = 0; i < PM_NUMCMPTS(dip); i++) 7982 *ip++ = PM_CURPOWER(dip, i); 7983 /* 7984 * There is a small window between stopping the 7985 * processes and setting pm_processes_stopped where 7986 * a driver could get hung up in a pm_raise_power() 7987 * call. Free any such driver now. 7988 */ 7989 pm_proceed(dip, PMP_RELEASE, -1, -1); 7990 } 7991 7992 return (DDI_WALK_CONTINUE); 7993 } 7994 7995 void 7996 pm_restore_direct_levels(void) 7997 { 7998 /* 7999 * If cpr didn't call pm_save_direct_levels, (because stopping user 8000 * threads failed) then we don't want to try to restore them 8001 */ 8002 if (!pm_processes_stopped) 8003 return; 8004 8005 ddi_walk_devs(ddi_root_node(), pm_restore_direct_lvl_walk, 0); 8006 pm_processes_stopped = 0; 8007 } 8008 8009 static int 8010 pm_restore_direct_lvl_walk(dev_info_t *dip, void *arg) 8011 { 8012 _NOTE(ARGUNUSED(arg)) 8013 PMD_FUNC(pmf, "restore_direct_lvl_walk") 8014 int i, nc, result; 8015 int *ip; 8016 8017 pm_info_t *info = PM_GET_PM_INFO(dip); 8018 if (!info) 8019 return (DDI_WALK_CONTINUE); 8020 8021 if (PM_ISDIRECT(dip) && !PM_ISBC(dip)) { 8022 if ((nc = PM_NUMCMPTS(dip)) > 2) { 8023 ip = &info->pmi_lp[nc - 1]; 8024 } else { 8025 ip = &info->pmi_levels[nc - 1]; 8026 } 8027 /* 8028 * Because fb drivers fail attempts to turn off the 8029 * fb when the monitor is on, but treat a request to 8030 * turn on the monitor as a request to turn on the 8031 * fb too, we process components in descending order 8032 * Because autopm is disabled and processes aren't 8033 * running, it is ok to examine current power outside 8034 * of the power lock 8035 */ 8036 for (i = nc - 1; i >= 0; i--, ip--) { 8037 if (PM_CURPOWER(dip, i) == *ip) 8038 continue; 8039 if (pm_set_power(dip, i, *ip, PM_LEVEL_EXACT, 8040 PM_CANBLOCK_BYPASS, 0, &result) != DDI_SUCCESS) { 8041 cmn_err(CE_WARN, "cpr: unable " 8042 "to restore power level of " 8043 "component %d of directly " 8044 "power manged device %s@%s" 8045 " to %d", 8046 i, PM_NAME(dip), 8047 PM_ADDR(dip), *ip); 8048 PMD(PMD_FAIL, ("%s: failed to restore " 8049 "%s@%s(%s#%d)[%d] exact(%d)->%d, " 8050 "errno %d\n", pmf, PM_DEVICE(dip), i, 8051 PM_CURPOWER(dip, i), *ip, result)) 8052 } 8053 } 8054 if (nc > 2) { 8055 kmem_free(info->pmi_lp, nc * sizeof (int)); 8056 info->pmi_lp = NULL; 8057 } 8058 } 8059 return (DDI_WALK_CONTINUE); 8060 } 8061 8062 /* 8063 * Stolen from the bootdev module 8064 * attempt to convert a path to a major number 8065 */ 8066 static major_t 8067 i_path_to_major(char *path, char *leaf_name) 8068 { 8069 extern major_t path_to_major(char *pathname); 8070 major_t maj; 8071 8072 if ((maj = path_to_major(path)) == DDI_MAJOR_T_NONE) { 8073 maj = ddi_name_to_major(leaf_name); 8074 } 8075 8076 return (maj); 8077 } 8078 8079 /* 8080 * When user calls rem_drv, we need to forget no-involuntary-power-cycles state 8081 * An entry in the list means that the device is detached, so we need to 8082 * adjust its ancestors as if they had just seen this attach, and any detached 8083 * ancestors need to have their list entries adjusted. 8084 */ 8085 void 8086 pm_driver_removed(major_t major) 8087 { 8088 static void i_pm_driver_removed(major_t major); 8089 8090 /* 8091 * Serialize removal of drivers. This is to keep ancestors of 8092 * a node that is being deleted from getting deleted and added back 8093 * with different counters. 8094 */ 8095 mutex_enter(&pm_remdrv_lock); 8096 i_pm_driver_removed(major); 8097 mutex_exit(&pm_remdrv_lock); 8098 } 8099 8100 /* 8101 * This routine is called recursively by pm_noinvol_process_ancestors() 8102 */ 8103 static void 8104 i_pm_driver_removed(major_t major) 8105 { 8106 PMD_FUNC(pmf, "driver_removed") 8107 static void adjust_ancestors(char *, int); 8108 static int pm_is_noinvol_ancestor(pm_noinvol_t *); 8109 static void pm_noinvol_process_ancestors(char *); 8110 pm_noinvol_t *ip, *pp = NULL; 8111 int wasvolpmd; 8112 ASSERT(major != DDI_MAJOR_T_NONE); 8113 PMD(PMD_NOINVOL, ("%s: %s\n", pmf, ddi_major_to_name(major))) 8114 again: 8115 rw_enter(&pm_noinvol_rwlock, RW_WRITER); 8116 for (ip = pm_noinvol_head; ip; pp = ip, ip = ip->ni_next) { 8117 if (major != ip->ni_major) 8118 continue; 8119 /* 8120 * If it is an ancestor of no-invol node, which is 8121 * not removed, skip it. This is to cover the case of 8122 * ancestor removed without removing its descendants. 8123 */ 8124 if (pm_is_noinvol_ancestor(ip)) { 8125 ip->ni_flags |= PMC_DRIVER_REMOVED; 8126 continue; 8127 } 8128 wasvolpmd = ip->ni_wasvolpmd; 8129 /* 8130 * remove the entry from the list 8131 */ 8132 if (pp) { 8133 PMD(PMD_NOINVOL, ("%s: freeing %s, prev is %s\n", 8134 pmf, ip->ni_path, pp->ni_path)) 8135 pp->ni_next = ip->ni_next; 8136 } else { 8137 PMD(PMD_NOINVOL, ("%s: free %s head\n", pmf, 8138 ip->ni_path)) 8139 ASSERT(pm_noinvol_head == ip); 8140 pm_noinvol_head = ip->ni_next; 8141 } 8142 rw_exit(&pm_noinvol_rwlock); 8143 adjust_ancestors(ip->ni_path, wasvolpmd); 8144 /* 8145 * Had an ancestor been removed before this node, it would have 8146 * been skipped. Adjust the no-invol counters for such skipped 8147 * ancestors. 8148 */ 8149 pm_noinvol_process_ancestors(ip->ni_path); 8150 kmem_free(ip->ni_path, ip->ni_size); 8151 kmem_free(ip, sizeof (*ip)); 8152 goto again; 8153 } 8154 rw_exit(&pm_noinvol_rwlock); 8155 } 8156 8157 /* 8158 * returns 1, if *aip is a ancestor of a no-invol node 8159 * 0, otherwise 8160 */ 8161 static int 8162 pm_is_noinvol_ancestor(pm_noinvol_t *aip) 8163 { 8164 pm_noinvol_t *ip; 8165 8166 ASSERT(strlen(aip->ni_path) != 0); 8167 for (ip = pm_noinvol_head; ip; ip = ip->ni_next) { 8168 if (ip == aip) 8169 continue; 8170 /* 8171 * To be an ancestor, the path must be an initial substring of 8172 * the descendent, and end just before a '/' in the 8173 * descendent's path. 8174 */ 8175 if ((strstr(ip->ni_path, aip->ni_path) == ip->ni_path) && 8176 (ip->ni_path[strlen(aip->ni_path)] == '/')) 8177 return (1); 8178 } 8179 return (0); 8180 } 8181 8182 #define PM_MAJOR(dip) ddi_name_to_major(ddi_binding_name(dip)) 8183 /* 8184 * scan through the pm_noinvolpm list adjusting ancestors of the current 8185 * node; Modifies string *path. 8186 */ 8187 static void 8188 adjust_ancestors(char *path, int wasvolpmd) 8189 { 8190 PMD_FUNC(pmf, "adjust_ancestors") 8191 char *cp; 8192 pm_noinvol_t *lp; 8193 pm_noinvol_t *pp = NULL; 8194 major_t locked = DDI_MAJOR_T_NONE; 8195 dev_info_t *dip; 8196 char *pathbuf; 8197 size_t pathbuflen = strlen(path) + 1; 8198 8199 /* 8200 * First we look up the ancestor's dip. If we find it, then we 8201 * adjust counts up the tree 8202 */ 8203 PMD(PMD_NOINVOL, ("%s: %s wasvolpmd %d\n", pmf, path, wasvolpmd)) 8204 pathbuf = kmem_alloc(pathbuflen, KM_SLEEP); 8205 (void) strcpy(pathbuf, path); 8206 cp = strrchr(pathbuf, '/'); 8207 if (cp == NULL) { 8208 /* if no ancestors, then nothing to do */ 8209 kmem_free(pathbuf, pathbuflen); 8210 return; 8211 } 8212 *cp = '\0'; 8213 dip = pm_name_to_dip(pathbuf, 1); 8214 if (dip != NULL) { 8215 locked = PM_MAJOR(dip); 8216 8217 (void) pm_noinvol_update(PM_BP_NOINVOL_REMDRV, 0, wasvolpmd, 8218 path, dip); 8219 8220 if (locked != DDI_MAJOR_T_NONE) 8221 ddi_release_devi(dip); 8222 } else { 8223 char *apath; 8224 size_t len = strlen(pathbuf) + 1; 8225 int lock_held = 1; 8226 8227 /* 8228 * Now check for ancestors that exist only in the list 8229 */ 8230 apath = kmem_alloc(len, KM_SLEEP); 8231 (void) strcpy(apath, pathbuf); 8232 rw_enter(&pm_noinvol_rwlock, RW_WRITER); 8233 for (lp = pm_noinvol_head; lp; pp = lp, lp = lp->ni_next) { 8234 /* 8235 * This can only happen once. Since we have to drop 8236 * the lock, we need to extract the relevant info. 8237 */ 8238 if (strcmp(pathbuf, lp->ni_path) == 0) { 8239 PMD(PMD_NOINVOL, ("%s: %s no %d -> %d\n", pmf, 8240 lp->ni_path, lp->ni_noinvolpm, 8241 lp->ni_noinvolpm - 1)) 8242 lp->ni_noinvolpm--; 8243 if (wasvolpmd && lp->ni_volpmd) { 8244 PMD(PMD_NOINVOL, ("%s: %s vol %d -> " 8245 "%d\n", pmf, lp->ni_path, 8246 lp->ni_volpmd, lp->ni_volpmd - 1)) 8247 lp->ni_volpmd--; 8248 } 8249 /* 8250 * remove the entry from the list, if there 8251 * are no more no-invol descendants and node 8252 * itself is not a no-invol node. 8253 */ 8254 if (!(lp->ni_noinvolpm || 8255 (lp->ni_flags & PMC_NO_INVOL))) { 8256 ASSERT(lp->ni_volpmd == 0); 8257 if (pp) { 8258 PMD(PMD_NOINVOL, ("%s: freeing " 8259 "%s, prev is %s\n", pmf, 8260 lp->ni_path, pp->ni_path)) 8261 pp->ni_next = lp->ni_next; 8262 } else { 8263 PMD(PMD_NOINVOL, ("%s: free %s " 8264 "head\n", pmf, lp->ni_path)) 8265 ASSERT(pm_noinvol_head == lp); 8266 pm_noinvol_head = lp->ni_next; 8267 } 8268 lock_held = 0; 8269 rw_exit(&pm_noinvol_rwlock); 8270 adjust_ancestors(apath, wasvolpmd); 8271 /* restore apath */ 8272 (void) strcpy(apath, pathbuf); 8273 kmem_free(lp->ni_path, lp->ni_size); 8274 kmem_free(lp, sizeof (*lp)); 8275 } 8276 break; 8277 } 8278 } 8279 if (lock_held) 8280 rw_exit(&pm_noinvol_rwlock); 8281 adjust_ancestors(apath, wasvolpmd); 8282 kmem_free(apath, len); 8283 } 8284 kmem_free(pathbuf, pathbuflen); 8285 } 8286 8287 /* 8288 * Do no-invol processing for any ancestors i.e. adjust counters of ancestors, 8289 * which were skipped even though their drivers were removed. 8290 */ 8291 static void 8292 pm_noinvol_process_ancestors(char *path) 8293 { 8294 pm_noinvol_t *lp; 8295 8296 rw_enter(&pm_noinvol_rwlock, RW_READER); 8297 for (lp = pm_noinvol_head; lp; lp = lp->ni_next) { 8298 if (strstr(path, lp->ni_path) && 8299 (lp->ni_flags & PMC_DRIVER_REMOVED)) { 8300 rw_exit(&pm_noinvol_rwlock); 8301 i_pm_driver_removed(lp->ni_major); 8302 return; 8303 } 8304 } 8305 rw_exit(&pm_noinvol_rwlock); 8306 } 8307 8308 /* 8309 * Returns true if (detached) device needs to be kept up because it exported the 8310 * "no-involuntary-power-cycles" property or we're pretending it did (console 8311 * fb case) or it is an ancestor of such a device and has used up the "one 8312 * free cycle" allowed when all such leaf nodes have voluntarily powered down 8313 * upon detach. In any event, we need an exact hit on the path or we return 8314 * false. 8315 */ 8316 int 8317 pm_noinvol_detached(char *path) 8318 { 8319 PMD_FUNC(pmf, "noinvol_detached") 8320 pm_noinvol_t *ip; 8321 int ret = 0; 8322 8323 rw_enter(&pm_noinvol_rwlock, RW_READER); 8324 for (ip = pm_noinvol_head; ip; ip = ip->ni_next) { 8325 if (strcmp(path, ip->ni_path) == 0) { 8326 if (ip->ni_flags & PMC_CONSOLE_FB) { 8327 PMD(PMD_NOINVOL | PMD_CFB, ("%s: inhibits CFB " 8328 "%s\n", pmf, path)) 8329 ret = 1; 8330 break; 8331 } 8332 #ifdef DEBUG 8333 if (ip->ni_noinvolpm != ip->ni_volpmd) 8334 PMD(PMD_NOINVOL, ("%s: (%d != %d) inhibits %s" 8335 "\n", pmf, ip->ni_noinvolpm, ip->ni_volpmd, 8336 path)) 8337 #endif 8338 ret = (ip->ni_noinvolpm != ip->ni_volpmd); 8339 break; 8340 } 8341 } 8342 rw_exit(&pm_noinvol_rwlock); 8343 return (ret); 8344 } 8345 8346 int 8347 pm_is_cfb(dev_info_t *dip) 8348 { 8349 return (dip == cfb_dip); 8350 } 8351 8352 #ifdef DEBUG 8353 /* 8354 * Return true if all components of the console frame buffer are at 8355 * "normal" power, i.e., fully on. For the case where the console is not 8356 * a framebuffer, we also return true 8357 */ 8358 int 8359 pm_cfb_is_up(void) 8360 { 8361 return (pm_cfb_comps_off == 0); 8362 } 8363 #endif 8364 8365 /* 8366 * Preventing scan from powering down the node by incrementing the 8367 * kidsupcnt. 8368 */ 8369 void 8370 pm_hold_power(dev_info_t *dip) 8371 { 8372 e_pm_hold_rele_power(dip, 1); 8373 } 8374 8375 /* 8376 * Releasing the hold by decrementing the kidsupcnt allowing scan 8377 * to power down the node if all conditions are met. 8378 */ 8379 void 8380 pm_rele_power(dev_info_t *dip) 8381 { 8382 e_pm_hold_rele_power(dip, -1); 8383 } 8384 8385 /* 8386 * A wrapper of pm_all_to_normal() to power up a dip 8387 * to its normal level 8388 */ 8389 int 8390 pm_powerup(dev_info_t *dip) 8391 { 8392 PMD_FUNC(pmf, "pm_powerup") 8393 8394 PMD(PMD_ALLNORM, ("%s: %s@%s(%s#%d)\n", pmf, PM_DEVICE(dip))) 8395 ASSERT(!(servicing_interrupt())); 8396 8397 /* 8398 * in case this node is not already participating pm 8399 */ 8400 if (!PM_GET_PM_INFO(dip)) { 8401 if (!DEVI_IS_ATTACHING(dip)) 8402 return (DDI_SUCCESS); 8403 if (pm_start(dip) != DDI_SUCCESS) 8404 return (DDI_FAILURE); 8405 if (!PM_GET_PM_INFO(dip)) 8406 return (DDI_SUCCESS); 8407 } 8408 8409 return (pm_all_to_normal(dip, PM_CANBLOCK_BLOCK)); 8410 } 8411 8412 int 8413 pm_rescan_walk(dev_info_t *dip, void *arg) 8414 { 8415 _NOTE(ARGUNUSED(arg)) 8416 8417 if (!PM_GET_PM_INFO(dip) || PM_ISBC(dip)) 8418 return (DDI_WALK_CONTINUE); 8419 8420 /* 8421 * Currently pm_cpr_callb/resume code is the only caller 8422 * and it needs to make sure that stopped scan get 8423 * reactivated. Otherwise, rescan walk needn't reactive 8424 * stopped scan. 8425 */ 8426 pm_scan_init(dip); 8427 8428 (void) pm_rescan(dip); 8429 return (DDI_WALK_CONTINUE); 8430 } 8431 8432 static dev_info_t * 8433 pm_get_next_descendent(dev_info_t *dip, dev_info_t *tdip) 8434 { 8435 dev_info_t *wdip, *pdip; 8436 8437 for (wdip = tdip; wdip != dip; wdip = pdip) { 8438 pdip = ddi_get_parent(wdip); 8439 if (pdip == dip) 8440 return (wdip); 8441 } 8442 return (NULL); 8443 } 8444 8445 int 8446 pm_busop_bus_power(dev_info_t *dip, void *impl_arg, pm_bus_power_op_t op, 8447 void *arg, void *result) 8448 { 8449 PMD_FUNC(pmf, "bp_bus_power") 8450 dev_info_t *cdip; 8451 pm_info_t *cinfo; 8452 pm_bp_child_pwrchg_t *bpc; 8453 pm_sp_misc_t *pspm; 8454 pm_bp_nexus_pwrup_t *bpn; 8455 pm_bp_child_pwrchg_t new_bpc; 8456 pm_bp_noinvol_t *bpi; 8457 dev_info_t *tdip; 8458 char *pathbuf; 8459 int ret = DDI_SUCCESS; 8460 int errno = 0; 8461 pm_component_t *cp; 8462 8463 PMD(PMD_SET, ("%s: %s@%s(%s#%d) %s\n", pmf, PM_DEVICE(dip), 8464 pm_decode_op(op))) 8465 switch (op) { 8466 case BUS_POWER_CHILD_PWRCHG: 8467 bpc = (pm_bp_child_pwrchg_t *)arg; 8468 pspm = (pm_sp_misc_t *)bpc->bpc_private; 8469 tdip = bpc->bpc_dip; 8470 cdip = pm_get_next_descendent(dip, tdip); 8471 cinfo = PM_GET_PM_INFO(cdip); 8472 if (cdip != tdip) { 8473 /* 8474 * If the node is an involved parent, it needs to 8475 * power up the node as it is needed. There is nothing 8476 * else the framework can do here. 8477 */ 8478 if (PM_WANTS_NOTIFICATION(cdip)) { 8479 PMD(PMD_SET, ("%s: call bus_power for " 8480 "%s@%s(%s#%d)\n", pmf, PM_DEVICE(cdip))) 8481 return ((*PM_BUS_POWER_FUNC(cdip))(cdip, 8482 impl_arg, op, arg, result)); 8483 } 8484 ASSERT(pspm->pspm_direction == PM_LEVEL_UPONLY || 8485 pspm->pspm_direction == PM_LEVEL_DOWNONLY || 8486 pspm->pspm_direction == PM_LEVEL_EXACT); 8487 /* 8488 * we presume that the parent needs to be up in 8489 * order for the child to change state (either 8490 * because it must already be on if the child is on 8491 * (and the pm_all_to_normal_nexus() will be a nop) 8492 * or because it will need to be on for the child 8493 * to come on; so we make the call regardless 8494 */ 8495 pm_hold_power(cdip); 8496 if (cinfo) { 8497 pm_canblock_t canblock = pspm->pspm_canblock; 8498 ret = pm_all_to_normal_nexus(cdip, canblock); 8499 if (ret != DDI_SUCCESS) { 8500 pm_rele_power(cdip); 8501 return (ret); 8502 } 8503 } 8504 PMD(PMD_SET, ("%s: walk down to %s@%s(%s#%d)\n", pmf, 8505 PM_DEVICE(cdip))) 8506 ret = pm_busop_bus_power(cdip, impl_arg, op, arg, 8507 result); 8508 pm_rele_power(cdip); 8509 } else { 8510 ret = pm_busop_set_power(cdip, impl_arg, op, arg, 8511 result); 8512 } 8513 return (ret); 8514 8515 case BUS_POWER_NEXUS_PWRUP: 8516 bpn = (pm_bp_nexus_pwrup_t *)arg; 8517 pspm = (pm_sp_misc_t *)bpn->bpn_private; 8518 8519 if (!e_pm_valid_info(dip, NULL) || 8520 !e_pm_valid_comp(dip, bpn->bpn_comp, &cp) || 8521 !e_pm_valid_power(dip, bpn->bpn_comp, bpn->bpn_level)) { 8522 PMD(PMD_SET, ("%s: %s@%s(%s#%d) has no pm info; EIO\n", 8523 pmf, PM_DEVICE(dip))) 8524 *pspm->pspm_errnop = EIO; 8525 *(int *)result = DDI_FAILURE; 8526 return (DDI_FAILURE); 8527 } 8528 8529 ASSERT(bpn->bpn_dip == dip); 8530 PMD(PMD_SET, ("%s: nexus powerup for %s@%s(%s#%d)\n", pmf, 8531 PM_DEVICE(dip))) 8532 new_bpc.bpc_dip = dip; 8533 pathbuf = kmem_alloc(MAXPATHLEN, KM_SLEEP); 8534 new_bpc.bpc_path = ddi_pathname(dip, pathbuf); 8535 new_bpc.bpc_comp = bpn->bpn_comp; 8536 new_bpc.bpc_olevel = PM_CURPOWER(dip, bpn->bpn_comp); 8537 new_bpc.bpc_nlevel = bpn->bpn_level; 8538 new_bpc.bpc_private = bpn->bpn_private; 8539 ((pm_sp_misc_t *)(new_bpc.bpc_private))->pspm_direction = 8540 PM_LEVEL_UPONLY; 8541 ((pm_sp_misc_t *)(new_bpc.bpc_private))->pspm_errnop = 8542 &errno; 8543 ret = pm_busop_set_power(dip, impl_arg, BUS_POWER_CHILD_PWRCHG, 8544 (void *)&new_bpc, result); 8545 kmem_free(pathbuf, MAXPATHLEN); 8546 return (ret); 8547 8548 case BUS_POWER_NOINVOL: 8549 bpi = (pm_bp_noinvol_t *)arg; 8550 tdip = bpi->bpni_dip; 8551 cdip = pm_get_next_descendent(dip, tdip); 8552 8553 /* In case of rem_drv, the leaf node has been removed */ 8554 if (cdip == NULL) 8555 return (DDI_SUCCESS); 8556 8557 cinfo = PM_GET_PM_INFO(cdip); 8558 if (cdip != tdip) { 8559 if (PM_WANTS_NOTIFICATION(cdip)) { 8560 PMD(PMD_NOINVOL, 8561 ("%s: call bus_power for %s@%s(%s#%d)\n", 8562 pmf, PM_DEVICE(cdip))) 8563 ret = (*PM_BUS_POWER_FUNC(cdip)) 8564 (cdip, NULL, op, arg, result); 8565 if ((cinfo) && (ret == DDI_SUCCESS)) 8566 (void) pm_noinvol_update_node(cdip, 8567 bpi); 8568 return (ret); 8569 } else { 8570 PMD(PMD_NOINVOL, 8571 ("%s: walk down to %s@%s(%s#%d)\n", pmf, 8572 PM_DEVICE(cdip))) 8573 ret = pm_busop_bus_power(cdip, NULL, op, 8574 arg, result); 8575 /* 8576 * Update the current node. 8577 */ 8578 if ((cinfo) && (ret == DDI_SUCCESS)) 8579 (void) pm_noinvol_update_node(cdip, 8580 bpi); 8581 return (ret); 8582 } 8583 } else { 8584 /* 8585 * For attach, detach, power up: 8586 * Do nothing for leaf node since its 8587 * counts are already updated. 8588 * For CFB and driver removal, since the 8589 * path and the target dip passed in is up to and incl. 8590 * the immediate ancestor, need to do the update. 8591 */ 8592 PMD(PMD_NOINVOL, ("%s: target %s@%s(%s#%d) is " 8593 "reached\n", pmf, PM_DEVICE(cdip))) 8594 if (cinfo && ((bpi->bpni_cmd == PM_BP_NOINVOL_REMDRV) || 8595 (bpi->bpni_cmd == PM_BP_NOINVOL_CFB))) 8596 (void) pm_noinvol_update_node(cdip, bpi); 8597 return (DDI_SUCCESS); 8598 } 8599 8600 default: 8601 PMD(PMD_SET, ("%s: operation %d is not supported!\n", pmf, op)) 8602 return (DDI_FAILURE); 8603 } 8604 } 8605 8606 static int 8607 pm_busop_set_power(dev_info_t *dip, void *impl_arg, pm_bus_power_op_t op, 8608 void *arg, void *resultp) 8609 { 8610 _NOTE(ARGUNUSED(impl_arg)) 8611 PMD_FUNC(pmf, "bp_set_power") 8612 pm_ppm_devlist_t *devl = NULL; 8613 int clevel, circ; 8614 #ifdef DEBUG 8615 int circ_db, ccirc_db; 8616 #endif 8617 int ret = DDI_SUCCESS; 8618 dev_info_t *cdip; 8619 pm_bp_child_pwrchg_t *bpc = (pm_bp_child_pwrchg_t *)arg; 8620 pm_sp_misc_t *pspm = (pm_sp_misc_t *)bpc->bpc_private; 8621 pm_canblock_t canblock = pspm->pspm_canblock; 8622 int scan = pspm->pspm_scan; 8623 int comp = bpc->bpc_comp; 8624 int olevel = bpc->bpc_olevel; 8625 int nlevel = bpc->bpc_nlevel; 8626 int comps_off_incr = 0; 8627 dev_info_t *pdip = ddi_get_parent(dip); 8628 int dodeps; 8629 int direction = pspm->pspm_direction; 8630 int *errnop = pspm->pspm_errnop; 8631 #ifdef PMDDEBUG 8632 char *dir = pm_decode_direction(direction); 8633 #endif 8634 int *iresp = (int *)resultp; 8635 time_t idletime, thresh; 8636 pm_component_t *cp = PM_CP(dip, comp); 8637 int work_type; 8638 8639 *iresp = DDI_SUCCESS; 8640 *errnop = 0; 8641 ASSERT(op == BUS_POWER_CHILD_PWRCHG); 8642 PMD(PMD_SET, ("%s: %s@%s(%s#%d) %s\n", pmf, PM_DEVICE(dip), 8643 pm_decode_op(op))) 8644 8645 /* 8646 * The following set of conditions indicate we are here to handle a 8647 * driver's pm_[raise|lower]_power request, but the device is being 8648 * power managed (PM_DIRECT_PM) by a user process. For that case 8649 * we want to pm_block and pass a status back to the caller based 8650 * on whether the controlling process's next activity on the device 8651 * matches the current request or not. This distinction tells 8652 * downstream functions to avoid calling into a driver or changing 8653 * the framework's power state. To actually block, we need: 8654 * 8655 * PM_ISDIRECT(dip) 8656 * no reason to block unless a process is directly controlling dev 8657 * direction != PM_LEVEL_EXACT 8658 * EXACT is used by controlling proc's PM_SET_CURRENT_POWER ioctl 8659 * !pm_processes_stopped 8660 * don't block if controlling proc already be stopped for cpr 8661 * canblock != PM_CANBLOCK_BYPASS 8662 * our caller must not have explicitly prevented blocking 8663 */ 8664 if (direction != PM_LEVEL_EXACT && canblock != PM_CANBLOCK_BYPASS) { 8665 PM_LOCK_DIP(dip); 8666 while (PM_ISDIRECT(dip) && !pm_processes_stopped) { 8667 /* releases dip lock */ 8668 ret = pm_busop_match_request(dip, bpc); 8669 if (ret == EAGAIN) { 8670 PM_LOCK_DIP(dip); 8671 continue; 8672 } 8673 return (*iresp = ret); 8674 } 8675 PM_UNLOCK_DIP(dip); 8676 } 8677 /* BC device is never scanned, so power will stick until we are done */ 8678 if (PM_ISBC(dip) && comp != 0 && nlevel != 0 && 8679 direction != PM_LEVEL_DOWNONLY) { 8680 int nrmpwr0 = pm_get_normal_power(dip, 0); 8681 if (pm_set_power(dip, 0, nrmpwr0, direction, 8682 canblock, 0, resultp) != DDI_SUCCESS) { 8683 /* *resultp set by pm_set_power */ 8684 return (DDI_FAILURE); 8685 } 8686 } 8687 if (PM_WANTS_NOTIFICATION(pdip)) { 8688 PMD(PMD_SET, ("%s: pre_notify %s@%s(%s#%d) for child " 8689 "%s@%s(%s#%d)\n", pmf, PM_DEVICE(pdip), PM_DEVICE(dip))) 8690 ret = (*PM_BUS_POWER_FUNC(pdip))(pdip, NULL, 8691 BUS_POWER_PRE_NOTIFICATION, bpc, resultp); 8692 if (ret != DDI_SUCCESS) { 8693 PMD(PMD_SET, ("%s: failed to pre_notify %s@%s(%s#%d)\n", 8694 pmf, PM_DEVICE(pdip))) 8695 return (DDI_FAILURE); 8696 } 8697 } else { 8698 /* 8699 * Since we don't know what the actual power level is, 8700 * we place a power hold on the parent no matter what 8701 * component and level is changing. 8702 */ 8703 pm_hold_power(pdip); 8704 } 8705 PM_LOCK_POWER(dip, &circ); 8706 clevel = PM_CURPOWER(dip, comp); 8707 /* 8708 * It's possible that a call was made to pm_update_maxpower() 8709 * on another thread before we took the lock above. So, we need to 8710 * make sure that this request isn't processed after the 8711 * change of power executed on behalf of pm_update_maxpower(). 8712 */ 8713 if (nlevel > pm_get_normal_power(dip, comp)) { 8714 PMD(PMD_SET, ("%s: requested level is higher than normal.\n", 8715 pmf)) 8716 ret = DDI_FAILURE; 8717 *iresp = DDI_FAILURE; 8718 goto post_notify; 8719 } 8720 PMD(PMD_SET, ("%s: %s@%s(%s#%d), cmp=%d, olvl=%d, nlvl=%d, clvl=%d, " 8721 "dir=%s\n", pmf, PM_DEVICE(dip), comp, bpc->bpc_olevel, nlevel, 8722 clevel, dir)) 8723 switch (direction) { 8724 case PM_LEVEL_UPONLY: 8725 /* Powering up */ 8726 if (clevel >= nlevel) { 8727 PMD(PMD_SET, ("%s: current level is already " 8728 "at or above the requested level.\n", pmf)) 8729 *iresp = DDI_SUCCESS; 8730 ret = DDI_SUCCESS; 8731 goto post_notify; 8732 } 8733 break; 8734 case PM_LEVEL_EXACT: 8735 /* specific level request */ 8736 if (clevel == nlevel && !PM_ISBC(dip)) { 8737 PMD(PMD_SET, ("%s: current level is already " 8738 "at the requested level.\n", pmf)) 8739 *iresp = DDI_SUCCESS; 8740 ret = DDI_SUCCESS; 8741 goto post_notify; 8742 } else if (PM_IS_CFB(dip) && (nlevel < clevel)) { 8743 PMD(PMD_CFB, ("%s: powerdown of console\n", pmf)) 8744 if (!pm_cfb_enabled) { 8745 PMD(PMD_ERROR | PMD_CFB, 8746 ("%s: !pm_cfb_enabled, fails\n", pmf)) 8747 *errnop = EINVAL; 8748 *iresp = DDI_FAILURE; 8749 ret = DDI_FAILURE; 8750 goto post_notify; 8751 } 8752 mutex_enter(&pm_cfb_lock); 8753 while (cfb_inuse) { 8754 mutex_exit(&pm_cfb_lock); 8755 if (delay_sig(1) == EINTR) { 8756 ret = DDI_FAILURE; 8757 *iresp = DDI_FAILURE; 8758 *errnop = EINTR; 8759 goto post_notify; 8760 } 8761 mutex_enter(&pm_cfb_lock); 8762 } 8763 mutex_exit(&pm_cfb_lock); 8764 } 8765 break; 8766 case PM_LEVEL_DOWNONLY: 8767 /* Powering down */ 8768 thresh = cur_threshold(dip, comp); 8769 idletime = gethrestime_sec() - cp->pmc_timestamp; 8770 if (scan && ((PM_KUC(dip) != 0) || 8771 (cp->pmc_busycount > 0) || 8772 ((idletime < thresh) && !PM_IS_PID(dip)))) { 8773 #ifdef DEBUG 8774 if (DEVI(dip)->devi_pm_kidsupcnt != 0) 8775 PMD(PMD_SET, ("%s: scan failed: " 8776 "kidsupcnt != 0\n", pmf)) 8777 if (cp->pmc_busycount > 0) 8778 PMD(PMD_SET, ("%s: scan failed: " 8779 "device become busy\n", pmf)) 8780 if (idletime < thresh) 8781 PMD(PMD_SET, ("%s: scan failed: device " 8782 "hasn't been idle long enough\n", pmf)) 8783 #endif 8784 *iresp = DDI_FAILURE; 8785 *errnop = EBUSY; 8786 ret = DDI_FAILURE; 8787 goto post_notify; 8788 } else if (clevel != PM_LEVEL_UNKNOWN && clevel <= nlevel) { 8789 PMD(PMD_SET, ("%s: current level is already at " 8790 "or below the requested level.\n", pmf)) 8791 *iresp = DDI_SUCCESS; 8792 ret = DDI_SUCCESS; 8793 goto post_notify; 8794 } 8795 break; 8796 } 8797 8798 if (PM_IS_CFB(dip) && (comps_off_incr = 8799 calc_cfb_comps_incr(dip, comp, clevel, nlevel)) > 0) { 8800 /* 8801 * Pre-adjust pm_cfb_comps_off if lowering a console fb 8802 * component from full power. Remember that we tried to 8803 * lower power in case it fails and we need to back out 8804 * the adjustment. 8805 */ 8806 update_comps_off(comps_off_incr, dip); 8807 PMD(PMD_CFB, ("%s: %s@%s(%s#%d)[%d] %d->%d cfb_comps_off->%d\n", 8808 pmf, PM_DEVICE(dip), comp, clevel, nlevel, 8809 pm_cfb_comps_off)) 8810 } 8811 8812 if ((*iresp = power_dev(dip, 8813 comp, nlevel, clevel, canblock, &devl)) == DDI_SUCCESS) { 8814 #ifdef DEBUG 8815 /* 8816 * All descendents of this node should already be powered off. 8817 */ 8818 if (PM_CURPOWER(dip, comp) == 0) { 8819 pm_desc_pwrchk_t pdpchk; 8820 pdpchk.pdpc_dip = dip; 8821 pdpchk.pdpc_par_involved = PM_WANTS_NOTIFICATION(dip); 8822 ndi_devi_enter(dip, &circ_db); 8823 for (cdip = ddi_get_child(dip); cdip != NULL; 8824 cdip = ddi_get_next_sibling(cdip)) { 8825 ndi_devi_enter(cdip, &ccirc_db); 8826 ddi_walk_devs(cdip, pm_desc_pwrchk_walk, 8827 (void *)&pdpchk); 8828 ndi_devi_exit(cdip, ccirc_db); 8829 } 8830 ndi_devi_exit(dip, circ_db); 8831 } 8832 #endif 8833 /* 8834 * Post-adjust pm_cfb_comps_off if we brought an fb component 8835 * back up to full power. 8836 */ 8837 if (PM_IS_CFB(dip) && comps_off_incr < 0) { 8838 update_comps_off(comps_off_incr, dip); 8839 PMD(PMD_CFB, ("%s: %s@%s(%s#%d)[%d] %d->%d " 8840 "cfb_comps_off->%d\n", pmf, PM_DEVICE(dip), 8841 comp, clevel, nlevel, pm_cfb_comps_off)) 8842 } 8843 dodeps = 0; 8844 if (POWERING_OFF(clevel, nlevel)) { 8845 if (PM_ISBC(dip)) { 8846 dodeps = (comp == 0); 8847 } else { 8848 int i; 8849 dodeps = 1; 8850 for (i = 0; i < PM_NUMCMPTS(dip); i++) { 8851 /* if some component still on */ 8852 if (PM_CURPOWER(dip, i)) { 8853 dodeps = 0; 8854 break; 8855 } 8856 } 8857 } 8858 if (dodeps) 8859 work_type = PM_DEP_WK_POWER_OFF; 8860 } else if (POWERING_ON(clevel, nlevel)) { 8861 if (PM_ISBC(dip)) { 8862 dodeps = (comp == 0); 8863 } else { 8864 int i; 8865 dodeps = 1; 8866 for (i = 0; i < PM_NUMCMPTS(dip); i++) { 8867 if (i == comp) 8868 continue; 8869 if (PM_CURPOWER(dip, i) > 0) { 8870 dodeps = 0; 8871 break; 8872 } 8873 } 8874 } 8875 if (dodeps) 8876 work_type = PM_DEP_WK_POWER_ON; 8877 } 8878 8879 if (dodeps) { 8880 char *pathbuf = kmem_alloc(MAXPATHLEN, KM_SLEEP); 8881 8882 (void) ddi_pathname(dip, pathbuf); 8883 pm_dispatch_to_dep_thread(work_type, pathbuf, NULL, 8884 PM_DEP_NOWAIT, NULL, 0); 8885 kmem_free(pathbuf, MAXPATHLEN); 8886 } 8887 if ((PM_CURPOWER(dip, comp) == nlevel) && pm_watchers()) { 8888 int old; 8889 8890 /* If old power cached during deadlock, use it. */ 8891 old = (cp->pmc_flags & PM_PHC_WHILE_SET_POWER ? 8892 cp->pmc_phc_pwr : olevel); 8893 mutex_enter(&pm_rsvp_lock); 8894 pm_enqueue_notify(PSC_HAS_CHANGED, dip, comp, nlevel, 8895 old, canblock); 8896 pm_enqueue_notify_others(&devl, canblock); 8897 mutex_exit(&pm_rsvp_lock); 8898 } else { 8899 pm_ppm_devlist_t *p; 8900 pm_ppm_devlist_t *next; 8901 for (p = devl; p != NULL; p = next) { 8902 next = p->ppd_next; 8903 kmem_free(p, sizeof (pm_ppm_devlist_t)); 8904 } 8905 devl = NULL; 8906 } 8907 8908 /* 8909 * If we are coming from a scan, don't do it again, 8910 * else we can have infinite loops. 8911 */ 8912 if (!scan) 8913 pm_rescan(dip); 8914 } else { 8915 /* if we incremented pm_comps_off_count, but failed */ 8916 if (comps_off_incr > 0) { 8917 update_comps_off(-comps_off_incr, dip); 8918 PMD(PMD_CFB, ("%s: %s@%s(%s#%d)[%d] %d->%d " 8919 "cfb_comps_off->%d\n", pmf, PM_DEVICE(dip), 8920 comp, clevel, nlevel, pm_cfb_comps_off)) 8921 } 8922 *errnop = EIO; 8923 } 8924 8925 post_notify: 8926 /* 8927 * This thread may have been in deadlock with pm_power_has_changed. 8928 * Before releasing power lock, clear the flag which marks this 8929 * condition. 8930 */ 8931 cp->pmc_flags &= ~PM_PHC_WHILE_SET_POWER; 8932 8933 /* 8934 * Update the old power level in the bus power structure with the 8935 * actual power level before the transition was made to the new level. 8936 * Some involved parents depend on this information to keep track of 8937 * their children's power transition. 8938 */ 8939 if (*iresp != DDI_FAILURE) 8940 bpc->bpc_olevel = clevel; 8941 8942 if (PM_WANTS_NOTIFICATION(pdip)) { 8943 ret = (*PM_BUS_POWER_FUNC(pdip))(pdip, NULL, 8944 BUS_POWER_POST_NOTIFICATION, bpc, resultp); 8945 PM_UNLOCK_POWER(dip, circ); 8946 PMD(PMD_SET, ("%s: post_notify %s@%s(%s#%d) for " 8947 "child %s@%s(%s#%d), ret=%d\n", pmf, PM_DEVICE(pdip), 8948 PM_DEVICE(dip), ret)) 8949 } else { 8950 nlevel = cur_power(cp); /* in case phc deadlock updated pwr */ 8951 PM_UNLOCK_POWER(dip, circ); 8952 /* 8953 * Now that we know what power transition has occurred 8954 * (if any), release the power hold. Leave the hold 8955 * in effect in the case of OFF->ON transition. 8956 */ 8957 if (!(clevel == 0 && nlevel > 0 && 8958 (!PM_ISBC(dip) || comp == 0))) 8959 pm_rele_power(pdip); 8960 /* 8961 * If the power transition was an ON->OFF transition, 8962 * remove the power hold from the parent. 8963 */ 8964 if ((clevel > 0 || clevel == PM_LEVEL_UNKNOWN) && 8965 nlevel == 0 && (!PM_ISBC(dip) || comp == 0)) 8966 pm_rele_power(pdip); 8967 } 8968 if (*iresp != DDI_SUCCESS || ret != DDI_SUCCESS) 8969 return (DDI_FAILURE); 8970 else 8971 return (DDI_SUCCESS); 8972 } 8973 8974 /* 8975 * If an app (SunVTS or Xsun) has taken control, then block until it 8976 * gives it up or makes the requested power level change, unless 8977 * we have other instructions about blocking. Returns DDI_SUCCESS, 8978 * DDI_FAILURE or EAGAIN (owner released device from directpm). 8979 */ 8980 static int 8981 pm_busop_match_request(dev_info_t *dip, void *arg) 8982 { 8983 PMD_FUNC(pmf, "bp_match_request") 8984 pm_bp_child_pwrchg_t *bpc = (pm_bp_child_pwrchg_t *)arg; 8985 pm_sp_misc_t *pspm = (pm_sp_misc_t *)bpc->bpc_private; 8986 int comp = bpc->bpc_comp; 8987 int nlevel = bpc->bpc_nlevel; 8988 pm_canblock_t canblock = pspm->pspm_canblock; 8989 int direction = pspm->pspm_direction; 8990 int clevel, circ; 8991 8992 ASSERT(PM_IAM_LOCKING_DIP(dip)); 8993 PM_LOCK_POWER(dip, &circ); 8994 clevel = PM_CURPOWER(dip, comp); 8995 PMD(PMD_SET, ("%s: %s@%s(%s#%d), cmp=%d, nlvl=%d, clvl=%d\n", 8996 pmf, PM_DEVICE(dip), comp, nlevel, clevel)) 8997 if (direction == PM_LEVEL_UPONLY) { 8998 if (clevel >= nlevel) { 8999 PM_UNLOCK_POWER(dip, circ); 9000 PM_UNLOCK_DIP(dip); 9001 return (DDI_SUCCESS); 9002 } 9003 } else if (clevel == nlevel) { 9004 PM_UNLOCK_POWER(dip, circ); 9005 PM_UNLOCK_DIP(dip); 9006 return (DDI_SUCCESS); 9007 } 9008 if (canblock == PM_CANBLOCK_FAIL) { 9009 PM_UNLOCK_POWER(dip, circ); 9010 PM_UNLOCK_DIP(dip); 9011 return (DDI_FAILURE); 9012 } 9013 if (canblock == PM_CANBLOCK_BLOCK) { 9014 /* 9015 * To avoid a deadlock, we must not hold the 9016 * power lock when we pm_block. 9017 */ 9018 PM_UNLOCK_POWER(dip, circ); 9019 PMD(PMD_SET, ("%s: blocking\n", pmf)) 9020 /* pm_block releases dip lock */ 9021 switch (pm_block(dip, comp, nlevel, clevel)) { 9022 case PMP_RELEASE: 9023 return (EAGAIN); 9024 case PMP_SUCCEED: 9025 return (DDI_SUCCESS); 9026 case PMP_FAIL: 9027 return (DDI_FAILURE); 9028 } 9029 } else { 9030 ASSERT(0); 9031 } 9032 _NOTE(NOTREACHED); 9033 return (DDI_FAILURE); /* keep gcc happy */ 9034 } 9035 9036 static int 9037 pm_all_to_normal_nexus(dev_info_t *dip, pm_canblock_t canblock) 9038 { 9039 PMD_FUNC(pmf, "all_to_normal_nexus") 9040 int *normal; 9041 int i, ncomps; 9042 size_t size; 9043 int changefailed = 0; 9044 int ret, result = DDI_SUCCESS; 9045 pm_bp_nexus_pwrup_t bpn; 9046 pm_sp_misc_t pspm; 9047 9048 ASSERT(PM_GET_PM_INFO(dip)); 9049 PMD(PMD_ALLNORM, ("%s: %s@%s(%s#%d)\n", pmf, PM_DEVICE(dip))) 9050 if (pm_get_norm_pwrs(dip, &normal, &size) != DDI_SUCCESS) { 9051 PMD(PMD_ALLNORM, ("%s: can't get norm pwrs\n", pmf)) 9052 return (DDI_FAILURE); 9053 } 9054 ncomps = PM_NUMCMPTS(dip); 9055 for (i = 0; i < ncomps; i++) { 9056 bpn.bpn_dip = dip; 9057 bpn.bpn_comp = i; 9058 bpn.bpn_level = normal[i]; 9059 pspm.pspm_canblock = canblock; 9060 pspm.pspm_scan = 0; 9061 bpn.bpn_private = &pspm; 9062 ret = pm_busop_bus_power(dip, NULL, BUS_POWER_NEXUS_PWRUP, 9063 (void *)&bpn, (void *)&result); 9064 if (ret != DDI_SUCCESS || result != DDI_SUCCESS) { 9065 PMD(PMD_FAIL | PMD_ALLNORM, ("%s: %s@%s(%s#%d)[%d] " 9066 "->%d failure result %d\n", pmf, PM_DEVICE(dip), 9067 i, normal[i], result)) 9068 changefailed++; 9069 } 9070 } 9071 kmem_free(normal, size); 9072 if (changefailed) { 9073 PMD(PMD_FAIL, ("%s: failed to set %d comps %s@%s(%s#%d) " 9074 "full power\n", pmf, changefailed, PM_DEVICE(dip))) 9075 return (DDI_FAILURE); 9076 } 9077 return (DDI_SUCCESS); 9078 } 9079 9080 int 9081 pm_noinvol_update(int subcmd, int volpmd, int wasvolpmd, char *path, 9082 dev_info_t *tdip) 9083 { 9084 PMD_FUNC(pmf, "noinvol_update") 9085 pm_bp_noinvol_t args; 9086 int ret; 9087 int result = DDI_SUCCESS; 9088 9089 args.bpni_path = path; 9090 args.bpni_dip = tdip; 9091 args.bpni_cmd = subcmd; 9092 args.bpni_wasvolpmd = wasvolpmd; 9093 args.bpni_volpmd = volpmd; 9094 PMD(PMD_NOINVOL, ("%s: update for path %s tdip %p subcmd %d " 9095 "volpmd %d wasvolpmd %d\n", pmf, 9096 path, (void *)tdip, subcmd, wasvolpmd, volpmd)) 9097 ret = pm_busop_bus_power(ddi_root_node(), NULL, BUS_POWER_NOINVOL, 9098 &args, &result); 9099 return (ret); 9100 } 9101 9102 void 9103 pm_noinvol_update_node(dev_info_t *dip, pm_bp_noinvol_t *req) 9104 { 9105 PMD_FUNC(pmf, "noinvol_update_node") 9106 9107 PMD(PMD_NOINVOL, ("%s: %s@%s(%s#%d)\n", pmf, PM_DEVICE(dip))) 9108 switch (req->bpni_cmd) { 9109 case PM_BP_NOINVOL_ATTACH: 9110 PMD(PMD_NOINVOL, ("%s: PM_PB_NOINVOL_ATTACH %s@%s(%s#%d) " 9111 "noinvol %d->%d\n", pmf, PM_DEVICE(dip), 9112 DEVI(dip)->devi_pm_noinvolpm, 9113 DEVI(dip)->devi_pm_noinvolpm - 1)) 9114 ASSERT(DEVI(dip)->devi_pm_noinvolpm); 9115 PM_LOCK_DIP(dip); 9116 DEVI(dip)->devi_pm_noinvolpm--; 9117 if (req->bpni_wasvolpmd) { 9118 PMD(PMD_NOINVOL, ("%s: PM_BP_NOINVOL_ATTACH " 9119 "%s@%s(%s#%d) volpmd %d->%d\n", pmf, 9120 PM_DEVICE(dip), DEVI(dip)->devi_pm_volpmd, 9121 DEVI(dip)->devi_pm_volpmd - 1)) 9122 if (DEVI(dip)->devi_pm_volpmd) 9123 DEVI(dip)->devi_pm_volpmd--; 9124 } 9125 PM_UNLOCK_DIP(dip); 9126 break; 9127 9128 case PM_BP_NOINVOL_DETACH: 9129 PMD(PMD_NOINVOL, ("%s: PM_BP_NOINVOL_DETACH %s@%s(%s#%d) " 9130 "noinvolpm %d->%d\n", pmf, PM_DEVICE(dip), 9131 DEVI(dip)->devi_pm_noinvolpm, 9132 DEVI(dip)->devi_pm_noinvolpm + 1)) 9133 PM_LOCK_DIP(dip); 9134 DEVI(dip)->devi_pm_noinvolpm++; 9135 if (req->bpni_wasvolpmd) { 9136 PMD(PMD_NOINVOL, ("%s: PM_BP_NOINVOL_DETACH " 9137 "%s@%s(%s#%d) volpmd %d->%d\n", pmf, 9138 PM_DEVICE(dip), DEVI(dip)->devi_pm_volpmd, 9139 DEVI(dip)->devi_pm_volpmd + 1)) 9140 DEVI(dip)->devi_pm_volpmd++; 9141 } 9142 PM_UNLOCK_DIP(dip); 9143 break; 9144 9145 case PM_BP_NOINVOL_REMDRV: 9146 PMD(PMD_NOINVOL, ("%s: PM_BP_NOINVOL_REMDRV %s@%s(%s#%d) " 9147 "noinvol %d->%d\n", pmf, PM_DEVICE(dip), 9148 DEVI(dip)->devi_pm_noinvolpm, 9149 DEVI(dip)->devi_pm_noinvolpm - 1)) 9150 ASSERT(DEVI(dip)->devi_pm_noinvolpm); 9151 PM_LOCK_DIP(dip); 9152 DEVI(dip)->devi_pm_noinvolpm--; 9153 if (req->bpni_wasvolpmd) { 9154 PMD(PMD_NOINVOL, 9155 ("%s: PM_BP_NOINVOL_REMDRV %s@%s(%s#%d) " 9156 "volpmd %d->%d\n", pmf, PM_DEVICE(dip), 9157 DEVI(dip)->devi_pm_volpmd, 9158 DEVI(dip)->devi_pm_volpmd - 1)) 9159 /* 9160 * A power up could come in between and 9161 * clear the volpmd, if that's the case, 9162 * volpmd would be clear. 9163 */ 9164 if (DEVI(dip)->devi_pm_volpmd) 9165 DEVI(dip)->devi_pm_volpmd--; 9166 } 9167 PM_UNLOCK_DIP(dip); 9168 break; 9169 9170 case PM_BP_NOINVOL_CFB: 9171 PMD(PMD_NOINVOL, 9172 ("%s: PM_BP_NOIVOL_CFB %s@%s(%s#%d) noinvol %d->%d\n", 9173 pmf, PM_DEVICE(dip), DEVI(dip)->devi_pm_noinvolpm, 9174 DEVI(dip)->devi_pm_noinvolpm + 1)) 9175 PM_LOCK_DIP(dip); 9176 DEVI(dip)->devi_pm_noinvolpm++; 9177 PM_UNLOCK_DIP(dip); 9178 break; 9179 9180 case PM_BP_NOINVOL_POWER: 9181 PMD(PMD_NOINVOL, 9182 ("%s: PM_BP_NOIVOL_PWR %s@%s(%s#%d) volpmd %d->%d\n", 9183 pmf, PM_DEVICE(dip), 9184 DEVI(dip)->devi_pm_volpmd, DEVI(dip)->devi_pm_volpmd - 9185 req->bpni_volpmd)) 9186 PM_LOCK_DIP(dip); 9187 DEVI(dip)->devi_pm_volpmd -= req->bpni_volpmd; 9188 PM_UNLOCK_DIP(dip); 9189 break; 9190 9191 default: 9192 break; 9193 } 9194 9195 } 9196 9197 #ifdef DEBUG 9198 static int 9199 pm_desc_pwrchk_walk(dev_info_t *dip, void *arg) 9200 { 9201 PMD_FUNC(pmf, "desc_pwrchk") 9202 pm_desc_pwrchk_t *pdpchk = (pm_desc_pwrchk_t *)arg; 9203 pm_info_t *info = PM_GET_PM_INFO(dip); 9204 int i; 9205 /* LINTED */ 9206 int curpwr, ce_level; 9207 9208 if (!info) 9209 return (DDI_WALK_CONTINUE); 9210 9211 PMD(PMD_SET, ("%s: %s@%s(%s#%d)\n", pmf, PM_DEVICE(dip))) 9212 for (i = 0; i < PM_NUMCMPTS(dip); i++) { 9213 /* LINTED */ 9214 if ((curpwr = PM_CURPOWER(dip, i)) == 0) 9215 continue; 9216 /* E_FUNC_SET_NOT_USED */ 9217 ce_level = (pdpchk->pdpc_par_involved == 0) ? CE_PANIC : 9218 CE_WARN; 9219 PMD(PMD_SET, ("%s: %s@%s(%s#%d) is powered off while desc " 9220 "%s@%s(%s#%d)[%d] is at %d\n", pmf, 9221 PM_DEVICE(pdpchk->pdpc_dip), PM_DEVICE(dip), i, curpwr)) 9222 cmn_err(ce_level, "!device %s@%s(%s#%d) is powered on, " 9223 "while its ancestor, %s@%s(%s#%d), is powering off!", 9224 PM_DEVICE(dip), PM_DEVICE(pdpchk->pdpc_dip)); 9225 } 9226 return (DDI_WALK_CONTINUE); 9227 } 9228 #endif 9229 9230 /* 9231 * Record the fact that one thread is borrowing the lock on a device node. 9232 * Use is restricted to the case where the lending thread will block until 9233 * the borrowing thread (always curthread) completes. 9234 */ 9235 void 9236 pm_borrow_lock(kthread_t *lender) 9237 { 9238 lock_loan_t *prev = &lock_loan_head; 9239 lock_loan_t *cur = (lock_loan_t *)kmem_zalloc(sizeof (*cur), KM_SLEEP); 9240 9241 cur->pmlk_borrower = curthread; 9242 cur->pmlk_lender = lender; 9243 mutex_enter(&pm_loan_lock); 9244 cur->pmlk_next = prev->pmlk_next; 9245 prev->pmlk_next = cur; 9246 mutex_exit(&pm_loan_lock); 9247 } 9248 9249 /* 9250 * Return the borrowed lock. A thread can borrow only one. 9251 */ 9252 void 9253 pm_return_lock(void) 9254 { 9255 lock_loan_t *cur; 9256 lock_loan_t *prev = &lock_loan_head; 9257 9258 mutex_enter(&pm_loan_lock); 9259 ASSERT(prev->pmlk_next != NULL); 9260 for (cur = prev->pmlk_next; cur; prev = cur, cur = cur->pmlk_next) 9261 if (cur->pmlk_borrower == curthread) 9262 break; 9263 9264 ASSERT(cur != NULL); 9265 prev->pmlk_next = cur->pmlk_next; 9266 mutex_exit(&pm_loan_lock); 9267 kmem_free(cur, sizeof (*cur)); 9268 } 9269 9270 #if defined(__x86) 9271 9272 #define CPR_RXR 0x1 9273 #define CPR_TXR 0x20 9274 #define CPR_DATAREG 0x3f8 9275 #define CPR_LSTAT 0x3fd 9276 #define CPR_INTRCTL 0x3f9 9277 9278 char 9279 pm_getchar(void) 9280 { 9281 while ((inb(CPR_LSTAT) & CPR_RXR) != CPR_RXR) 9282 drv_usecwait(10); 9283 9284 return (inb(CPR_DATAREG)); 9285 9286 } 9287 9288 void 9289 pm_putchar(char c) 9290 { 9291 while ((inb(CPR_LSTAT) & CPR_TXR) == 0) 9292 drv_usecwait(10); 9293 9294 outb(CPR_DATAREG, c); 9295 } 9296 9297 void 9298 pm_printf(char *s) 9299 { 9300 while (*s) { 9301 pm_putchar(*s++); 9302 } 9303 } 9304 9305 #endif 9306 9307 int 9308 pm_ppm_searchlist(pm_searchargs_t *sp) 9309 { 9310 power_req_t power_req; 9311 int result = 0; 9312 /* LINTED */ 9313 int ret; 9314 9315 power_req.request_type = PMR_PPM_SEARCH_LIST; 9316 power_req.req.ppm_search_list_req.searchlist = sp; 9317 ASSERT(DEVI(ddi_root_node())->devi_pm_ppm); 9318 ret = pm_ctlops((dev_info_t *)DEVI(ddi_root_node())->devi_pm_ppm, 9319 ddi_root_node(), DDI_CTLOPS_POWER, &power_req, &result); 9320 PMD(PMD_SX, ("pm_ppm_searchlist returns %d, result %d\n", 9321 ret, result)) 9322 return (result); 9323 } 9324