xref: /titanic_44/usr/src/uts/common/io/pm.c (revision 0e7515250c8395f368aa45fb9acae7c4f8f8b786)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2009 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  */
25 
26 /*
27  * pm	This driver now only handles the ioctl interface.  The scanning
28  *	and policy stuff now lives in common/os/sunpm.c.
29  *	Not DDI compliant
30  */
31 
32 #include <sys/types.h>
33 #include <sys/errno.h>
34 #include <sys/modctl.h>
35 #include <sys/callb.h>		/* callback registration for cpu_deep_idle */
36 #include <sys/conf.h>		/* driver flags and functions */
37 #include <sys/open.h>		/* OTYP_CHR definition */
38 #include <sys/stat.h>		/* S_IFCHR definition */
39 #include <sys/pathname.h>	/* name -> dev_info xlation */
40 #include <sys/kmem.h>		/* memory alloc stuff */
41 #include <sys/debug.h>
42 #include <sys/pm.h>
43 #include <sys/ddi.h>
44 #include <sys/sunddi.h>
45 #include <sys/epm.h>
46 #include <sys/vfs.h>
47 #include <sys/mode.h>
48 #include <sys/mkdev.h>
49 #include <sys/promif.h>
50 #include <sys/consdev.h>
51 #include <sys/ddi_impldefs.h>
52 #include <sys/poll.h>
53 #include <sys/note.h>
54 #include <sys/taskq.h>
55 #include <sys/policy.h>
56 #include <sys/cpu_pm.h>
57 
58 /*
59  * Minor number is instance<<8 + clone minor from range 1-254; (0 reserved
60  * for "original")
61  */
62 #define	PM_MINOR_TO_CLONE(minor) ((minor) & (PM_MAX_CLONE -1))
63 
64 #define	PM_NUMCMPTS(dip) (DEVI(dip)->devi_pm_num_components)
65 #define	PM_IS_CFB(dip) (DEVI(dip)->devi_pm_flags & PMC_CONSOLE_FB)
66 #define	PM_MAJOR(dip) ddi_driver_major(dip)
67 #define	PM_RELE(dip) ddi_release_devi(dip)
68 
69 #define	PM_IDLEDOWN_TIME	10
70 #define	MAXSMBIOSSTRLEN 64	/* from SMBIOS spec */
71 #define	MAXCOPYBUF 	(MAXSMBIOSSTRLEN + 1)
72 
73 extern kmutex_t	pm_scan_lock;	/* protects autopm_enable, pm_scans_disabled */
74 extern kmutex_t	pm_clone_lock;	/* protects pm_clones array */
75 extern int	autopm_enabled;
76 extern pm_cpupm_t cpupm;
77 extern pm_cpupm_t cpupm_default_mode;
78 extern int	pm_default_idle_threshold;
79 extern int	pm_system_idle_threshold;
80 extern int	pm_cpu_idle_threshold;
81 extern kcondvar_t pm_clones_cv[PM_MAX_CLONE];
82 extern uint_t	pm_poll_cnt[PM_MAX_CLONE];
83 extern int	autoS3_enabled;
84 extern void	pm_record_thresh(pm_thresh_rec_t *);
85 extern void	pm_register_watcher(int, dev_info_t *);
86 extern int	pm_get_current_power(dev_info_t *, int, int *);
87 extern int	pm_interest_registered(int);
88 extern void	pm_all_to_default_thresholds(void);
89 extern int	pm_current_threshold(dev_info_t *, int, int *);
90 extern void	pm_deregister_watcher(int, dev_info_t *);
91 extern void	pm_unrecord_threshold(char *);
92 extern int	pm_S3_enabled;
93 extern int	pm_ppm_searchlist(pm_searchargs_t *);
94 extern psce_t	*pm_psc_clone_to_direct(int);
95 extern psce_t	*pm_psc_clone_to_interest(int);
96 
97 /*
98  * The soft state of the power manager.  Since there will only
99  * one of these, just reference it through a static pointer.
100  */
101 static struct pmstate {
102 	dev_info_t	*pm_dip;		/* ptr to our dev_info node */
103 	int		pm_instance;		/* for ddi_get_instance() */
104 	timeout_id_t	pm_idledown_id;		/* pm idledown timeout id */
105 	uchar_t		pm_clones[PM_MAX_CLONE]; /* uniqueify multiple opens */
106 	struct cred	*pm_cred[PM_MAX_CLONE];	/* cred for each unique open */
107 } pm_state = { NULL, -1, (timeout_id_t)0 };
108 typedef struct pmstate *pm_state_t;
109 static pm_state_t pmstp = &pm_state;
110 
111 static int	pm_open(dev_t *, int, int, cred_t *);
112 static int	pm_close(dev_t, int, int, cred_t *);
113 static int	pm_ioctl(dev_t, int, intptr_t, int, cred_t *, int *);
114 static int	pm_chpoll(dev_t, short, int, short *, struct pollhead **);
115 
116 static struct cb_ops pm_cb_ops = {
117 	pm_open,	/* open */
118 	pm_close,	/* close */
119 	nodev,		/* strategy */
120 	nodev,		/* print */
121 	nodev,		/* dump */
122 	nodev,		/* read */
123 	nodev,		/* write */
124 	pm_ioctl,	/* ioctl */
125 	nodev,		/* devmap */
126 	nodev,		/* mmap */
127 	nodev,		/* segmap */
128 	pm_chpoll,	/* poll */
129 	ddi_prop_op,	/* prop_op */
130 	NULL,		/* streamtab */
131 	D_NEW | D_MP	/* driver compatibility flag */
132 };
133 
134 static int pm_getinfo(dev_info_t *dip, ddi_info_cmd_t infocmd, void *arg,
135     void **result);
136 static int pm_attach(dev_info_t *dip, ddi_attach_cmd_t cmd);
137 static int pm_detach(dev_info_t *dip, ddi_detach_cmd_t cmd);
138 
139 static struct dev_ops pm_ops = {
140 	DEVO_REV,		/* devo_rev */
141 	0,			/* refcnt */
142 	pm_getinfo,		/* info */
143 	nulldev,		/* identify */
144 	nulldev,		/* probe */
145 	pm_attach,		/* attach */
146 	pm_detach,		/* detach */
147 	nodev,			/* reset */
148 	&pm_cb_ops,		/* driver operations */
149 	NULL,			/* bus operations */
150 	NULL,			/* power */
151 	ddi_quiesce_not_needed,		/* quiesce */
152 };
153 
154 static struct modldrv modldrv = {
155 	&mod_driverops,
156 	"power management driver",
157 	&pm_ops
158 };
159 
160 static struct modlinkage modlinkage = {
161 	MODREV_1, &modldrv, 0
162 };
163 
164 /* Local functions */
165 #ifdef DEBUG
166 static int	print_info(dev_info_t *, void *);
167 
168 #endif
169 
170 int
_init(void)171 _init(void)
172 {
173 	return (mod_install(&modlinkage));
174 }
175 
176 int
_fini(void)177 _fini(void)
178 {
179 	return (mod_remove(&modlinkage));
180 }
181 
182 int
_info(struct modinfo * modinfop)183 _info(struct modinfo *modinfop)
184 {
185 	return (mod_info(&modlinkage, modinfop));
186 }
187 
188 static int
pm_attach(dev_info_t * dip,ddi_attach_cmd_t cmd)189 pm_attach(dev_info_t *dip, ddi_attach_cmd_t cmd)
190 {
191 	int		i;
192 
193 	switch (cmd) {
194 
195 	case DDI_ATTACH:
196 		if (pmstp->pm_instance != -1)	/* Only allow one instance */
197 			return (DDI_FAILURE);
198 		pmstp->pm_instance = ddi_get_instance(dip);
199 		if (ddi_create_minor_node(dip, "pm", S_IFCHR,
200 		    (pmstp->pm_instance << 8) + 0,
201 		    DDI_PSEUDO, 0) != DDI_SUCCESS) {
202 			return (DDI_FAILURE);
203 		}
204 		pmstp->pm_dip = dip;	/* pm_init and getinfo depend on it */
205 
206 		for (i = 0; i < PM_MAX_CLONE; i++)
207 			cv_init(&pm_clones_cv[i], NULL, CV_DEFAULT, NULL);
208 
209 		ddi_report_dev(dip);
210 		return (DDI_SUCCESS);
211 
212 	default:
213 		return (DDI_FAILURE);
214 	}
215 }
216 
217 /* ARGSUSED */
218 static int
pm_detach(dev_info_t * dip,ddi_detach_cmd_t cmd)219 pm_detach(dev_info_t *dip, ddi_detach_cmd_t cmd)
220 {
221 	int i;
222 
223 	switch (cmd) {
224 	case DDI_DETACH:
225 		/*
226 		 * Don't detach while idledown timeout is pending.  Note that
227 		 * we already know we're not in pm_ioctl() due to framework
228 		 * synchronization, so this is a sufficient test
229 		 */
230 		if (pmstp->pm_idledown_id)
231 			return (DDI_FAILURE);
232 
233 		for (i = 0; i < PM_MAX_CLONE; i++)
234 			cv_destroy(&pm_clones_cv[i]);
235 
236 		ddi_remove_minor_node(dip, NULL);
237 		pmstp->pm_instance = -1;
238 		return (DDI_SUCCESS);
239 
240 	default:
241 		return (DDI_FAILURE);
242 	}
243 }
244 
245 static int
pm_close_direct_pm_device(dev_info_t * dip,void * arg)246 pm_close_direct_pm_device(dev_info_t *dip, void *arg)
247 {
248 	int clone;
249 	char *pathbuf;
250 	pm_info_t *info = PM_GET_PM_INFO(dip);
251 
252 	clone = *((int *)arg);
253 
254 	if (!info)
255 		return (DDI_WALK_CONTINUE);
256 
257 	pathbuf = kmem_alloc(MAXPATHLEN, KM_SLEEP);
258 	PM_LOCK_DIP(dip);
259 	if (clone == info->pmi_clone) {
260 		PMD(PMD_CLOSE, ("pm_close: found %s@%s(%s#%d)\n",
261 		    PM_DEVICE(dip)))
262 		ASSERT(PM_ISDIRECT(dip));
263 		info->pmi_dev_pm_state &= ~PM_DIRECT;
264 		PM_UNLOCK_DIP(dip);
265 		pm_proceed(dip, PMP_RELEASE, -1, -1);
266 		/* Bring ourselves up if there is a keeper that is up */
267 		(void) ddi_pathname(dip, pathbuf);
268 		pm_dispatch_to_dep_thread(PM_DEP_WK_BRINGUP_SELF, NULL,
269 		    pathbuf, PM_DEP_NOWAIT, NULL, 0);
270 		PM_LOCK_DIP(dip);
271 		info->pmi_clone = 0;
272 		PM_UNLOCK_DIP(dip);
273 	} else {
274 		PM_UNLOCK_DIP(dip);
275 	}
276 	kmem_free(pathbuf, MAXPATHLEN);
277 
278 	/* restart autopm on device released from direct pm */
279 	pm_rescan(dip);
280 
281 	return (DDI_WALK_CONTINUE);
282 }
283 
284 #define	PM_REQ		1
285 #define	NOSTRUCT	2
286 #define	DIP		3
287 #define	NODIP		4
288 #define	NODEP		5
289 #define	DEP		6
290 #define	PM_PSC		7
291 #define	PM_SRCH		8
292 
293 #define	CHECKPERMS	0x001
294 #define	SU		0x002
295 #define	SG		0x004
296 #define	OWNER		0x008
297 
298 #define	INWHO		0x001
299 #define	INDATAINT	0x002
300 #define	INDATASTRING	0x004
301 #define	INDEP		0x008
302 #define	INDATAOUT	0x010
303 #define	INDATA	(INDATAOUT | INDATAINT | INDATASTRING | INDEP)
304 
305 struct pm_cmd_info {
306 	int cmd;		/* command code */
307 	char *name;		/* printable string */
308 	int supported;		/* true if still supported */
309 	int str_type;		/* PM_REQ or NOSTRUCT */
310 	int inargs;		/* INWHO, INDATAINT, INDATASTRING, INDEP, */
311 				/* INDATAOUT */
312 	int diptype;		/* DIP or NODIP */
313 	int deptype;		/* DEP or NODEP */
314 	int permission;		/* SU, GU, or CHECKPERMS */
315 };
316 
317 #ifdef DEBUG
318 char *pm_cmd_string;
319 int pm_cmd;
320 #endif
321 
322 /*
323  * Returns true if permission granted by credentials
324  */
325 static int
pm_perms(int perm,cred_t * cr)326 pm_perms(int perm, cred_t *cr)
327 {
328 	if (perm == 0)			/* no restrictions */
329 		return (1);
330 	if (perm == CHECKPERMS)		/* ok for now (is checked later) */
331 		return (1);
332 	if ((perm & SU) && secpolicy_power_mgmt(cr) == 0) /* privileged? */
333 		return (1);
334 	if ((perm & SG) && (crgetgid(cr) == 0))	/* group 0 is ok */
335 		return (1);
336 	return (0);
337 }
338 
339 #ifdef DEBUG
340 static int
print_info(dev_info_t * dip,void * arg)341 print_info(dev_info_t *dip, void *arg)
342 {
343 	_NOTE(ARGUNUSED(arg))
344 	pm_info_t	*info;
345 	int		i, j;
346 	struct pm_component *cp;
347 	extern int pm_cur_power(pm_component_t *cp);
348 
349 	info = PM_GET_PM_INFO(dip);
350 	if (!info)
351 		return (DDI_WALK_CONTINUE);
352 	cmn_err(CE_CONT, "pm_info for %s\n", ddi_node_name(dip));
353 	for (i = 0; i < PM_NUMCMPTS(dip); i++) {
354 		cp = PM_CP(dip, i);
355 		cmn_err(CE_CONT, "\tThresholds[%d] =",  i);
356 		for (j = 0; j < cp->pmc_comp.pmc_numlevels; j++)
357 			cmn_err(CE_CONT, " %d", cp->pmc_comp.pmc_thresh[i]);
358 		cmn_err(CE_CONT, "\n");
359 		cmn_err(CE_CONT, "\tCurrent power[%d] = %d\n", i,
360 		    pm_cur_power(cp));
361 	}
362 	if (PM_ISDIRECT(dip))
363 		cmn_err(CE_CONT, "\tDirect power management\n");
364 	return (DDI_WALK_CONTINUE);
365 }
366 #endif
367 
368 /*
369  * command, name, supported, str_type, inargs, diptype, deptype, permission
370  */
371 static struct pm_cmd_info pmci[] = {
372 	{PM_SCHEDULE, "PM_SCHEDULE", 0},
373 	{PM_GET_IDLE_TIME, "PM_GET_IDLE_TIME", 0},
374 	{PM_GET_NUM_CMPTS, "PM_GET_NUM_CMPTS", 0},
375 	{PM_GET_THRESHOLD, "PM_GET_THRESHOLD", 0},
376 	{PM_SET_THRESHOLD, "PM_SET_THRESHOLD", 0},
377 	{PM_GET_NORM_PWR, "PM_GET_NORM_PWR", 0},
378 	{PM_SET_CUR_PWR, "PM_SET_CUR_PWR", 0},
379 	{PM_GET_CUR_PWR, "PM_GET_CUR_PWR", 0},
380 	{PM_GET_NUM_DEPS, "PM_GET_NUM_DEPS", 0},
381 	{PM_GET_DEP, "PM_GET_DEP", 0},
382 	{PM_ADD_DEP, "PM_ADD_DEP", 0},
383 	{PM_REM_DEP, "PM_REM_DEP", 0},
384 	{PM_REM_DEVICE, "PM_REM_DEVICE", 0},
385 	{PM_REM_DEVICES, "PM_REM_DEVICES", 0},
386 	{PM_REPARSE_PM_PROPS, "PM_REPARSE_PM_PROPS", 1, PM_REQ, INWHO, DIP,
387 	    NODEP},
388 	{PM_DISABLE_AUTOPM, "PM_DISABLE_AUTOPM", 0},
389 	{PM_REENABLE_AUTOPM, "PM_REENABLE_AUTOPM", 0},
390 	{PM_SET_NORM_PWR, "PM_SET_NORM_PWR", 0 },
391 	{PM_SET_DEVICE_THRESHOLD, "PM_SET_DEVICE_THRESHOLD", 1, PM_REQ,
392 	    INWHO, NODIP, NODEP, SU},
393 	{PM_GET_SYSTEM_THRESHOLD, "PM_GET_SYSTEM_THRESHOLD", 1, NOSTRUCT},
394 	{PM_GET_DEFAULT_SYSTEM_THRESHOLD, "PM_GET_DEFAULT_SYSTEM_THRESHOLD",
395 	    1, NOSTRUCT},
396 	{PM_SET_SYSTEM_THRESHOLD, "PM_SET_SYSTEM_THRESHOLD", 1, NOSTRUCT,
397 	    0, 0, 0, SU},
398 	{PM_START_PM, "PM_START_PM", 1, NOSTRUCT, 0, 0, 0, SU},
399 	{PM_STOP_PM, "PM_STOP_PM", 1, NOSTRUCT, 0, 0, 0, SU},
400 	{PM_RESET_PM, "PM_RESET_PM", 1, NOSTRUCT, 0, 0, 0, SU},
401 	{PM_GET_STATS, "PM_GET_STATS", 1, PM_REQ, INWHO | INDATAOUT,
402 	    DIP, NODEP},
403 	{PM_GET_DEVICE_THRESHOLD, "PM_GET_DEVICE_THRESHOLD", 1, PM_REQ, INWHO,
404 	    DIP, NODEP},
405 	{PM_GET_POWER_NAME, "PM_GET_POWER_NAME", 1, PM_REQ, INWHO | INDATAOUT,
406 	    DIP, NODEP},
407 	{PM_GET_POWER_LEVELS, "PM_GET_POWER_LEVELS", 1, PM_REQ,
408 	    INWHO | INDATAOUT, DIP, NODEP},
409 	{PM_GET_NUM_COMPONENTS, "PM_GET_NUM_COMPONENTS", 1, PM_REQ, INWHO,
410 	    DIP, NODEP},
411 	{PM_GET_COMPONENT_NAME, "PM_GET_COMPONENT_NAME", 1, PM_REQ,
412 	    INWHO | INDATAOUT, DIP, NODEP},
413 	{PM_GET_NUM_POWER_LEVELS, "PM_GET_NUM_POWER_LEVELS", 1, PM_REQ, INWHO,
414 	    DIP, NODEP},
415 	{PM_GET_STATE_CHANGE, "PM_GET_STATE_CHANGE", 1, PM_PSC},
416 	{PM_GET_STATE_CHANGE_WAIT, "PM_GET_STATE_CHANGE_WAIT", 1, PM_PSC},
417 	{PM_DIRECT_PM, "PM_DIRECT_PM", 1, PM_REQ, INWHO, DIP, NODEP,
418 	    (SU | SG)},
419 	{PM_RELEASE_DIRECT_PM, "PM_RELEASE_DIRECT_PM", 1, PM_REQ, INWHO,
420 	    DIP, NODEP},
421 	{PM_DIRECT_NOTIFY, "PM_DIRECT_NOTIFY", 1, PM_PSC},
422 	{PM_DIRECT_NOTIFY_WAIT, "PM_DIRECT_NOTIFY_WAIT", 1, PM_PSC},
423 	{PM_RESET_DEVICE_THRESHOLD, "PM_RESET_DEVICE_THRESHOLD", 1, PM_REQ,
424 	    INWHO, DIP, NODEP, SU},
425 	{PM_GET_PM_STATE, "PM_GET_PM_STATE", 1, NOSTRUCT},
426 	{PM_GET_AUTOS3_STATE, "PM_GET_AUTOS3_STATE", 1, NOSTRUCT},
427 	{PM_GET_S3_SUPPORT_STATE, "PM_GET_S3_SUPPORT_STATE", 1, NOSTRUCT},
428 	{PM_GET_DEVICE_TYPE, "PM_GET_DEVICE_TYPE", 1, PM_REQ, INWHO,
429 	    DIP, NODEP},
430 	{PM_SET_COMPONENT_THRESHOLDS, "PM_SET_COMPONENT_THRESHOLDS", 1, PM_REQ,
431 	    INWHO | INDATAINT, NODIP, NODEP, SU},
432 	{PM_GET_COMPONENT_THRESHOLDS, "PM_GET_COMPONENT_THRESHOLDS", 1, PM_REQ,
433 	    INWHO | INDATAOUT, DIP, NODEP},
434 	{PM_IDLE_DOWN, "PM_IDLE_DOWN", 1, NOSTRUCT, 0, 0, 0, SU},
435 	{PM_GET_DEVICE_THRESHOLD_BASIS, "PM_GET_DEVICE_THRESHOLD_BASIS", 1,
436 	    PM_REQ, INWHO, DIP, NODEP},
437 	{PM_SET_CURRENT_POWER, "PM_SET_CURRENT_POWER", 1, PM_REQ, INWHO, DIP,
438 	    NODEP},
439 	{PM_GET_CURRENT_POWER, "PM_GET_CURRENT_POWER", 1, PM_REQ, INWHO, DIP,
440 	    NODEP},
441 	{PM_GET_FULL_POWER, "PM_GET_FULL_POWER", 1, PM_REQ, INWHO, DIP,
442 	    NODEP},
443 	{PM_ADD_DEPENDENT, "PM_ADD_DEPENDENT", 1, PM_REQ, INWHO | INDATASTRING,
444 	    DIP, DEP, SU},
445 	{PM_GET_TIME_IDLE, "PM_GET_TIME_IDLE", 1, PM_REQ, INWHO, DIP, NODEP},
446 	{PM_ADD_DEPENDENT_PROPERTY, "PM_ADD_DEPENDENT_PROPERTY", 1, PM_REQ,
447 	    INWHO | INDATASTRING, NODIP, DEP, SU},
448 	{PM_START_CPUPM, "PM_START_CPUPM", 1, NOSTRUCT, 0, 0, 0, SU},
449 	{PM_START_CPUPM_EV, "PM_START_CPUPM_EV", 1, NOSTRUCT, 0,
450 	    0, 0, SU},
451 	{PM_START_CPUPM_POLL, "PM_START_CPUPM_POLL", 1, NOSTRUCT, 0,
452 	    0, 0, SU},
453 	{PM_STOP_CPUPM, "PM_STOP_CPUPM", 1, NOSTRUCT, 0, 0, 0, SU},
454 	{PM_GET_CPU_THRESHOLD, "PM_GET_CPU_THRESHOLD", 1, NOSTRUCT},
455 	{PM_SET_CPU_THRESHOLD, "PM_SET_CPU_THRESHOLD", 1, NOSTRUCT,
456 	    0, 0, 0, SU},
457 	{PM_GET_CPUPM_STATE, "PM_GET_CPUPM_STATE", 1, NOSTRUCT},
458 	{PM_START_AUTOS3, "PM_START_AUTOS3", 1, NOSTRUCT, 0, 0, 0, SU},
459 	{PM_STOP_AUTOS3, "PM_STOP_AUTOS3", 1, NOSTRUCT, 0, 0, 0, SU},
460 	{PM_ENABLE_S3, "PM_ENABLE_S3", 1, NOSTRUCT, 0, 0, 0, SU},
461 	{PM_DISABLE_S3, "PM_DISABLE_S3", 1, NOSTRUCT, 0, 0, 0, SU},
462 	{PM_ENTER_S3, "PM_ENTER_S3", 1, NOSTRUCT, 0, 0, 0, SU},
463 	{PM_SEARCH_LIST, "PM_SEARCH_LIST", 1, PM_SRCH, 0, 0, 0, SU},
464 	{PM_GET_CMD_NAME, "PM_GET_CMD_NAME", 1, PM_REQ, INDATAOUT, NODIP,
465 	    NODEP, 0},
466 	{PM_DISABLE_CPU_DEEP_IDLE, "PM_DISABLE_CPU_DEEP_IDLE", 1, NOSTRUCT, 0,
467 	    0, 0, SU},
468 	{PM_ENABLE_CPU_DEEP_IDLE, "PM_START_CPU_DEEP_IDLE", 1, NOSTRUCT, 0,
469 	    0, 0, SU},
470 	{PM_DEFAULT_CPU_DEEP_IDLE, "PM_DFLT_CPU_DEEP_IDLE", 1, NOSTRUCT, 0,
471 	    0, 0, SU},
472 	{0, NULL}
473 };
474 
475 struct pm_cmd_info *
pc_info(int cmd)476 pc_info(int cmd)
477 {
478 	struct pm_cmd_info *pcip;
479 
480 	for (pcip = pmci; pcip->name; pcip++) {
481 		if (cmd == pcip->cmd)
482 			return (pcip);
483 	}
484 	return (NULL);
485 }
486 
487 static char *
pm_decode_cmd(int cmd)488 pm_decode_cmd(int cmd)
489 {
490 	static char invbuf[64];
491 	struct pm_cmd_info *pcip = pc_info(cmd);
492 	if (pcip != NULL)
493 		return (pcip->name);
494 	(void) sprintf(invbuf, "ioctl: invalid command %d\n", cmd);
495 	return (invbuf);
496 }
497 
498 /*
499  * Allocate scan resource, create taskq, then dispatch scan,
500  * called only if autopm is enabled.
501  */
502 int
pm_start_pm_walk(dev_info_t * dip,void * arg)503 pm_start_pm_walk(dev_info_t *dip, void *arg)
504 {
505 	int cmd = *((int *)arg);
506 #ifdef PMDDEBUG
507 	char *cmdstr = pm_decode_cmd(cmd);
508 #endif
509 
510 	if (!PM_GET_PM_INFO(dip) || PM_ISBC(dip))
511 		return (DDI_WALK_CONTINUE);
512 
513 	switch (cmd) {
514 	case PM_START_CPUPM:
515 	case PM_START_CPUPM_POLL:
516 		if (!PM_ISCPU(dip))
517 			return (DDI_WALK_CONTINUE);
518 		mutex_enter(&pm_scan_lock);
519 		if (!PM_CPUPM_DISABLED && !PM_EVENT_CPUPM)
520 			pm_scan_init(dip);
521 		mutex_exit(&pm_scan_lock);
522 		break;
523 	case PM_START_PM:
524 		mutex_enter(&pm_scan_lock);
525 		if (PM_ISCPU(dip) && (PM_CPUPM_DISABLED || PM_EVENT_CPUPM)) {
526 			mutex_exit(&pm_scan_lock);
527 			return (DDI_WALK_CONTINUE);
528 		}
529 		if (autopm_enabled)
530 			pm_scan_init(dip);
531 		mutex_exit(&pm_scan_lock);
532 		break;
533 	}
534 
535 	/*
536 	 * Start doing pm on device: ensure pm_scan data structure initiated,
537 	 * no need to guarantee a successful scan run.
538 	 */
539 	PMD(PMD_SCAN | PMD_IOCTL, ("ioctl: %s: scan %s@%s(%s#%d)\n", cmdstr,
540 	    PM_DEVICE(dip)))
541 	pm_rescan(dip);
542 
543 	return (DDI_WALK_CONTINUE);
544 }
545 
546 /*
547  * Bring devices to full power level, then stop scan
548  */
549 int
pm_stop_pm_walk(dev_info_t * dip,void * arg)550 pm_stop_pm_walk(dev_info_t *dip, void *arg)
551 {
552 	pm_info_t *info = PM_GET_PM_INFO(dip);
553 	int cmd = *((int *)arg);
554 #ifdef PMDDEBUG
555 	char *cmdstr = pm_decode_cmd(cmd);
556 #endif
557 
558 	if (!info)
559 		return (DDI_WALK_CONTINUE);
560 
561 	switch (cmd) {
562 	case PM_STOP_PM:
563 		/*
564 		 * If CPU devices are being managed independently, then don't
565 		 * stop them as part of PM_STOP_PM. Only stop them as part of
566 		 * PM_STOP_CPUPM and PM_RESET_PM.
567 		 */
568 		if (PM_ISCPU(dip) && PM_POLLING_CPUPM)
569 			return (DDI_WALK_CONTINUE);
570 		break;
571 	case PM_STOP_CPUPM:
572 		/*
573 		 * If stopping CPU devices and this device is not marked
574 		 * as a CPU device, then skip.
575 		 */
576 		if (!PM_ISCPU(dip))
577 			return (DDI_WALK_CONTINUE);
578 		break;
579 	}
580 
581 	/*
582 	 * Stop the current scan, and then bring it back to normal power.
583 	 */
584 	if (!PM_ISBC(dip)) {
585 		PMD(PMD_SCAN | PMD_IOCTL, ("ioctl: %s: stop scan for "
586 		    "%s@%s(%s#%d)\n", cmdstr, PM_DEVICE(dip)))
587 		pm_scan_stop(dip);
588 	}
589 
590 	if (!PM_ISBC(dip) && !PM_ISDIRECT(dip) &&
591 	    !pm_all_at_normal(dip)) {
592 		PM_LOCK_DIP(dip);
593 		if (info->pmi_dev_pm_state & PM_DETACHING) {
594 			PMD(PMD_ALLNORM, ("ioctl: %s: deferring "
595 			    "all_to_normal because %s@%s(%s#%d) is detaching\n",
596 			    cmdstr, PM_DEVICE(dip)))
597 			info->pmi_dev_pm_state |= PM_ALLNORM_DEFERRED;
598 			PM_UNLOCK_DIP(dip);
599 			return (DDI_WALK_CONTINUE);
600 		}
601 		PM_UNLOCK_DIP(dip);
602 		if (pm_all_to_normal(dip, PM_CANBLOCK_FAIL) != DDI_SUCCESS) {
603 			PMD(PMD_ERROR, ("ioctl: %s: could not bring %s@%s"
604 			    "(%s#%d) to normal\n", cmdstr, PM_DEVICE(dip)))
605 		}
606 	}
607 
608 	return (DDI_WALK_CONTINUE);
609 }
610 
611 static int
pm_start_idledown(dev_info_t * dip,void * arg)612 pm_start_idledown(dev_info_t *dip, void *arg)
613 {
614 	int		flag = (int)(intptr_t)arg;
615 	pm_scan_t	*scanp = PM_GET_PM_SCAN(dip);
616 
617 	if (!scanp)
618 		return (DDI_WALK_CONTINUE);
619 
620 	PM_LOCK_DIP(dip);
621 	scanp->ps_idle_down |= flag;
622 	PM_UNLOCK_DIP(dip);
623 	pm_rescan(dip);
624 
625 	return (DDI_WALK_CONTINUE);
626 }
627 
628 /*ARGSUSED*/
629 static int
pm_end_idledown(dev_info_t * dip,void * ignore)630 pm_end_idledown(dev_info_t *dip, void *ignore)
631 {
632 	pm_scan_t	*scanp = PM_GET_PM_SCAN(dip);
633 
634 	if (!scanp)
635 		return (DDI_WALK_CONTINUE);
636 
637 	PM_LOCK_DIP(dip);
638 	/*
639 	 * The PMID_TIMERS bits are place holder till idledown expires.
640 	 * The bits are also the base for regenerating PMID_SCANS bits.
641 	 * While it's up to scan thread to clear up the PMID_SCANS bits
642 	 * after each scan run, PMID_TIMERS ensure aggressive scan down
643 	 * performance throughout the idledown period.
644 	 */
645 	scanp->ps_idle_down &= ~PMID_TIMERS;
646 	PM_UNLOCK_DIP(dip);
647 
648 	return (DDI_WALK_CONTINUE);
649 }
650 
651 /*ARGSUSED*/
652 static void
pm_end_idledown_walk(void * ignore)653 pm_end_idledown_walk(void *ignore)
654 {
655 	PMD(PMD_IDLEDOWN, ("ioctl: end_idledown: idledown_id(%lx) timer is "
656 	    "off\n", (ulong_t)pmstp->pm_idledown_id));
657 
658 	mutex_enter(&pm_scan_lock);
659 	pmstp->pm_idledown_id = 0;
660 	mutex_exit(&pm_scan_lock);
661 
662 	ddi_walk_devs(ddi_root_node(), pm_end_idledown, NULL);
663 }
664 
665 /*
666  * pm_timeout_idledown - keep idledown effect for 10 seconds.
667  *
668  * Return 0 if another competing caller scheduled idledown timeout,
669  * otherwise, return idledown timeout_id.
670  */
671 static timeout_id_t
pm_timeout_idledown(void)672 pm_timeout_idledown(void)
673 {
674 	timeout_id_t	to_id;
675 
676 	/*
677 	 * Keep idle-down in effect for either 10 seconds
678 	 * or length of a scan interval, which ever is greater.
679 	 */
680 	mutex_enter(&pm_scan_lock);
681 	if (pmstp->pm_idledown_id != 0) {
682 		to_id = pmstp->pm_idledown_id;
683 		pmstp->pm_idledown_id = 0;
684 		mutex_exit(&pm_scan_lock);
685 		(void) untimeout(to_id);
686 		mutex_enter(&pm_scan_lock);
687 		if (pmstp->pm_idledown_id != 0) {
688 			PMD(PMD_IDLEDOWN, ("ioctl: timeout_idledown: "
689 			    "another caller got it, idledown_id(%lx)!\n",
690 			    (ulong_t)pmstp->pm_idledown_id))
691 			mutex_exit(&pm_scan_lock);
692 			return (0);
693 		}
694 	}
695 	pmstp->pm_idledown_id = timeout(pm_end_idledown_walk, NULL,
696 	    PM_IDLEDOWN_TIME * hz);
697 	PMD(PMD_IDLEDOWN, ("ioctl: timeout_idledown: idledown_id(%lx)\n",
698 	    (ulong_t)pmstp->pm_idledown_id))
699 	mutex_exit(&pm_scan_lock);
700 
701 	return (pmstp->pm_idledown_id);
702 }
703 
704 static int
pm_chpoll(dev_t dev,short events,int anyyet,short * reventsp,struct pollhead ** phpp)705 pm_chpoll(dev_t dev, short events, int anyyet, short *reventsp,
706 	struct pollhead **phpp)
707 {
708 	extern struct pollhead pm_pollhead;	/* common/os/sunpm.c */
709 	int	clone;
710 
711 	clone = PM_MINOR_TO_CLONE(getminor(dev));
712 	PMD(PMD_IOCTL, ("ioctl: pm_chpoll: clone %d\n", clone))
713 	if ((events & (POLLIN | POLLRDNORM)) && pm_poll_cnt[clone]) {
714 		*reventsp |= (POLLIN | POLLRDNORM);
715 		PMD(PMD_IOCTL, ("ioctl: pm_chpoll: reventsp set\n"))
716 	} else {
717 		*reventsp = 0;
718 		if (!anyyet) {
719 			PMD(PMD_IOCTL, ("ioctl: pm_chpoll: not anyyet\n"))
720 			*phpp = &pm_pollhead;
721 		}
722 #ifdef DEBUG
723 		else {
724 			PMD(PMD_IOCTL, ("ioctl: pm_chpoll: anyyet\n"))
725 		}
726 #endif
727 	}
728 	return (0);
729 }
730 
731 /*
732  * called by pm_dicard_entries to free up the memory. It also decrements
733  * pm_poll_cnt, if direct is non zero.
734  */
735 static void
pm_free_entries(psce_t * pscep,int clone,int direct)736 pm_free_entries(psce_t *pscep, int clone, int direct)
737 {
738 	pm_state_change_t	*p;
739 
740 	if (pscep) {
741 		p = pscep->psce_out;
742 		while (p->size) {
743 			if (direct) {
744 				PMD(PMD_IOCTL, ("ioctl: discard: "
745 				    "pm_poll_cnt[%d] is %d before "
746 				    "ASSERT\n", clone,
747 				    pm_poll_cnt[clone]))
748 				ASSERT(pm_poll_cnt[clone]);
749 				pm_poll_cnt[clone]--;
750 			}
751 			kmem_free(p->physpath, p->size);
752 			p->size = 0;
753 			if (p == pscep->psce_last)
754 				p = pscep->psce_first;
755 			else
756 				p++;
757 		}
758 		pscep->psce_out = pscep->psce_first;
759 		pscep->psce_in = pscep->psce_first;
760 		mutex_exit(&pscep->psce_lock);
761 	}
762 }
763 
764 /*
765  * Discard entries for this clone. Calls pm_free_entries to free up memory.
766  */
767 static void
pm_discard_entries(int clone)768 pm_discard_entries(int clone)
769 {
770 	psce_t	*pscep;
771 	int			direct = 0;
772 
773 	mutex_enter(&pm_clone_lock);
774 	if ((pscep = pm_psc_clone_to_direct(clone)) != NULL)
775 		direct = 1;
776 	pm_free_entries(pscep, clone, direct);
777 	pscep = pm_psc_clone_to_interest(clone);
778 	pm_free_entries(pscep, clone, 0);
779 	mutex_exit(&pm_clone_lock);
780 }
781 
782 
783 static void
pm_set_idle_threshold(dev_info_t * dip,int thresh,int flag)784 pm_set_idle_threshold(dev_info_t *dip, int thresh, int flag)
785 {
786 	if (!PM_ISBC(dip) && !PM_ISDIRECT(dip)) {
787 		switch (DEVI(dip)->devi_pm_flags & PMC_THRESH_ALL) {
788 		case PMC_DEF_THRESH:
789 		case PMC_CPU_THRESH:
790 			PMD(PMD_IOCTL, ("ioctl: set_idle_threshold: set "
791 			    "%s@%s(%s#%d) default thresh to 0t%d\n",
792 			    PM_DEVICE(dip), thresh))
793 			pm_set_device_threshold(dip, thresh, flag);
794 			break;
795 		default:
796 			break;
797 		}
798 	}
799 }
800 
801 static int
pm_set_idle_thresh_walk(dev_info_t * dip,void * arg)802 pm_set_idle_thresh_walk(dev_info_t *dip, void *arg)
803 {
804 	int cmd = *((int *)arg);
805 
806 	if (!PM_GET_PM_INFO(dip))
807 		return (DDI_WALK_CONTINUE);
808 
809 	switch (cmd) {
810 	case PM_SET_SYSTEM_THRESHOLD:
811 		if (DEVI(dip)->devi_pm_flags & PMC_CPU_THRESH)
812 			break;
813 		pm_set_idle_threshold(dip, pm_system_idle_threshold,
814 		    PMC_DEF_THRESH);
815 		pm_rescan(dip);
816 		break;
817 	case PM_SET_CPU_THRESHOLD:
818 		if (!PM_ISCPU(dip))
819 			break;
820 		pm_set_idle_threshold(dip, pm_cpu_idle_threshold,
821 		    PMC_CPU_THRESH);
822 		pm_rescan(dip);
823 		break;
824 	}
825 
826 	return (DDI_WALK_CONTINUE);
827 }
828 
829 /*ARGSUSED*/
830 static int
pm_getinfo(dev_info_t * dip,ddi_info_cmd_t infocmd,void * arg,void ** result)831 pm_getinfo(dev_info_t *dip, ddi_info_cmd_t infocmd, void *arg, void **result)
832 {
833 	dev_t	dev;
834 	int	instance;
835 
836 	switch (infocmd) {
837 	case DDI_INFO_DEVT2DEVINFO:
838 		if (pmstp->pm_instance == -1)
839 			return (DDI_FAILURE);
840 		*result = pmstp->pm_dip;
841 		return (DDI_SUCCESS);
842 
843 	case DDI_INFO_DEVT2INSTANCE:
844 		dev = (dev_t)arg;
845 		instance = getminor(dev) >> 8;
846 		*result = (void *)(uintptr_t)instance;
847 		return (DDI_SUCCESS);
848 
849 	default:
850 		return (DDI_FAILURE);
851 	}
852 }
853 
854 
855 /*ARGSUSED1*/
856 static int
pm_open(dev_t * devp,int flag,int otyp,cred_t * cr)857 pm_open(dev_t *devp, int flag, int otyp, cred_t *cr)
858 {
859 	int		clone;
860 
861 	if (otyp != OTYP_CHR)
862 		return (EINVAL);
863 
864 	mutex_enter(&pm_clone_lock);
865 	for (clone = 1; clone < PM_MAX_CLONE; clone++)
866 		if (!pmstp->pm_clones[clone])
867 			break;
868 
869 	if (clone == PM_MAX_CLONE) {
870 		mutex_exit(&pm_clone_lock);
871 		return (ENXIO);
872 	}
873 	pmstp->pm_cred[clone] = cr;
874 	crhold(cr);
875 
876 	*devp = makedevice(getmajor(*devp), (pmstp->pm_instance << 8) + clone);
877 	pmstp->pm_clones[clone] = 1;
878 	mutex_exit(&pm_clone_lock);
879 
880 	return (0);
881 }
882 
883 /*ARGSUSED1*/
884 static int
pm_close(dev_t dev,int flag,int otyp,cred_t * cr)885 pm_close(dev_t dev, int flag, int otyp, cred_t *cr)
886 {
887 	int clone;
888 
889 	if (otyp != OTYP_CHR)
890 		return (EINVAL);
891 
892 	clone = PM_MINOR_TO_CLONE(getminor(dev));
893 	PMD(PMD_CLOSE, ("pm_close: minor %x, clone %x\n", getminor(dev),
894 	    clone))
895 
896 	/*
897 	 * Walk the entire device tree to find the corresponding
898 	 * device and operate on it.
899 	 */
900 	ddi_walk_devs(ddi_root_node(), pm_close_direct_pm_device,
901 	    (void *) &clone);
902 
903 	crfree(pmstp->pm_cred[clone]);
904 	pmstp->pm_cred[clone] = 0;
905 	pmstp->pm_clones[clone] = 0;
906 	pm_discard_entries(clone);
907 	ASSERT(pm_poll_cnt[clone] == 0);
908 	pm_deregister_watcher(clone, NULL);
909 	return (0);
910 }
911 
912 /*ARGSUSED*/
913 static int
pm_ioctl(dev_t dev,int cmd,intptr_t arg,int mode,cred_t * cr,int * rval_p)914 pm_ioctl(dev_t dev, int cmd, intptr_t arg, int mode, cred_t *cr, int *rval_p)
915 {
916 	struct pm_cmd_info *pc_info(int);
917 	struct pm_cmd_info *pcip = pc_info(cmd);
918 	pm_req_t	req;
919 	dev_info_t	*dip = NULL;
920 	pm_info_t	*info = NULL;
921 	int		clone;
922 	char		*cmdstr = pm_decode_cmd(cmd);
923 	/*
924 	 * To keep devinfo nodes from going away while we're holding a
925 	 * pointer to their dip, pm_name_to_dip() optionally holds
926 	 * the devinfo node.  If we've done that, we set dipheld
927 	 * so we know at the end of the ioctl processing to release the
928 	 * node again.
929 	 */
930 	int		dipheld = 0;
931 	int		icount = 0;
932 	int		i;
933 	int		comps;
934 	size_t		lencopied;
935 	int		ret = ENOTTY;
936 	int		curpower;
937 	char		who[MAXNAMELEN];
938 	size_t		wholen;			/* copyinstr length */
939 	size_t		deplen = MAXNAMELEN;
940 	char		*dep, i_dep_buf[MAXNAMELEN];
941 	char		pathbuf[MAXNAMELEN];
942 	struct pm_component *cp;
943 #ifdef	_MULTI_DATAMODEL
944 	pm_state_change32_t		*pscp32;
945 	pm_state_change32_t		psc32;
946 	pm_searchargs32_t		psa32;
947 	size_t				copysize32;
948 #endif
949 	pm_state_change_t		*pscp;
950 	pm_state_change_t		psc;
951 	pm_searchargs_t		psa;
952 	char		listname[MAXCOPYBUF];
953 	char		manufacturer[MAXCOPYBUF];
954 	char		product[MAXCOPYBUF];
955 	size_t		copysize;
956 
957 	PMD(PMD_IOCTL, ("ioctl: %s: begin\n", cmdstr))
958 
959 #ifdef DEBUG
960 	if (cmd == 666) {
961 		ddi_walk_devs(ddi_root_node(), print_info, NULL);
962 		return (0);
963 	}
964 	ret = 0x0badcafe;			/* sanity checking */
965 	pm_cmd = cmd;				/* for ASSERT debugging */
966 	pm_cmd_string = cmdstr;	/* for ASSERT debugging */
967 #endif
968 
969 
970 	if (pcip == NULL) {
971 		PMD(PMD_ERROR, ("ioctl: unknown command %d\n", cmd))
972 		return (ENOTTY);
973 	}
974 	if (pcip == NULL || pcip->supported == 0) {
975 		PMD(PMD_ERROR, ("ioctl: command %s no longer supported\n",
976 		    pcip->name))
977 		return (ENOTTY);
978 	}
979 
980 	wholen = 0;
981 	dep = i_dep_buf;
982 	i_dep_buf[0] = 0;
983 	clone = PM_MINOR_TO_CLONE(getminor(dev));
984 	if (!pm_perms(pcip->permission, pmstp->pm_cred[clone])) {
985 		ret = EPERM;
986 		return (ret);
987 	}
988 	switch (pcip->str_type) {
989 	case PM_REQ:
990 	{
991 #ifdef	_MULTI_DATAMODEL
992 		if ((mode & DATAMODEL_MASK) == DATAMODEL_ILP32) {
993 			pm_req32_t	req32;
994 
995 			if (ddi_copyin((caddr_t)arg, &req32,
996 			    sizeof (req32), mode) != 0) {
997 				PMD(PMD_ERROR, ("ioctl: %s: ddi_copyin "
998 				    "EFAULT\n\n", cmdstr))
999 				ret = EFAULT;
1000 				break;
1001 			}
1002 			req.component = req32.component;
1003 			req.value = req32.value;
1004 			req.datasize = req32.datasize;
1005 			if (pcip->inargs & INWHO) {
1006 				ret = copyinstr((char *)(uintptr_t)
1007 				    req32.physpath, who, MAXNAMELEN, &wholen);
1008 				if (ret) {
1009 					PMD(PMD_ERROR, ("ioctl: %s: "
1010 					    "copyinstr fails returning %d\n",
1011 					    cmdstr, ret))
1012 					break;
1013 				}
1014 				req.physpath = who;
1015 				PMD(PMD_IOCTL, ("ioctl: %s: physpath=%s\n",
1016 				    cmdstr, req.physpath))
1017 			}
1018 			if (pcip->inargs & INDATA) {
1019 				req.data = (void *)(uintptr_t)req32.data;
1020 				req.datasize = req32.datasize;
1021 			} else {
1022 				req.data = NULL;
1023 				req.datasize = 0;
1024 			}
1025 			switch (pcip->diptype) {
1026 			case DIP:
1027 				if (!(dip =
1028 				    pm_name_to_dip(req.physpath, 1))) {
1029 					PMD(PMD_ERROR, ("ioctl: %s: "
1030 					    "pm_name_to_dip for %s failed\n",
1031 					    cmdstr, req.physpath))
1032 					return (ENODEV);
1033 				}
1034 				ASSERT(!dipheld);
1035 				dipheld++;
1036 				break;
1037 			case NODIP:
1038 				break;
1039 			default:
1040 				/*
1041 				 * Internal error, invalid ioctl description
1042 				 * force debug entry even if pm_debug not set
1043 				 */
1044 #ifdef	DEBUG
1045 				pm_log("invalid diptype %d for cmd %d (%s)\n",
1046 				    pcip->diptype, cmd, pcip->name);
1047 #endif
1048 				ASSERT(0);
1049 				return (EIO);
1050 			}
1051 			if (pcip->inargs & INDATAINT) {
1052 				int32_t int32buf;
1053 				int32_t *i32p;
1054 				int *ip;
1055 				icount = req32.datasize / sizeof (int32_t);
1056 				if (icount <= 0) {
1057 					PMD(PMD_ERROR, ("ioctl: %s: datasize"
1058 					    " 0 or neg EFAULT\n\n", cmdstr))
1059 					ret = EFAULT;
1060 					break;
1061 				}
1062 				ASSERT(!(pcip->inargs & INDATASTRING));
1063 				req.datasize = icount * sizeof (int);
1064 				req.data = kmem_alloc(req.datasize, KM_SLEEP);
1065 				ip = req.data;
1066 				ret = 0;
1067 				for (i = 0,
1068 				    i32p = (int32_t *)(uintptr_t)req32.data;
1069 				    i < icount; i++, i32p++) {
1070 					if (ddi_copyin((void *)i32p, &int32buf,
1071 					    sizeof (int32_t), mode)) {
1072 						kmem_free(req.data,
1073 						    req.datasize);
1074 						PMD(PMD_ERROR, ("ioctl: %s: "
1075 						    "entry %d EFAULT\n",
1076 						    cmdstr, i))
1077 						ret = EFAULT;
1078 						break;
1079 					}
1080 					*ip++ = (int)int32buf;
1081 				}
1082 				if (ret)
1083 					break;
1084 			}
1085 			if (pcip->inargs & INDATASTRING) {
1086 				ASSERT(!(pcip->inargs & INDATAINT));
1087 				ASSERT(pcip->deptype == DEP);
1088 				if (req32.data != NULL) {
1089 					if (copyinstr((void *)(uintptr_t)
1090 					    req32.data, dep, deplen, NULL)) {
1091 						PMD(PMD_ERROR, ("ioctl: %s: "
1092 						    "0x%p dep size %lx, EFAULT"
1093 						    "\n", cmdstr,
1094 						    (void *)req.data, deplen))
1095 						ret = EFAULT;
1096 						break;
1097 					}
1098 #ifdef DEBUG
1099 					else {
1100 						PMD(PMD_DEP, ("ioctl: %s: "
1101 						    "dep %s\n", cmdstr, dep))
1102 					}
1103 #endif
1104 				} else {
1105 					PMD(PMD_ERROR, ("ioctl: %s: no "
1106 					    "dependent\n", cmdstr))
1107 					ret = EINVAL;
1108 					break;
1109 				}
1110 			}
1111 		} else
1112 #endif /* _MULTI_DATAMODEL */
1113 		{
1114 			if (ddi_copyin((caddr_t)arg,
1115 			    &req, sizeof (req), mode) != 0) {
1116 				PMD(PMD_ERROR, ("ioctl: %s: ddi_copyin "
1117 				    "EFAULT\n\n", cmdstr))
1118 				ret = EFAULT;
1119 				break;
1120 			}
1121 			if (pcip->inargs & INWHO) {
1122 				ret = copyinstr((char *)req.physpath, who,
1123 				    MAXNAMELEN, &wholen);
1124 				if (ret) {
1125 					PMD(PMD_ERROR, ("ioctl: %s copyinstr"
1126 					    " fails returning %d\n", cmdstr,
1127 					    ret))
1128 					break;
1129 				}
1130 				req.physpath = who;
1131 				PMD(PMD_IOCTL, ("ioctl: %s: physpath=%s\n",
1132 				    cmdstr, req.physpath))
1133 			}
1134 			if (!(pcip->inargs & INDATA)) {
1135 				req.data = NULL;
1136 				req.datasize = 0;
1137 			}
1138 			switch (pcip->diptype) {
1139 			case DIP:
1140 				if (!(dip =
1141 				    pm_name_to_dip(req.physpath, 1))) {
1142 					PMD(PMD_ERROR, ("ioctl: %s: "
1143 					    "pm_name_to_dip for %s failed\n",
1144 					    cmdstr, req.physpath))
1145 					return (ENODEV);
1146 				}
1147 				ASSERT(!dipheld);
1148 				dipheld++;
1149 				break;
1150 			case NODIP:
1151 				break;
1152 			default:
1153 				/*
1154 				 * Internal error, invalid ioctl description
1155 				 * force debug entry even if pm_debug not set
1156 				 */
1157 #ifdef	DEBUG
1158 				pm_log("invalid diptype %d for cmd %d (%s)\n",
1159 				    pcip->diptype, cmd, pcip->name);
1160 #endif
1161 				ASSERT(0);
1162 				return (EIO);
1163 			}
1164 			if (pcip->inargs & INDATAINT) {
1165 				int *ip;
1166 
1167 				ASSERT(!(pcip->inargs & INDATASTRING));
1168 				ip = req.data;
1169 				icount = req.datasize / sizeof (int);
1170 				if (icount <= 0) {
1171 					PMD(PMD_ERROR, ("ioctl: %s: datasize"
1172 					    " 0 or neg EFAULT\n\n", cmdstr))
1173 					ret = EFAULT;
1174 					break;
1175 				}
1176 				req.data = kmem_alloc(req.datasize, KM_SLEEP);
1177 				if (ddi_copyin((caddr_t)ip, req.data,
1178 				    req.datasize, mode) != 0) {
1179 					PMD(PMD_ERROR, ("ioctl: %s: ddi_copyin "
1180 					    "EFAULT\n\n", cmdstr))
1181 					ret = EFAULT;
1182 					break;
1183 				}
1184 			}
1185 			if (pcip->inargs & INDATASTRING) {
1186 				ASSERT(!(pcip->inargs & INDATAINT));
1187 				ASSERT(pcip->deptype == DEP);
1188 				if (req.data != NULL) {
1189 					if (copyinstr((caddr_t)req.data,
1190 					    dep, deplen, NULL)) {
1191 						PMD(PMD_ERROR, ("ioctl: %s: "
1192 						    "0x%p dep size %lu, "
1193 						    "EFAULT\n", cmdstr,
1194 						    (void *)req.data, deplen))
1195 						ret = EFAULT;
1196 						break;
1197 					}
1198 #ifdef DEBUG
1199 					else {
1200 						PMD(PMD_DEP, ("ioctl: %s: "
1201 						    "dep %s\n", cmdstr, dep))
1202 					}
1203 #endif
1204 				} else {
1205 					PMD(PMD_ERROR, ("ioctl: %s: no "
1206 					    "dependent\n", cmdstr))
1207 					ret = EINVAL;
1208 					break;
1209 				}
1210 			}
1211 		}
1212 		/*
1213 		 * Now we've got all the args in for the commands that
1214 		 * use the new pm_req struct.
1215 		 */
1216 		switch (cmd) {
1217 		case PM_REPARSE_PM_PROPS:
1218 		{
1219 			struct dev_ops	*drv;
1220 			struct cb_ops	*cb;
1221 			void		*propval;
1222 			int length;
1223 			/*
1224 			 * This ioctl is provided only for the ddivs pm test.
1225 			 * We only do it to a driver which explicitly allows
1226 			 * us to do so by exporting a pm-reparse-ok property.
1227 			 * We only care whether the property exists or not.
1228 			 */
1229 			if ((drv = ddi_get_driver(dip)) == NULL) {
1230 				ret = EINVAL;
1231 				break;
1232 			}
1233 			if ((cb = drv->devo_cb_ops) != NULL) {
1234 				if ((*cb->cb_prop_op)(DDI_DEV_T_ANY, dip,
1235 				    PROP_LEN_AND_VAL_ALLOC, (DDI_PROP_CANSLEEP |
1236 				    DDI_PROP_DONTPASS | DDI_PROP_NOTPROM),
1237 				    "pm-reparse-ok", (caddr_t)&propval,
1238 				    &length) != DDI_SUCCESS) {
1239 					ret = EINVAL;
1240 					break;
1241 				}
1242 			} else if (ddi_prop_op(DDI_DEV_T_ANY, dip,
1243 			    PROP_LEN_AND_VAL_ALLOC, (DDI_PROP_CANSLEEP |
1244 			    DDI_PROP_DONTPASS | DDI_PROP_NOTPROM),
1245 			    "pm-reparse-ok", (caddr_t)&propval,
1246 			    &length) != DDI_SUCCESS) {
1247 				ret = EINVAL;
1248 				break;
1249 			}
1250 			kmem_free(propval, length);
1251 			ret =  e_new_pm_props(dip);
1252 			break;
1253 		}
1254 
1255 		case PM_GET_DEVICE_THRESHOLD:
1256 		{
1257 			PM_LOCK_DIP(dip);
1258 			if (!PM_GET_PM_INFO(dip) || PM_ISBC(dip)) {
1259 				PM_UNLOCK_DIP(dip);
1260 				PMD(PMD_ERROR, ("ioctl: %s: ENODEV\n",
1261 				    cmdstr))
1262 				ret = ENODEV;
1263 				break;
1264 			}
1265 			*rval_p = DEVI(dip)->devi_pm_dev_thresh;
1266 			PM_UNLOCK_DIP(dip);
1267 			ret = 0;
1268 			break;
1269 		}
1270 
1271 		case PM_DIRECT_PM:
1272 		{
1273 			int has_dep;
1274 			if ((info = PM_GET_PM_INFO(dip)) == NULL) {
1275 				PMD(PMD_ERROR | PMD_DPM, ("ioctl: %s: "
1276 				    "ENODEV\n", cmdstr))
1277 				ret = ENODEV;
1278 				break;
1279 			}
1280 			/*
1281 			 * Check to see if we are there is a dependency on
1282 			 * this kept device, if so, return EBUSY.
1283 			 */
1284 			(void) ddi_pathname(dip, pathbuf);
1285 			pm_dispatch_to_dep_thread(PM_DEP_WK_CHECK_KEPT,
1286 			    NULL, pathbuf, PM_DEP_WAIT, &has_dep, 0);
1287 			if (has_dep) {
1288 				PMD(PMD_ERROR | PMD_DPM, ("%s EBUSY\n",
1289 				    cmdstr))
1290 				ret = EBUSY;
1291 				break;
1292 			}
1293 			PM_LOCK_DIP(dip);
1294 			if (PM_ISDIRECT(dip) || (info->pmi_clone != 0)) {
1295 				PMD(PMD_ERROR | PMD_DPM, ("ioctl: %s: "
1296 				    "%s@%s(%s#%d): EBUSY\n", cmdstr,
1297 				    PM_DEVICE(dip)))
1298 				PM_UNLOCK_DIP(dip);
1299 				ret = EBUSY;
1300 				break;
1301 			}
1302 			info->pmi_dev_pm_state |= PM_DIRECT;
1303 			info->pmi_clone = clone;
1304 			PM_UNLOCK_DIP(dip);
1305 			PMD(PMD_DPM, ("ioctl: %s: info %p, pmi_clone %d\n",
1306 			    cmdstr, (void *)info, clone))
1307 			mutex_enter(&pm_clone_lock);
1308 			pm_register_watcher(clone, dip);
1309 			mutex_exit(&pm_clone_lock);
1310 			ret = 0;
1311 			break;
1312 		}
1313 
1314 		case PM_RELEASE_DIRECT_PM:
1315 		{
1316 			if ((info = PM_GET_PM_INFO(dip)) == NULL) {
1317 				PMD(PMD_ERROR | PMD_DPM, ("ioctl: %s: "
1318 				    "ENODEV\n", cmdstr))
1319 				ret = ENODEV;
1320 				break;
1321 			}
1322 			PM_LOCK_DIP(dip);
1323 			if (info->pmi_clone != clone) {
1324 				PMD(PMD_ERROR | PMD_DPM, ("ioctl: %s: "
1325 				    "%s@%s(%s#%d) EINVAL\n", cmdstr,
1326 				    PM_DEVICE(dip)))
1327 				ret = EINVAL;
1328 				PM_UNLOCK_DIP(dip);
1329 				break;
1330 			}
1331 			ASSERT(PM_ISDIRECT(dip));
1332 			info->pmi_dev_pm_state &= ~PM_DIRECT;
1333 			PM_UNLOCK_DIP(dip);
1334 			/* Bring ourselves up if there is a keeper. */
1335 			(void) ddi_pathname(dip, pathbuf);
1336 			pm_dispatch_to_dep_thread(PM_DEP_WK_BRINGUP_SELF,
1337 			    NULL, pathbuf, PM_DEP_WAIT, NULL, 0);
1338 			pm_discard_entries(clone);
1339 			pm_deregister_watcher(clone, dip);
1340 			/*
1341 			 * Now we could let the other threads that are
1342 			 * trying to do a DIRECT_PM thru
1343 			 */
1344 			PM_LOCK_DIP(dip);
1345 			info->pmi_clone = 0;
1346 			PM_UNLOCK_DIP(dip);
1347 			pm_proceed(dip, PMP_RELEASE, -1, -1);
1348 			PMD(PMD_RESCAN | PMD_DPM, ("ioctl: %s: rescan\n",
1349 			    cmdstr))
1350 			pm_rescan(dip);
1351 			ret = 0;
1352 			break;
1353 		}
1354 
1355 		case PM_SET_CURRENT_POWER:
1356 		{
1357 			int comp = req.component;
1358 			int  value = req.value;
1359 			PMD(PMD_DPM, ("ioctl: %s: %s component %d to value "
1360 			    "%d\n", cmdstr, req.physpath, comp, value))
1361 			if (!e_pm_valid_comp(dip, comp, NULL) ||
1362 			    !e_pm_valid_power(dip, comp, value)) {
1363 				PMD(PMD_ERROR | PMD_DPM, ("ioctl: %s: "
1364 				    "physpath=%s, comp=%d, level=%d, fails\n",
1365 				    cmdstr, req.physpath, comp, value))
1366 				ret = EINVAL;
1367 				break;
1368 			}
1369 
1370 			if ((info = PM_GET_PM_INFO(dip)) == NULL) {
1371 				PMD(PMD_ERROR | PMD_DPM, ("ioctl: %s: "
1372 				    "ENODEV\n", cmdstr))
1373 				ret = ENODEV;
1374 				break;
1375 			}
1376 			if (info->pmi_clone != clone) {
1377 				PMD(PMD_ERROR | PMD_DPM, ("ioctl: %s: "
1378 				    "(not owner) %s fails; clone %d, owner %d"
1379 				    "\n", cmdstr, req.physpath, clone,
1380 				    info->pmi_clone))
1381 				ret = EINVAL;
1382 				break;
1383 			}
1384 			ASSERT(PM_ISDIRECT(dip));
1385 
1386 			if (pm_set_power(dip, comp, value, PM_LEVEL_EXACT,
1387 			    PM_CANBLOCK_BLOCK, 0, &ret) != DDI_SUCCESS) {
1388 				PMD(PMD_ERROR | PMD_DPM, ("ioctl: %s: "
1389 				    "pm_set_power for %s fails, errno=%d\n",
1390 				    cmdstr, req.physpath, ret))
1391 				break;
1392 			}
1393 
1394 			pm_proceed(dip, PMP_SETPOWER, comp, value);
1395 
1396 			/*
1397 			 * Power down all idle components if console framebuffer
1398 			 * is powered off.
1399 			 */
1400 			if (PM_IS_CFB(dip) && (pm_system_idle_threshold ==
1401 			    pm_default_idle_threshold)) {
1402 				dev_info_t	*root = ddi_root_node();
1403 				if (PM_ISBC(dip)) {
1404 					if (comp == 0 && value == 0 &&
1405 					    (pm_timeout_idledown() != 0)) {
1406 						ddi_walk_devs(root,
1407 						    pm_start_idledown,
1408 						    (void *)PMID_CFB);
1409 					}
1410 				} else {
1411 					int count = 0;
1412 					for (i = 0; i < PM_NUMCMPTS(dip); i++) {
1413 						ret = pm_get_current_power(dip,
1414 						    i, &curpower);
1415 						if (ret == DDI_SUCCESS &&
1416 						    curpower == 0)
1417 							count++;
1418 					}
1419 					if ((count == PM_NUMCMPTS(dip)) &&
1420 					    (pm_timeout_idledown() != 0)) {
1421 						ddi_walk_devs(root,
1422 						    pm_start_idledown,
1423 						    (void *)PMID_CFB);
1424 					}
1425 				}
1426 			}
1427 
1428 			PMD(PMD_RESCAN | PMD_DPM, ("ioctl: %s: rescan\n",
1429 			    cmdstr))
1430 			pm_rescan(dip);
1431 			*rval_p = 0;
1432 			ret = 0;
1433 			break;
1434 		}
1435 
1436 		case PM_GET_FULL_POWER:
1437 		{
1438 			int normal;
1439 			ASSERT(dip);
1440 			PMD(PMD_NORM, ("ioctl: %s: %s component %d\n",
1441 			    cmdstr, req.physpath, req.component))
1442 			normal =  pm_get_normal_power(dip, req.component);
1443 
1444 			if (normal == DDI_FAILURE) {
1445 				PMD(PMD_ERROR | PMD_NORM, ("ioctl: %s: "
1446 				    "returns EINVAL\n", cmdstr))
1447 				ret = EINVAL;
1448 				break;
1449 			}
1450 			*rval_p = normal;
1451 			PMD(PMD_NORM, ("ioctl: %s: returns %d\n",
1452 			    cmdstr, normal))
1453 			ret = 0;
1454 			break;
1455 		}
1456 
1457 		case PM_GET_CURRENT_POWER:
1458 		{
1459 			if (pm_get_current_power(dip, req.component,
1460 			    rval_p) != DDI_SUCCESS) {
1461 				PMD(PMD_ERROR | PMD_DPM, ("ioctl: %s "
1462 				    "EINVAL\n", cmdstr))
1463 				ret = EINVAL;
1464 				break;
1465 			}
1466 			PMD(PMD_DPM, ("ioctl: %s: %s comp %d returns %d\n",
1467 			    cmdstr, req.physpath, req.component, *rval_p))
1468 			if (*rval_p == PM_LEVEL_UNKNOWN)
1469 				ret = EAGAIN;
1470 			else
1471 				ret = 0;
1472 			break;
1473 		}
1474 
1475 		case PM_GET_TIME_IDLE:
1476 		{
1477 			time_t timestamp;
1478 			int comp = req.component;
1479 			pm_component_t *cp;
1480 			if (!e_pm_valid_comp(dip, comp, &cp)) {
1481 				PMD(PMD_ERROR, ("ioctl: %s: %s@%s(%s#%d) "
1482 				    "component %d > numcmpts - 1 %d--EINVAL\n",
1483 				    cmdstr, PM_DEVICE(dip), comp,
1484 				    PM_NUMCMPTS(dip) - 1))
1485 				ret = EINVAL;
1486 				break;
1487 			}
1488 			timestamp = cp->pmc_timestamp;
1489 			if (timestamp) {
1490 				time_t now;
1491 				(void) drv_getparm(TIME, &now);
1492 				*rval_p = (now - timestamp);
1493 			} else {
1494 				*rval_p = 0;
1495 			}
1496 			ret = 0;
1497 			break;
1498 		}
1499 
1500 		case PM_ADD_DEPENDENT:
1501 		{
1502 			dev_info_t	*kept_dip;
1503 
1504 			PMD(PMD_KEEPS, ("%s, kept %s, keeper %s\n", cmdstr,
1505 			    dep, req.physpath))
1506 
1507 			/*
1508 			 * hold and install kept while processing dependency
1509 			 * keeper (in .physpath) has already been held.
1510 			 */
1511 			if (dep[0] == '\0') {
1512 				PMD(PMD_ERROR, ("kept NULL or null\n"))
1513 				ret = EINVAL;
1514 				break;
1515 			} else if ((kept_dip =
1516 			    pm_name_to_dip(dep, 1)) == NULL) {
1517 				PMD(PMD_ERROR, ("no dip for kept %s\n", dep))
1518 				ret = ENODEV;
1519 				break;
1520 			} else if (kept_dip == dip) {
1521 				PMD(PMD_ERROR, ("keeper(%s, %p) - kept(%s, %p) "
1522 				    "self-dependency not allowed.\n",
1523 				    dep, (void *)kept_dip, req.physpath,
1524 				    (void *) dip))
1525 				PM_RELE(dip);	/* release "double" hold */
1526 				ret = EINVAL;
1527 				break;
1528 			}
1529 			ASSERT(!(strcmp(req.physpath, (char *)dep) == 0));
1530 
1531 			/*
1532 			 * record dependency, then walk through device tree
1533 			 * independently on behalf of kept and keeper to
1534 			 * establish newly created dependency.
1535 			 */
1536 			pm_dispatch_to_dep_thread(PM_DEP_WK_RECORD_KEEPER,
1537 			    req.physpath, dep, PM_DEP_WAIT, NULL, 0);
1538 
1539 			/*
1540 			 * release kept after establishing dependency, keeper
1541 			 * is released as part of ioctl exit processing.
1542 			 */
1543 			PM_RELE(kept_dip);
1544 			*rval_p = 0;
1545 			ret = 0;
1546 			break;
1547 		}
1548 
1549 		case PM_ADD_DEPENDENT_PROPERTY:
1550 		{
1551 			char *keeper, *kept;
1552 
1553 			if (dep[0] == '\0') {
1554 				PMD(PMD_ERROR, ("ioctl: %s: dep NULL or "
1555 				    "null\n", cmdstr))
1556 				ret = EINVAL;
1557 				break;
1558 			}
1559 			kept = dep;
1560 			keeper = req.physpath;
1561 			/*
1562 			 * record keeper - kept dependency, then walk through
1563 			 * device tree to find out all attached keeper, walk
1564 			 * through again to apply dependency to all the
1565 			 * potential kept.
1566 			 */
1567 			pm_dispatch_to_dep_thread(
1568 			    PM_DEP_WK_RECORD_KEEPER_PROP, keeper, kept,
1569 			    PM_DEP_WAIT, NULL, 0);
1570 
1571 			*rval_p = 0;
1572 			ret = 0;
1573 			break;
1574 		}
1575 
1576 		case PM_SET_DEVICE_THRESHOLD:
1577 		{
1578 			pm_thresh_rec_t *rp;
1579 			pm_pte_t *ep;	/* threshold header storage */
1580 			int *tp;	/* threshold storage */
1581 			size_t size;
1582 			extern int pm_thresh_specd(dev_info_t *);
1583 
1584 			/*
1585 			 * The header struct plus one entry struct plus one
1586 			 * threshold plus the length of the string
1587 			 */
1588 			size = sizeof (pm_thresh_rec_t) +
1589 			    (sizeof (pm_pte_t) * 1) +
1590 			    (1 * sizeof (int)) +
1591 			    strlen(req.physpath) + 1;
1592 
1593 			rp = kmem_zalloc(size, KM_SLEEP);
1594 			rp->ptr_size = size;
1595 			rp->ptr_numcomps = 0;	/* means device threshold */
1596 			ep = (pm_pte_t *)((intptr_t)rp + sizeof (*rp));
1597 			rp->ptr_entries = ep;
1598 			tp = (int *)((intptr_t)ep +
1599 			    (1 * sizeof (pm_pte_t)));
1600 			ep->pte_numthresh = 1;
1601 			ep->pte_thresh = tp;
1602 			*tp++ = req.value;
1603 			(void) strcat((char *)tp, req.physpath);
1604 			rp->ptr_physpath = (char *)tp;
1605 			ASSERT((intptr_t)tp + strlen(req.physpath) + 1 ==
1606 			    (intptr_t)rp + rp->ptr_size);
1607 			PMD(PMD_THRESH, ("ioctl: %s: record thresh %d for "
1608 			    "%s\n", cmdstr, req.value, req.physpath))
1609 			pm_record_thresh(rp);
1610 			/*
1611 			 * Don't free rp, pm_record_thresh() keeps it.
1612 			 * We don't try to apply it ourselves because we'd need
1613 			 * to know too much about locking.  Since we don't
1614 			 * hold a lock the entry could be removed before
1615 			 * we get here
1616 			 */
1617 			ASSERT(dip == NULL);
1618 			ret = 0;		/* can't fail now */
1619 			if (!(dip = pm_name_to_dip(req.physpath, 1))) {
1620 				break;
1621 			}
1622 			(void) pm_thresh_specd(dip);
1623 			PMD(PMD_DHR, ("ioctl: %s: releasing %s@%s(%s#%d)\n",
1624 			    cmdstr, PM_DEVICE(dip)))
1625 			PM_RELE(dip);
1626 			break;
1627 		}
1628 
1629 		case PM_RESET_DEVICE_THRESHOLD:
1630 		{
1631 			/*
1632 			 * This only applies to a currently attached and power
1633 			 * managed node
1634 			 */
1635 			/*
1636 			 * We don't do this to old-style drivers
1637 			 */
1638 			info = PM_GET_PM_INFO(dip);
1639 			if (info == NULL) {
1640 				PMD(PMD_ERROR, ("ioctl: %s: %s not power "
1641 				    "managed\n", cmdstr, req.physpath))
1642 				ret = EINVAL;
1643 				break;
1644 			}
1645 			if (PM_ISBC(dip)) {
1646 				PMD(PMD_ERROR, ("ioctl: %s: %s is BC\n",
1647 				    cmdstr, req.physpath))
1648 				ret = EINVAL;
1649 				break;
1650 			}
1651 			pm_unrecord_threshold(req.physpath);
1652 			if (DEVI(dip)->devi_pm_flags & PMC_CPU_THRESH)
1653 				pm_set_device_threshold(dip,
1654 				    pm_cpu_idle_threshold, PMC_CPU_THRESH);
1655 			else
1656 				pm_set_device_threshold(dip,
1657 				    pm_system_idle_threshold, PMC_DEF_THRESH);
1658 			ret = 0;
1659 			break;
1660 		}
1661 
1662 		case PM_GET_NUM_COMPONENTS:
1663 		{
1664 			ret = 0;
1665 			*rval_p = PM_NUMCMPTS(dip);
1666 			break;
1667 		}
1668 
1669 		case PM_GET_DEVICE_TYPE:
1670 		{
1671 			ret = 0;
1672 			if ((info = PM_GET_PM_INFO(dip)) == NULL) {
1673 				PMD(PMD_ERROR, ("ioctl: %s: "
1674 				    "PM_NO_PM_COMPONENTS\n", cmdstr))
1675 				*rval_p = PM_NO_PM_COMPONENTS;
1676 				break;
1677 			}
1678 			if (PM_ISBC(dip)) {
1679 				*rval_p = PM_CREATE_COMPONENTS;
1680 			} else {
1681 				*rval_p = PM_AUTOPM;
1682 			}
1683 			break;
1684 		}
1685 
1686 		case PM_SET_COMPONENT_THRESHOLDS:
1687 		{
1688 			int comps = 0;
1689 			int *end = (int *)req.data + icount;
1690 			pm_thresh_rec_t *rp;
1691 			pm_pte_t *ep;	/* threshold header storage */
1692 			int *tp;	/* threshold storage */
1693 			int *ip;
1694 			int j;
1695 			size_t size;
1696 			extern int pm_thresh_specd(dev_info_t *);
1697 			extern int pm_valid_thresh(dev_info_t *,
1698 			    pm_thresh_rec_t *);
1699 
1700 			for (ip = req.data; *ip; ip++) {
1701 				if (ip >= end) {
1702 					ret = EFAULT;
1703 					break;
1704 				}
1705 				comps++;
1706 				/* skip over indicated number of entries */
1707 				for (j = *ip; j; j--) {
1708 					if (++ip >= end) {
1709 						ret = EFAULT;
1710 						break;
1711 					}
1712 				}
1713 				if (ret)
1714 					break;
1715 			}
1716 			if (ret)
1717 				break;
1718 			if ((intptr_t)ip != (intptr_t)end - sizeof (int)) {
1719 				/* did not exactly fill buffer */
1720 				ret = EINVAL;
1721 				break;
1722 			}
1723 			if (comps == 0) {
1724 				PMD(PMD_ERROR, ("ioctl: %s: %s 0 components"
1725 				    "--EINVAL\n", cmdstr, req.physpath))
1726 				ret = EINVAL;
1727 				break;
1728 			}
1729 			/*
1730 			 * The header struct plus one entry struct per component
1731 			 * plus the size of the lists minus the counts
1732 			 * plus the length of the string
1733 			 */
1734 			size = sizeof (pm_thresh_rec_t) +
1735 			    (sizeof (pm_pte_t) * comps) + req.datasize -
1736 			    ((comps + 1) * sizeof (int)) +
1737 			    strlen(req.physpath) + 1;
1738 
1739 			rp = kmem_zalloc(size, KM_SLEEP);
1740 			rp->ptr_size = size;
1741 			rp->ptr_numcomps = comps;
1742 			ep = (pm_pte_t *)((intptr_t)rp + sizeof (*rp));
1743 			rp->ptr_entries = ep;
1744 			tp = (int *)((intptr_t)ep +
1745 			    (comps * sizeof (pm_pte_t)));
1746 			for (ip = req.data; *ip; ep++) {
1747 				ep->pte_numthresh = *ip;
1748 				ep->pte_thresh = tp;
1749 				for (j = *ip++; j; j--) {
1750 					*tp++ = *ip++;
1751 				}
1752 			}
1753 			(void) strcat((char *)tp, req.physpath);
1754 			rp->ptr_physpath = (char *)tp;
1755 			ASSERT((intptr_t)end == (intptr_t)ip + sizeof (int));
1756 			ASSERT((intptr_t)tp + strlen(req.physpath) + 1 ==
1757 			    (intptr_t)rp + rp->ptr_size);
1758 
1759 			ASSERT(dip == NULL);
1760 			/*
1761 			 * If this is not a currently power managed node,
1762 			 * then we can't check for validity of the thresholds
1763 			 */
1764 			if (!(dip = pm_name_to_dip(req.physpath, 1))) {
1765 				/* don't free rp, pm_record_thresh uses it */
1766 				pm_record_thresh(rp);
1767 				PMD(PMD_ERROR, ("ioctl: %s: pm_name_to_dip "
1768 				    "for %s failed\n", cmdstr, req.physpath))
1769 				ret = 0;
1770 				break;
1771 			}
1772 			ASSERT(!dipheld);
1773 			dipheld++;
1774 
1775 			if (!pm_valid_thresh(dip, rp)) {
1776 				PMD(PMD_ERROR, ("ioctl: %s: invalid thresh "
1777 				    "for %s@%s(%s#%d)\n", cmdstr,
1778 				    PM_DEVICE(dip)))
1779 				kmem_free(rp, size);
1780 				ret = EINVAL;
1781 				break;
1782 			}
1783 			/*
1784 			 * We don't just apply it ourselves because we'd need
1785 			 * to know too much about locking.  Since we don't
1786 			 * hold a lock the entry could be removed before
1787 			 * we get here
1788 			 */
1789 			pm_record_thresh(rp);
1790 			(void) pm_thresh_specd(dip);
1791 			ret = 0;
1792 			break;
1793 		}
1794 
1795 		case PM_GET_COMPONENT_THRESHOLDS:
1796 		{
1797 			int musthave;
1798 			int numthresholds = 0;
1799 			int wordsize;
1800 			int numcomps;
1801 			caddr_t uaddr = req.data;	/* user address */
1802 			int val;	/* int value to be copied out */
1803 			int32_t val32;	/* int32 value to be copied out */
1804 			caddr_t vaddr;	/* address to copyout from */
1805 			int j;
1806 
1807 #ifdef	_MULTI_DATAMODEL
1808 			if ((mode & DATAMODEL_MASK) == DATAMODEL_ILP32) {
1809 				wordsize = sizeof (int32_t);
1810 			} else
1811 #endif /* _MULTI_DATAMODEL */
1812 			{
1813 				wordsize = sizeof (int);
1814 			}
1815 
1816 			ASSERT(dip);
1817 
1818 			numcomps = PM_NUMCMPTS(dip);
1819 			for (i = 0; i < numcomps; i++) {
1820 				cp = PM_CP(dip, i);
1821 				numthresholds += cp->pmc_comp.pmc_numlevels - 1;
1822 			}
1823 			musthave = (numthresholds + numcomps + 1) *  wordsize;
1824 			if (req.datasize < musthave) {
1825 				PMD(PMD_ERROR, ("ioctl: %s: size %ld, need "
1826 				    "%d--EINVAL\n", cmdstr, req.datasize,
1827 				    musthave))
1828 				ret = EINVAL;
1829 				break;
1830 			}
1831 			PM_LOCK_DIP(dip);
1832 			for (i = 0; i < numcomps; i++) {
1833 				int *thp;
1834 				cp = PM_CP(dip, i);
1835 				thp = cp->pmc_comp.pmc_thresh;
1836 				/* first copyout the count */
1837 				if (wordsize == sizeof (int32_t)) {
1838 					val32 = cp->pmc_comp.pmc_numlevels - 1;
1839 					vaddr = (caddr_t)&val32;
1840 				} else {
1841 					val = cp->pmc_comp.pmc_numlevels - 1;
1842 					vaddr = (caddr_t)&val;
1843 				}
1844 				if (ddi_copyout(vaddr, (void *)uaddr,
1845 				    wordsize, mode) != 0) {
1846 					PM_UNLOCK_DIP(dip);
1847 					PMD(PMD_ERROR, ("ioctl: %s: %s@%s"
1848 					    "(%s#%d) vaddr %p EFAULT\n",
1849 					    cmdstr, PM_DEVICE(dip),
1850 					    (void*)vaddr))
1851 					ret = EFAULT;
1852 					break;
1853 				}
1854 				vaddr = uaddr;
1855 				vaddr += wordsize;
1856 				uaddr = (caddr_t)vaddr;
1857 				/* then copyout each threshold value */
1858 				for (j = 0; j < cp->pmc_comp.pmc_numlevels - 1;
1859 				    j++) {
1860 					if (wordsize == sizeof (int32_t)) {
1861 						val32 = thp[j + 1];
1862 						vaddr = (caddr_t)&val32;
1863 					} else {
1864 						val = thp[i + 1];
1865 						vaddr = (caddr_t)&val;
1866 					}
1867 					if (ddi_copyout(vaddr, (void *) uaddr,
1868 					    wordsize, mode) != 0) {
1869 						PM_UNLOCK_DIP(dip);
1870 						PMD(PMD_ERROR, ("ioctl: %s: "
1871 						    "%s@%s(%s#%d) uaddr %p "
1872 						    "EFAULT\n", cmdstr,
1873 						    PM_DEVICE(dip),
1874 						    (void *)uaddr))
1875 						ret = EFAULT;
1876 						break;
1877 					}
1878 					vaddr = uaddr;
1879 					vaddr += wordsize;
1880 					uaddr = (caddr_t)vaddr;
1881 				}
1882 			}
1883 			if (ret)
1884 				break;
1885 			/* last copyout a terminating 0 count */
1886 			if (wordsize == sizeof (int32_t)) {
1887 				val32 = 0;
1888 				vaddr = (caddr_t)&val32;
1889 			} else {
1890 				ASSERT(wordsize == sizeof (int));
1891 				val = 0;
1892 				vaddr = (caddr_t)&val;
1893 			}
1894 			if (ddi_copyout(vaddr, uaddr, wordsize, mode) != 0) {
1895 				PM_UNLOCK_DIP(dip);
1896 				PMD(PMD_ERROR, ("ioctl: %s: %s@%s(%s#%d) "
1897 				    "vaddr %p (0 count) EFAULT\n", cmdstr,
1898 				    PM_DEVICE(dip), (void *)vaddr))
1899 				ret = EFAULT;
1900 				break;
1901 			}
1902 			/* finished, so don't need to increment addresses */
1903 			PM_UNLOCK_DIP(dip);
1904 			ret = 0;
1905 			break;
1906 		}
1907 
1908 		case PM_GET_STATS:
1909 		{
1910 			time_t now;
1911 			time_t *timestamp;
1912 			extern int pm_cur_power(pm_component_t *cp);
1913 			int musthave;
1914 			int wordsize;
1915 
1916 #ifdef	_MULTI_DATAMODEL
1917 			if ((mode & DATAMODEL_MASK) == DATAMODEL_ILP32) {
1918 				wordsize = sizeof (int32_t);
1919 			} else
1920 #endif /* _MULTI_DATAMODEL */
1921 			{
1922 				wordsize = sizeof (int);
1923 			}
1924 
1925 			comps = PM_NUMCMPTS(dip);
1926 			if (comps == 0 || PM_GET_PM_INFO(dip) == NULL) {
1927 				PMD(PMD_ERROR, ("ioctl: %s: %s no components"
1928 				    " or not power managed--EINVAL\n", cmdstr,
1929 				    req.physpath))
1930 				ret = EINVAL;
1931 				break;
1932 			}
1933 			musthave = comps * 2 * wordsize;
1934 			if (req.datasize < musthave) {
1935 				PMD(PMD_ERROR, ("ioctl: %s: size %lu, need "
1936 				    "%d--EINVAL\n", cmdstr, req.datasize,
1937 				    musthave))
1938 				ret = EINVAL;
1939 				break;
1940 			}
1941 
1942 			PM_LOCK_DIP(dip);
1943 			(void) drv_getparm(TIME, &now);
1944 			timestamp = kmem_zalloc(comps * sizeof (time_t),
1945 			    KM_SLEEP);
1946 			pm_get_timestamps(dip, timestamp);
1947 			/*
1948 			 * First the current power levels
1949 			 */
1950 			for (i = 0; i < comps; i++) {
1951 				int curpwr;
1952 				int32_t curpwr32;
1953 				caddr_t cpaddr;
1954 
1955 				cp = PM_CP(dip, i);
1956 				if (wordsize == sizeof (int)) {
1957 					curpwr = pm_cur_power(cp);
1958 					cpaddr = (caddr_t)&curpwr;
1959 				} else {
1960 					ASSERT(wordsize == sizeof (int32_t));
1961 					curpwr32 = pm_cur_power(cp);
1962 					cpaddr = (caddr_t)&curpwr32;
1963 				}
1964 				if (ddi_copyout(cpaddr, (void *) req.data,
1965 				    wordsize, mode) != 0) {
1966 					PM_UNLOCK_DIP(dip);
1967 					PMD(PMD_ERROR, ("ioctl: %s: %s@%s"
1968 					    "(%s#%d) req.data %p EFAULT\n",
1969 					    cmdstr, PM_DEVICE(dip),
1970 					    (void *)req.data))
1971 					ASSERT(!dipheld);
1972 					return (EFAULT);
1973 				}
1974 				cpaddr = (caddr_t)req.data;
1975 				cpaddr += wordsize;
1976 				req.data = cpaddr;
1977 			}
1978 			/*
1979 			 * Then the times remaining
1980 			 */
1981 			for (i = 0; i < comps; i++) {
1982 				int retval;
1983 				int32_t retval32;
1984 				caddr_t rvaddr;
1985 				int curpwr;
1986 
1987 				cp = PM_CP(dip, i);
1988 				curpwr = cp->pmc_cur_pwr;
1989 				if (curpwr == 0 || timestamp[i] == 0) {
1990 					PMD(PMD_STATS, ("ioctl: %s: "
1991 					    "cur_pwer %x, timestamp %lx\n",
1992 					    cmdstr, curpwr, timestamp[i]))
1993 					retval = INT_MAX;
1994 				} else {
1995 					int thresh;
1996 					(void) pm_current_threshold(dip, i,
1997 					    &thresh);
1998 					retval = thresh - (now - timestamp[i]);
1999 					PMD(PMD_STATS, ("ioctl: %s: current "
2000 					    "thresh %x, now %lx, timestamp %lx,"
2001 					    " retval %x\n", cmdstr, thresh, now,
2002 					    timestamp[i], retval))
2003 				}
2004 				if (wordsize == sizeof (int)) {
2005 					rvaddr = (caddr_t)&retval;
2006 				} else {
2007 					ASSERT(wordsize == sizeof (int32_t));
2008 					retval32 = retval;
2009 					rvaddr = (caddr_t)&retval32;
2010 				}
2011 				if (ddi_copyout(rvaddr, (void *) req.data,
2012 				    wordsize, mode) != 0) {
2013 					PM_UNLOCK_DIP(dip);
2014 					PMD(PMD_ERROR, ("ioctl: %s: %s@%s"
2015 					    "(%s#%d) req.data %p EFAULT\n",
2016 					    cmdstr, PM_DEVICE(dip),
2017 					    (void *)req.data))
2018 					ASSERT(!dipheld);
2019 					kmem_free(timestamp,
2020 					    comps * sizeof (time_t));
2021 					return (EFAULT);
2022 				}
2023 				rvaddr = (caddr_t)req.data;
2024 				rvaddr += wordsize;
2025 				req.data = (int *)rvaddr;
2026 			}
2027 			PM_UNLOCK_DIP(dip);
2028 			*rval_p = comps;
2029 			ret = 0;
2030 			kmem_free(timestamp, comps * sizeof (time_t));
2031 			break;
2032 		}
2033 
2034 		case PM_GET_CMD_NAME:
2035 		{
2036 			PMD(PMD_IOCTL, ("%s: %s\n", cmdstr,
2037 			    pm_decode_cmd(req.value)))
2038 			if (ret = copyoutstr(pm_decode_cmd(req.value),
2039 			    (char *)req.data, req.datasize, &lencopied)) {
2040 				PMD(PMD_ERROR, ("ioctl: %s: %s@%s(%s#%d) "
2041 				    "copyoutstr %p failed--EFAULT\n", cmdstr,
2042 				    PM_DEVICE(dip), (void *)req.data))
2043 				break;
2044 			}
2045 			*rval_p = lencopied;
2046 			ret = 0;
2047 			break;
2048 		}
2049 
2050 		case PM_GET_COMPONENT_NAME:
2051 		{
2052 			ASSERT(dip);
2053 			if (!e_pm_valid_comp(dip, req.component, &cp)) {
2054 				PMD(PMD_ERROR, ("ioctl: %s: %s@%s(%s#%d) "
2055 				    "component %d > numcmpts - 1 %d--EINVAL\n",
2056 				    cmdstr, PM_DEVICE(dip), req.component,
2057 				    PM_NUMCMPTS(dip) - 1))
2058 				ret = EINVAL;
2059 				break;
2060 			}
2061 			if (ret = copyoutstr(cp->pmc_comp.pmc_name,
2062 			    (char *)req.data, req.datasize, &lencopied)) {
2063 				PMD(PMD_ERROR, ("ioctl: %s: %s@%s(%s#%d) "
2064 				    "copyoutstr %p failed--EFAULT\n", cmdstr,
2065 				    PM_DEVICE(dip), (void *)req.data))
2066 				break;
2067 			}
2068 			*rval_p = lencopied;
2069 			ret = 0;
2070 			break;
2071 		}
2072 
2073 		case PM_GET_POWER_NAME:
2074 		{
2075 			int i;
2076 
2077 			ASSERT(dip);
2078 			if (!e_pm_valid_comp(dip, req.component, &cp)) {
2079 				PMD(PMD_ERROR, ("ioctl: %s: %s@%s(%s#%d) "
2080 				    "component %d > numcmpts - 1 %d--EINVAL\n",
2081 				    cmdstr, PM_DEVICE(dip), req.component,
2082 				    PM_NUMCMPTS(dip) - 1))
2083 				ret = EINVAL;
2084 				break;
2085 			}
2086 			if ((i = req.value) < 0 ||
2087 			    i > cp->pmc_comp.pmc_numlevels - 1) {
2088 				PMD(PMD_ERROR, ("ioctl: %s: %s@%s(%s#%d) "
2089 				    "value %d > num_levels - 1 %d--EINVAL\n",
2090 				    cmdstr, PM_DEVICE(dip), req.value,
2091 				    cp->pmc_comp.pmc_numlevels - 1))
2092 				ret = EINVAL;
2093 				break;
2094 			}
2095 			dep = cp->pmc_comp.pmc_lnames[req.value];
2096 			if (ret = copyoutstr(dep,
2097 			    req.data, req.datasize, &lencopied)) {
2098 				PMD(PMD_ERROR, ("ioctl: %s: %s@%s(%s#%d) "
2099 				    "copyoutstr %p failed--EFAULT\n", cmdstr,
2100 				    PM_DEVICE(dip), (void *)req.data))
2101 				break;
2102 			}
2103 			*rval_p = lencopied;
2104 			ret = 0;
2105 			break;
2106 		}
2107 
2108 		case PM_GET_POWER_LEVELS:
2109 		{
2110 			int musthave;
2111 			int numlevels;
2112 			int wordsize;
2113 
2114 #ifdef	_MULTI_DATAMODEL
2115 			if ((mode & DATAMODEL_MASK) == DATAMODEL_ILP32) {
2116 				wordsize = sizeof (int32_t);
2117 			} else
2118 #endif /* _MULTI_DATAMODEL */
2119 			{
2120 				wordsize = sizeof (int);
2121 			}
2122 			ASSERT(dip);
2123 
2124 			if (!e_pm_valid_comp(dip, req.component, &cp)) {
2125 				PMD(PMD_ERROR, ("ioctl: %s: %s@%s(%s#%d) "
2126 				    "has %d components, component %d requested"
2127 				    "--EINVAL\n", cmdstr, PM_DEVICE(dip),
2128 				    PM_NUMCMPTS(dip), req.component))
2129 				ret = EINVAL;
2130 				break;
2131 			}
2132 			numlevels = cp->pmc_comp.pmc_numlevels;
2133 			musthave = numlevels *  wordsize;
2134 			if (req.datasize < musthave) {
2135 				PMD(PMD_ERROR, ("ioctl: %s: size %lu, need "
2136 				    "%d--EINVAL\n", cmdstr, req.datasize,
2137 				    musthave))
2138 				ret = EINVAL;
2139 				break;
2140 			}
2141 			PM_LOCK_DIP(dip);
2142 			for (i = 0; i < numlevels; i++) {
2143 				int level;
2144 				int32_t level32;
2145 				caddr_t laddr;
2146 
2147 				if (wordsize == sizeof (int)) {
2148 					level = cp->pmc_comp.pmc_lvals[i];
2149 					laddr = (caddr_t)&level;
2150 				} else {
2151 					level32 = cp->pmc_comp.pmc_lvals[i];
2152 					laddr = (caddr_t)&level32;
2153 				}
2154 				if (ddi_copyout(laddr, (void *) req.data,
2155 				    wordsize, mode) != 0) {
2156 					PM_UNLOCK_DIP(dip);
2157 					PMD(PMD_ERROR, ("ioctl: %s: %s@%s"
2158 					    "(%s#%d) laddr %p EFAULT\n",
2159 					    cmdstr, PM_DEVICE(dip),
2160 					    (void *)laddr))
2161 					ASSERT(!dipheld);
2162 					return (EFAULT);
2163 				}
2164 				laddr = (caddr_t)req.data;
2165 				laddr += wordsize;
2166 				req.data = (int *)laddr;
2167 			}
2168 			PM_UNLOCK_DIP(dip);
2169 			*rval_p = numlevels;
2170 			ret = 0;
2171 			break;
2172 		}
2173 
2174 
2175 		case PM_GET_NUM_POWER_LEVELS:
2176 		{
2177 			if (!e_pm_valid_comp(dip, req.component, &cp)) {
2178 				PMD(PMD_ERROR, ("ioctl: %s: %s@%s(%s#%d) "
2179 				    "component %d > numcmpts - 1 %d--EINVAL\n",
2180 				    cmdstr, PM_DEVICE(dip), req.component,
2181 				    PM_NUMCMPTS(dip) - 1))
2182 				ret = EINVAL;
2183 				break;
2184 			}
2185 			*rval_p = cp->pmc_comp.pmc_numlevels;
2186 			ret = 0;
2187 			break;
2188 		}
2189 
2190 		case PM_GET_DEVICE_THRESHOLD_BASIS:
2191 		{
2192 			ret = 0;
2193 			PM_LOCK_DIP(dip);
2194 			if ((info = PM_GET_PM_INFO(dip)) == NULL) {
2195 				PM_UNLOCK_DIP(dip);
2196 				PMD(PMD_ERROR, ("ioctl: %s: "
2197 				    "PM_NO_PM_COMPONENTS\n", cmdstr))
2198 				*rval_p = PM_NO_PM_COMPONENTS;
2199 				break;
2200 			}
2201 			if (PM_ISDIRECT(dip)) {
2202 				PM_UNLOCK_DIP(dip);
2203 				*rval_p = PM_DIRECTLY_MANAGED;
2204 				break;
2205 			}
2206 			switch (DEVI(dip)->devi_pm_flags & PMC_THRESH_ALL) {
2207 			case PMC_DEF_THRESH:
2208 			case PMC_NEXDEF_THRESH:
2209 				*rval_p = PM_DEFAULT_THRESHOLD;
2210 				break;
2211 			case PMC_DEV_THRESH:
2212 				*rval_p = PM_DEVICE_THRESHOLD;
2213 				break;
2214 			case PMC_COMP_THRESH:
2215 				*rval_p = PM_COMPONENT_THRESHOLD;
2216 				break;
2217 			case PMC_CPU_THRESH:
2218 				*rval_p = PM_CPU_THRESHOLD;
2219 				break;
2220 			default:
2221 				if (PM_ISBC(dip)) {
2222 					*rval_p = PM_OLD_THRESHOLD;
2223 					break;
2224 				}
2225 				PMD(PMD_ERROR, ("ioctl: %s: default, not "
2226 				    "BC--EINVAL", cmdstr))
2227 				ret = EINVAL;
2228 				break;
2229 			}
2230 			PM_UNLOCK_DIP(dip);
2231 			break;
2232 		}
2233 		default:
2234 			/*
2235 			 * Internal error, invalid ioctl description
2236 			 * force debug entry even if pm_debug not set
2237 			 */
2238 #ifdef	DEBUG
2239 			pm_log("invalid diptype %d for cmd %d (%s)\n",
2240 			    pcip->diptype, cmd, pcip->name);
2241 #endif
2242 			ASSERT(0);
2243 			return (EIO);
2244 		}
2245 		break;
2246 	}
2247 
2248 	case PM_PSC:
2249 	{
2250 		/*
2251 		 * Commands that require pm_state_change_t as arg
2252 		 */
2253 #ifdef	_MULTI_DATAMODEL
2254 		if ((mode & DATAMODEL_MASK) == DATAMODEL_ILP32) {
2255 			pscp32 = (pm_state_change32_t *)arg;
2256 			if (ddi_copyin((caddr_t)arg, &psc32,
2257 			    sizeof (psc32), mode) != 0) {
2258 				PMD(PMD_ERROR, ("ioctl: %s: ddi_copyin "
2259 				    "EFAULT\n\n", cmdstr))
2260 				ASSERT(!dipheld);
2261 				return (EFAULT);
2262 			}
2263 			psc.physpath = (caddr_t)(uintptr_t)psc32.physpath;
2264 			psc.size = psc32.size;
2265 		} else
2266 #endif /* _MULTI_DATAMODEL */
2267 		{
2268 			pscp = (pm_state_change_t *)arg;
2269 			if (ddi_copyin((caddr_t)arg, &psc,
2270 			    sizeof (psc), mode) != 0) {
2271 				PMD(PMD_ERROR, ("ioctl: %s: ddi_copyin "
2272 				    "EFAULT\n\n", cmdstr))
2273 				ASSERT(!dipheld);
2274 				return (EFAULT);
2275 			}
2276 		}
2277 		switch (cmd) {
2278 
2279 		case PM_GET_STATE_CHANGE:
2280 		case PM_GET_STATE_CHANGE_WAIT:
2281 		{
2282 			psce_t			*pscep;
2283 			pm_state_change_t	*p;
2284 			caddr_t			physpath;
2285 			size_t			physlen;
2286 
2287 			/*
2288 			 * We want to know if any device has changed state.
2289 			 * We look up by clone.  In case we have another thread
2290 			 * from the same process, we loop.
2291 			 * pm_psc_clone_to_interest() returns a locked entry.
2292 			 * We create an internal copy of the event entry prior
2293 			 * to copyout to user space because we don't want to
2294 			 * hold the psce_lock while doing copyout as we might
2295 			 * hit page fault  which eventually brings us back
2296 			 * here requesting the same lock.
2297 			 */
2298 			mutex_enter(&pm_clone_lock);
2299 			if (!pm_interest_registered(clone))
2300 				pm_register_watcher(clone, NULL);
2301 			while ((pscep =
2302 			    pm_psc_clone_to_interest(clone)) == NULL) {
2303 				if (cmd == PM_GET_STATE_CHANGE) {
2304 					PMD(PMD_IOCTL, ("ioctl: %s: "
2305 					    "EWOULDBLOCK\n", cmdstr))
2306 					mutex_exit(&pm_clone_lock);
2307 					ASSERT(!dipheld);
2308 					return (EWOULDBLOCK);
2309 				} else {
2310 					if (cv_wait_sig(&pm_clones_cv[clone],
2311 					    &pm_clone_lock) == 0) {
2312 						mutex_exit(&pm_clone_lock);
2313 						PMD(PMD_ERROR, ("ioctl: %s "
2314 						    "EINTR\n", cmdstr))
2315 						ASSERT(!dipheld);
2316 						return (EINTR);
2317 					}
2318 				}
2319 			}
2320 			mutex_exit(&pm_clone_lock);
2321 
2322 			physlen = pscep->psce_out->size;
2323 			physpath = NULL;
2324 			/*
2325 			 * If we were unable to store the path while bringing
2326 			 * up the console fb upon entering the prom, we give
2327 			 * a "" name with the overrun event set
2328 			 */
2329 			if (physlen == (size_t)-1) {	/* kmemalloc failed */
2330 				physpath = kmem_zalloc(1, KM_SLEEP);
2331 				physlen = 1;
2332 			}
2333 			if ((psc.physpath == NULL) || (psc.size < physlen)) {
2334 				PMD(PMD_ERROR, ("ioctl: %s: EFAULT\n", cmdstr))
2335 				mutex_exit(&pscep->psce_lock);
2336 				ret = EFAULT;
2337 				break;
2338 			}
2339 			if (physpath == NULL) {
2340 				physpath = kmem_zalloc(physlen, KM_SLEEP);
2341 				bcopy((const void *) pscep->psce_out->physpath,
2342 				    (void *) physpath, physlen);
2343 			}
2344 
2345 			p = pscep->psce_out;
2346 #ifdef	_MULTI_DATAMODEL
2347 			if ((mode & DATAMODEL_MASK) == DATAMODEL_ILP32) {
2348 #ifdef DEBUG
2349 				size_t usrcopysize;
2350 #endif
2351 				psc32.flags = (ushort_t)p->flags;
2352 				psc32.event = (ushort_t)p->event;
2353 				psc32.timestamp = (int32_t)p->timestamp;
2354 				psc32.component = (int32_t)p->component;
2355 				psc32.old_level = (int32_t)p->old_level;
2356 				psc32.new_level = (int32_t)p->new_level;
2357 				copysize32 = ((intptr_t)&psc32.size -
2358 				    (intptr_t)&psc32.component);
2359 #ifdef DEBUG
2360 				usrcopysize = ((intptr_t)&pscp32->size -
2361 				    (intptr_t)&pscp32->component);
2362 				ASSERT(usrcopysize == copysize32);
2363 #endif
2364 			} else
2365 #endif /* _MULTI_DATAMODEL */
2366 			{
2367 				psc.flags = p->flags;
2368 				psc.event = p->event;
2369 				psc.timestamp = p->timestamp;
2370 				psc.component = p->component;
2371 				psc.old_level = p->old_level;
2372 				psc.new_level = p->new_level;
2373 				copysize = ((long)&p->size -
2374 				    (long)&p->component);
2375 			}
2376 			if (p->size != (size_t)-1)
2377 				kmem_free(p->physpath, p->size);
2378 			p->size = 0;
2379 			p->physpath = NULL;
2380 			if (pscep->psce_out == pscep->psce_last)
2381 				p = pscep->psce_first;
2382 			else
2383 				p++;
2384 			pscep->psce_out = p;
2385 			mutex_exit(&pscep->psce_lock);
2386 
2387 			ret = copyoutstr(physpath, psc.physpath,
2388 			    physlen, &lencopied);
2389 			kmem_free(physpath, physlen);
2390 			if (ret) {
2391 				PMD(PMD_ERROR, ("ioctl: %s: copyoutstr %p "
2392 				    "failed--EFAULT\n", cmdstr,
2393 				    (void *)psc.physpath))
2394 				break;
2395 			}
2396 
2397 #ifdef	_MULTI_DATAMODEL
2398 			if ((mode & DATAMODEL_MASK) == DATAMODEL_ILP32) {
2399 				if (ddi_copyout(&psc32.component,
2400 				    &pscp32->component, copysize32, mode)
2401 				    != 0) {
2402 					PMD(PMD_ERROR, ("ioctl: %s: copyout "
2403 					    "failed--EFAULT\n", cmdstr))
2404 					ret = EFAULT;
2405 					break;
2406 				}
2407 			} else
2408 #endif	/* _MULTI_DATAMODEL */
2409 			{
2410 				if (ddi_copyout(&psc.component,
2411 				    &pscp->component, copysize, mode) != 0) {
2412 					PMD(PMD_ERROR, ("ioctl: %s: copyout "
2413 					    "failed--EFAULT\n", cmdstr))
2414 					ret = EFAULT;
2415 					break;
2416 				}
2417 			}
2418 			ret = 0;
2419 			break;
2420 		}
2421 
2422 		case PM_DIRECT_NOTIFY:
2423 		case PM_DIRECT_NOTIFY_WAIT:
2424 		{
2425 			psce_t			*pscep;
2426 			pm_state_change_t	*p;
2427 			caddr_t			physpath;
2428 			size_t			physlen;
2429 			/*
2430 			 * We want to know if any direct device of ours has
2431 			 * something we should know about.  We look up by clone.
2432 			 * In case we have another thread from the same process,
2433 			 * we loop.
2434 			 * pm_psc_clone_to_direct() returns a locked entry.
2435 			 */
2436 			mutex_enter(&pm_clone_lock);
2437 			while (pm_poll_cnt[clone] == 0 ||
2438 			    (pscep = pm_psc_clone_to_direct(clone)) == NULL) {
2439 				if (cmd == PM_DIRECT_NOTIFY) {
2440 					PMD(PMD_IOCTL, ("ioctl: %s: "
2441 					    "EWOULDBLOCK\n", cmdstr))
2442 					mutex_exit(&pm_clone_lock);
2443 					ASSERT(!dipheld);
2444 					return (EWOULDBLOCK);
2445 				} else {
2446 					if (cv_wait_sig(&pm_clones_cv[clone],
2447 					    &pm_clone_lock) == 0) {
2448 						mutex_exit(&pm_clone_lock);
2449 						PMD(PMD_ERROR, ("ioctl: %s: "
2450 						    "EINTR\n", cmdstr))
2451 						ASSERT(!dipheld);
2452 						return (EINTR);
2453 					}
2454 				}
2455 			}
2456 			mutex_exit(&pm_clone_lock);
2457 			physlen = pscep->psce_out->size;
2458 			if ((psc.physpath == NULL) || (psc.size < physlen)) {
2459 				mutex_exit(&pscep->psce_lock);
2460 				PMD(PMD_ERROR, ("ioctl: %s: EFAULT\n",
2461 				    cmdstr))
2462 				ret = EFAULT;
2463 				break;
2464 			}
2465 			physpath = kmem_zalloc(physlen, KM_SLEEP);
2466 			bcopy((const void *) pscep->psce_out->physpath,
2467 			    (void *) physpath, physlen);
2468 
2469 			p = pscep->psce_out;
2470 #ifdef	_MULTI_DATAMODEL
2471 			if ((mode & DATAMODEL_MASK) == DATAMODEL_ILP32) {
2472 #ifdef DEBUG
2473 				size_t usrcopysize;
2474 #endif
2475 				psc32.component = (int32_t)p->component;
2476 				psc32.flags = (ushort_t)p->flags;
2477 				psc32.event = (ushort_t)p->event;
2478 				psc32.timestamp = (int32_t)p->timestamp;
2479 				psc32.old_level = (int32_t)p->old_level;
2480 				psc32.new_level = (int32_t)p->new_level;
2481 				copysize32 = (intptr_t)&psc32.size -
2482 				    (intptr_t)&psc32.component;
2483 				PMD(PMD_DPM, ("ioctl: %s: PDN32 %s, comp %d "
2484 				    "%d -> %d\n", cmdstr, physpath,
2485 				    p->component, p->old_level, p->new_level))
2486 #ifdef DEBUG
2487 				usrcopysize = (intptr_t)&pscp32->size -
2488 				    (intptr_t)&pscp32->component;
2489 				ASSERT(usrcopysize == copysize32);
2490 #endif
2491 			} else
2492 #endif
2493 			{
2494 				psc.component = p->component;
2495 				psc.flags = p->flags;
2496 				psc.event = p->event;
2497 				psc.timestamp = p->timestamp;
2498 				psc.old_level = p->old_level;
2499 				psc.new_level = p->new_level;
2500 				copysize = (intptr_t)&p->size -
2501 				    (intptr_t)&p->component;
2502 				PMD(PMD_DPM, ("ioctl: %s: PDN %s, comp %d "
2503 				    "%d -> %d\n", cmdstr, physpath,
2504 				    p->component, p->old_level, p->new_level))
2505 			}
2506 			mutex_enter(&pm_clone_lock);
2507 			PMD(PMD_IOCTL, ("ioctl: %s: pm_poll_cnt[%d] is %d "
2508 			    "before decrement\n", cmdstr, clone,
2509 			    pm_poll_cnt[clone]))
2510 			pm_poll_cnt[clone]--;
2511 			mutex_exit(&pm_clone_lock);
2512 			kmem_free(p->physpath, p->size);
2513 			p->size = 0;
2514 			p->physpath = NULL;
2515 			if (pscep->psce_out == pscep->psce_last)
2516 				p = pscep->psce_first;
2517 			else
2518 				p++;
2519 			pscep->psce_out = p;
2520 			mutex_exit(&pscep->psce_lock);
2521 
2522 			ret = copyoutstr(physpath, psc.physpath,
2523 			    physlen, &lencopied);
2524 			kmem_free(physpath, physlen);
2525 			if (ret) {
2526 				PMD(PMD_ERROR, ("ioctl: %s: copyoutstr %p "
2527 				    "failed--EFAULT\n", cmdstr,
2528 				    (void *)psc.physpath))
2529 				break;
2530 			}
2531 
2532 #ifdef	_MULTI_DATAMODEL
2533 			if ((mode & DATAMODEL_MASK) == DATAMODEL_ILP32) {
2534 				if (ddi_copyout(&psc32.component,
2535 				    &pscp32->component, copysize32, mode)
2536 				    != 0) {
2537 					PMD(PMD_ERROR, ("ioctl: %s: copyout "
2538 					    "failed--EFAULT\n", cmdstr))
2539 					ret = EFAULT;
2540 					break;
2541 				}
2542 			} else
2543 #endif	/* _MULTI_DATAMODEL */
2544 			{
2545 				if (ddi_copyout(&psc.component,
2546 				    &pscp->component, copysize, mode) != 0) {
2547 					PMD(PMD_ERROR, ("ioctl: %s: copyout "
2548 					    "failed--EFAULT\n", cmdstr))
2549 					ret = EFAULT;
2550 					break;
2551 				}
2552 			}
2553 			ret = 0;
2554 			break;
2555 		}
2556 		default:
2557 			/*
2558 			 * Internal error, invalid ioctl description
2559 			 * force debug entry even if pm_debug not set
2560 			 */
2561 #ifdef	DEBUG
2562 			pm_log("invalid diptype %d for cmd %d (%s)\n",
2563 			    pcip->diptype, cmd, pcip->name);
2564 #endif
2565 			ASSERT(0);
2566 			return (EIO);
2567 		}
2568 		break;
2569 	}
2570 
2571 	case PM_SRCH:		/* command that takes a pm_searchargs_t arg */
2572 	{
2573 		/*
2574 		 * If no ppm, then there is nothing to search.
2575 		 */
2576 		if (DEVI(ddi_root_node())->devi_pm_ppm == NULL) {
2577 			ret = ENODEV;
2578 			break;
2579 		}
2580 
2581 #ifdef	_MULTI_DATAMODEL
2582 		if ((mode & DATAMODEL_MASK) == DATAMODEL_ILP32) {
2583 			if (ddi_copyin((caddr_t)arg, &psa32,
2584 			    sizeof (psa32), mode) != 0) {
2585 				PMD(PMD_ERROR, ("ioctl: %s: ddi_copyin "
2586 				    "EFAULT\n\n", cmdstr))
2587 				return (EFAULT);
2588 			}
2589 			if (copyinstr((void *)(uintptr_t)psa32.pms_listname,
2590 			    listname, MAXCOPYBUF, NULL)) {
2591 				PMD(PMD_ERROR, ("ioctl: %s: 0x%p MAXCOPYBUF "
2592 				    "%d, " "EFAULT\n", cmdstr,
2593 				    (void *)(uintptr_t)psa32.pms_listname,
2594 				    MAXCOPYBUF))
2595 				ret = EFAULT;
2596 				break;
2597 			}
2598 			if (copyinstr((void *)(uintptr_t)psa32.pms_manufacturer,
2599 			    manufacturer, MAXCOPYBUF, NULL)) {
2600 				PMD(PMD_ERROR, ("ioctl: %s: 0x%p MAXCOPYBUF "
2601 				    "%d, " "EFAULT\n", cmdstr,
2602 				    (void *)(uintptr_t)psa32.pms_manufacturer,
2603 				    MAXCOPYBUF))
2604 				ret = EFAULT;
2605 				break;
2606 			}
2607 			if (copyinstr((void *)(uintptr_t)psa32.pms_product,
2608 			    product, MAXCOPYBUF, NULL)) {
2609 				PMD(PMD_ERROR, ("ioctl: %s: 0x%p MAXCOPYBUF "
2610 				    "%d, " "EFAULT\n", cmdstr,
2611 				    (void *)(uintptr_t)psa32.pms_product,
2612 				    MAXCOPYBUF))
2613 				ret = EFAULT;
2614 				break;
2615 			}
2616 		} else
2617 #endif /* _MULTI_DATAMODEL */
2618 		{
2619 			if (ddi_copyin((caddr_t)arg, &psa,
2620 			    sizeof (psa), mode) != 0) {
2621 				PMD(PMD_ERROR, ("ioctl: %s: ddi_copyin "
2622 				    "EFAULT\n\n", cmdstr))
2623 				return (EFAULT);
2624 			}
2625 			if (copyinstr(psa.pms_listname,
2626 			    listname, MAXCOPYBUF, NULL)) {
2627 				PMD(PMD_ERROR, ("ioctl: %s: 0x%p MAXCOPYBUF "
2628 				    "%d, " "EFAULT\n", cmdstr,
2629 				    (void *)psa.pms_listname, MAXCOPYBUF))
2630 				ret = EFAULT;
2631 				break;
2632 			}
2633 			if (copyinstr(psa.pms_manufacturer,
2634 			    manufacturer, MAXCOPYBUF, NULL)) {
2635 				PMD(PMD_ERROR, ("ioctl: %s: 0x%p MAXCOPYBUF "
2636 				    "%d, " "EFAULT\n", cmdstr,
2637 				    (void *)psa.pms_manufacturer, MAXCOPYBUF))
2638 				ret = EFAULT;
2639 				break;
2640 			}
2641 			if (copyinstr(psa.pms_product,
2642 			    product, MAXCOPYBUF, NULL)) {
2643 				PMD(PMD_ERROR, ("ioctl: %s: 0x%p MAXCOPYBUF "
2644 				    "%d, " "EFAULT\n", cmdstr,
2645 				    (void *)psa.pms_product, MAXCOPYBUF))
2646 				ret = EFAULT;
2647 				break;
2648 			}
2649 		}
2650 		psa.pms_listname = listname;
2651 		psa.pms_manufacturer = manufacturer;
2652 		psa.pms_product = product;
2653 		switch (cmd) {
2654 		case PM_SEARCH_LIST:
2655 			ret = pm_ppm_searchlist(&psa);
2656 			break;
2657 
2658 		default:
2659 			/*
2660 			 * Internal error, invalid ioctl description
2661 			 * force debug entry even if pm_debug not set
2662 			 */
2663 #ifdef	DEBUG
2664 			pm_log("invalid diptype %d for cmd %d (%s)\n",
2665 			    pcip->diptype, cmd, pcip->name);
2666 #endif
2667 			ASSERT(0);
2668 			return (EIO);
2669 		}
2670 		break;
2671 	}
2672 
2673 	case NOSTRUCT:
2674 	{
2675 		switch (cmd) {
2676 		case PM_START_PM:
2677 		case PM_START_CPUPM:
2678 		case PM_START_CPUPM_EV:
2679 		case PM_START_CPUPM_POLL:
2680 		{
2681 			pm_cpupm_t	new_mode = PM_CPUPM_NOTSET;
2682 			pm_cpupm_t	old_mode = PM_CPUPM_NOTSET;
2683 			int		r;
2684 
2685 			mutex_enter(&pm_scan_lock);
2686 			if ((cmd == PM_START_PM && autopm_enabled) ||
2687 			    (cmd == PM_START_CPUPM && PM_DEFAULT_CPUPM) ||
2688 			    (cmd == PM_START_CPUPM_EV && PM_EVENT_CPUPM) ||
2689 			    (cmd == PM_START_CPUPM_POLL && PM_POLLING_CPUPM)) {
2690 				mutex_exit(&pm_scan_lock);
2691 				PMD(PMD_ERROR, ("ioctl: %s: EBUSY\n", cmdstr))
2692 				ret = EBUSY;
2693 				break;
2694 			}
2695 
2696 			if (cmd == PM_START_PM) {
2697 				autopm_enabled = 1;
2698 			} else if (cmd == PM_START_CPUPM) {
2699 				old_mode = cpupm;
2700 				new_mode = cpupm = cpupm_default_mode;
2701 			} else if (cmd == PM_START_CPUPM_EV) {
2702 				old_mode = cpupm;
2703 				new_mode = cpupm = PM_CPUPM_EVENT;
2704 			} else if (cmd == PM_START_CPUPM_POLL) {
2705 				old_mode = cpupm;
2706 				new_mode = cpupm = PM_CPUPM_POLLING;
2707 			}
2708 
2709 			mutex_exit(&pm_scan_lock);
2710 
2711 			/*
2712 			 * If we are changing CPUPM modes, and it is active,
2713 			 * then stop it from operating in the old mode.
2714 			 */
2715 			if (old_mode == PM_CPUPM_POLLING) {
2716 				int c = PM_STOP_CPUPM;
2717 				ddi_walk_devs(ddi_root_node(), pm_stop_pm_walk,
2718 				    &c);
2719 			} else if (old_mode == PM_CPUPM_EVENT) {
2720 				r = cpupm_set_policy(CPUPM_POLICY_DISABLED);
2721 
2722 				/*
2723 				 * Disabling CPUPM policy should always
2724 				 * succeed
2725 				 */
2726 				ASSERT(r == 0);
2727 			}
2728 
2729 			/*
2730 			 * If we are changing to event based CPUPM, enable it.
2731 			 * In the event it's not supported, fall back to
2732 			 * polling based CPUPM.
2733 			 */
2734 			if (new_mode == PM_CPUPM_EVENT &&
2735 			    cpupm_set_policy(CPUPM_POLICY_ELASTIC) < 0) {
2736 				mutex_enter(&pm_scan_lock);
2737 				new_mode = cpupm = PM_CPUPM_POLLING;
2738 				cmd = PM_START_CPUPM_POLL;
2739 				mutex_exit(&pm_scan_lock);
2740 			}
2741 			if (new_mode == PM_CPUPM_POLLING ||
2742 			    cmd == PM_START_PM) {
2743 				ddi_walk_devs(ddi_root_node(), pm_start_pm_walk,
2744 				    &cmd);
2745 			}
2746 			ret = 0;
2747 			break;
2748 		}
2749 
2750 		case PM_RESET_PM:
2751 		case PM_STOP_PM:
2752 		case PM_STOP_CPUPM:
2753 		{
2754 			extern void pm_discard_thresholds(void);
2755 			pm_cpupm_t old_mode = PM_CPUPM_NOTSET;
2756 
2757 			mutex_enter(&pm_scan_lock);
2758 			if ((cmd == PM_STOP_PM && !autopm_enabled) ||
2759 			    (cmd == PM_STOP_CPUPM && PM_CPUPM_DISABLED)) {
2760 				mutex_exit(&pm_scan_lock);
2761 				PMD(PMD_ERROR, ("ioctl: %s: EINVAL\n",
2762 				    cmdstr))
2763 				ret = EINVAL;
2764 				break;
2765 			}
2766 
2767 			if (cmd == PM_STOP_PM) {
2768 				autopm_enabled = 0;
2769 				pm_S3_enabled = 0;
2770 				autoS3_enabled = 0;
2771 			} else if (cmd == PM_STOP_CPUPM) {
2772 				old_mode = cpupm;
2773 				cpupm = PM_CPUPM_DISABLE;
2774 			} else {
2775 				autopm_enabled = 0;
2776 				autoS3_enabled = 0;
2777 				old_mode = cpupm;
2778 				cpupm = PM_CPUPM_NOTSET;
2779 			}
2780 			mutex_exit(&pm_scan_lock);
2781 
2782 			/*
2783 			 * bring devices to full power level, stop scan
2784 			 * If CPUPM was operating in event driven mode, disable
2785 			 * that.
2786 			 */
2787 			if (old_mode == PM_CPUPM_EVENT) {
2788 				(void) cpupm_set_policy(CPUPM_POLICY_DISABLED);
2789 			}
2790 			ddi_walk_devs(ddi_root_node(), pm_stop_pm_walk, &cmd);
2791 			ret = 0;
2792 			if (cmd == PM_STOP_PM || cmd == PM_STOP_CPUPM)
2793 				break;
2794 			/*
2795 			 * Now do only PM_RESET_PM stuff.
2796 			 */
2797 			pm_system_idle_threshold = pm_default_idle_threshold;
2798 			pm_cpu_idle_threshold = 0;
2799 			pm_discard_thresholds();
2800 			pm_all_to_default_thresholds();
2801 			pm_dispatch_to_dep_thread(PM_DEP_WK_REMOVE_DEP,
2802 			    NULL, NULL, PM_DEP_WAIT, NULL, 0);
2803 			break;
2804 		}
2805 
2806 		case PM_GET_SYSTEM_THRESHOLD:
2807 		{
2808 			*rval_p = pm_system_idle_threshold;
2809 			ret = 0;
2810 			break;
2811 		}
2812 
2813 		case PM_GET_DEFAULT_SYSTEM_THRESHOLD:
2814 		{
2815 			*rval_p = pm_default_idle_threshold;
2816 			ret = 0;
2817 			break;
2818 		}
2819 
2820 		case PM_GET_CPU_THRESHOLD:
2821 		{
2822 			*rval_p = pm_cpu_idle_threshold;
2823 			ret = 0;
2824 			break;
2825 		}
2826 
2827 		case PM_SET_SYSTEM_THRESHOLD:
2828 		case PM_SET_CPU_THRESHOLD:
2829 		{
2830 			if ((int)arg < 0) {
2831 				PMD(PMD_ERROR, ("ioctl: %s: arg 0x%x < 0"
2832 				    "--EINVAL\n", cmdstr, (int)arg))
2833 				ret = EINVAL;
2834 				break;
2835 			}
2836 			PMD(PMD_IOCTL, ("ioctl: %s: 0x%x 0t%d\n", cmdstr,
2837 			    (int)arg, (int)arg))
2838 			if (cmd == PM_SET_SYSTEM_THRESHOLD)
2839 				pm_system_idle_threshold = (int)arg;
2840 			else {
2841 				pm_cpu_idle_threshold = (int)arg;
2842 			}
2843 			ddi_walk_devs(ddi_root_node(), pm_set_idle_thresh_walk,
2844 			    (void *) &cmd);
2845 
2846 			ret = 0;
2847 			break;
2848 		}
2849 
2850 		case PM_IDLE_DOWN:
2851 		{
2852 			if (pm_timeout_idledown() != 0) {
2853 				ddi_walk_devs(ddi_root_node(),
2854 				    pm_start_idledown, (void *)PMID_IOC);
2855 			}
2856 			ret = 0;
2857 			break;
2858 		}
2859 
2860 		case PM_GET_PM_STATE:
2861 		{
2862 			if (autopm_enabled) {
2863 				*rval_p = PM_SYSTEM_PM_ENABLED;
2864 			} else {
2865 				*rval_p = PM_SYSTEM_PM_DISABLED;
2866 			}
2867 			ret = 0;
2868 			break;
2869 		}
2870 
2871 		case PM_GET_CPUPM_STATE:
2872 		{
2873 			if (PM_POLLING_CPUPM || PM_EVENT_CPUPM)
2874 				*rval_p = PM_CPU_PM_ENABLED;
2875 			else if (PM_CPUPM_DISABLED)
2876 				*rval_p = PM_CPU_PM_DISABLED;
2877 			else
2878 				*rval_p = PM_CPU_PM_NOTSET;
2879 			ret = 0;
2880 			break;
2881 		}
2882 
2883 		case PM_GET_AUTOS3_STATE:
2884 		{
2885 			if (autoS3_enabled) {
2886 				*rval_p = PM_AUTOS3_ENABLED;
2887 			} else {
2888 				*rval_p = PM_AUTOS3_DISABLED;
2889 			}
2890 			ret = 0;
2891 			break;
2892 		}
2893 
2894 		case PM_GET_S3_SUPPORT_STATE:
2895 		{
2896 			if (pm_S3_enabled) {
2897 				*rval_p = PM_S3_SUPPORT_ENABLED;
2898 			} else {
2899 				*rval_p = PM_S3_SUPPORT_DISABLED;
2900 			}
2901 			ret = 0;
2902 			break;
2903 		}
2904 
2905 		/*
2906 		 * pmconfig tells us if the platform supports S3
2907 		 */
2908 		case PM_ENABLE_S3:
2909 		{
2910 			mutex_enter(&pm_scan_lock);
2911 			if (pm_S3_enabled) {
2912 				mutex_exit(&pm_scan_lock);
2913 				PMD(PMD_ERROR, ("ioctl: %s: EBUSY\n",
2914 				    cmdstr))
2915 				ret = EBUSY;
2916 				break;
2917 			}
2918 			pm_S3_enabled = 1;
2919 			mutex_exit(&pm_scan_lock);
2920 			ret = 0;
2921 			break;
2922 		}
2923 
2924 		case PM_DISABLE_S3:
2925 		{
2926 			mutex_enter(&pm_scan_lock);
2927 			pm_S3_enabled = 0;
2928 			mutex_exit(&pm_scan_lock);
2929 			ret = 0;
2930 			break;
2931 		}
2932 
2933 		case PM_START_AUTOS3:
2934 		{
2935 			mutex_enter(&pm_scan_lock);
2936 			if (autoS3_enabled) {
2937 				mutex_exit(&pm_scan_lock);
2938 				PMD(PMD_ERROR, ("ioctl: %s: EBUSY\n",
2939 				    cmdstr))
2940 				ret = EBUSY;
2941 				break;
2942 			}
2943 			autoS3_enabled = 1;
2944 			mutex_exit(&pm_scan_lock);
2945 			ret = 0;
2946 			break;
2947 		}
2948 
2949 		case PM_STOP_AUTOS3:
2950 		{
2951 			mutex_enter(&pm_scan_lock);
2952 			autoS3_enabled = 0;
2953 			mutex_exit(&pm_scan_lock);
2954 			ret = 0;
2955 			break;
2956 		}
2957 
2958 		case PM_ENABLE_CPU_DEEP_IDLE:
2959 		{
2960 			if (callb_execute_class(CB_CL_CPU_DEEP_IDLE,
2961 			    PM_ENABLE_CPU_DEEP_IDLE) == NULL)
2962 				ret = 0;
2963 			else
2964 				ret = EBUSY;
2965 			break;
2966 		}
2967 		case PM_DISABLE_CPU_DEEP_IDLE:
2968 		{
2969 			if (callb_execute_class(CB_CL_CPU_DEEP_IDLE,
2970 			    PM_DISABLE_CPU_DEEP_IDLE) == NULL)
2971 				ret = 0;
2972 			else
2973 				ret = EINVAL;
2974 			break;
2975 		}
2976 		case PM_DEFAULT_CPU_DEEP_IDLE:
2977 		{
2978 			if (callb_execute_class(CB_CL_CPU_DEEP_IDLE,
2979 			    PM_DEFAULT_CPU_DEEP_IDLE) == NULL)
2980 				ret = 0;
2981 			else
2982 				ret = EBUSY;
2983 			break;
2984 		}
2985 
2986 		default:
2987 			/*
2988 			 * Internal error, invalid ioctl description
2989 			 * force debug entry even if pm_debug not set
2990 			 */
2991 #ifdef	DEBUG
2992 			pm_log("invalid diptype %d for cmd %d (%s)\n",
2993 			    pcip->diptype, cmd, pcip->name);
2994 #endif
2995 			ASSERT(0);
2996 			return (EIO);
2997 		}
2998 		break;
2999 	}
3000 
3001 default:
3002 		/*
3003 		 * Internal error, invalid ioctl description
3004 		 * force debug entry even if pm_debug not set
3005 		 */
3006 #ifdef	DEBUG
3007 		pm_log("ioctl: invalid str_type %d for cmd %d (%s)\n",
3008 		    pcip->str_type, cmd, pcip->name);
3009 #endif
3010 		ASSERT(0);
3011 		return (EIO);
3012 	}
3013 	ASSERT(ret != 0x0badcafe);	/* some cmd in wrong case! */
3014 	if (dipheld) {
3015 		ASSERT(dip);
3016 		PMD(PMD_DHR, ("ioctl: %s: releasing %s@%s(%s#%d) for "
3017 		    "exiting pm_ioctl\n", cmdstr, PM_DEVICE(dip)))
3018 		PM_RELE(dip);
3019 	}
3020 	PMD(PMD_IOCTL, ("ioctl: %s: end, ret=%d\n", cmdstr, ret))
3021 	return (ret);
3022 }
3023