xref: /titanic_51/usr/src/uts/i86pc/io/pciex/npe.c (revision 89b43686db1fe9681d80a7cf5662730cb9378cae)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 
22 /*
23  * Copyright 2010 Sun Microsystems, Inc.  All rights reserved.
24  * Use is subject to license terms.
25  */
26 
27 /*
28  *	Host to PCI-Express local bus driver
29  */
30 
31 #include <sys/conf.h>
32 #include <sys/modctl.h>
33 #include <sys/file.h>
34 #include <sys/pci_impl.h>
35 #include <sys/pcie_impl.h>
36 #include <sys/sysmacros.h>
37 #include <sys/ddi_intr.h>
38 #include <sys/sunndi.h>
39 #include <sys/sunddi.h>
40 #include <sys/ddifm.h>
41 #include <sys/ndifm.h>
42 #include <sys/fm/util.h>
43 #include <sys/hotplug/pci/pcie_hp.h>
44 #include <io/pci/pci_tools_ext.h>
45 #include <io/pci/pci_common.h>
46 #include <io/pciex/pcie_nvidia.h>
47 
48 /*
49  * Helper Macros
50  */
51 #define	NPE_IS_HANDLE_FOR_STDCFG_ACC(hp) \
52 	((hp) != NULL &&						\
53 	((ddi_acc_hdl_t *)(hp))->ah_platform_private != NULL &&		\
54 	(((ddi_acc_impl_t *)((ddi_acc_hdl_t *)(hp))->			\
55 	ah_platform_private)->						\
56 	    ahi_acc_attr &(DDI_ACCATTR_CPU_VADDR|DDI_ACCATTR_CONFIG_SPACE)) \
57 		== DDI_ACCATTR_CONFIG_SPACE)
58 
59 /*
60  * Bus Operation functions
61  */
62 static int	npe_bus_map(dev_info_t *, dev_info_t *, ddi_map_req_t *,
63 		    off_t, off_t, caddr_t *);
64 static int	npe_ctlops(dev_info_t *, dev_info_t *, ddi_ctl_enum_t,
65 		    void *, void *);
66 static int	npe_intr_ops(dev_info_t *, dev_info_t *, ddi_intr_op_t,
67 		    ddi_intr_handle_impl_t *, void *);
68 static int	npe_fm_init(dev_info_t *, dev_info_t *, int,
69 		    ddi_iblock_cookie_t *);
70 
71 static int	npe_fm_callback(dev_info_t *, ddi_fm_error_t *, const void *);
72 
73 /*
74  * Disable URs and Received MA for all PCIe devices.  Until x86 SW is changed so
75  * that random drivers do not do PIO accesses on devices that it does not own,
76  * these error bits must be disabled.  SERR must also be disabled if URs have
77  * been masked.
78  */
79 uint32_t	npe_aer_uce_mask = PCIE_AER_UCE_UR;
80 uint32_t	npe_aer_ce_mask = 0;
81 uint32_t	npe_aer_suce_mask = PCIE_AER_SUCE_RCVD_MA;
82 
83 struct bus_ops npe_bus_ops = {
84 	BUSO_REV,
85 	npe_bus_map,
86 	NULL,
87 	NULL,
88 	NULL,
89 	i_ddi_map_fault,
90 	ddi_dma_map,
91 	ddi_dma_allochdl,
92 	ddi_dma_freehdl,
93 	ddi_dma_bindhdl,
94 	ddi_dma_unbindhdl,
95 	ddi_dma_flush,
96 	ddi_dma_win,
97 	ddi_dma_mctl,
98 	npe_ctlops,
99 	ddi_bus_prop_op,
100 	0,			/* (*bus_get_eventcookie)();	*/
101 	0,			/* (*bus_add_eventcall)();	*/
102 	0,			/* (*bus_remove_eventcall)();	*/
103 	0,			/* (*bus_post_event)();		*/
104 	0,			/* (*bus_intr_ctl)(); */
105 	0,			/* (*bus_config)(); */
106 	0,			/* (*bus_unconfig)(); */
107 	npe_fm_init,		/* (*bus_fm_init)(); */
108 	NULL,			/* (*bus_fm_fini)(); */
109 	NULL,			/* (*bus_fm_access_enter)(); */
110 	NULL,			/* (*bus_fm_access_exit)(); */
111 	NULL,			/* (*bus_power)(); */
112 	npe_intr_ops,		/* (*bus_intr_op)(); */
113 	pcie_hp_common_ops	/* (*bus_hp_op)(); */
114 };
115 
116 static int	npe_open(dev_t *, int, int, cred_t *);
117 static int	npe_close(dev_t, int, int, cred_t *);
118 static int	npe_ioctl(dev_t, int, intptr_t, int, cred_t *, int *);
119 
120 struct cb_ops npe_cb_ops = {
121 	npe_open,			/* open */
122 	npe_close,			/* close */
123 	nodev,				/* strategy */
124 	nodev,				/* print */
125 	nodev,				/* dump */
126 	nodev,				/* read */
127 	nodev,				/* write */
128 	npe_ioctl,			/* ioctl */
129 	nodev,				/* devmap */
130 	nodev,				/* mmap */
131 	nodev,				/* segmap */
132 	nochpoll,			/* poll */
133 	pcie_prop_op,			/* cb_prop_op */
134 	NULL,				/* streamtab */
135 	D_NEW | D_MP | D_HOTPLUG,	/* Driver compatibility flag */
136 	CB_REV,				/* rev */
137 	nodev,				/* int (*cb_aread)() */
138 	nodev				/* int (*cb_awrite)() */
139 };
140 
141 
142 /*
143  * Device Node Operation functions
144  */
145 static int	npe_attach(dev_info_t *devi, ddi_attach_cmd_t cmd);
146 static int	npe_detach(dev_info_t *devi, ddi_detach_cmd_t cmd);
147 static int	npe_info(dev_info_t *, ddi_info_cmd_t, void *, void **);
148 
149 struct dev_ops npe_ops = {
150 	DEVO_REV,		/* devo_rev */
151 	0,			/* refcnt  */
152 	npe_info,		/* info */
153 	nulldev,		/* identify */
154 	nulldev,		/* probe */
155 	npe_attach,		/* attach */
156 	npe_detach,		/* detach */
157 	nulldev,		/* reset */
158 	&npe_cb_ops,		/* driver operations */
159 	&npe_bus_ops,		/* bus operations */
160 	NULL,			/* power */
161 	ddi_quiesce_not_needed,		/* quiesce */
162 };
163 
164 /*
165  * Internal routines in support of particular npe_ctlops.
166  */
167 static int npe_removechild(dev_info_t *child);
168 static int npe_initchild(dev_info_t *child);
169 
170 /*
171  * External support routine
172  */
173 extern void	npe_query_acpi_mcfg(dev_info_t *dip);
174 extern void	npe_ck804_fix_aer_ptr(ddi_acc_handle_t cfg_hdl);
175 extern int	npe_disable_empty_bridges_workaround(dev_info_t *child);
176 extern void	npe_nvidia_error_workaround(ddi_acc_handle_t cfg_hdl);
177 extern void	npe_intel_error_workaround(ddi_acc_handle_t cfg_hdl);
178 extern boolean_t npe_is_mmcfg_supported(dev_info_t *dip);
179 extern void	npe_enable_htmsi_children(dev_info_t *dip);
180 extern int	npe_save_htconfig_children(dev_info_t *dip);
181 extern int	npe_restore_htconfig_children(dev_info_t *dip);
182 
183 /*
184  * Module linkage information for the kernel.
185  */
186 static struct modldrv modldrv = {
187 	&mod_driverops,				/* Type of module */
188 	"Host to PCIe nexus driver",		/* Name of module */
189 	&npe_ops,				/* driver ops */
190 };
191 
192 static struct modlinkage modlinkage = {
193 	MODREV_1,
194 	(void *)&modldrv,
195 	NULL
196 };
197 
198 /* Save minimal state. */
199 void *npe_statep;
200 
201 int
202 _init(void)
203 {
204 	int e;
205 
206 	/*
207 	 * Initialize per-pci bus soft state pointer.
208 	 */
209 	e = ddi_soft_state_init(&npe_statep, sizeof (pci_state_t), 1);
210 	if (e != 0)
211 		return (e);
212 
213 	if ((e = mod_install(&modlinkage)) != 0)
214 		ddi_soft_state_fini(&npe_statep);
215 
216 	return (e);
217 }
218 
219 
220 int
221 _fini(void)
222 {
223 	int rc;
224 
225 	rc = mod_remove(&modlinkage);
226 	if (rc != 0)
227 		return (rc);
228 
229 	ddi_soft_state_fini(&npe_statep);
230 	return (rc);
231 }
232 
233 
234 int
235 _info(struct modinfo *modinfop)
236 {
237 	return (mod_info(&modlinkage, modinfop));
238 }
239 
240 /*ARGSUSED*/
241 static int
242 npe_info(dev_info_t *dip, ddi_info_cmd_t cmd, void *arg, void **result)
243 {
244 	minor_t		minor = getminor((dev_t)arg);
245 	int		instance = PCI_MINOR_NUM_TO_INSTANCE(minor);
246 	pci_state_t	*pcip = ddi_get_soft_state(npe_statep, instance);
247 	int		ret = DDI_SUCCESS;
248 
249 	switch (cmd) {
250 	case DDI_INFO_DEVT2INSTANCE:
251 		*result = (void *)(intptr_t)instance;
252 		break;
253 	case DDI_INFO_DEVT2DEVINFO:
254 		if (pcip == NULL) {
255 			ret = DDI_FAILURE;
256 			break;
257 		}
258 
259 		*result = (void *)pcip->pci_dip;
260 		break;
261 	default:
262 		ret = DDI_FAILURE;
263 		break;
264 	}
265 
266 	return (ret);
267 }
268 
269 /*ARGSUSED*/
270 static int
271 npe_attach(dev_info_t *devi, ddi_attach_cmd_t cmd)
272 {
273 	int		instance = ddi_get_instance(devi);
274 	pci_state_t	*pcip = NULL;
275 
276 	if (cmd == DDI_RESUME) {
277 		/*
278 		 * the system might still be able to resume even if this fails
279 		 */
280 		(void) npe_restore_htconfig_children(devi);
281 		return (DDI_SUCCESS);
282 	}
283 
284 	/*
285 	 * We must do this here in order to ensure that all top level devices
286 	 * get their HyperTransport MSI mapping regs programmed first.
287 	 * "Memory controller" and "hostbridge" class devices are leaf devices
288 	 * that may affect MSI translation functionality for devices
289 	 * connected to the same link/bus.
290 	 *
291 	 * This will also program HT MSI mapping registers on root buses
292 	 * devices (basically sitting on an HT bus) that are not dependent
293 	 * on the aforementioned HT devices for MSI translation.
294 	 */
295 	npe_enable_htmsi_children(devi);
296 
297 	if (ddi_prop_update_string(DDI_DEV_T_NONE, devi, "device_type",
298 	    "pciex") != DDI_PROP_SUCCESS) {
299 		cmn_err(CE_WARN, "npe:  'device_type' prop create failed");
300 	}
301 
302 	if (ddi_soft_state_zalloc(npe_statep, instance) == DDI_SUCCESS)
303 		pcip = ddi_get_soft_state(npe_statep, instance);
304 
305 	if (pcip == NULL)
306 		return (DDI_FAILURE);
307 
308 	pcip->pci_dip = devi;
309 	pcip->pci_soft_state = PCI_SOFT_STATE_CLOSED;
310 
311 	if (pcie_init(devi, NULL) != DDI_SUCCESS)
312 		goto fail1;
313 
314 	/* Second arg: initialize for pci_express root nexus */
315 	if (pcitool_init(devi, B_TRUE) != DDI_SUCCESS)
316 		goto fail2;
317 
318 	pcip->pci_fmcap = DDI_FM_EREPORT_CAPABLE | DDI_FM_ERRCB_CAPABLE |
319 	    DDI_FM_ACCCHK_CAPABLE | DDI_FM_DMACHK_CAPABLE;
320 	ddi_fm_init(devi, &pcip->pci_fmcap, &pcip->pci_fm_ibc);
321 
322 	if (pcip->pci_fmcap & DDI_FM_ERRCB_CAPABLE) {
323 		ddi_fm_handler_register(devi, npe_fm_callback, NULL);
324 	}
325 
326 	PCIE_DIP2PFD(devi) = kmem_zalloc(sizeof (pf_data_t), KM_SLEEP);
327 	pcie_rc_init_pfd(devi, PCIE_DIP2PFD(devi));
328 
329 	npe_query_acpi_mcfg(devi);
330 	ddi_report_dev(devi);
331 	pcie_fab_init_bus(devi, PCIE_BUS_FINAL);
332 
333 	return (DDI_SUCCESS);
334 
335 fail2:
336 	(void) pcie_uninit(devi);
337 fail1:
338 	pcie_rc_fini_bus(devi);
339 	ddi_soft_state_free(npe_statep, instance);
340 
341 	return (DDI_FAILURE);
342 }
343 
344 /*ARGSUSED*/
345 static int
346 npe_detach(dev_info_t *devi, ddi_detach_cmd_t cmd)
347 {
348 	int instance = ddi_get_instance(devi);
349 	pci_state_t *pcip;
350 
351 	pcip = ddi_get_soft_state(npe_statep, ddi_get_instance(devi));
352 
353 	switch (cmd) {
354 	case DDI_DETACH:
355 		pcie_fab_fini_bus(devi, PCIE_BUS_INITIAL);
356 
357 		/* Uninitialize pcitool support. */
358 		pcitool_uninit(devi);
359 
360 		if (pcie_uninit(devi) != DDI_SUCCESS)
361 			return (DDI_FAILURE);
362 
363 		if (pcip->pci_fmcap & DDI_FM_ERRCB_CAPABLE)
364 			ddi_fm_handler_unregister(devi);
365 
366 		pcie_rc_fini_pfd(PCIE_DIP2PFD(devi));
367 		kmem_free(PCIE_DIP2PFD(devi), sizeof (pf_data_t));
368 
369 		ddi_fm_fini(devi);
370 		ddi_soft_state_free(npe_statep, instance);
371 		return (DDI_SUCCESS);
372 
373 	case DDI_SUSPEND:
374 		/*
375 		 * the system might still be able to suspend/resume even if
376 		 * this fails
377 		 */
378 		(void) npe_save_htconfig_children(devi);
379 		return (DDI_SUCCESS);
380 	default:
381 		return (DDI_FAILURE);
382 	}
383 }
384 
385 /*
386  * Configure the access handle for standard configuration space
387  * access (see pci_fm_acc_setup for code that initializes the
388  * access-function pointers).
389  */
390 static int
391 npe_setup_std_pcicfg_acc(dev_info_t *rdip, ddi_map_req_t *mp,
392     ddi_acc_hdl_t *hp, off_t offset, off_t len)
393 {
394 	int ret;
395 
396 	if ((ret = pci_fm_acc_setup(hp, offset, len)) ==
397 	    DDI_SUCCESS) {
398 		if (DDI_FM_ACC_ERR_CAP(ddi_fm_capable(rdip)) &&
399 		    mp->map_handlep->ah_acc.devacc_attr_access
400 		    != DDI_DEFAULT_ACC) {
401 			ndi_fmc_insert(rdip, ACC_HANDLE,
402 			    (void *)mp->map_handlep, NULL);
403 		}
404 	}
405 	return (ret);
406 }
407 
408 static int
409 npe_bus_map(dev_info_t *dip, dev_info_t *rdip, ddi_map_req_t *mp,
410     off_t offset, off_t len, caddr_t *vaddrp)
411 {
412 	int 		rnumber;
413 	int		length;
414 	int		space;
415 	ddi_acc_impl_t	*ap;
416 	ddi_acc_hdl_t	*hp;
417 	ddi_map_req_t	mr;
418 	pci_regspec_t	pci_reg;
419 	pci_regspec_t	*pci_rp;
420 	struct regspec	reg;
421 	pci_acc_cfblk_t	*cfp;
422 	int		retval;
423 	int64_t		*ecfginfo;
424 	uint_t		nelem;
425 
426 	mr = *mp; /* Get private copy of request */
427 	mp = &mr;
428 
429 	/*
430 	 * check for register number
431 	 */
432 	switch (mp->map_type) {
433 	case DDI_MT_REGSPEC:
434 		pci_reg = *(pci_regspec_t *)(mp->map_obj.rp);
435 		pci_rp = &pci_reg;
436 		if (pci_common_get_reg_prop(rdip, pci_rp) != DDI_SUCCESS)
437 			return (DDI_FAILURE);
438 		break;
439 	case DDI_MT_RNUMBER:
440 		rnumber = mp->map_obj.rnumber;
441 		/*
442 		 * get ALL "reg" properties for dip, select the one of
443 		 * of interest. In x86, "assigned-addresses" property
444 		 * is identical to the "reg" property, so there is no
445 		 * need to cross check the two to determine the physical
446 		 * address of the registers.
447 		 * This routine still performs some validity checks to
448 		 * make sure that everything is okay.
449 		 */
450 		if (ddi_prop_lookup_int_array(DDI_DEV_T_ANY, rdip,
451 		    DDI_PROP_DONTPASS, "reg", (int **)&pci_rp,
452 		    (uint_t *)&length) != DDI_PROP_SUCCESS)
453 			return (DDI_FAILURE);
454 
455 		/*
456 		 * validate the register number.
457 		 */
458 		length /= (sizeof (pci_regspec_t) / sizeof (int));
459 		if (rnumber >= length) {
460 			ddi_prop_free(pci_rp);
461 			return (DDI_FAILURE);
462 		}
463 
464 		/*
465 		 * copy the required entry.
466 		 */
467 		pci_reg = pci_rp[rnumber];
468 
469 		/*
470 		 * free the memory allocated by ddi_prop_lookup_int_array
471 		 */
472 		ddi_prop_free(pci_rp);
473 
474 		pci_rp = &pci_reg;
475 		if (pci_common_get_reg_prop(rdip, pci_rp) != DDI_SUCCESS)
476 			return (DDI_FAILURE);
477 		mp->map_type = DDI_MT_REGSPEC;
478 		break;
479 	default:
480 		return (DDI_ME_INVAL);
481 	}
482 
483 	space = pci_rp->pci_phys_hi & PCI_REG_ADDR_M;
484 
485 	/*
486 	 * check for unmap and unlock of address space
487 	 */
488 	if ((mp->map_op == DDI_MO_UNMAP) || (mp->map_op == DDI_MO_UNLOCK)) {
489 		switch (space) {
490 		case PCI_ADDR_IO:
491 			reg.regspec_bustype = 1;
492 			break;
493 
494 		case PCI_ADDR_CONFIG:
495 			/*
496 			 * If this is an unmap/unlock of a standard config
497 			 * space mapping (memory-mapped config space mappings
498 			 * would have the DDI_ACCATTR_CPU_VADDR bit set in the
499 			 * acc_attr), undo that setup here.
500 			 */
501 			if (NPE_IS_HANDLE_FOR_STDCFG_ACC(mp->map_handlep)) {
502 
503 				if (DDI_FM_ACC_ERR_CAP(ddi_fm_capable(rdip)) &&
504 				    mp->map_handlep->ah_acc.devacc_attr_access
505 				    != DDI_DEFAULT_ACC) {
506 					ndi_fmc_remove(rdip, ACC_HANDLE,
507 					    (void *)mp->map_handlep);
508 				}
509 				return (DDI_SUCCESS);
510 			}
511 
512 			pci_rp->pci_size_low = PCIE_CONF_HDR_SIZE;
513 
514 			/* FALLTHROUGH */
515 		case PCI_ADDR_MEM64:
516 			/*
517 			 * MEM64 requires special treatment on map, to check
518 			 * that the device is below 4G.  On unmap, however,
519 			 * we can assume that everything is OK... the map
520 			 * must have succeeded.
521 			 */
522 			/* FALLTHROUGH */
523 		case PCI_ADDR_MEM32:
524 			reg.regspec_bustype = 0;
525 			break;
526 
527 		default:
528 			return (DDI_FAILURE);
529 		}
530 
531 		/*
532 		 * Adjust offset and length
533 		 * A non-zero length means override the one in the regspec.
534 		 */
535 		pci_rp->pci_phys_low += (uint_t)offset;
536 		if (len != 0)
537 			pci_rp->pci_size_low = len;
538 
539 		reg.regspec_addr = pci_rp->pci_phys_low;
540 		reg.regspec_size = pci_rp->pci_size_low;
541 
542 		mp->map_obj.rp = &reg;
543 		retval = ddi_map(dip, mp, (off_t)0, (off_t)0, vaddrp);
544 		if (DDI_FM_ACC_ERR_CAP(ddi_fm_capable(rdip)) &&
545 		    mp->map_handlep->ah_acc.devacc_attr_access !=
546 		    DDI_DEFAULT_ACC) {
547 			ndi_fmc_remove(rdip, ACC_HANDLE,
548 			    (void *)mp->map_handlep);
549 		}
550 		return (retval);
551 
552 	}
553 
554 	/* check for user mapping request - not legal for Config */
555 	if (mp->map_op == DDI_MO_MAP_HANDLE && space == PCI_ADDR_CONFIG) {
556 		cmn_err(CE_NOTE, "npe: Config mapping request from user\n");
557 		return (DDI_FAILURE);
558 	}
559 
560 
561 	/*
562 	 * Note that pci_fm_acc_setup() is called to serve two purposes
563 	 * i) enable legacy PCI I/O style config space access
564 	 * ii) register with FMA
565 	 */
566 	if (space == PCI_ADDR_CONFIG) {
567 
568 		/* Can't map config space without a handle */
569 		hp = (ddi_acc_hdl_t *)mp->map_handlep;
570 		if (hp == NULL)
571 			return (DDI_FAILURE);
572 
573 		/* record the device address for future reference */
574 		cfp = (pci_acc_cfblk_t *)&hp->ah_bus_private;
575 		cfp->c_busnum = PCI_REG_BUS_G(pci_rp->pci_phys_hi);
576 		cfp->c_devnum = PCI_REG_DEV_G(pci_rp->pci_phys_hi);
577 		cfp->c_funcnum = PCI_REG_FUNC_G(pci_rp->pci_phys_hi);
578 
579 		*vaddrp = (caddr_t)offset;
580 
581 		/* Check if MMCFG is supported */
582 		if (!npe_is_mmcfg_supported(rdip)) {
583 			return (npe_setup_std_pcicfg_acc(rdip, mp, hp,
584 			    offset, len));
585 		}
586 
587 
588 		if (ddi_prop_lookup_int64_array(DDI_DEV_T_ANY, rdip, 0,
589 		    "ecfg", &ecfginfo, &nelem) == DDI_PROP_SUCCESS) {
590 
591 			if (nelem != 4 ||
592 			    cfp->c_busnum < ecfginfo[2] ||
593 			    cfp->c_busnum > ecfginfo[3]) {
594 				/*
595 				 * Invalid property or Doesn't contain the
596 				 * requested bus; fall back to standard
597 				 * (I/O-based) config access.
598 				 */
599 				ddi_prop_free(ecfginfo);
600 				return (npe_setup_std_pcicfg_acc(rdip, mp, hp,
601 				    offset, len));
602 			} else {
603 				pci_rp->pci_phys_low = ecfginfo[0];
604 
605 				ddi_prop_free(ecfginfo);
606 
607 				pci_rp->pci_phys_low += ((cfp->c_busnum << 20) |
608 				    (cfp->c_devnum) << 15 |
609 				    (cfp->c_funcnum << 12));
610 
611 				pci_rp->pci_size_low = PCIE_CONF_HDR_SIZE;
612 			}
613 		} else {
614 			/*
615 			 * Couldn't find the MMCFG property -- fall back to
616 			 * standard config access
617 			 */
618 			return (npe_setup_std_pcicfg_acc(rdip, mp, hp,
619 			    offset, len));
620 		}
621 	}
622 
623 	length = pci_rp->pci_size_low;
624 
625 	/*
626 	 * range check
627 	 */
628 	if ((offset >= length) || (len > length) || (offset + len > length))
629 		return (DDI_FAILURE);
630 
631 	/*
632 	 * Adjust offset and length
633 	 * A non-zero length means override the one in the regspec.
634 	 */
635 	pci_rp->pci_phys_low += (uint_t)offset;
636 	if (len != 0)
637 		pci_rp->pci_size_low = len;
638 
639 	/*
640 	 * convert the pci regsec into the generic regspec used by the
641 	 * parent root nexus driver.
642 	 */
643 	switch (space) {
644 	case PCI_ADDR_IO:
645 		reg.regspec_bustype = 1;
646 		break;
647 	case PCI_ADDR_CONFIG:
648 	case PCI_ADDR_MEM64:
649 		/*
650 		 * We can't handle 64-bit devices that are mapped above
651 		 * 4G or that are larger than 4G.
652 		 */
653 		if (pci_rp->pci_phys_mid != 0 || pci_rp->pci_size_hi != 0)
654 			return (DDI_FAILURE);
655 		/*
656 		 * Other than that, we can treat them as 32-bit mappings
657 		 */
658 		/* FALLTHROUGH */
659 	case PCI_ADDR_MEM32:
660 		reg.regspec_bustype = 0;
661 		break;
662 	default:
663 		return (DDI_FAILURE);
664 	}
665 
666 	reg.regspec_addr = pci_rp->pci_phys_low;
667 	reg.regspec_size = pci_rp->pci_size_low;
668 
669 	mp->map_obj.rp = &reg;
670 	retval = ddi_map(dip, mp, (off_t)0, (off_t)0, vaddrp);
671 	if (retval == DDI_SUCCESS) {
672 		/*
673 		 * For config space gets force use of cautious access routines.
674 		 * These will handle default and protected mode accesses too.
675 		 */
676 		if (space == PCI_ADDR_CONFIG) {
677 			ap = (ddi_acc_impl_t *)mp->map_handlep;
678 			ap->ahi_acc_attr &= ~DDI_ACCATTR_DIRECT;
679 			ap->ahi_acc_attr |= DDI_ACCATTR_CONFIG_SPACE;
680 			ap->ahi_get8 = i_ddi_caut_get8;
681 			ap->ahi_get16 = i_ddi_caut_get16;
682 			ap->ahi_get32 = i_ddi_caut_get32;
683 			ap->ahi_get64 = i_ddi_caut_get64;
684 			ap->ahi_rep_get8 = i_ddi_caut_rep_get8;
685 			ap->ahi_rep_get16 = i_ddi_caut_rep_get16;
686 			ap->ahi_rep_get32 = i_ddi_caut_rep_get32;
687 			ap->ahi_rep_get64 = i_ddi_caut_rep_get64;
688 		}
689 		if (DDI_FM_ACC_ERR_CAP(ddi_fm_capable(rdip)) &&
690 		    mp->map_handlep->ah_acc.devacc_attr_access !=
691 		    DDI_DEFAULT_ACC) {
692 			ndi_fmc_insert(rdip, ACC_HANDLE,
693 			    (void *)mp->map_handlep, NULL);
694 		}
695 	}
696 	return (retval);
697 }
698 
699 
700 
701 /*ARGSUSED*/
702 static int
703 npe_ctlops(dev_info_t *dip, dev_info_t *rdip,
704 	ddi_ctl_enum_t ctlop, void *arg, void *result)
705 {
706 	int		rn;
707 	int		totreg;
708 	uint_t		reglen;
709 	pci_regspec_t	*drv_regp;
710 	struct attachspec *asp;
711 	struct detachspec *dsp;
712 	pci_state_t	*pci_p = ddi_get_soft_state(npe_statep,
713 	    ddi_get_instance(dip));
714 
715 	switch (ctlop) {
716 	case DDI_CTLOPS_REPORTDEV:
717 		if (rdip == (dev_info_t *)0)
718 			return (DDI_FAILURE);
719 		cmn_err(CE_CONT, "?PCI Express-device: %s@%s, %s%d\n",
720 		    ddi_node_name(rdip), ddi_get_name_addr(rdip),
721 		    ddi_driver_name(rdip), ddi_get_instance(rdip));
722 		return (DDI_SUCCESS);
723 
724 	case DDI_CTLOPS_INITCHILD:
725 		return (npe_initchild((dev_info_t *)arg));
726 
727 	case DDI_CTLOPS_UNINITCHILD:
728 		return (npe_removechild((dev_info_t *)arg));
729 
730 	case DDI_CTLOPS_SIDDEV:
731 		return (DDI_SUCCESS);
732 
733 	case DDI_CTLOPS_REGSIZE:
734 	case DDI_CTLOPS_NREGS:
735 		if (rdip == (dev_info_t *)0)
736 			return (DDI_FAILURE);
737 
738 		*(int *)result = 0;
739 		if (ddi_prop_lookup_int_array(DDI_DEV_T_ANY, rdip,
740 		    DDI_PROP_DONTPASS, "reg", (int **)&drv_regp,
741 		    &reglen) != DDI_PROP_SUCCESS) {
742 			return (DDI_FAILURE);
743 		}
744 
745 		totreg = (reglen * sizeof (int)) / sizeof (pci_regspec_t);
746 		if (ctlop == DDI_CTLOPS_NREGS)
747 			*(int *)result = totreg;
748 		else if (ctlop == DDI_CTLOPS_REGSIZE) {
749 			rn = *(int *)arg;
750 			if (rn >= totreg) {
751 				ddi_prop_free(drv_regp);
752 				return (DDI_FAILURE);
753 			}
754 			*(off_t *)result = drv_regp[rn].pci_size_low;
755 		}
756 		ddi_prop_free(drv_regp);
757 
758 		return (DDI_SUCCESS);
759 
760 	case DDI_CTLOPS_POWER:
761 	{
762 		power_req_t	*reqp = (power_req_t *)arg;
763 		/*
764 		 * We currently understand reporting of PCI_PM_IDLESPEED
765 		 * capability. Everything else is passed up.
766 		 */
767 		if ((reqp->request_type == PMR_REPORT_PMCAP) &&
768 		    (reqp->req.report_pmcap_req.cap ==  PCI_PM_IDLESPEED))
769 			return (DDI_SUCCESS);
770 
771 		break;
772 	}
773 
774 	case DDI_CTLOPS_PEEK:
775 	case DDI_CTLOPS_POKE:
776 		return (pci_common_peekpoke(dip, rdip, ctlop, arg, result));
777 
778 	/* X86 systems support PME wakeup from suspended state */
779 	case DDI_CTLOPS_ATTACH:
780 		if (!pcie_is_child(dip, rdip))
781 			return (DDI_SUCCESS);
782 
783 		asp = (struct attachspec *)arg;
784 		if ((asp->when == DDI_POST) && (asp->result == DDI_SUCCESS)) {
785 			pf_init(rdip, (void *)pci_p->pci_fm_ibc, asp->cmd);
786 			(void) pcie_postattach_child(rdip);
787 		}
788 
789 		/* only do this for immediate children */
790 		if (asp->cmd == DDI_RESUME && asp->when == DDI_PRE &&
791 		    ddi_get_parent(rdip) == dip)
792 			if (pci_pre_resume(rdip) != DDI_SUCCESS) {
793 				/* Not good, better stop now. */
794 				cmn_err(CE_PANIC,
795 				    "Couldn't pre-resume device %p",
796 				    (void *) dip);
797 				/* NOTREACHED */
798 			}
799 
800 		return (DDI_SUCCESS);
801 
802 	case DDI_CTLOPS_DETACH:
803 		if (!pcie_is_child(dip, rdip))
804 			return (DDI_SUCCESS);
805 
806 		dsp = (struct detachspec *)arg;
807 
808 		if (dsp->when == DDI_PRE)
809 			pf_fini(rdip, dsp->cmd);
810 
811 		/* only do this for immediate children */
812 		if (dsp->cmd == DDI_SUSPEND && dsp->when == DDI_POST &&
813 		    ddi_get_parent(rdip) == dip)
814 			if (pci_post_suspend(rdip) != DDI_SUCCESS)
815 				return (DDI_FAILURE);
816 
817 		return (DDI_SUCCESS);
818 
819 	default:
820 		break;
821 	}
822 
823 	return (ddi_ctlops(dip, rdip, ctlop, arg, result));
824 
825 }
826 
827 
828 /*
829  * npe_intr_ops
830  */
831 static int
832 npe_intr_ops(dev_info_t *pdip, dev_info_t *rdip, ddi_intr_op_t intr_op,
833     ddi_intr_handle_impl_t *hdlp, void *result)
834 {
835 	return (pci_common_intr_ops(pdip, rdip, intr_op, hdlp, result));
836 }
837 
838 
839 static int
840 npe_initchild(dev_info_t *child)
841 {
842 	char		name[80];
843 	pcie_bus_t	*bus_p;
844 	uint32_t	regs;
845 	ddi_acc_handle_t	cfg_hdl;
846 
847 	/*
848 	 * Do not bind drivers to empty bridges.
849 	 * Fail above, if the bridge is found to be hotplug capable
850 	 */
851 	if (npe_disable_empty_bridges_workaround(child) == 1)
852 		return (DDI_FAILURE);
853 
854 	if (pci_common_name_child(child, name, 80) != DDI_SUCCESS)
855 		return (DDI_FAILURE);
856 
857 	ddi_set_name_addr(child, name);
858 
859 	/*
860 	 * Pseudo nodes indicate a prototype node with per-instance
861 	 * properties to be merged into the real h/w device node.
862 	 * The interpretation of the unit-address is DD[,F]
863 	 * where DD is the device id and F is the function.
864 	 */
865 	if (ndi_dev_is_persistent_node(child) == 0) {
866 		extern int pci_allow_pseudo_children;
867 
868 		ddi_set_parent_data(child, NULL);
869 
870 		/*
871 		 * Try to merge the properties from this prototype
872 		 * node into real h/w nodes.
873 		 */
874 		if (ndi_merge_node(child, pci_common_name_child) ==
875 		    DDI_SUCCESS) {
876 			/*
877 			 * Merged ok - return failure to remove the node.
878 			 */
879 			ddi_set_name_addr(child, NULL);
880 			return (DDI_FAILURE);
881 		}
882 
883 		/* workaround for DDIVS to run under PCI Express */
884 		if (pci_allow_pseudo_children) {
885 			/*
886 			 * If the "interrupts" property doesn't exist,
887 			 * this must be the ddivs no-intr case, and it returns
888 			 * DDI_SUCCESS instead of DDI_FAILURE.
889 			 */
890 			if (ddi_prop_get_int(DDI_DEV_T_ANY, child,
891 			    DDI_PROP_DONTPASS, "interrupts", -1) == -1)
892 				return (DDI_SUCCESS);
893 			/*
894 			 * Create the ddi_parent_private_data for a pseudo
895 			 * child.
896 			 */
897 			pci_common_set_parent_private_data(child);
898 			return (DDI_SUCCESS);
899 		}
900 
901 		/*
902 		 * The child was not merged into a h/w node,
903 		 * but there's not much we can do with it other
904 		 * than return failure to cause the node to be removed.
905 		 */
906 		cmn_err(CE_WARN, "!%s@%s: %s.conf properties not merged",
907 		    ddi_get_name(child), ddi_get_name_addr(child),
908 		    ddi_get_name(child));
909 		ddi_set_name_addr(child, NULL);
910 		return (DDI_NOT_WELL_FORMED);
911 	}
912 
913 	if (ddi_prop_get_int(DDI_DEV_T_ANY, child, DDI_PROP_DONTPASS,
914 	    "interrupts", -1) != -1)
915 		pci_common_set_parent_private_data(child);
916 	else
917 		ddi_set_parent_data(child, NULL);
918 
919 	/* Disable certain errors on PCIe drivers for x86 platforms */
920 	regs = pcie_get_aer_uce_mask() | npe_aer_uce_mask;
921 	pcie_set_aer_uce_mask(regs);
922 	regs = pcie_get_aer_ce_mask() | npe_aer_ce_mask;
923 	pcie_set_aer_ce_mask(regs);
924 	regs = pcie_get_aer_suce_mask() | npe_aer_suce_mask;
925 	pcie_set_aer_suce_mask(regs);
926 
927 	/*
928 	 * If URs are disabled, mask SERRs as well, otherwise the system will
929 	 * still be notified of URs
930 	 */
931 	if (npe_aer_uce_mask & PCIE_AER_UCE_UR)
932 		pcie_set_serr_mask(1);
933 
934 	if (pci_config_setup(child, &cfg_hdl) == DDI_SUCCESS) {
935 		npe_ck804_fix_aer_ptr(cfg_hdl);
936 		npe_nvidia_error_workaround(cfg_hdl);
937 		npe_intel_error_workaround(cfg_hdl);
938 		pci_config_teardown(&cfg_hdl);
939 	}
940 
941 	bus_p = PCIE_DIP2BUS(child);
942 	if (bus_p) {
943 		uint16_t device_id = (uint16_t)(bus_p->bus_dev_ven_id >> 16);
944 		uint16_t vendor_id = (uint16_t)(bus_p->bus_dev_ven_id & 0xFFFF);
945 		uint16_t rev_id = bus_p->bus_rev_id;
946 
947 		/* Disable AER for certain NVIDIA Chipsets */
948 		if ((vendor_id == NVIDIA_VENDOR_ID) &&
949 		    (device_id == NVIDIA_CK804_DEVICE_ID) &&
950 		    (rev_id < NVIDIA_CK804_AER_VALID_REVID))
951 			bus_p->bus_aer_off = 0;
952 
953 		pcie_init_dom(child);
954 		(void) pcie_initchild(child);
955 	}
956 
957 	return (DDI_SUCCESS);
958 }
959 
960 
961 static int
962 npe_removechild(dev_info_t *dip)
963 {
964 	pcie_uninitchild(dip);
965 
966 	ddi_set_name_addr(dip, NULL);
967 
968 	/*
969 	 * Strip the node to properly convert it back to prototype form
970 	 */
971 	ddi_remove_minor_node(dip, NULL);
972 
973 	ddi_prop_remove_all(dip);
974 
975 	return (DDI_SUCCESS);
976 }
977 
978 static int
979 npe_open(dev_t *devp, int flags, int otyp, cred_t *credp)
980 {
981 	minor_t		minor = getminor(*devp);
982 	int		instance = PCI_MINOR_NUM_TO_INSTANCE(minor);
983 	pci_state_t	*pci_p = ddi_get_soft_state(npe_statep, instance);
984 	int	rv;
985 
986 	/*
987 	 * Make sure the open is for the right file type.
988 	 */
989 	if (otyp != OTYP_CHR)
990 		return (EINVAL);
991 
992 	if (pci_p == NULL)
993 		return (ENXIO);
994 
995 	mutex_enter(&pci_p->pci_mutex);
996 	switch (PCI_MINOR_NUM_TO_PCI_DEVNUM(minor)) {
997 	case PCI_TOOL_REG_MINOR_NUM:
998 	case PCI_TOOL_INTR_MINOR_NUM:
999 		break;
1000 	default:
1001 		/* Handle devctl ioctls */
1002 		rv = pcie_open(pci_p->pci_dip, devp, flags, otyp, credp);
1003 		mutex_exit(&pci_p->pci_mutex);
1004 		return (rv);
1005 	}
1006 
1007 	/* Handle pcitool ioctls */
1008 	if (flags & FEXCL) {
1009 		if (pci_p->pci_soft_state != PCI_SOFT_STATE_CLOSED) {
1010 			mutex_exit(&pci_p->pci_mutex);
1011 			cmn_err(CE_NOTE, "npe_open: busy");
1012 			return (EBUSY);
1013 		}
1014 		pci_p->pci_soft_state = PCI_SOFT_STATE_OPEN_EXCL;
1015 	} else {
1016 		if (pci_p->pci_soft_state == PCI_SOFT_STATE_OPEN_EXCL) {
1017 			mutex_exit(&pci_p->pci_mutex);
1018 			cmn_err(CE_NOTE, "npe_open: busy");
1019 			return (EBUSY);
1020 		}
1021 		pci_p->pci_soft_state = PCI_SOFT_STATE_OPEN;
1022 	}
1023 	mutex_exit(&pci_p->pci_mutex);
1024 
1025 	return (0);
1026 }
1027 
1028 static int
1029 npe_close(dev_t dev, int flags, int otyp, cred_t *credp)
1030 {
1031 	minor_t		minor = getminor(dev);
1032 	int		instance = PCI_MINOR_NUM_TO_INSTANCE(minor);
1033 	pci_state_t	*pci_p = ddi_get_soft_state(npe_statep, instance);
1034 	int	rv;
1035 
1036 	if (pci_p == NULL)
1037 		return (ENXIO);
1038 
1039 	mutex_enter(&pci_p->pci_mutex);
1040 
1041 	switch (PCI_MINOR_NUM_TO_PCI_DEVNUM(minor)) {
1042 	case PCI_TOOL_REG_MINOR_NUM:
1043 	case PCI_TOOL_INTR_MINOR_NUM:
1044 		break;
1045 	default:
1046 		/* Handle devctl ioctls */
1047 		rv = pcie_close(pci_p->pci_dip, dev, flags, otyp, credp);
1048 		mutex_exit(&pci_p->pci_mutex);
1049 		return (rv);
1050 	}
1051 
1052 	/* Handle pcitool ioctls */
1053 	pci_p->pci_soft_state = PCI_SOFT_STATE_CLOSED;
1054 	mutex_exit(&pci_p->pci_mutex);
1055 	return (0);
1056 }
1057 
1058 static int
1059 npe_ioctl(dev_t dev, int cmd, intptr_t arg, int mode, cred_t *credp, int *rvalp)
1060 {
1061 	minor_t		minor = getminor(dev);
1062 	int		instance = PCI_MINOR_NUM_TO_INSTANCE(minor);
1063 	pci_state_t	*pci_p = ddi_get_soft_state(npe_statep, instance);
1064 	int		ret = ENOTTY;
1065 
1066 	if (pci_p == NULL)
1067 		return (ENXIO);
1068 
1069 	switch (PCI_MINOR_NUM_TO_PCI_DEVNUM(minor)) {
1070 	case PCI_TOOL_REG_MINOR_NUM:
1071 	case PCI_TOOL_INTR_MINOR_NUM:
1072 		/* To handle pcitool related ioctls */
1073 		ret =  pci_common_ioctl(pci_p->pci_dip, dev, cmd, arg, mode,
1074 		    credp, rvalp);
1075 		break;
1076 	default:
1077 		/* To handle devctl and hotplug related ioctls */
1078 		ret = pcie_ioctl(pci_p->pci_dip, dev, cmd, arg, mode, credp,
1079 		    rvalp);
1080 		break;
1081 	}
1082 
1083 	return (ret);
1084 }
1085 
1086 /*ARGSUSED*/
1087 static int
1088 npe_fm_init(dev_info_t *dip, dev_info_t *tdip, int cap,
1089     ddi_iblock_cookie_t *ibc)
1090 {
1091 	pci_state_t  *pcip = ddi_get_soft_state(npe_statep,
1092 	    ddi_get_instance(dip));
1093 
1094 	ASSERT(ibc != NULL);
1095 	*ibc = pcip->pci_fm_ibc;
1096 
1097 	return (pcip->pci_fmcap);
1098 }
1099 
1100 /*ARGSUSED*/
1101 static int
1102 npe_fm_callback(dev_info_t *dip, ddi_fm_error_t *derr, const void *no_used)
1103 {
1104 	/*
1105 	 * On current x86 systems, npe's callback does not get called for failed
1106 	 * loads.  If in the future this feature is used, the fault PA should be
1107 	 * logged in the derr->fme_bus_specific field.  The appropriate PCIe
1108 	 * error handling code should be called and needs to be coordinated with
1109 	 * safe access handling.
1110 	 */
1111 
1112 	return (DDI_FM_OK);
1113 }
1114