xref: /titanic_51/usr/src/uts/sun4u/io/px/px_hlib.c (revision 1a7c1b724419d3cb5fa6eea75123c6b2060ba31b)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License, Version 1.0 only
6  * (the "License").  You may not use this file except in compliance
7  * with the License.
8  *
9  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
10  * or http://www.opensolaris.org/os/licensing.
11  * See the License for the specific language governing permissions
12  * and limitations under the License.
13  *
14  * When distributing Covered Code, include this CDDL HEADER in each
15  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
16  * If applicable, add the following below this CDDL HEADER, with the
17  * fields enclosed by brackets "[]" replaced with your own identifying
18  * information: Portions Copyright [yyyy] [name of copyright owner]
19  *
20  * CDDL HEADER END
21  */
22 /*
23  * Copyright 2005 Sun Microsystems, Inc.  All rights reserved.
24  * Use is subject to license terms.
25  */
26 
27 #pragma ident	"%Z%%M%	%I%	%E% SMI"
28 
29 #include <sys/types.h>
30 #include <sys/cmn_err.h>
31 #include <sys/vmsystm.h>
32 #include <sys/vmem.h>
33 #include <sys/machsystm.h>	/* lddphys() */
34 #include <sys/iommutsb.h>
35 #include <sys/pci.h>
36 #include <pcie_pwr.h>
37 #include <px_obj.h>
38 #include "px_regs.h"
39 #include "px_csr.h"
40 #include "px_lib4u.h"
41 
42 /*
43  * Registers that need to be saved and restored during suspend/resume.
44  */
45 
46 /*
47  * Registers in the PEC Module.
48  * LPU_RESET should be set to 0ull during resume
49  */
50 static uint64_t	pec_config_state_regs[] = {
51 	PEC_CORE_AND_BLOCK_INTERRUPT_ENABLE,
52 	ILU_ERROR_LOG_ENABLE,
53 	ILU_INTERRUPT_ENABLE,
54 	TLU_CONTROL,
55 	TLU_OTHER_EVENT_LOG_ENABLE,
56 	TLU_OTHER_EVENT_INTERRUPT_ENABLE,
57 	TLU_DEVICE_CONTROL,
58 	TLU_LINK_CONTROL,
59 	TLU_UNCORRECTABLE_ERROR_LOG_ENABLE,
60 	TLU_UNCORRECTABLE_ERROR_INTERRUPT_ENABLE,
61 	TLU_CORRECTABLE_ERROR_LOG_ENABLE,
62 	TLU_CORRECTABLE_ERROR_INTERRUPT_ENABLE,
63 	LPU_LINK_LAYER_INTERRUPT_MASK,
64 	LPU_PHY_INTERRUPT_MASK,
65 	LPU_RECEIVE_PHY_INTERRUPT_MASK,
66 	LPU_TRANSMIT_PHY_INTERRUPT_MASK,
67 	LPU_GIGABLAZE_GLUE_INTERRUPT_MASK,
68 	LPU_LTSSM_INTERRUPT_MASK,
69 	LPU_RESET,
70 	LPU_DEBUG_CONFIG,
71 	LPU_INTERRUPT_MASK,
72 	LPU_LINK_LAYER_CONFIG,
73 	LPU_FLOW_CONTROL_UPDATE_CONTROL,
74 	LPU_TXLINK_FREQUENT_NAK_LATENCY_TIMER_THRESHOLD,
75 	LPU_TXLINK_REPLAY_TIMER_THRESHOLD,
76 	LPU_REPLAY_BUFFER_MAX_ADDRESS,
77 	LPU_TXLINK_RETRY_FIFO_POINTER,
78 	LPU_LTSSM_CONFIG2,
79 	LPU_LTSSM_CONFIG3,
80 	LPU_LTSSM_CONFIG4,
81 	LPU_LTSSM_CONFIG5,
82 	DMC_CORE_AND_BLOCK_INTERRUPT_ENABLE,
83 	DMC_DEBUG_SELECT_FOR_PORT_A,
84 	DMC_DEBUG_SELECT_FOR_PORT_B
85 };
86 #define	PEC_SIZE (sizeof (pec_config_state_regs))
87 #define	PEC_KEYS (PEC_SIZE / sizeof (uint64_t))
88 
89 /*
90  * Registers for the MMU module.
91  * MMU_TTE_CACHE_INVALIDATE needs to be cleared. (-1ull)
92  */
93 static uint64_t mmu_config_state_regs[] = {
94 	MMU_TSB_CONTROL,
95 	MMU_CONTROL_AND_STATUS,
96 	MMU_ERROR_LOG_ENABLE,
97 	MMU_INTERRUPT_ENABLE
98 };
99 #define	MMU_SIZE (sizeof (mmu_config_state_regs))
100 #define	MMU_KEYS (MMU_SIZE / sizeof (uint64_t))
101 
102 /*
103  * Registers for the IB Module
104  */
105 static uint64_t ib_config_state_regs[] = {
106 	IMU_ERROR_LOG_ENABLE,
107 	IMU_INTERRUPT_ENABLE
108 };
109 #define	IB_SIZE (sizeof (ib_config_state_regs))
110 #define	IB_KEYS (IB_SIZE / sizeof (uint64_t))
111 #define	IB_MAP_SIZE (INTERRUPT_MAPPING_ENTRIES * sizeof (uint64_t))
112 
113 /*
114  * Registers for the CB module.
115  * JBC_ERROR_STATUS_CLEAR needs to be cleared. (-1ull)
116  */
117 static uint64_t	cb_config_state_regs[] = {
118 	JBUS_PARITY_CONTROL,
119 	JBC_FATAL_RESET_ENABLE,
120 	JBC_CORE_AND_BLOCK_INTERRUPT_ENABLE,
121 	JBC_ERROR_LOG_ENABLE,
122 	JBC_INTERRUPT_ENABLE
123 };
124 #define	CB_SIZE (sizeof (cb_config_state_regs))
125 #define	CB_KEYS (CB_SIZE / sizeof (uint64_t))
126 
127 static uint64_t	msiq_config_other_regs[] = {
128 	ERR_COR_MAPPING,
129 	ERR_NONFATAL_MAPPING,
130 	ERR_FATAL_MAPPING,
131 	PM_PME_MAPPING,
132 	PME_TO_ACK_MAPPING,
133 	MSI_32_BIT_ADDRESS,
134 	MSI_64_BIT_ADDRESS
135 };
136 #define	MSIQ_OTHER_SIZE	(sizeof (msiq_config_other_regs))
137 #define	MSIQ_OTHER_KEYS	(MSIQ_OTHER_SIZE / sizeof (uint64_t))
138 
139 #define	MSIQ_STATE_SIZE		(EVENT_QUEUE_STATE_ENTRIES * sizeof (uint64_t))
140 #define	MSIQ_MAPPING_SIZE	(MSI_MAPPING_ENTRIES * sizeof (uint64_t))
141 
142 static uint64_t msiq_suspend(devhandle_t dev_hdl, pxu_t *pxu_p);
143 static void msiq_resume(devhandle_t dev_hdl, pxu_t *pxu_p);
144 
145 /*
146  * Initialize the module, but do not enable interrupts.
147  */
148 /* ARGSUSED */
149 void
150 hvio_cb_init(caddr_t xbc_csr_base, pxu_t *pxu_p)
151 {
152 	uint64_t val;
153 
154 	/* Check if we need to enable inverted parity */
155 	val = (1ULL << JBUS_PARITY_CONTROL_P_EN);
156 	CSR_XS(xbc_csr_base, JBUS_PARITY_CONTROL, val);
157 	DBG(DBG_CB, NULL, "hvio_cb_init, JBUS_PARITY_CONTROL: 0x%llx\n",
158 	    CSR_XR(xbc_csr_base, JBUS_PARITY_CONTROL));
159 
160 	val = (1 << JBC_FATAL_RESET_ENABLE_SPARE_P_INT_EN) |
161 	    (1 << JBC_FATAL_RESET_ENABLE_MB_PEA_P_INT_EN) |
162 	    (1 << JBC_FATAL_RESET_ENABLE_CPE_P_INT_EN) |
163 	    (1 << JBC_FATAL_RESET_ENABLE_APE_P_INT_EN) |
164 	    (1 << JBC_FATAL_RESET_ENABLE_PIO_CPE_INT_EN) |
165 	    (1 << JBC_FATAL_RESET_ENABLE_JTCEEW_P_INT_EN) |
166 	    (1 << JBC_FATAL_RESET_ENABLE_JTCEEI_P_INT_EN) |
167 	    (1 << JBC_FATAL_RESET_ENABLE_JTCEER_P_INT_EN);
168 	CSR_XS(xbc_csr_base, JBC_FATAL_RESET_ENABLE, val);
169 	DBG(DBG_CB, NULL, "hvio_cb_init, JBC_FATAL_RESET_ENABLE: 0x%llx\n",
170 		CSR_XR(xbc_csr_base, JBC_FATAL_RESET_ENABLE));
171 
172 	/*
173 	 * Enable merge, jbc and dmc interrupts.
174 	 */
175 	CSR_XS(xbc_csr_base, JBC_CORE_AND_BLOCK_INTERRUPT_ENABLE, -1ull);
176 	DBG(DBG_CB, NULL,
177 	    "hvio_cb_init, JBC_CORE_AND_BLOCK_INTERRUPT_ENABLE: 0x%llx\n",
178 	    CSR_XR(xbc_csr_base, JBC_CORE_AND_BLOCK_INTERRUPT_ENABLE));
179 
180 	/*
181 	 * CSR_V CB's interrupt regs (log, enable, status, clear)
182 	 */
183 	DBG(DBG_CB, NULL, "hvio_cb_init, JBC_ERROR_LOG_ENABLE: 0x%llx\n",
184 	    CSR_XR(xbc_csr_base, JBC_ERROR_LOG_ENABLE));
185 
186 	DBG(DBG_CB, NULL, "hvio_cb_init, JBC_INTERRUPT_ENABLE: 0x%llx\n",
187 	    CSR_XR(xbc_csr_base, JBC_INTERRUPT_ENABLE));
188 
189 	DBG(DBG_CB, NULL, "hvio_cb_init, JBC_INTERRUPT_STATUS: 0x%llx\n",
190 	    CSR_XR(xbc_csr_base, JBC_INTERRUPT_STATUS));
191 
192 	DBG(DBG_CB, NULL, "hvio_cb_init, JBC_ERROR_STATUS_CLEAR: 0x%llx\n",
193 	    CSR_XR(xbc_csr_base, JBC_ERROR_STATUS_CLEAR));
194 }
195 
196 /*
197  * Initialize the module, but do not enable interrupts.
198  */
199 /* ARGSUSED */
200 void
201 hvio_ib_init(caddr_t csr_base, pxu_t *pxu_p)
202 {
203 	/*
204 	 * CSR_V IB's interrupt regs (log, enable, status, clear)
205 	 */
206 	DBG(DBG_IB, NULL, "hvio_ib_init - IMU_ERROR_LOG_ENABLE: 0x%llx\n",
207 	    CSR_XR(csr_base, IMU_ERROR_LOG_ENABLE));
208 
209 	DBG(DBG_IB, NULL, "hvio_ib_init - IMU_INTERRUPT_ENABLE: 0x%llx\n",
210 	    CSR_XR(csr_base, IMU_INTERRUPT_ENABLE));
211 
212 	DBG(DBG_IB, NULL, "hvio_ib_init - IMU_INTERRUPT_STATUS: 0x%llx\n",
213 	    CSR_XR(csr_base, IMU_INTERRUPT_STATUS));
214 
215 	DBG(DBG_IB, NULL, "hvio_ib_init - IMU_ERROR_STATUS_CLEAR: 0x%llx\n",
216 	    CSR_XR(csr_base, IMU_ERROR_STATUS_CLEAR));
217 }
218 
219 /*
220  * Initialize the module, but do not enable interrupts.
221  */
222 /* ARGSUSED */
223 static void
224 ilu_init(caddr_t csr_base, pxu_t *pxu_p)
225 {
226 	/*
227 	 * CSR_V ILU's interrupt regs (log, enable, status, clear)
228 	 */
229 	DBG(DBG_ILU, NULL, "ilu_init - ILU_ERROR_LOG_ENABLE: 0x%llx\n",
230 	    CSR_XR(csr_base, ILU_ERROR_LOG_ENABLE));
231 
232 	DBG(DBG_ILU, NULL, "ilu_init - ILU_INTERRUPT_ENABLE: 0x%llx\n",
233 	    CSR_XR(csr_base, ILU_INTERRUPT_ENABLE));
234 
235 	DBG(DBG_ILU, NULL, "ilu_init - ILU_INTERRUPT_STATUS: 0x%llx\n",
236 	    CSR_XR(csr_base, ILU_INTERRUPT_STATUS));
237 
238 	DBG(DBG_ILU, NULL, "ilu_init - ILU_ERROR_STATUS_CLEAR: 0x%llx\n",
239 	    CSR_XR(csr_base, ILU_ERROR_STATUS_CLEAR));
240 }
241 
242 /*
243  * Initialize the module, but do not enable interrupts.
244  */
245 /* ARGSUSED */
246 static void
247 tlu_init(caddr_t csr_base, pxu_t *pxu_p)
248 {
249 	uint64_t val;
250 
251 	/*
252 	 * CSR_V TLU_CONTROL Expect OBP ???
253 	 */
254 
255 	/*
256 	 * L0s entry default timer value - 7.0 us
257 	 * Completion timeout select default value - 67.1 ms and
258 	 * OBP will set this value.
259 	 *
260 	 * Configuration - Bit 0 should always be 0 for upstream port.
261 	 * Bit 1 is clock - how is this related to the clock bit in TLU
262 	 * Link Control register?  Both are hardware dependent and likely
263 	 * set by OBP.
264 	 *
265 	 * Disable non-posted write bit - ordering by setting
266 	 * NPWR_EN bit to force serialization of writes.
267 	 */
268 	val = CSR_XR(csr_base, TLU_CONTROL);
269 	val |= (TLU_CONTROL_L0S_TIM_DEFAULT << TLU_CONTROL_L0S_TIM) |
270 	    (1ull << TLU_CONTROL_NPWR_EN) | TLU_CONTROL_CONFIG_DEFAULT;
271 
272 	/*
273 	 * Set Detect.Quiet. This will disable automatic link
274 	 * re-training, if the link goes down e.g. power management
275 	 * turns off power to the downstream device. This will enable
276 	 * Fire to go to Drain state, after link down. The drain state
277 	 * forces a reset to the FC state machine, which is required for
278 	 * proper link re-training.
279 	 */
280 	val |= (1ull << TLU_REMAIN_DETECT_QUIET);
281 	CSR_XS(csr_base, TLU_CONTROL, val);
282 	DBG(DBG_TLU, NULL, "tlu_init - TLU_CONTROL: 0x%llx\n",
283 	    CSR_XR(csr_base, TLU_CONTROL));
284 
285 	/*
286 	 * CSR_V TLU_STATUS Expect HW 0x4
287 	 */
288 
289 	/*
290 	 * Only bit [7:0] are currently defined.  Bits [2:0]
291 	 * are the state, which should likely be in state active,
292 	 * 100b.  Bit three is 'recovery', which is not understood.
293 	 * All other bits are reserved.
294 	 */
295 	DBG(DBG_TLU, NULL, "tlu_init - TLU_STATUS: 0x%llx\n",
296 	    CSR_XR(csr_base, TLU_STATUS));
297 
298 	/*
299 	 * CSR_V TLU_PME_TURN_OFF_GENERATE Expect HW 0x0
300 	 */
301 	DBG(DBG_TLU, NULL, "tlu_init - TLU_PME_TURN_OFF_GENERATE: 0x%llx\n",
302 	    CSR_XR(csr_base, TLU_PME_TURN_OFF_GENERATE));
303 
304 	/*
305 	 * CSR_V TLU_INGRESS_CREDITS_INITIAL Expect HW 0x10000200C0
306 	 */
307 
308 	/*
309 	 * Ingress credits initial register.  Bits [39:32] should be
310 	 * 0x10, bits [19:12] should be 0x20, and bits [11:0] should
311 	 * be 0xC0.  These are the reset values, and should be set by
312 	 * HW.
313 	 */
314 	DBG(DBG_TLU, NULL, "tlu_init - TLU_INGRESS_CREDITS_INITIAL: 0x%llx\n",
315 	    CSR_XR(csr_base, TLU_INGRESS_CREDITS_INITIAL));
316 
317 	/*
318 	 * CSR_V TLU_DIAGNOSTIC Expect HW 0x0
319 	 */
320 
321 	/*
322 	 * Diagnostic register - always zero unless we are debugging.
323 	 */
324 	DBG(DBG_TLU, NULL, "tlu_init - TLU_DIAGNOSTIC: 0x%llx\n",
325 	    CSR_XR(csr_base, TLU_DIAGNOSTIC));
326 
327 	/*
328 	 * CSR_V TLU_EGRESS_CREDITS_CONSUMED Expect HW 0x0
329 	 */
330 	DBG(DBG_TLU, NULL, "tlu_init - TLU_EGRESS_CREDITS_CONSUMED: 0x%llx\n",
331 	    CSR_XR(csr_base, TLU_EGRESS_CREDITS_CONSUMED));
332 
333 	/*
334 	 * CSR_V TLU_EGRESS_CREDIT_LIMIT Expect HW 0x0
335 	 */
336 	DBG(DBG_TLU, NULL, "tlu_init - TLU_EGRESS_CREDIT_LIMIT: 0x%llx\n",
337 	    CSR_XR(csr_base, TLU_EGRESS_CREDIT_LIMIT));
338 
339 	/*
340 	 * CSR_V TLU_EGRESS_RETRY_BUFFER Expect HW 0x0
341 	 */
342 	DBG(DBG_TLU, NULL, "tlu_init - TLU_EGRESS_RETRY_BUFFER: 0x%llx\n",
343 	    CSR_XR(csr_base, TLU_EGRESS_RETRY_BUFFER));
344 
345 	/*
346 	 * CSR_V TLU_INGRESS_CREDITS_ALLOCATED Expected HW 0x0
347 	 */
348 	DBG(DBG_TLU, NULL,
349 	    "tlu_init - TLU_INGRESS_CREDITS_ALLOCATED: 0x%llx\n",
350 	    CSR_XR(csr_base, TLU_INGRESS_CREDITS_ALLOCATED));
351 
352 	/*
353 	 * CSR_V TLU_INGRESS_CREDITS_RECEIVED Expected HW 0x0
354 	 */
355 	DBG(DBG_TLU, NULL,
356 	    "tlu_init - TLU_INGRESS_CREDITS_RECEIVED: 0x%llx\n",
357 	    CSR_XR(csr_base, TLU_INGRESS_CREDITS_RECEIVED));
358 
359 	/*
360 	 * CSR_V TLU's interrupt regs (log, enable, status, clear)
361 	 */
362 	DBG(DBG_TLU, NULL,
363 	    "tlu_init - TLU_OTHER_EVENT_LOG_ENABLE: 0x%llx\n",
364 	    CSR_XR(csr_base, TLU_OTHER_EVENT_LOG_ENABLE));
365 
366 	DBG(DBG_TLU, NULL,
367 	    "tlu_init - TLU_OTHER_EVENT_INTERRUPT_ENABLE: 0x%llx\n",
368 	    CSR_XR(csr_base, TLU_OTHER_EVENT_INTERRUPT_ENABLE));
369 
370 	DBG(DBG_TLU, NULL,
371 	    "tlu_init - TLU_OTHER_EVENT_INTERRUPT_STATUS: 0x%llx\n",
372 	    CSR_XR(csr_base, TLU_OTHER_EVENT_INTERRUPT_STATUS));
373 
374 	DBG(DBG_TLU, NULL,
375 	    "tlu_init - TLU_OTHER_EVENT_STATUS_CLEAR: 0x%llx\n",
376 	    CSR_XR(csr_base, TLU_OTHER_EVENT_STATUS_CLEAR));
377 
378 	/*
379 	 * CSR_V TLU_RECEIVE_OTHER_EVENT_HEADER1_LOG Expect HW 0x0
380 	 */
381 	DBG(DBG_TLU, NULL,
382 	    "tlu_init - TLU_RECEIVE_OTHER_EVENT_HEADER1_LOG: 0x%llx\n",
383 	    CSR_XR(csr_base, TLU_RECEIVE_OTHER_EVENT_HEADER1_LOG));
384 
385 	/*
386 	 * CSR_V TLU_RECEIVE_OTHER_EVENT_HEADER2_LOG Expect HW 0x0
387 	 */
388 	DBG(DBG_TLU, NULL,
389 	    "tlu_init - TLU_RECEIVE_OTHER_EVENT_HEADER2_LOG: 0x%llx\n",
390 	    CSR_XR(csr_base, TLU_RECEIVE_OTHER_EVENT_HEADER2_LOG));
391 
392 	/*
393 	 * CSR_V TLU_TRANSMIT_OTHER_EVENT_HEADER1_LOG Expect HW 0x0
394 	 */
395 	DBG(DBG_TLU, NULL,
396 	    "tlu_init - TLU_TRANSMIT_OTHER_EVENT_HEADER1_LOG: 0x%llx\n",
397 	    CSR_XR(csr_base, TLU_TRANSMIT_OTHER_EVENT_HEADER1_LOG));
398 
399 	/*
400 	 * CSR_V TLU_TRANSMIT_OTHER_EVENT_HEADER2_LOG Expect HW 0x0
401 	 */
402 	DBG(DBG_TLU, NULL,
403 	    "tlu_init - TLU_TRANSMIT_OTHER_EVENT_HEADER2_LOG: 0x%llx\n",
404 	    CSR_XR(csr_base, TLU_TRANSMIT_OTHER_EVENT_HEADER2_LOG));
405 
406 	/*
407 	 * CSR_V TLU_PERFORMANCE_COUNTER_SELECT Expect HW 0x0
408 	 */
409 	DBG(DBG_TLU, NULL,
410 	    "tlu_init - TLU_PERFORMANCE_COUNTER_SELECT: 0x%llx\n",
411 	    CSR_XR(csr_base, TLU_PERFORMANCE_COUNTER_SELECT));
412 
413 	/*
414 	 * CSR_V TLU_PERFORMANCE_COUNTER_ZERO Expect HW 0x0
415 	 */
416 	DBG(DBG_TLU, NULL,
417 	    "tlu_init - TLU_PERFORMANCE_COUNTER_ZERO: 0x%llx\n",
418 	    CSR_XR(csr_base, TLU_PERFORMANCE_COUNTER_ZERO));
419 
420 	/*
421 	 * CSR_V TLU_PERFORMANCE_COUNTER_ONE Expect HW 0x0
422 	 */
423 	DBG(DBG_TLU, NULL, "tlu_init - TLU_PERFORMANCE_COUNTER_ONE: 0x%llx\n",
424 	    CSR_XR(csr_base, TLU_PERFORMANCE_COUNTER_ONE));
425 
426 	/*
427 	 * CSR_V TLU_PERFORMANCE_COUNTER_TWO Expect HW 0x0
428 	 */
429 	DBG(DBG_TLU, NULL, "tlu_init - TLU_PERFORMANCE_COUNTER_TWO: 0x%llx\n",
430 	    CSR_XR(csr_base, TLU_PERFORMANCE_COUNTER_TWO));
431 
432 	/*
433 	 * CSR_V TLU_DEBUG_SELECT_A Expect HW 0x0
434 	 */
435 
436 	DBG(DBG_TLU, NULL, "tlu_init - TLU_DEBUG_SELECT_A: 0x%llx\n",
437 	    CSR_XR(csr_base, TLU_DEBUG_SELECT_A));
438 
439 	/*
440 	 * CSR_V TLU_DEBUG_SELECT_B Expect HW 0x0
441 	 */
442 	DBG(DBG_TLU, NULL, "tlu_init - TLU_DEBUG_SELECT_B: 0x%llx\n",
443 	    CSR_XR(csr_base, TLU_DEBUG_SELECT_B));
444 
445 	/*
446 	 * CSR_V TLU_DEVICE_CAPABILITIES Expect HW 0xFC2
447 	 */
448 	DBG(DBG_TLU, NULL, "tlu_init - TLU_DEVICE_CAPABILITIES: 0x%llx\n",
449 	    CSR_XR(csr_base, TLU_DEVICE_CAPABILITIES));
450 
451 	/*
452 	 * CSR_V TLU_DEVICE_CONTROL Expect HW 0x0
453 	 */
454 
455 	/*
456 	 * Bits [14:12] are the Max Read Request Size, which is always 64
457 	 * bytes which is 000b.  Bits [7:5] are Max Payload Size, which
458 	 * start at 128 bytes which is 000b.  This may be revisited if
459 	 * init_child finds greater values.
460 	 */
461 	val = 0x0ull;
462 	CSR_XS(csr_base, TLU_DEVICE_CONTROL, val);
463 	DBG(DBG_TLU, NULL, "tlu_init - TLU_DEVICE_CONTROL: 0x%llx\n",
464 	    CSR_XR(csr_base, TLU_DEVICE_CONTROL));
465 
466 	/*
467 	 * CSR_V TLU_DEVICE_STATUS Expect HW 0x0
468 	 */
469 	DBG(DBG_TLU, NULL, "tlu_init - TLU_DEVICE_STATUS: 0x%llx\n",
470 	    CSR_XR(csr_base, TLU_DEVICE_STATUS));
471 
472 	/*
473 	 * CSR_V TLU_LINK_CAPABILITIES Expect HW 0x15C81
474 	 */
475 	DBG(DBG_TLU, NULL, "tlu_init - TLU_LINK_CAPABILITIES: 0x%llx\n",
476 	    CSR_XR(csr_base, TLU_LINK_CAPABILITIES));
477 
478 	/*
479 	 * CSR_V TLU_LINK_CONTROL Expect OBP 0x40
480 	 */
481 
482 	/*
483 	 * The CLOCK bit should be set by OBP if the hardware dictates,
484 	 * and if it is set then ASPM should be used since then L0s exit
485 	 * latency should be lower than L1 exit latency.
486 	 *
487 	 * Note that we will not enable power management during bringup
488 	 * since it has not been test and is creating some problems in
489 	 * simulation.
490 	 */
491 	val = (1ull << TLU_LINK_CONTROL_CLOCK);
492 
493 	CSR_XS(csr_base, TLU_LINK_CONTROL, val);
494 	DBG(DBG_TLU, NULL, "tlu_init - TLU_LINK_CONTROL: 0x%llx\n",
495 	    CSR_XR(csr_base, TLU_LINK_CONTROL));
496 
497 	/*
498 	 * CSR_V TLU_LINK_STATUS Expect OBP 0x1011
499 	 */
500 
501 	/*
502 	 * Not sure if HW or OBP will be setting this read only
503 	 * register.  Bit 12 is Clock, and it should always be 1
504 	 * signifying that the component uses the same physical
505 	 * clock as the platform.  Bits [9:4] are for the width,
506 	 * with the expected value above signifying a x1 width.
507 	 * Bits [3:0] are the speed, with 1b signifying 2.5 Gb/s,
508 	 * the only speed as yet supported by the PCI-E spec.
509 	 */
510 	DBG(DBG_TLU, NULL, "tlu_init - TLU_LINK_STATUS: 0x%llx\n",
511 	    CSR_XR(csr_base, TLU_LINK_STATUS));
512 
513 	/*
514 	 * CSR_V TLU_SLOT_CAPABILITIES Expect OBP ???
515 	 */
516 
517 	/*
518 	 * Power Limits for the slots.  Will be platform
519 	 * dependent, and OBP will need to set after consulting
520 	 * with the HW guys.
521 	 *
522 	 * Bits [16:15] are power limit scale, which most likely
523 	 * will be 0b signifying 1x.  Bits [14:7] are the Set
524 	 * Power Limit Value, which is a number which is multiplied
525 	 * by the power limit scale to get the actual power limit.
526 	 */
527 	DBG(DBG_TLU, NULL, "tlu_init - TLU_SLOT_CAPABILITIES: 0x%llx\n",
528 	    CSR_XR(csr_base, TLU_SLOT_CAPABILITIES));
529 
530 	/*
531 	 * CSR_V TLU_UNCORRECTABLE_ERROR_LOG_ENABLE Expect Kernel 0x17F011
532 	 */
533 	DBG(DBG_TLU, NULL,
534 	    "tlu_init - TLU_UNCORRECTABLE_ERROR_LOG_ENABLE: 0x%llx\n",
535 	    CSR_XR(csr_base, TLU_UNCORRECTABLE_ERROR_LOG_ENABLE));
536 
537 	/*
538 	 * CSR_V TLU_UNCORRECTABLE_ERROR_INTERRUPT_ENABLE Expect
539 	 * Kernel 0x17F0110017F011
540 	 */
541 	DBG(DBG_TLU, NULL,
542 	    "tlu_init - TLU_UNCORRECTABLE_ERROR_INTERRUPT_ENABLE: 0x%llx\n",
543 	    CSR_XR(csr_base, TLU_UNCORRECTABLE_ERROR_INTERRUPT_ENABLE));
544 
545 	/*
546 	 * CSR_V TLU_UNCORRECTABLE_ERROR_INTERRUPT_STATUS Expect HW 0x0
547 	 */
548 	DBG(DBG_TLU, NULL,
549 	    "tlu_init - TLU_UNCORRECTABLE_ERROR_INTERRUPT_STATUS: 0x%llx\n",
550 	    CSR_XR(csr_base, TLU_UNCORRECTABLE_ERROR_INTERRUPT_STATUS));
551 
552 	/*
553 	 * CSR_V TLU_UNCORRECTABLE_ERROR_STATUS_CLEAR Expect HW 0x0
554 	 */
555 	DBG(DBG_TLU, NULL,
556 	    "tlu_init - TLU_UNCORRECTABLE_ERROR_STATUS_CLEAR: 0x%llx\n",
557 	    CSR_XR(csr_base, TLU_UNCORRECTABLE_ERROR_STATUS_CLEAR));
558 
559 	/*
560 	 * CSR_V TLU_RECEIVE_UNCORRECTABLE_ERROR_HEADER1_LOG HW 0x0
561 	 */
562 	DBG(DBG_TLU, NULL,
563 	    "tlu_init - TLU_RECEIVE_UNCORRECTABLE_ERROR_HEADER1_LOG: 0x%llx\n",
564 	    CSR_XR(csr_base, TLU_RECEIVE_UNCORRECTABLE_ERROR_HEADER1_LOG));
565 
566 	/*
567 	 * CSR_V TLU_RECEIVE_UNCORRECTABLE_ERROR_HEADER2_LOG HW 0x0
568 	 */
569 	DBG(DBG_TLU, NULL,
570 	    "tlu_init - TLU_RECEIVE_UNCORRECTABLE_ERROR_HEADER2_LOG: 0x%llx\n",
571 	    CSR_XR(csr_base, TLU_RECEIVE_UNCORRECTABLE_ERROR_HEADER2_LOG));
572 
573 	/*
574 	 * CSR_V TLU_TRANSMIT_UNCORRECTABLE_ERROR_HEADER1_LOG HW 0x0
575 	 */
576 	DBG(DBG_TLU, NULL,
577 	    "tlu_init - TLU_TRANSMIT_UNCORRECTABLE_ERROR_HEADER1_LOG: 0x%llx\n",
578 	    CSR_XR(csr_base, TLU_TRANSMIT_UNCORRECTABLE_ERROR_HEADER1_LOG));
579 
580 	/*
581 	 * CSR_V TLU_TRANSMIT_UNCORRECTABLE_ERROR_HEADER2_LOG HW 0x0
582 	 */
583 	DBG(DBG_TLU, NULL,
584 	    "tlu_init - TLU_TRANSMIT_UNCORRECTABLE_ERROR_HEADER2_LOG: 0x%llx\n",
585 	    CSR_XR(csr_base, TLU_TRANSMIT_UNCORRECTABLE_ERROR_HEADER2_LOG));
586 
587 
588 	/*
589 	 * CSR_V TLU's CE interrupt regs (log, enable, status, clear)
590 	 * Plus header logs
591 	 */
592 
593 	/*
594 	 * CSR_V TLU_CORRECTABLE_ERROR_LOG_ENABLE Expect Kernel 0x11C1
595 	 */
596 	DBG(DBG_TLU, NULL,
597 	    "tlu_init - TLU_CORRECTABLE_ERROR_LOG_ENABLE: 0x%llx\n",
598 	    CSR_XR(csr_base, TLU_CORRECTABLE_ERROR_LOG_ENABLE));
599 
600 	/*
601 	 * CSR_V TLU_CORRECTABLE_ERROR_INTERRUPT_ENABLE Kernel 0x11C1000011C1
602 	 */
603 	DBG(DBG_TLU, NULL,
604 	    "tlu_init - TLU_CORRECTABLE_ERROR_INTERRUPT_ENABLE: 0x%llx\n",
605 	    CSR_XR(csr_base, TLU_CORRECTABLE_ERROR_INTERRUPT_ENABLE));
606 
607 	/*
608 	 * CSR_V TLU_CORRECTABLE_ERROR_INTERRUPT_STATUS Expect HW 0x0
609 	 */
610 	DBG(DBG_TLU, NULL,
611 	    "tlu_init - TLU_CORRECTABLE_ERROR_INTERRUPT_STATUS: 0x%llx\n",
612 	    CSR_XR(csr_base, TLU_CORRECTABLE_ERROR_INTERRUPT_STATUS));
613 
614 	/*
615 	 * CSR_V TLU_CORRECTABLE_ERROR_STATUS_CLEAR Expect HW 0x0
616 	 */
617 	DBG(DBG_TLU, NULL,
618 	    "tlu_init - TLU_CORRECTABLE_ERROR_STATUS_CLEAR: 0x%llx\n",
619 	    CSR_XR(csr_base, TLU_CORRECTABLE_ERROR_STATUS_CLEAR));
620 }
621 
622 /* ARGSUSED */
623 static void
624 lpu_init(caddr_t csr_base, pxu_t *pxu_p)
625 {
626 	/* Variables used to set the ACKNAK Latency Timer and Replay Timer */
627 	int link_width, max_payload;
628 
629 	uint64_t val;
630 
631 	/*
632 	 * ACKNAK Latency Threshold Table.
633 	 * See Fire PRM 2.0 section 1.2.12.2, table 1-17.
634 	 */
635 	int acknak_timer_table[LINK_MAX_PKT_ARR_SIZE][LINK_WIDTH_ARR_SIZE] = {
636 		{0xED,   0x49,  0x43,  0x30},
637 		{0x1A0,  0x76,  0x6B,  0x48},
638 		{0x22F,  0x9A,  0x56,  0x56},
639 		{0x42F,  0x11A, 0x96,  0x96},
640 		{0x82F,  0x21A, 0x116, 0x116},
641 		{0x102F, 0x41A, 0x216, 0x216}
642 	};
643 
644 	/*
645 	 * TxLink Replay Timer Latency Table
646 	 * See Fire PRM 2.0 sections 1.2.12.3, table 1-18.
647 	 */
648 	int replay_timer_table[LINK_MAX_PKT_ARR_SIZE][LINK_WIDTH_ARR_SIZE] = {
649 		{0x379,  0x112, 0xFC,  0xB4},
650 		{0x618,  0x1BA, 0x192, 0x10E},
651 		{0x831,  0x242, 0x143, 0x143},
652 		{0xFB1,  0x422, 0x233, 0x233},
653 		{0x1EB0, 0x7E1, 0x412, 0x412},
654 		{0x3CB0, 0xF61, 0x7D2, 0x7D2}
655 	};
656 
657 	/*
658 	 * Get the Link Width.  See table above LINK_WIDTH_ARR_SIZE #define
659 	 * Only Link Widths of x1, x4, and x8 are supported.
660 	 * If any width is reported other than x8, set default to x8.
661 	 */
662 	link_width = CSR_FR(csr_base, TLU_LINK_STATUS, WIDTH);
663 	DBG(DBG_LPU, NULL, "lpu_init - Link Width: x%d\n", link_width);
664 
665 	/*
666 	 * Convert link_width to match timer array configuration.
667 	 */
668 	switch (link_width) {
669 	case 1:
670 		link_width = 0;
671 		break;
672 	case 4:
673 		link_width = 1;
674 		break;
675 	case 8:
676 		link_width = 2;
677 		break;
678 	case 16:
679 		link_width = 3;
680 		break;
681 	default:
682 		link_width = 0;
683 	}
684 
685 	/*
686 	 * Get the Max Payload Size.
687 	 * See table above LINK_MAX_PKT_ARR_SIZE #define
688 	 */
689 	max_payload = ((CSR_FR(csr_base, TLU_CONTROL, CONFIG) &
690 	    TLU_CONTROL_MPS_MASK) >> TLU_CONTROL_MPS_SHIFT);
691 
692 	DBG(DBG_LPU, NULL, "lpu_init - May Payload: %d\n",
693 	    (0x80 << max_payload));
694 
695 	/* Make sure the packet size is not greater than 4096 */
696 	max_payload = (max_payload >= LINK_MAX_PKT_ARR_SIZE) ?
697 	    (LINK_MAX_PKT_ARR_SIZE - 1) : max_payload;
698 
699 	/*
700 	 * CSR_V LPU_ID Expect HW 0x0
701 	 */
702 
703 	/*
704 	 * This register has link id, phy id and gigablaze id.
705 	 * Should be set by HW.
706 	 */
707 	DBG(DBG_LPU, NULL, "lpu_init - LPU_ID: 0x%llx\n",
708 	    CSR_XR(csr_base, LPU_ID));
709 
710 	/*
711 	 * CSR_V LPU_RESET Expect Kernel 0x0
712 	 */
713 
714 	/*
715 	 * No reason to have any reset bits high until an error is
716 	 * detected on the link.
717 	 */
718 	val = 0ull;
719 	CSR_XS(csr_base, LPU_RESET, val);
720 	DBG(DBG_LPU, NULL, "lpu_init - LPU_RESET: 0x%llx\n",
721 	    CSR_XR(csr_base, LPU_RESET));
722 
723 	/*
724 	 * CSR_V LPU_DEBUG_STATUS Expect HW 0x0
725 	 */
726 
727 	/*
728 	 * Bits [15:8] are Debug B, and bit [7:0] are Debug A.
729 	 * They are read-only.  What do the 8 bits mean, and
730 	 * how do they get set if they are read only?
731 	 */
732 	DBG(DBG_LPU, NULL, "lpu_init - LPU_DEBUG_STATUS: 0x%llx\n",
733 	    CSR_XR(csr_base, LPU_DEBUG_STATUS));
734 
735 	/*
736 	 * CSR_V LPU_DEBUG_CONFIG Expect Kernel 0x0
737 	 */
738 	DBG(DBG_LPU, NULL, "lpu_init - LPU_DEBUG_CONFIG: 0x%llx\n",
739 	    CSR_XR(csr_base, LPU_DEBUG_CONFIG));
740 
741 	/*
742 	 * CSR_V LPU_LTSSM_CONTROL Expect HW 0x0
743 	 */
744 	DBG(DBG_LPU, NULL, "lpu_init - LPU_LTSSM_CONTROL: 0x%llx\n",
745 	    CSR_XR(csr_base, LPU_LTSSM_CONTROL));
746 
747 	/*
748 	 * CSR_V LPU_LINK_STATUS Expect HW 0x101
749 	 */
750 
751 	/*
752 	 * This register has bits [9:4] for link width, and the
753 	 * default 0x10, means a width of x16.  The problem is
754 	 * this width is not supported according to the TLU
755 	 * link status register.
756 	 */
757 	DBG(DBG_LPU, NULL, "lpu_init - LPU_LINK_STATUS: 0x%llx\n",
758 	    CSR_XR(csr_base, LPU_LINK_STATUS));
759 
760 	/*
761 	 * CSR_V LPU_INTERRUPT_STATUS Expect HW 0x0
762 	 */
763 	DBG(DBG_LPU, NULL, "lpu_init - LPU_INTERRUPT_STATUS: 0x%llx\n",
764 	    CSR_XR(csr_base, LPU_INTERRUPT_STATUS));
765 
766 	/*
767 	 * CSR_V LPU_INTERRUPT_MASK Expect HW 0x0
768 	 */
769 	DBG(DBG_LPU, NULL, "lpu_init - LPU_INTERRUPT_MASK: 0x%llx\n",
770 	    CSR_XR(csr_base, LPU_INTERRUPT_MASK));
771 
772 	/*
773 	 * CSR_V LPU_LINK_PERFORMANCE_COUNTER_SELECT Expect HW 0x0
774 	 */
775 	DBG(DBG_LPU, NULL,
776 	    "lpu_init - LPU_LINK_PERFORMANCE_COUNTER_SELECT: 0x%llx\n",
777 	    CSR_XR(csr_base, LPU_LINK_PERFORMANCE_COUNTER_SELECT));
778 
779 	/*
780 	 * CSR_V LPU_LINK_PERFORMANCE_COUNTER_CONTROL Expect HW 0x0
781 	 */
782 	DBG(DBG_LPU, NULL,
783 	    "lpu_init - LPU_LINK_PERFORMANCE_COUNTER_CONTROL: 0x%llx\n",
784 	    CSR_XR(csr_base, LPU_LINK_PERFORMANCE_COUNTER_CONTROL));
785 
786 	/*
787 	 * CSR_V LPU_LINK_PERFORMANCE_COUNTER1 Expect HW 0x0
788 	 */
789 	DBG(DBG_LPU, NULL,
790 	    "lpu_init - LPU_LINK_PERFORMANCE_COUNTER1: 0x%llx\n",
791 	    CSR_XR(csr_base, LPU_LINK_PERFORMANCE_COUNTER1));
792 
793 	/*
794 	 * CSR_V LPU_LINK_PERFORMANCE_COUNTER1_TEST Expect HW 0x0
795 	 */
796 	DBG(DBG_LPU, NULL,
797 	    "lpu_init - LPU_LINK_PERFORMANCE_COUNTER1_TEST: 0x%llx\n",
798 	    CSR_XR(csr_base, LPU_LINK_PERFORMANCE_COUNTER1_TEST));
799 
800 	/*
801 	 * CSR_V LPU_LINK_PERFORMANCE_COUNTER2 Expect HW 0x0
802 	 */
803 	DBG(DBG_LPU, NULL,
804 	    "lpu_init - LPU_LINK_PERFORMANCE_COUNTER2: 0x%llx\n",
805 	    CSR_XR(csr_base, LPU_LINK_PERFORMANCE_COUNTER2));
806 
807 	/*
808 	 * CSR_V LPU_LINK_PERFORMANCE_COUNTER2_TEST Expect HW 0x0
809 	 */
810 	DBG(DBG_LPU, NULL,
811 	    "lpu_init - LPU_LINK_PERFORMANCE_COUNTER2_TEST: 0x%llx\n",
812 	    CSR_XR(csr_base, LPU_LINK_PERFORMANCE_COUNTER2_TEST));
813 
814 	/*
815 	 * CSR_V LPU_LINK_LAYER_CONFIG Expect HW 0x100
816 	 */
817 
818 	/*
819 	 * This is another place where Max Payload can be set,
820 	 * this time for the link layer.  It will be set to
821 	 * 128B, which is the default, but this will need to
822 	 * be revisited.
823 	 */
824 	val = (1ull << LPU_LINK_LAYER_CONFIG_VC0_EN);
825 	CSR_XS(csr_base, LPU_LINK_LAYER_CONFIG, val);
826 	DBG(DBG_LPU, NULL, "lpu_init - LPU_LINK_LAYER_CONFIG: 0x%llx\n",
827 	    CSR_XR(csr_base, LPU_LINK_LAYER_CONFIG));
828 
829 	/*
830 	 * CSR_V LPU_LINK_LAYER_STATUS Expect OBP 0x5
831 	 */
832 
833 	/*
834 	 * Another R/W status register.  Bit 3, DL up Status, will
835 	 * be set high.  The link state machine status bits [2:0]
836 	 * are set to 0x1, but the status bits are not defined in the
837 	 * PRM.  What does 0x1 mean, what others values are possible
838 	 * and what are thier meanings?
839 	 *
840 	 * This register has been giving us problems in simulation.
841 	 * It has been mentioned that software should not program
842 	 * any registers with WE bits except during debug.  So
843 	 * this register will no longer be programmed.
844 	 */
845 
846 	DBG(DBG_LPU, NULL, "lpu_init - LPU_LINK_LAYER_STATUS: 0x%llx\n",
847 	    CSR_XR(csr_base, LPU_LINK_LAYER_STATUS));
848 
849 	/*
850 	 * CSR_V LPU_LINK_LAYER_INTERRUPT_AND_STATUS_TEST Expect HW 0x0
851 	 */
852 	DBG(DBG_LPU, NULL,
853 	    "lpu_init - LPU_LINK_LAYER_INTERRUPT_AND_STATUS_TEST: 0x%llx\n",
854 	    CSR_XR(csr_base, LPU_LINK_LAYER_INTERRUPT_AND_STATUS_TEST));
855 
856 	/*
857 	 * CSR_V LPU Link Layer interrupt regs (mask, status)
858 	 */
859 	DBG(DBG_LPU, NULL,
860 	    "lpu_init - LPU_LINK_LAYER_INTERRUPT_MASK: 0x%llx\n",
861 	    CSR_XR(csr_base, LPU_LINK_LAYER_INTERRUPT_MASK));
862 
863 	DBG(DBG_LPU, NULL,
864 	    "lpu_init - LPU_LINK_LAYER_INTERRUPT_AND_STATUS: 0x%llx\n",
865 	    CSR_XR(csr_base, LPU_LINK_LAYER_INTERRUPT_AND_STATUS));
866 
867 	/*
868 	 * CSR_V LPU_FLOW_CONTROL_UPDATE_CONTROL Expect OBP 0x7
869 	 */
870 
871 	/*
872 	 * The PRM says that only the first two bits will be set
873 	 * high by default, which will enable flow control for
874 	 * posted and non-posted updates, but NOT completetion
875 	 * updates.
876 	 */
877 	val = (1ull << LPU_FLOW_CONTROL_UPDATE_CONTROL_FC0_U_NP_EN) |
878 	    (1ull << LPU_FLOW_CONTROL_UPDATE_CONTROL_FC0_U_P_EN);
879 	CSR_XS(csr_base, LPU_FLOW_CONTROL_UPDATE_CONTROL, val);
880 	DBG(DBG_LPU, NULL,
881 	    "lpu_init - LPU_FLOW_CONTROL_UPDATE_CONTROL: 0x%llx\n",
882 	    CSR_XR(csr_base, LPU_FLOW_CONTROL_UPDATE_CONTROL));
883 
884 	/*
885 	 * CSR_V LPU_LINK_LAYER_FLOW_CONTROL_UPDATE_TIMEOUT_VALUE
886 	 * Expect OBP 0x1D4C
887 	 */
888 
889 	/*
890 	 * This should be set by OBP.  We'll check to make sure.
891 	 */
892 	DBG(DBG_LPU, NULL, "lpu_init - "
893 	    "LPU_LINK_LAYER_FLOW_CONTROL_UPDATE_TIMEOUT_VALUE: 0x%llx\n",
894 	    CSR_XR(csr_base,
895 	    LPU_LINK_LAYER_FLOW_CONTROL_UPDATE_TIMEOUT_VALUE));
896 
897 	/*
898 	 * CSR_V LPU_LINK_LAYER_VC0_FLOW_CONTROL_UPDATE_TIMER0 Expect OBP ???
899 	 */
900 
901 	/*
902 	 * This register has Flow Control Update Timer values for
903 	 * non-posted and posted requests, bits [30:16] and bits
904 	 * [14:0], respectively.  These are read-only to SW so
905 	 * either HW or OBP needs to set them.
906 	 */
907 	DBG(DBG_LPU, NULL, "lpu_init - "
908 	    "LPU_LINK_LAYER_VC0_FLOW_CONTROL_UPDATE_TIMER0: 0x%llx\n",
909 	    CSR_XR(csr_base,
910 	    LPU_LINK_LAYER_VC0_FLOW_CONTROL_UPDATE_TIMER0));
911 
912 	/*
913 	 * CSR_V LPU_LINK_LAYER_VC0_FLOW_CONTROL_UPDATE_TIMER1 Expect OBP ???
914 	 */
915 
916 	/*
917 	 * Same as timer0 register above, except for bits [14:0]
918 	 * have the timer values for completetions.  Read-only to
919 	 * SW; OBP or HW need to set it.
920 	 */
921 	DBG(DBG_LPU, NULL, "lpu_init - "
922 	    "LPU_LINK_LAYER_VC0_FLOW_CONTROL_UPDATE_TIMER1: 0x%llx\n",
923 	    CSR_XR(csr_base,
924 	    LPU_LINK_LAYER_VC0_FLOW_CONTROL_UPDATE_TIMER1));
925 
926 	/*
927 	 * CSR_V LPU_TXLINK_FREQUENT_NAK_LATENCY_TIMER_THRESHOLD
928 	 */
929 	val = acknak_timer_table[max_payload][link_width];
930 	CSR_XS(csr_base, LPU_TXLINK_FREQUENT_NAK_LATENCY_TIMER_THRESHOLD, val);
931 
932 	DBG(DBG_LPU, NULL, "lpu_init - "
933 	    "LPU_TXLINK_FREQUENT_NAK_LATENCY_TIMER_THRESHOLD: 0x%llx\n",
934 	    CSR_XR(csr_base, LPU_TXLINK_FREQUENT_NAK_LATENCY_TIMER_THRESHOLD));
935 
936 	/*
937 	 * CSR_V LPU_TXLINK_ACKNAK_LATENCY_TIMER Expect HW 0x0
938 	 */
939 	DBG(DBG_LPU, NULL,
940 	    "lpu_init - LPU_TXLINK_ACKNAK_LATENCY_TIMER: 0x%llx\n",
941 	    CSR_XR(csr_base, LPU_TXLINK_ACKNAK_LATENCY_TIMER));
942 
943 	/*
944 	 * CSR_V LPU_TXLINK_REPLAY_TIMER_THRESHOLD
945 	 */
946 	val = replay_timer_table[max_payload][link_width];
947 	CSR_XS(csr_base, LPU_TXLINK_REPLAY_TIMER_THRESHOLD, val);
948 
949 	DBG(DBG_LPU, NULL,
950 	    "lpu_init - LPU_TXLINK_REPLAY_TIMER_THRESHOLD: 0x%llx\n",
951 	    CSR_XR(csr_base, LPU_TXLINK_REPLAY_TIMER_THRESHOLD));
952 
953 	/*
954 	 * CSR_V LPU_TXLINK_REPLAY_TIMER Expect HW 0x0
955 	 */
956 	DBG(DBG_LPU, NULL, "lpu_init - LPU_TXLINK_REPLAY_TIMER: 0x%llx\n",
957 	    CSR_XR(csr_base, LPU_TXLINK_REPLAY_TIMER));
958 
959 	/*
960 	 * CSR_V LPU_TXLINK_REPLAY_NUMBER_STATUS Expect OBP 0x3
961 	 */
962 	DBG(DBG_LPU, NULL,
963 	    "lpu_init - LPU_TXLINK_REPLAY_NUMBER_STATUS: 0x%llx\n",
964 	    CSR_XR(csr_base, LPU_TXLINK_REPLAY_NUMBER_STATUS));
965 
966 	/*
967 	 * CSR_V LPU_REPLAY_BUFFER_MAX_ADDRESS Expect OBP 0xB3F
968 	 */
969 	DBG(DBG_LPU, NULL,
970 	    "lpu_init - LPU_REPLAY_BUFFER_MAX_ADDRESS: 0x%llx\n",
971 	    CSR_XR(csr_base, LPU_REPLAY_BUFFER_MAX_ADDRESS));
972 
973 	/*
974 	 * CSR_V LPU_TXLINK_RETRY_FIFO_POINTER Expect OBP 0xFFFF0000
975 	 */
976 	val = ((LPU_TXLINK_RETRY_FIFO_POINTER_RTRY_FIFO_TLPTR_DEFAULT <<
977 	    LPU_TXLINK_RETRY_FIFO_POINTER_RTRY_FIFO_TLPTR) |
978 	    (LPU_TXLINK_RETRY_FIFO_POINTER_RTRY_FIFO_HDPTR_DEFAULT <<
979 	    LPU_TXLINK_RETRY_FIFO_POINTER_RTRY_FIFO_HDPTR));
980 
981 	CSR_XS(csr_base, LPU_TXLINK_RETRY_FIFO_POINTER, val);
982 	DBG(DBG_LPU, NULL,
983 	    "lpu_init - LPU_TXLINK_RETRY_FIFO_POINTER: 0x%llx\n",
984 	    CSR_XR(csr_base, LPU_TXLINK_RETRY_FIFO_POINTER));
985 
986 	/*
987 	 * CSR_V LPU_TXLINK_RETRY_FIFO_R_W_POINTER Expect OBP 0x0
988 	 */
989 	DBG(DBG_LPU, NULL,
990 	    "lpu_init - LPU_TXLINK_RETRY_FIFO_R_W_POINTER: 0x%llx\n",
991 	    CSR_XR(csr_base, LPU_TXLINK_RETRY_FIFO_R_W_POINTER));
992 
993 	/*
994 	 * CSR_V LPU_TXLINK_RETRY_FIFO_CREDIT Expect HW 0x1580
995 	 */
996 	DBG(DBG_LPU, NULL,
997 	    "lpu_init - LPU_TXLINK_RETRY_FIFO_CREDIT: 0x%llx\n",
998 	    CSR_XR(csr_base, LPU_TXLINK_RETRY_FIFO_CREDIT));
999 
1000 	/*
1001 	 * CSR_V LPU_TXLINK_SEQUENCE_COUNTER Expect OBP 0xFFF0000
1002 	 */
1003 	DBG(DBG_LPU, NULL, "lpu_init - LPU_TXLINK_SEQUENCE_COUNTER: 0x%llx\n",
1004 	    CSR_XR(csr_base, LPU_TXLINK_SEQUENCE_COUNTER));
1005 
1006 	/*
1007 	 * CSR_V LPU_TXLINK_ACK_SENT_SEQUENCE_NUMBER Expect HW 0xFFF
1008 	 */
1009 	DBG(DBG_LPU, NULL,
1010 	    "lpu_init - LPU_TXLINK_ACK_SENT_SEQUENCE_NUMBER: 0x%llx\n",
1011 	    CSR_XR(csr_base, LPU_TXLINK_ACK_SENT_SEQUENCE_NUMBER));
1012 
1013 	/*
1014 	 * CSR_V LPU_TXLINK_SEQUENCE_COUNT_FIFO_MAX_ADDR Expect OBP 0x157
1015 	 */
1016 
1017 	/*
1018 	 * Test only register.  Will not be programmed.
1019 	 */
1020 	DBG(DBG_LPU, NULL,
1021 	    "lpu_init - LPU_TXLINK_SEQUENCE_COUNT_FIFO_MAX_ADDR: 0x%llx\n",
1022 	    CSR_XR(csr_base, LPU_TXLINK_SEQUENCE_COUNT_FIFO_MAX_ADDR));
1023 
1024 	/*
1025 	 * CSR_V LPU_TXLINK_SEQUENCE_COUNT_FIFO_POINTERS Expect HW 0xFFF0000
1026 	 */
1027 
1028 	/*
1029 	 * Test only register.  Will not be programmed.
1030 	 */
1031 	DBG(DBG_LPU, NULL,
1032 	    "lpu_init - LPU_TXLINK_SEQUENCE_COUNT_FIFO_POINTERS: 0x%llx\n",
1033 	    CSR_XR(csr_base, LPU_TXLINK_SEQUENCE_COUNT_FIFO_POINTERS));
1034 
1035 	/*
1036 	 * CSR_V LPU_TXLINK_SEQUENCE_COUNT_R_W_POINTERS Expect HW 0x0
1037 	 */
1038 	DBG(DBG_LPU, NULL,
1039 	    "lpu_init - LPU_TXLINK_SEQUENCE_COUNT_R_W_POINTERS: 0x%llx\n",
1040 	    CSR_XR(csr_base, LPU_TXLINK_SEQUENCE_COUNT_R_W_POINTERS));
1041 
1042 	/*
1043 	 * CSR_V LPU_TXLINK_TEST_CONTROL Expect HW 0x0
1044 	 */
1045 	DBG(DBG_LPU, NULL, "lpu_init - LPU_TXLINK_TEST_CONTROL: 0x%llx\n",
1046 	    CSR_XR(csr_base, LPU_TXLINK_TEST_CONTROL));
1047 
1048 	/*
1049 	 * CSR_V LPU_TXLINK_MEMORY_ADDRESS_CONTROL Expect HW 0x0
1050 	 */
1051 
1052 	/*
1053 	 * Test only register.  Will not be programmed.
1054 	 */
1055 	DBG(DBG_LPU, NULL,
1056 	    "lpu_init - LPU_TXLINK_MEMORY_ADDRESS_CONTROL: 0x%llx\n",
1057 	    CSR_XR(csr_base, LPU_TXLINK_MEMORY_ADDRESS_CONTROL));
1058 
1059 	/*
1060 	 * CSR_V LPU_TXLINK_MEMORY_DATA_LOAD0 Expect HW 0x0
1061 	 */
1062 	DBG(DBG_LPU, NULL,
1063 	    "lpu_init - LPU_TXLINK_MEMORY_DATA_LOAD0: 0x%llx\n",
1064 	    CSR_XR(csr_base, LPU_TXLINK_MEMORY_DATA_LOAD0));
1065 
1066 	/*
1067 	 * CSR_V LPU_TXLINK_MEMORY_DATA_LOAD1 Expect HW 0x0
1068 	 */
1069 	DBG(DBG_LPU, NULL,
1070 	    "lpu_init - LPU_TXLINK_MEMORY_DATA_LOAD1: 0x%llx\n",
1071 	    CSR_XR(csr_base, LPU_TXLINK_MEMORY_DATA_LOAD1));
1072 
1073 	/*
1074 	 * CSR_V LPU_TXLINK_MEMORY_DATA_LOAD2 Expect HW 0x0
1075 	 */
1076 	DBG(DBG_LPU, NULL,
1077 	    "lpu_init - LPU_TXLINK_MEMORY_DATA_LOAD2: 0x%llx\n",
1078 	    CSR_XR(csr_base, LPU_TXLINK_MEMORY_DATA_LOAD2));
1079 
1080 	/*
1081 	 * CSR_V LPU_TXLINK_MEMORY_DATA_LOAD3 Expect HW 0x0
1082 	 */
1083 	DBG(DBG_LPU, NULL,
1084 	    "lpu_init - LPU_TXLINK_MEMORY_DATA_LOAD3: 0x%llx\n",
1085 	    CSR_XR(csr_base, LPU_TXLINK_MEMORY_DATA_LOAD3));
1086 
1087 	/*
1088 	 * CSR_V LPU_TXLINK_MEMORY_DATA_LOAD4 Expect HW 0x0
1089 	 */
1090 	DBG(DBG_LPU, NULL,
1091 	    "lpu_init - LPU_TXLINK_MEMORY_DATA_LOAD4: 0x%llx\n",
1092 	    CSR_XR(csr_base, LPU_TXLINK_MEMORY_DATA_LOAD4));
1093 
1094 	/*
1095 	 * CSR_V LPU_TXLINK_RETRY_DATA_COUNT Expect HW 0x0
1096 	 */
1097 
1098 	/*
1099 	 * Test only register.  Will not be programmed.
1100 	 */
1101 	DBG(DBG_LPU, NULL, "lpu_init - LPU_TXLINK_RETRY_DATA_COUNT: 0x%llx\n",
1102 	    CSR_XR(csr_base, LPU_TXLINK_RETRY_DATA_COUNT));
1103 
1104 	/*
1105 	 * CSR_V LPU_TXLINK_SEQUENCE_BUFFER_COUNT Expect HW 0x0
1106 	 */
1107 
1108 	/*
1109 	 * Test only register.  Will not be programmed.
1110 	 */
1111 	DBG(DBG_LPU, NULL,
1112 	    "lpu_init - LPU_TXLINK_SEQUENCE_BUFFER_COUNT: 0x%llx\n",
1113 	    CSR_XR(csr_base, LPU_TXLINK_SEQUENCE_BUFFER_COUNT));
1114 
1115 	/*
1116 	 * CSR_V LPU_TXLINK_SEQUENCE_BUFFER_BOTTOM_DATA Expect HW 0x0
1117 	 */
1118 
1119 	/*
1120 	 * Test only register.
1121 	 */
1122 	DBG(DBG_LPU, NULL,
1123 	    "lpu_init - LPU_TXLINK_SEQUENCE_BUFFER_BOTTOM_DATA: 0x%llx\n",
1124 	    CSR_XR(csr_base, LPU_TXLINK_SEQUENCE_BUFFER_BOTTOM_DATA));
1125 
1126 	/*
1127 	 * CSR_V LPU_RXLINK_NEXT_RECEIVE_SEQUENCE_1_COUNTER Expect HW 0x0
1128 	 */
1129 	DBG(DBG_LPU, NULL, "lpu_init - "
1130 	    "LPU_RXLINK_NEXT_RECEIVE_SEQUENCE_1_COUNTER: 0x%llx\n",
1131 	    CSR_XR(csr_base, LPU_RXLINK_NEXT_RECEIVE_SEQUENCE_1_COUNTER));
1132 
1133 	/*
1134 	 * CSR_V LPU_RXLINK_UNSUPPORTED_DLLP_RECEIVED Expect HW 0x0
1135 	 */
1136 
1137 	/*
1138 	 * test only register.
1139 	 */
1140 	DBG(DBG_LPU, NULL,
1141 	    "lpu_init - LPU_RXLINK_UNSUPPORTED_DLLP_RECEIVED: 0x%llx\n",
1142 	    CSR_XR(csr_base, LPU_RXLINK_UNSUPPORTED_DLLP_RECEIVED));
1143 
1144 	/*
1145 	 * CSR_V LPU_RXLINK_TEST_CONTROL Expect HW 0x0
1146 	 */
1147 
1148 	/*
1149 	 * test only register.
1150 	 */
1151 	DBG(DBG_LPU, NULL, "lpu_init - LPU_RXLINK_TEST_CONTROL: 0x%llx\n",
1152 	    CSR_XR(csr_base, LPU_RXLINK_TEST_CONTROL));
1153 
1154 	/*
1155 	 * CSR_V LPU_PHYSICAL_LAYER_CONFIGURATION Expect HW 0x10
1156 	 */
1157 	DBG(DBG_LPU, NULL,
1158 	    "lpu_init - LPU_PHYSICAL_LAYER_CONFIGURATION: 0x%llx\n",
1159 	    CSR_XR(csr_base, LPU_PHYSICAL_LAYER_CONFIGURATION));
1160 
1161 	/*
1162 	 * CSR_V LPU_PHY_LAYER_STATUS Expect HW 0x0
1163 	 */
1164 	DBG(DBG_LPU, NULL, "lpu_init - LPU_PHY_LAYER_STATUS: 0x%llx\n",
1165 	    CSR_XR(csr_base, LPU_PHY_LAYER_STATUS));
1166 
1167 	/*
1168 	 * CSR_V LPU_PHY_INTERRUPT_AND_STATUS_TEST Expect HW 0x0
1169 	 */
1170 	DBG(DBG_LPU, NULL,
1171 	    "lpu_init - LPU_PHY_INTERRUPT_AND_STATUS_TEST: 0x%llx\n",
1172 	    CSR_XR(csr_base, LPU_PHY_INTERRUPT_AND_STATUS_TEST));
1173 
1174 	/*
1175 	 * CSR_V LPU PHY LAYER interrupt regs (mask, status)
1176 	 */
1177 	DBG(DBG_LPU, NULL, "lpu_init - LPU_PHY_INTERRUPT_MASK: 0x%llx\n",
1178 	    CSR_XR(csr_base, LPU_PHY_INTERRUPT_MASK));
1179 
1180 	DBG(DBG_LPU, NULL,
1181 	    "lpu_init - LPU_PHY_LAYER_INTERRUPT_AND_STATUS: 0x%llx\n",
1182 	    CSR_XR(csr_base, LPU_PHY_LAYER_INTERRUPT_AND_STATUS));
1183 
1184 	/*
1185 	 * CSR_V LPU_RECEIVE_PHY_CONFIG Expect HW 0x0
1186 	 */
1187 
1188 	/*
1189 	 * This also needs some explanation.  What is the best value
1190 	 * for the water mark?  Test mode enables which test mode?
1191 	 * Programming model needed for the Receiver Reset Lane N
1192 	 * bits.
1193 	 */
1194 	DBG(DBG_LPU, NULL, "lpu_init - LPU_RECEIVE_PHY_CONFIG: 0x%llx\n",
1195 	    CSR_XR(csr_base, LPU_RECEIVE_PHY_CONFIG));
1196 
1197 	/*
1198 	 * CSR_V LPU_RECEIVE_PHY_STATUS1 Expect HW 0x0
1199 	 */
1200 	DBG(DBG_LPU, NULL, "lpu_init - LPU_RECEIVE_PHY_STATUS1: 0x%llx\n",
1201 	    CSR_XR(csr_base, LPU_RECEIVE_PHY_STATUS1));
1202 
1203 	/*
1204 	 * CSR_V LPU_RECEIVE_PHY_STATUS2 Expect HW 0x0
1205 	 */
1206 	DBG(DBG_LPU, NULL, "lpu_init - LPU_RECEIVE_PHY_STATUS2: 0x%llx\n",
1207 	    CSR_XR(csr_base, LPU_RECEIVE_PHY_STATUS2));
1208 
1209 	/*
1210 	 * CSR_V LPU_RECEIVE_PHY_STATUS3 Expect HW 0x0
1211 	 */
1212 	DBG(DBG_LPU, NULL, "lpu_init - LPU_RECEIVE_PHY_STATUS3: 0x%llx\n",
1213 	    CSR_XR(csr_base, LPU_RECEIVE_PHY_STATUS3));
1214 
1215 	/*
1216 	 * CSR_V LPU_RECEIVE_PHY_INTERRUPT_AND_STATUS_TEST Expect HW 0x0
1217 	 */
1218 	DBG(DBG_LPU, NULL,
1219 	    "lpu_init - LPU_RECEIVE_PHY_INTERRUPT_AND_STATUS_TEST: 0x%llx\n",
1220 	    CSR_XR(csr_base, LPU_RECEIVE_PHY_INTERRUPT_AND_STATUS_TEST));
1221 
1222 	/*
1223 	 * CSR_V LPU RX LAYER interrupt regs (mask, status)
1224 	 */
1225 	DBG(DBG_LPU, NULL,
1226 	    "lpu_init - LPU_RECEIVE_PHY_INTERRUPT_MASK: 0x%llx\n",
1227 	    CSR_XR(csr_base, LPU_RECEIVE_PHY_INTERRUPT_MASK));
1228 
1229 	DBG(DBG_LPU, NULL,
1230 	    "lpu_init - LPU_RECEIVE_PHY_INTERRUPT_AND_STATUS: 0x%llx\n",
1231 	    CSR_XR(csr_base, LPU_RECEIVE_PHY_INTERRUPT_AND_STATUS));
1232 
1233 	/*
1234 	 * CSR_V LPU_TRANSMIT_PHY_CONFIG Expect HW 0x0
1235 	 */
1236 	DBG(DBG_LPU, NULL, "lpu_init - LPU_TRANSMIT_PHY_CONFIG: 0x%llx\n",
1237 	    CSR_XR(csr_base, LPU_TRANSMIT_PHY_CONFIG));
1238 
1239 	/*
1240 	 * CSR_V LPU_TRANSMIT_PHY_STATUS Expect HW 0x0
1241 	 */
1242 	DBG(DBG_LPU, NULL, "lpu_init - LPU_TRANSMIT_PHY_STATUS: 0x%llx\n",
1243 		CSR_XR(csr_base, LPU_TRANSMIT_PHY_STATUS));
1244 
1245 	/*
1246 	 * CSR_V LPU_TRANSMIT_PHY_INTERRUPT_AND_STATUS_TEST Expect HW 0x0
1247 	 */
1248 	DBG(DBG_LPU, NULL,
1249 	    "lpu_init - LPU_TRANSMIT_PHY_INTERRUPT_AND_STATUS_TEST: 0x%llx\n",
1250 	    CSR_XR(csr_base,
1251 	    LPU_TRANSMIT_PHY_INTERRUPT_AND_STATUS_TEST));
1252 
1253 	/*
1254 	 * CSR_V LPU TX LAYER interrupt regs (mask, status)
1255 	 */
1256 	DBG(DBG_LPU, NULL,
1257 	    "lpu_init - LPU_TRANSMIT_PHY_INTERRUPT_MASK: 0x%llx\n",
1258 	    CSR_XR(csr_base, LPU_TRANSMIT_PHY_INTERRUPT_MASK));
1259 
1260 	DBG(DBG_LPU, NULL,
1261 	    "lpu_init - LPU_TRANSMIT_PHY_INTERRUPT_AND_STATUS: 0x%llx\n",
1262 	    CSR_XR(csr_base, LPU_TRANSMIT_PHY_INTERRUPT_AND_STATUS));
1263 
1264 	/*
1265 	 * CSR_V LPU_TRANSMIT_PHY_STATUS_2 Expect HW 0x0
1266 	 */
1267 	DBG(DBG_LPU, NULL, "lpu_init - LPU_TRANSMIT_PHY_STATUS_2: 0x%llx\n",
1268 	    CSR_XR(csr_base, LPU_TRANSMIT_PHY_STATUS_2));
1269 
1270 	/*
1271 	 * CSR_V LPU_LTSSM_CONFIG1 Expect OBP 0x205
1272 	 */
1273 
1274 	/*
1275 	 * The new PRM has values for LTSSM 8 ns timeout value and
1276 	 * LTSSM 20 ns timeout value.  But what do these values mean?
1277 	 * Most of the other bits are questions as well.
1278 	 *
1279 	 * As such we will use the reset value.
1280 	 */
1281 	DBG(DBG_LPU, NULL, "lpu_init - LPU_LTSSM_CONFIG1: 0x%llx\n",
1282 	    CSR_XR(csr_base, LPU_LTSSM_CONFIG1));
1283 
1284 	/*
1285 	 * CSR_V LPU_LTSSM_CONFIG2 Expect OBP 0x2DC6C0
1286 	 */
1287 
1288 	/*
1289 	 * Again, what does '12 ms timeout value mean'?
1290 	 */
1291 	val = (LPU_LTSSM_CONFIG2_LTSSM_12_TO_DEFAULT <<
1292 	    LPU_LTSSM_CONFIG2_LTSSM_12_TO);
1293 	CSR_XS(csr_base, LPU_LTSSM_CONFIG2, val);
1294 	DBG(DBG_LPU, NULL, "lpu_init - LPU_LTSSM_CONFIG2: 0x%llx\n",
1295 	    CSR_XR(csr_base, LPU_LTSSM_CONFIG2));
1296 
1297 	/*
1298 	 * CSR_V LPU_LTSSM_CONFIG3 Expect OBP 0x7A120
1299 	 */
1300 	val = (LPU_LTSSM_CONFIG3_LTSSM_2_TO_DEFAULT <<
1301 	    LPU_LTSSM_CONFIG3_LTSSM_2_TO);
1302 	CSR_XS(csr_base, LPU_LTSSM_CONFIG3, val);
1303 	DBG(DBG_LPU, NULL, "lpu_init - LPU_LTSSM_CONFIG3: 0x%llx\n",
1304 	    CSR_XR(csr_base, LPU_LTSSM_CONFIG3));
1305 
1306 	/*
1307 	 * CSR_V LPU_LTSSM_CONFIG4 Expect OBP 0x21300
1308 	 */
1309 	val = ((LPU_LTSSM_CONFIG4_DATA_RATE_DEFAULT <<
1310 	    LPU_LTSSM_CONFIG4_DATA_RATE) |
1311 		(LPU_LTSSM_CONFIG4_N_FTS_DEFAULT <<
1312 		LPU_LTSSM_CONFIG4_N_FTS));
1313 	CSR_XS(csr_base, LPU_LTSSM_CONFIG4, val);
1314 	DBG(DBG_LPU, NULL, "lpu_init - LPU_LTSSM_CONFIG4: 0x%llx\n",
1315 	    CSR_XR(csr_base, LPU_LTSSM_CONFIG4));
1316 
1317 	/*
1318 	 * CSR_V LPU_LTSSM_CONFIG5 Expect OBP 0x0
1319 	 */
1320 	val = 0ull;
1321 	CSR_XS(csr_base, LPU_LTSSM_CONFIG5, val);
1322 	DBG(DBG_LPU, NULL, "lpu_init - LPU_LTSSM_CONFIG5: 0x%llx\n",
1323 	    CSR_XR(csr_base, LPU_LTSSM_CONFIG5));
1324 
1325 	/*
1326 	 * CSR_V LPU_LTSSM_STATUS1 Expect OBP 0x0
1327 	 */
1328 
1329 	/*
1330 	 * LTSSM Status registers are test only.
1331 	 */
1332 	DBG(DBG_LPU, NULL, "lpu_init - LPU_LTSSM_STATUS1: 0x%llx\n",
1333 	    CSR_XR(csr_base, LPU_LTSSM_STATUS1));
1334 
1335 	/*
1336 	 * CSR_V LPU_LTSSM_STATUS2 Expect OBP 0x0
1337 	 */
1338 	DBG(DBG_LPU, NULL, "lpu_init - LPU_LTSSM_STATUS2: 0x%llx\n",
1339 	    CSR_XR(csr_base, LPU_LTSSM_STATUS2));
1340 
1341 	/*
1342 	 * CSR_V LPU_LTSSM_INTERRUPT_AND_STATUS_TEST Expect HW 0x0
1343 	 */
1344 	DBG(DBG_LPU, NULL,
1345 	    "lpu_init - LPU_LTSSM_INTERRUPT_AND_STATUS_TEST: 0x%llx\n",
1346 	    CSR_XR(csr_base, LPU_LTSSM_INTERRUPT_AND_STATUS_TEST));
1347 
1348 	/*
1349 	 * CSR_V LPU LTSSM  LAYER interrupt regs (mask, status)
1350 	 */
1351 	DBG(DBG_LPU, NULL, "lpu_init - LPU_LTSSM_INTERRUPT_MASK: 0x%llx\n",
1352 	    CSR_XR(csr_base, LPU_LTSSM_INTERRUPT_MASK));
1353 
1354 	DBG(DBG_LPU, NULL,
1355 	    "lpu_init - LPU_LTSSM_INTERRUPT_AND_STATUS: 0x%llx\n",
1356 	    CSR_XR(csr_base, LPU_LTSSM_INTERRUPT_AND_STATUS));
1357 
1358 	/*
1359 	 * CSR_V LPU_LTSSM_STATUS_WRITE_ENABLE Expect OBP 0x0
1360 	 */
1361 	DBG(DBG_LPU, NULL,
1362 	    "lpu_init - LPU_LTSSM_STATUS_WRITE_ENABLE: 0x%llx\n",
1363 	    CSR_XR(csr_base, LPU_LTSSM_STATUS_WRITE_ENABLE));
1364 
1365 	/*
1366 	 * CSR_V LPU_GIGABLAZE_GLUE_CONFIG1 Expect OBP 0x88407
1367 	 */
1368 	DBG(DBG_LPU, NULL, "lpu_init - LPU_GIGABLAZE_GLUE_CONFIG1: 0x%llx\n",
1369 	    CSR_XR(csr_base, LPU_GIGABLAZE_GLUE_CONFIG1));
1370 
1371 	/*
1372 	 * CSR_V LPU_GIGABLAZE_GLUE_CONFIG2 Expect OBP 0x35
1373 	 */
1374 	DBG(DBG_LPU, NULL, "lpu_init - LPU_GIGABLAZE_GLUE_CONFIG2: 0x%llx\n",
1375 	    CSR_XR(csr_base, LPU_GIGABLAZE_GLUE_CONFIG2));
1376 
1377 	/*
1378 	 * CSR_V LPU_GIGABLAZE_GLUE_CONFIG3 Expect OBP 0x4400FA
1379 	 */
1380 	DBG(DBG_LPU, NULL, "lpu_init - LPU_GIGABLAZE_GLUE_CONFIG3: 0x%llx\n",
1381 	    CSR_XR(csr_base, LPU_GIGABLAZE_GLUE_CONFIG3));
1382 
1383 	/*
1384 	 * CSR_V LPU_GIGABLAZE_GLUE_CONFIG4 Expect OBP 0x1E848
1385 	 */
1386 	DBG(DBG_LPU, NULL, "lpu_init - LPU_GIGABLAZE_GLUE_CONFIG4: 0x%llx\n",
1387 	    CSR_XR(csr_base, LPU_GIGABLAZE_GLUE_CONFIG4));
1388 
1389 	/*
1390 	 * CSR_V LPU_GIGABLAZE_GLUE_STATUS Expect OBP 0x0
1391 	 */
1392 	DBG(DBG_LPU, NULL, "lpu_init - LPU_GIGABLAZE_GLUE_STATUS: 0x%llx\n",
1393 	    CSR_XR(csr_base, LPU_GIGABLAZE_GLUE_STATUS));
1394 
1395 	/*
1396 	 * CSR_V LPU_GIGABLAZE_GLUE_INTERRUPT_AND_STATUS_TEST Expect OBP 0x0
1397 	 */
1398 	DBG(DBG_LPU, NULL, "lpu_init - "
1399 	    "LPU_GIGABLAZE_GLUE_INTERRUPT_AND_STATUS_TEST: 0x%llx\n",
1400 	    CSR_XR(csr_base,
1401 	    LPU_GIGABLAZE_GLUE_INTERRUPT_AND_STATUS_TEST));
1402 
1403 	/*
1404 	 * CSR_V LPU GIGABLASE LAYER interrupt regs (mask, status)
1405 	 */
1406 	DBG(DBG_LPU, NULL,
1407 	    "lpu_init - LPU_GIGABLAZE_GLUE_INTERRUPT_MASK: 0x%llx\n",
1408 	    CSR_XR(csr_base, LPU_GIGABLAZE_GLUE_INTERRUPT_MASK));
1409 
1410 	DBG(DBG_LPU, NULL,
1411 	    "lpu_init - LPU_GIGABLAZE_GLUE_INTERRUPT_AND_STATUS: 0x%llx\n",
1412 	    CSR_XR(csr_base, LPU_GIGABLAZE_GLUE_INTERRUPT_AND_STATUS));
1413 
1414 	/*
1415 	 * CSR_V LPU_GIGABLAZE_GLUE_POWER_DOWN1 Expect HW 0x0
1416 	 */
1417 	DBG(DBG_LPU, NULL,
1418 	    "lpu_init - LPU_GIGABLAZE_GLUE_POWER_DOWN1: 0x%llx\n",
1419 	    CSR_XR(csr_base, LPU_GIGABLAZE_GLUE_POWER_DOWN1));
1420 
1421 	/*
1422 	 * CSR_V LPU_GIGABLAZE_GLUE_POWER_DOWN2 Expect HW 0x0
1423 	 */
1424 	DBG(DBG_LPU, NULL,
1425 	    "lpu_init - LPU_GIGABLAZE_GLUE_POWER_DOWN2: 0x%llx\n",
1426 	    CSR_XR(csr_base, LPU_GIGABLAZE_GLUE_POWER_DOWN2));
1427 
1428 	/*
1429 	 * CSR_V LPU_GIGABLAZE_GLUE_CONFIG5 Expect OBP 0x0
1430 	 */
1431 	DBG(DBG_LPU, NULL, "lpu_init - LPU_GIGABLAZE_GLUE_CONFIG5: 0x%llx\n",
1432 	    CSR_XR(csr_base, LPU_GIGABLAZE_GLUE_CONFIG5));
1433 }
1434 
1435 /* ARGSUSED */
1436 static void
1437 dmc_init(caddr_t csr_base, pxu_t *pxu_p)
1438 {
1439 	uint64_t val;
1440 
1441 /*
1442  * CSR_V DMC_CORE_AND_BLOCK_INTERRUPT_ENABLE Expect OBP 0x8000000000000003
1443  */
1444 
1445 	val = -1ull;
1446 	CSR_XS(csr_base, DMC_CORE_AND_BLOCK_INTERRUPT_ENABLE, val);
1447 	DBG(DBG_DMC, NULL,
1448 	    "dmc_init - DMC_CORE_AND_BLOCK_INTERRUPT_ENABLE: 0x%llx\n",
1449 	    CSR_XR(csr_base, DMC_CORE_AND_BLOCK_INTERRUPT_ENABLE));
1450 
1451 	/*
1452 	 * CSR_V DMC_CORE_AND_BLOCK_ERROR_STATUS Expect HW 0x0
1453 	 */
1454 	DBG(DBG_DMC, NULL,
1455 	    "dmc_init - DMC_CORE_AND_BLOCK_ERROR_STATUS: 0x%llx\n",
1456 	    CSR_XR(csr_base, DMC_CORE_AND_BLOCK_ERROR_STATUS));
1457 
1458 	/*
1459 	 * CSR_V DMC_DEBUG_SELECT_FOR_PORT_A Expect HW 0x0
1460 	 */
1461 	val = 0x0ull;
1462 	CSR_XS(csr_base, DMC_DEBUG_SELECT_FOR_PORT_A, val);
1463 	DBG(DBG_DMC, NULL, "dmc_init - DMC_DEBUG_SELECT_FOR_PORT_A: 0x%llx\n",
1464 	    CSR_XR(csr_base, DMC_DEBUG_SELECT_FOR_PORT_A));
1465 
1466 	/*
1467 	 * CSR_V DMC_DEBUG_SELECT_FOR_PORT_B Expect HW 0x0
1468 	 */
1469 	val = 0x0ull;
1470 	CSR_XS(csr_base, DMC_DEBUG_SELECT_FOR_PORT_B, val);
1471 	DBG(DBG_DMC, NULL, "dmc_init - DMC_DEBUG_SELECT_FOR_PORT_B: 0x%llx\n",
1472 	    CSR_XR(csr_base, DMC_DEBUG_SELECT_FOR_PORT_B));
1473 }
1474 
1475 void
1476 hvio_pec_init(caddr_t csr_base, pxu_t *pxu_p)
1477 {
1478 	uint64_t val;
1479 
1480 	ilu_init(csr_base, pxu_p);
1481 	tlu_init(csr_base, pxu_p);
1482 	lpu_init(csr_base, pxu_p);
1483 	dmc_init(csr_base, pxu_p);
1484 
1485 /*
1486  * CSR_V PEC_CORE_AND_BLOCK_INTERRUPT_ENABLE Expect Kernel 0x800000000000000F
1487  */
1488 
1489 	val = -1ull;
1490 	CSR_XS(csr_base, PEC_CORE_AND_BLOCK_INTERRUPT_ENABLE, val);
1491 	DBG(DBG_PEC, NULL,
1492 	    "hvio_pec_init - PEC_CORE_AND_BLOCK_INTERRUPT_ENABLE: 0x%llx\n",
1493 	    CSR_XR(csr_base, PEC_CORE_AND_BLOCK_INTERRUPT_ENABLE));
1494 
1495 	/*
1496 	 * CSR_V PEC_CORE_AND_BLOCK_INTERRUPT_STATUS Expect HW 0x0
1497 	 */
1498 	DBG(DBG_PEC, NULL,
1499 	    "hvio_pec_init - PEC_CORE_AND_BLOCK_INTERRUPT_STATUS: 0x%llx\n",
1500 	    CSR_XR(csr_base, PEC_CORE_AND_BLOCK_INTERRUPT_STATUS));
1501 }
1502 
1503 /*
1504  * Initialize the module, but do not enable interrupts.
1505  */
1506 void
1507 hvio_mmu_init(caddr_t csr_base, pxu_t *pxu_p)
1508 {
1509 	uint64_t	val, i, tsb_ctrl, obp_tsb_pa, *base_tte_addr;
1510 	uint_t		obp_tsb_entries, obp_tsb_size;
1511 
1512 	bzero(pxu_p->tsb_vaddr, pxu_p->tsb_size);
1513 
1514 	/*
1515 	 * Preserve OBP's TSB
1516 	 */
1517 	val = CSR_XR(csr_base, MMU_TSB_CONTROL);
1518 
1519 	tsb_ctrl = CSR_XR(csr_base, MMU_TSB_CONTROL);
1520 
1521 	obp_tsb_pa = tsb_ctrl &  0x7FFFFFFE000;
1522 	obp_tsb_size = tsb_ctrl & 0xF;
1523 
1524 	obp_tsb_entries = MMU_TSBSIZE_TO_TSBENTRIES(obp_tsb_size);
1525 
1526 	base_tte_addr = pxu_p->tsb_vaddr +
1527 		((pxu_p->tsb_size >> 3) - obp_tsb_entries);
1528 
1529 	for (i = 0; i < obp_tsb_entries; i++) {
1530 		uint64_t tte = lddphys(obp_tsb_pa + i * 8);
1531 
1532 		if (!MMU_TTE_VALID(tte))
1533 			continue;
1534 
1535 		base_tte_addr[i] = tte;
1536 	}
1537 
1538 	/*
1539 	 * Invalidate the TLB through the diagnostic register.
1540 	 */
1541 
1542 	CSR_XS(csr_base, MMU_TTE_CACHE_INVALIDATE, -1ull);
1543 
1544 	/*
1545 	 * Configure the Fire MMU TSB Control Register.  Determine
1546 	 * the encoding for either 8KB pages (0) or 64KB pages (1).
1547 	 *
1548 	 * Write the most significant 30 bits of the TSB physical address
1549 	 * and the encoded TSB table size.
1550 	 */
1551 	for (i = 8; i && (pxu_p->tsb_size < (0x2000 << i)); i--);
1552 
1553 	val = (((((va_to_pa(pxu_p->tsb_vaddr)) >> 13) << 13) |
1554 	    ((MMU_PAGE_SHIFT == 13) ? 0 : 1) << 8) | i);
1555 
1556 	CSR_XS(csr_base, MMU_TSB_CONTROL, val);
1557 
1558 	/*
1559 	 * Enable the MMU, set the "TSB Cache Snoop Enable",
1560 	 * the "Cache Mode", the "Bypass Enable" and
1561 	 * the "Translation Enable" bits.
1562 	 */
1563 	val = CSR_XR(csr_base, MMU_CONTROL_AND_STATUS);
1564 	val |= ((1ull << MMU_CONTROL_AND_STATUS_SE)
1565 	    | (MMU_CONTROL_AND_STATUS_CM_MASK << MMU_CONTROL_AND_STATUS_CM)
1566 	    | (1ull << MMU_CONTROL_AND_STATUS_BE)
1567 	    | (1ull << MMU_CONTROL_AND_STATUS_TE));
1568 
1569 	CSR_XS(csr_base, MMU_CONTROL_AND_STATUS, val);
1570 
1571 	/*
1572 	 * Read the register here to ensure that the previous writes to
1573 	 * the Fire MMU registers have been flushed.  (Technically, this
1574 	 * is not entirely necessary here as we will likely do later reads
1575 	 * during Fire initialization, but it is a small price to pay for
1576 	 * more modular code.)
1577 	 */
1578 	(void) CSR_XR(csr_base, MMU_CONTROL_AND_STATUS);
1579 
1580 	/*
1581 	 * CSR_V TLU's UE interrupt regs (log, enable, status, clear)
1582 	 * Plus header logs
1583 	 */
1584 	DBG(DBG_MMU, NULL, "mmu_init - MMU_ERROR_LOG_ENABLE: 0x%llx\n",
1585 	    CSR_XR(csr_base, MMU_ERROR_LOG_ENABLE));
1586 
1587 	DBG(DBG_MMU, NULL, "mmu_init - MMU_INTERRUPT_ENABLE: 0x%llx\n",
1588 	    CSR_XR(csr_base, MMU_INTERRUPT_ENABLE));
1589 
1590 	DBG(DBG_MMU, NULL, "mmu_init - MMU_INTERRUPT_STATUS: 0x%llx\n",
1591 	    CSR_XR(csr_base, MMU_INTERRUPT_STATUS));
1592 
1593 	DBG(DBG_MMU, NULL, "mmu_init - MMU_ERROR_STATUS_CLEAR: 0x%llx\n",
1594 	    CSR_XR(csr_base, MMU_ERROR_STATUS_CLEAR));
1595 }
1596 
1597 /*
1598  * Generic IOMMU Servies
1599  */
1600 
1601 /* ARGSUSED */
1602 uint64_t
1603 hvio_iommu_map(devhandle_t dev_hdl, pxu_t *pxu_p, tsbid_t tsbid,
1604     pages_t pages, io_attributes_t io_attributes,
1605     void *addr, size_t pfn_index, int flag)
1606 {
1607 	tsbindex_t	tsb_index = PCI_TSBID_TO_TSBINDEX(tsbid);
1608 	uint64_t	attr = MMU_TTE_V;
1609 	int		i;
1610 
1611 	if (io_attributes & PCI_MAP_ATTR_WRITE)
1612 		attr |= MMU_TTE_W;
1613 
1614 	if (flag == MMU_MAP_MP) {
1615 		ddi_dma_impl_t  *mp = (ddi_dma_impl_t *)addr;
1616 
1617 		for (i = 0; i < pages; i++, pfn_index++, tsb_index++) {
1618 			px_iopfn_t	pfn = PX_GET_MP_PFN(mp, pfn_index);
1619 
1620 			pxu_p->tsb_vaddr[tsb_index] =
1621 			    MMU_PTOB(pfn) | attr;
1622 		}
1623 	} else {
1624 		caddr_t a = (caddr_t)addr;
1625 
1626 		for (i = 0; i < pages; i++, a += MMU_PAGE_SIZE, tsb_index++) {
1627 			px_iopfn_t pfn = hat_getpfnum(kas.a_hat, a);
1628 
1629 			pxu_p->tsb_vaddr[tsb_index] =
1630 			    MMU_PTOB(pfn) | attr;
1631 		}
1632 	}
1633 
1634 	return (H_EOK);
1635 }
1636 
1637 /* ARGSUSED */
1638 uint64_t
1639 hvio_iommu_demap(devhandle_t dev_hdl, pxu_t *pxu_p, tsbid_t tsbid,
1640     pages_t pages)
1641 {
1642 	tsbindex_t	tsb_index = PCI_TSBID_TO_TSBINDEX(tsbid);
1643 	int		i;
1644 
1645 	for (i = 0; i < pages; i++, tsb_index++) {
1646 		pxu_p->tsb_vaddr[tsb_index] = MMU_INVALID_TTE;
1647 	}
1648 
1649 	return (H_EOK);
1650 }
1651 
1652 /* ARGSUSED */
1653 uint64_t
1654 hvio_iommu_getmap(devhandle_t dev_hdl, pxu_t *pxu_p, tsbid_t tsbid,
1655     io_attributes_t *attributes_p, r_addr_t *r_addr_p)
1656 {
1657 	tsbindex_t	tsb_index = PCI_TSBID_TO_TSBINDEX(tsbid);
1658 	uint64_t	*tte_addr;
1659 	uint64_t	ret = H_EOK;
1660 
1661 	tte_addr = (uint64_t *)(pxu_p->tsb_vaddr) + tsb_index;
1662 
1663 	if (*tte_addr & MMU_TTE_V) {
1664 		*r_addr_p = MMU_TTETOPA(*tte_addr);
1665 		*attributes_p = (*tte_addr & MMU_TTE_W) ?
1666 		    PCI_MAP_ATTR_WRITE:PCI_MAP_ATTR_READ;
1667 	} else {
1668 		*r_addr_p = 0;
1669 		*attributes_p = 0;
1670 		ret = H_ENOMAP;
1671 	}
1672 
1673 	return (ret);
1674 }
1675 
1676 /* ARGSUSED */
1677 uint64_t
1678 hvio_iommu_getbypass(devhandle_t dev_hdl, r_addr_t ra,
1679     io_attributes_t io_attributes, io_addr_t *io_addr_p)
1680 {
1681 	uint64_t	pfn = MMU_BTOP(ra);
1682 
1683 	*io_addr_p = MMU_BYPASS_BASE | ra |
1684 	    (pf_is_memory(pfn) ? 0 : MMU_BYPASS_NONCACHE);
1685 
1686 	return (H_EOK);
1687 }
1688 
1689 /*
1690  * Generic IO Interrupt Servies
1691  */
1692 
1693 /*
1694  * Converts a device specific interrupt number given by the
1695  * arguments devhandle and devino into a system specific ino.
1696  */
1697 /* ARGSUSED */
1698 uint64_t
1699 hvio_intr_devino_to_sysino(devhandle_t dev_hdl, pxu_t *pxu_p, devino_t devino,
1700     sysino_t *sysino)
1701 {
1702 	if (devino > INTERRUPT_MAPPING_ENTRIES) {
1703 		DBG(DBG_IB, NULL, "ino %x is invalid\n", devino);
1704 		return (H_ENOINTR);
1705 	}
1706 
1707 	*sysino = DEVINO_TO_SYSINO(pxu_p->portid, devino);
1708 
1709 	return (H_EOK);
1710 }
1711 
1712 /*
1713  * Returns state in intr_valid_state if the interrupt defined by sysino
1714  * is valid (enabled) or not-valid (disabled).
1715  */
1716 uint64_t
1717 hvio_intr_getvalid(devhandle_t dev_hdl, sysino_t sysino,
1718     intr_valid_state_t *intr_valid_state)
1719 {
1720 	if (CSRA_BR((caddr_t)dev_hdl, INTERRUPT_MAPPING,
1721 	    SYSINO_TO_DEVINO(sysino), ENTRIES_V)) {
1722 		*intr_valid_state = INTR_VALID;
1723 	} else {
1724 		*intr_valid_state = INTR_NOTVALID;
1725 	}
1726 
1727 	return (H_EOK);
1728 }
1729 
1730 /*
1731  * Sets the 'valid' state of the interrupt defined by
1732  * the argument sysino to the state defined by the
1733  * argument intr_valid_state.
1734  */
1735 uint64_t
1736 hvio_intr_setvalid(devhandle_t dev_hdl, sysino_t sysino,
1737     intr_valid_state_t intr_valid_state)
1738 {
1739 	switch (intr_valid_state) {
1740 	case INTR_VALID:
1741 		CSRA_BS((caddr_t)dev_hdl, INTERRUPT_MAPPING,
1742 		    SYSINO_TO_DEVINO(sysino), ENTRIES_V);
1743 		break;
1744 	case INTR_NOTVALID:
1745 		CSRA_BC((caddr_t)dev_hdl, INTERRUPT_MAPPING,
1746 		    SYSINO_TO_DEVINO(sysino), ENTRIES_V);
1747 		break;
1748 	default:
1749 		return (EINVAL);
1750 	}
1751 
1752 	return (H_EOK);
1753 }
1754 
1755 /*
1756  * Returns the current state of the interrupt given by the sysino
1757  * argument.
1758  */
1759 uint64_t
1760 hvio_intr_getstate(devhandle_t dev_hdl, sysino_t sysino,
1761     intr_state_t *intr_state)
1762 {
1763 	intr_state_t state;
1764 
1765 	state = CSRA_FR((caddr_t)dev_hdl, INTERRUPT_CLEAR,
1766 	    SYSINO_TO_DEVINO(sysino), ENTRIES_INT_STATE);
1767 
1768 	switch (state) {
1769 	case INTERRUPT_IDLE_STATE:
1770 		*intr_state = INTR_IDLE_STATE;
1771 		break;
1772 	case INTERRUPT_RECEIVED_STATE:
1773 		*intr_state = INTR_RECEIVED_STATE;
1774 		break;
1775 	case INTERRUPT_PENDING_STATE:
1776 		*intr_state = INTR_DELIVERED_STATE;
1777 		break;
1778 	default:
1779 		return (EINVAL);
1780 	}
1781 
1782 	return (H_EOK);
1783 
1784 }
1785 
1786 /*
1787  * Sets the current state of the interrupt given by the sysino
1788  * argument to the value given in the argument intr_state.
1789  *
1790  * Note: Setting the state to INTR_IDLE clears any pending
1791  * interrupt for sysino.
1792  */
1793 uint64_t
1794 hvio_intr_setstate(devhandle_t dev_hdl, sysino_t sysino,
1795     intr_state_t intr_state)
1796 {
1797 	intr_state_t state;
1798 
1799 	switch (intr_state) {
1800 	case INTR_IDLE_STATE:
1801 		state = INTERRUPT_IDLE_STATE;
1802 		break;
1803 	case INTR_DELIVERED_STATE:
1804 		state = INTERRUPT_PENDING_STATE;
1805 		break;
1806 	default:
1807 		return (EINVAL);
1808 	}
1809 
1810 	CSRA_FS((caddr_t)dev_hdl, INTERRUPT_CLEAR,
1811 	    SYSINO_TO_DEVINO(sysino), ENTRIES_INT_STATE, state);
1812 
1813 	return (H_EOK);
1814 }
1815 
1816 /*
1817  * Returns the cpuid that is the current target of the
1818  * interrupt given by the sysino argument.
1819  *
1820  * The cpuid value returned is undefined if the target
1821  * has not been set via intr_settarget.
1822  */
1823 uint64_t
1824 hvio_intr_gettarget(devhandle_t dev_hdl, sysino_t sysino, cpuid_t *cpuid)
1825 {
1826 	*cpuid = CSRA_FR((caddr_t)dev_hdl, INTERRUPT_MAPPING,
1827 	    SYSINO_TO_DEVINO(sysino), ENTRIES_T_JPID);
1828 
1829 	return (H_EOK);
1830 }
1831 
1832 /*
1833  * Set the target cpu for the interrupt defined by the argument
1834  * sysino to the target cpu value defined by the argument cpuid.
1835  */
1836 uint64_t
1837 hvio_intr_settarget(devhandle_t dev_hdl, sysino_t sysino, cpuid_t cpuid)
1838 {
1839 
1840 	uint64_t	val, intr_controller;
1841 	uint32_t	ino = SYSINO_TO_DEVINO(sysino);
1842 
1843 	/*
1844 	 * For now, we assign interrupt controller in a round
1845 	 * robin fashion.  Later, we may need to come up with
1846 	 * a more efficient assignment algorithm.
1847 	 */
1848 	intr_controller = 0x1ull << (cpuid % 4);
1849 
1850 	val = (((cpuid & INTERRUPT_MAPPING_ENTRIES_T_JPID_MASK) <<
1851 	    INTERRUPT_MAPPING_ENTRIES_T_JPID) |
1852 	    ((intr_controller & INTERRUPT_MAPPING_ENTRIES_INT_CNTRL_NUM_MASK)
1853 	    << INTERRUPT_MAPPING_ENTRIES_INT_CNTRL_NUM));
1854 
1855 	/* For EQ interrupts, set DATA MONDO bit */
1856 	if ((ino >= PX_DEFAULT_MSIQ_1ST_DEVINO) &&
1857 	    (ino < (PX_DEFAULT_MSIQ_1ST_DEVINO + PX_DEFAULT_MSIQ_CNT)))
1858 		val |= (0x1ull << INTERRUPT_MAPPING_ENTRIES_MDO_MODE);
1859 
1860 	CSRA_XS((caddr_t)dev_hdl, INTERRUPT_MAPPING, ino, val);
1861 
1862 	return (H_EOK);
1863 }
1864 
1865 /*
1866  * MSIQ Functions:
1867  */
1868 uint64_t
1869 hvio_msiq_init(devhandle_t dev_hdl, pxu_t *pxu_p)
1870 {
1871 	CSRA_XS((caddr_t)dev_hdl, EVENT_QUEUE_BASE_ADDRESS, 0,
1872 	    (uint64_t)pxu_p->msiq_mapped_p);
1873 	DBG(DBG_IB, NULL,
1874 	    "hvio_msiq_init: EVENT_QUEUE_BASE_ADDRESS 0x%llx\n",
1875 	    CSR_XR((caddr_t)dev_hdl, EVENT_QUEUE_BASE_ADDRESS));
1876 
1877 	CSRA_XS((caddr_t)dev_hdl, INTERRUPT_MONDO_DATA_0, 0,
1878 	    (uint64_t)ID_TO_IGN(pxu_p->portid) << INO_BITS);
1879 	DBG(DBG_IB, NULL, "hvio_msiq_init: "
1880 	    "INTERRUPT_MONDO_DATA_0: 0x%llx\n",
1881 	    CSR_XR((caddr_t)dev_hdl, INTERRUPT_MONDO_DATA_0));
1882 
1883 	return (H_EOK);
1884 }
1885 
1886 uint64_t
1887 hvio_msiq_getvalid(devhandle_t dev_hdl, msiqid_t msiq_id,
1888     pci_msiq_valid_state_t *msiq_valid_state)
1889 {
1890 	uint32_t	eq_state;
1891 	uint64_t	ret = H_EOK;
1892 
1893 	eq_state = CSRA_FR((caddr_t)dev_hdl, EVENT_QUEUE_STATE,
1894 	    msiq_id, ENTRIES_STATE);
1895 
1896 	switch (eq_state) {
1897 	case EQ_IDLE_STATE:
1898 		*msiq_valid_state = PCI_MSIQ_INVALID;
1899 		break;
1900 	case EQ_ACTIVE_STATE:
1901 	case EQ_ERROR_STATE:
1902 		*msiq_valid_state = PCI_MSIQ_VALID;
1903 		break;
1904 	default:
1905 		ret = H_EIO;
1906 		break;
1907 	}
1908 
1909 	return (ret);
1910 }
1911 
1912 uint64_t
1913 hvio_msiq_setvalid(devhandle_t dev_hdl, msiqid_t msiq_id,
1914     pci_msiq_valid_state_t msiq_valid_state)
1915 {
1916 	uint64_t	ret = H_EOK;
1917 
1918 	switch (msiq_valid_state) {
1919 	case PCI_MSIQ_INVALID:
1920 		CSRA_BS((caddr_t)dev_hdl, EVENT_QUEUE_CONTROL_CLEAR,
1921 		    msiq_id, ENTRIES_DIS);
1922 		break;
1923 	case PCI_MSIQ_VALID:
1924 		CSRA_BS((caddr_t)dev_hdl, EVENT_QUEUE_CONTROL_SET,
1925 		    msiq_id, ENTRIES_EN);
1926 		break;
1927 	default:
1928 		ret = H_EINVAL;
1929 		break;
1930 	}
1931 
1932 	return (ret);
1933 }
1934 
1935 uint64_t
1936 hvio_msiq_getstate(devhandle_t dev_hdl, msiqid_t msiq_id,
1937     pci_msiq_state_t *msiq_state)
1938 {
1939 	uint32_t	eq_state;
1940 	uint64_t	ret = H_EOK;
1941 
1942 	eq_state = CSRA_FR((caddr_t)dev_hdl, EVENT_QUEUE_STATE,
1943 	    msiq_id, ENTRIES_STATE);
1944 
1945 	switch (eq_state) {
1946 	case EQ_IDLE_STATE:
1947 	case EQ_ACTIVE_STATE:
1948 		*msiq_state = PCI_MSIQ_STATE_IDLE;
1949 		break;
1950 	case EQ_ERROR_STATE:
1951 		*msiq_state = PCI_MSIQ_STATE_ERROR;
1952 		break;
1953 	default:
1954 		ret = H_EIO;
1955 	}
1956 
1957 	return (ret);
1958 }
1959 
1960 uint64_t
1961 hvio_msiq_setstate(devhandle_t dev_hdl, msiqid_t msiq_id,
1962     pci_msiq_state_t msiq_state)
1963 {
1964 	uint32_t	eq_state;
1965 	uint64_t	ret = H_EOK;
1966 
1967 	eq_state = CSRA_FR((caddr_t)dev_hdl, EVENT_QUEUE_STATE,
1968 	    msiq_id, ENTRIES_STATE);
1969 
1970 	switch (eq_state) {
1971 	case EQ_IDLE_STATE:
1972 		if (msiq_state == PCI_MSIQ_STATE_ERROR)
1973 			ret = H_EIO;
1974 		break;
1975 	case EQ_ACTIVE_STATE:
1976 		if (msiq_state == PCI_MSIQ_STATE_ERROR)
1977 			CSRA_BS((caddr_t)dev_hdl, EVENT_QUEUE_CONTROL_SET,
1978 			    msiq_id, ENTRIES_ENOVERR);
1979 		else
1980 			ret = H_EIO;
1981 		break;
1982 	case EQ_ERROR_STATE:
1983 		if (msiq_state == PCI_MSIQ_STATE_IDLE)
1984 			CSRA_BS((caddr_t)dev_hdl, EVENT_QUEUE_CONTROL_CLEAR,
1985 			    msiq_id, ENTRIES_E2I);
1986 		else
1987 			ret = H_EIO;
1988 		break;
1989 	default:
1990 		ret = H_EIO;
1991 	}
1992 
1993 	return (ret);
1994 }
1995 
1996 uint64_t
1997 hvio_msiq_gethead(devhandle_t dev_hdl, msiqid_t msiq_id,
1998     msiqhead_t *msiq_head)
1999 {
2000 	*msiq_head = CSRA_FR((caddr_t)dev_hdl, EVENT_QUEUE_HEAD,
2001 	    msiq_id, ENTRIES_HEAD);
2002 
2003 	return (H_EOK);
2004 }
2005 
2006 uint64_t
2007 hvio_msiq_sethead(devhandle_t dev_hdl, msiqid_t msiq_id,
2008     msiqhead_t msiq_head)
2009 {
2010 	CSRA_FS((caddr_t)dev_hdl, EVENT_QUEUE_HEAD, msiq_id,
2011 	    ENTRIES_HEAD, msiq_head);
2012 
2013 	return (H_EOK);
2014 }
2015 
2016 uint64_t
2017 hvio_msiq_gettail(devhandle_t dev_hdl, msiqid_t msiq_id,
2018     msiqtail_t *msiq_tail)
2019 {
2020 	*msiq_tail = CSRA_FR((caddr_t)dev_hdl, EVENT_QUEUE_TAIL,
2021 	    msiq_id, ENTRIES_TAIL);
2022 
2023 	return (H_EOK);
2024 }
2025 
2026 /*
2027  * MSI Functions:
2028  */
2029 uint64_t
2030 hvio_msi_init(devhandle_t dev_hdl, uint64_t addr32, uint64_t addr64)
2031 {
2032 	/* PCI MEM 32 resources to perform 32 bit MSI transactions */
2033 	CSRA_FS((caddr_t)dev_hdl, MSI_32_BIT_ADDRESS, 0,
2034 	    ADDR, (uint64_t)addr32 >> MSI_32_BIT_ADDRESS_ADDR);
2035 	DBG(DBG_IB, NULL, "hvio_msiq_init: MSI_32_BIT_ADDRESS: 0x%llx\n",
2036 	    CSR_XR((caddr_t)dev_hdl, MSI_32_BIT_ADDRESS));
2037 
2038 	/* Reserve PCI MEM 64 resources to perform 64 bit MSI transactions */
2039 	CSRA_FS((caddr_t)dev_hdl, MSI_64_BIT_ADDRESS, 0,
2040 	    ADDR, (uint64_t)addr64 >> MSI_64_BIT_ADDRESS_ADDR);
2041 	DBG(DBG_IB, NULL, "hvio_msiq_init: MSI_64_BIT_ADDRESS: 0x%llx\n",
2042 	    CSR_XR((caddr_t)dev_hdl, MSI_64_BIT_ADDRESS));
2043 
2044 	return (H_EOK);
2045 }
2046 
2047 uint64_t
2048 hvio_msi_getmsiq(devhandle_t dev_hdl, msinum_t msi_num,
2049     msiqid_t *msiq_id)
2050 {
2051 	*msiq_id = CSRA_FR((caddr_t)dev_hdl, MSI_MAPPING,
2052 	    msi_num, ENTRIES_EQNUM);
2053 
2054 	return (H_EOK);
2055 }
2056 
2057 uint64_t
2058 hvio_msi_setmsiq(devhandle_t dev_hdl, msinum_t msi_num,
2059     msiqid_t msiq_id)
2060 {
2061 	CSRA_FS((caddr_t)dev_hdl, MSI_MAPPING, msi_num,
2062 	    ENTRIES_EQNUM, msiq_id);
2063 
2064 	return (H_EOK);
2065 }
2066 
2067 uint64_t
2068 hvio_msi_getvalid(devhandle_t dev_hdl, msinum_t msi_num,
2069     pci_msi_valid_state_t *msi_valid_state)
2070 {
2071 	*msi_valid_state = CSRA_BR((caddr_t)dev_hdl, MSI_MAPPING,
2072 	    msi_num, ENTRIES_V);
2073 
2074 	return (H_EOK);
2075 }
2076 
2077 uint64_t
2078 hvio_msi_setvalid(devhandle_t dev_hdl, msinum_t msi_num,
2079     pci_msi_valid_state_t msi_valid_state)
2080 {
2081 	uint64_t	ret = H_EOK;
2082 
2083 	switch (msi_valid_state) {
2084 	case PCI_MSI_VALID:
2085 		CSRA_BS((caddr_t)dev_hdl, MSI_MAPPING, msi_num,
2086 		    ENTRIES_V);
2087 		break;
2088 	case PCI_MSI_INVALID:
2089 		CSRA_BC((caddr_t)dev_hdl, MSI_MAPPING, msi_num,
2090 		    ENTRIES_V);
2091 		break;
2092 	default:
2093 		ret = H_EINVAL;
2094 	}
2095 
2096 	return (ret);
2097 }
2098 
2099 uint64_t
2100 hvio_msi_getstate(devhandle_t dev_hdl, msinum_t msi_num,
2101     pci_msi_state_t *msi_state)
2102 {
2103 	*msi_state = CSRA_BR((caddr_t)dev_hdl, MSI_MAPPING,
2104 	    msi_num, ENTRIES_EQWR_N);
2105 
2106 	return (H_EOK);
2107 }
2108 
2109 uint64_t
2110 hvio_msi_setstate(devhandle_t dev_hdl, msinum_t msi_num,
2111     pci_msi_state_t msi_state)
2112 {
2113 	uint64_t	ret = H_EOK;
2114 
2115 	switch (msi_state) {
2116 	case PCI_MSI_STATE_IDLE:
2117 		CSRA_BS((caddr_t)dev_hdl, MSI_CLEAR, msi_num,
2118 		    ENTRIES_EQWR_N);
2119 		break;
2120 	case PCI_MSI_STATE_DELIVERED:
2121 	default:
2122 		ret = H_EINVAL;
2123 		break;
2124 	}
2125 
2126 	return (ret);
2127 }
2128 
2129 /*
2130  * MSG Functions:
2131  */
2132 uint64_t
2133 hvio_msg_getmsiq(devhandle_t dev_hdl, pcie_msg_type_t msg_type,
2134     msiqid_t *msiq_id)
2135 {
2136 	uint64_t	ret = H_EOK;
2137 
2138 	switch (msg_type) {
2139 	case PCIE_PME_MSG:
2140 		*msiq_id = CSR_FR((caddr_t)dev_hdl, PM_PME_MAPPING, EQNUM);
2141 		break;
2142 	case PCIE_PME_ACK_MSG:
2143 		*msiq_id = CSR_FR((caddr_t)dev_hdl, PME_TO_ACK_MAPPING,
2144 		    EQNUM);
2145 		break;
2146 	case PCIE_CORR_MSG:
2147 		*msiq_id = CSR_FR((caddr_t)dev_hdl, ERR_COR_MAPPING, EQNUM);
2148 		break;
2149 	case PCIE_NONFATAL_MSG:
2150 		*msiq_id = CSR_FR((caddr_t)dev_hdl, ERR_NONFATAL_MAPPING,
2151 		    EQNUM);
2152 		break;
2153 	case PCIE_FATAL_MSG:
2154 		*msiq_id = CSR_FR((caddr_t)dev_hdl, ERR_FATAL_MAPPING, EQNUM);
2155 		break;
2156 	default:
2157 		ret = H_EINVAL;
2158 		break;
2159 	}
2160 
2161 	return (ret);
2162 }
2163 
2164 uint64_t
2165 hvio_msg_setmsiq(devhandle_t dev_hdl, pcie_msg_type_t msg_type,
2166     msiqid_t msiq_id)
2167 {
2168 	uint64_t	ret = H_EOK;
2169 
2170 	switch (msg_type) {
2171 	case PCIE_PME_MSG:
2172 		CSR_FS((caddr_t)dev_hdl, PM_PME_MAPPING, EQNUM, msiq_id);
2173 		break;
2174 	case PCIE_PME_ACK_MSG:
2175 		CSR_FS((caddr_t)dev_hdl, PME_TO_ACK_MAPPING, EQNUM, msiq_id);
2176 		break;
2177 	case PCIE_CORR_MSG:
2178 		CSR_FS((caddr_t)dev_hdl, ERR_COR_MAPPING, EQNUM, msiq_id);
2179 		break;
2180 	case PCIE_NONFATAL_MSG:
2181 		CSR_FS((caddr_t)dev_hdl, ERR_NONFATAL_MAPPING, EQNUM, msiq_id);
2182 		break;
2183 	case PCIE_FATAL_MSG:
2184 		CSR_FS((caddr_t)dev_hdl, ERR_FATAL_MAPPING, EQNUM, msiq_id);
2185 		break;
2186 	default:
2187 		ret = H_EINVAL;
2188 		break;
2189 	}
2190 
2191 	return (ret);
2192 }
2193 
2194 uint64_t
2195 hvio_msg_getvalid(devhandle_t dev_hdl, pcie_msg_type_t msg_type,
2196     pcie_msg_valid_state_t *msg_valid_state)
2197 {
2198 	uint64_t	ret = H_EOK;
2199 
2200 	switch (msg_type) {
2201 	case PCIE_PME_MSG:
2202 		*msg_valid_state = CSR_BR((caddr_t)dev_hdl, PM_PME_MAPPING, V);
2203 		break;
2204 	case PCIE_PME_ACK_MSG:
2205 		*msg_valid_state = CSR_BR((caddr_t)dev_hdl,
2206 		    PME_TO_ACK_MAPPING, V);
2207 		break;
2208 	case PCIE_CORR_MSG:
2209 		*msg_valid_state = CSR_BR((caddr_t)dev_hdl, ERR_COR_MAPPING, V);
2210 		break;
2211 	case PCIE_NONFATAL_MSG:
2212 		*msg_valid_state = CSR_BR((caddr_t)dev_hdl,
2213 		    ERR_NONFATAL_MAPPING, V);
2214 		break;
2215 	case PCIE_FATAL_MSG:
2216 		*msg_valid_state = CSR_BR((caddr_t)dev_hdl, ERR_FATAL_MAPPING,
2217 		    V);
2218 		break;
2219 	default:
2220 		ret = H_EINVAL;
2221 		break;
2222 	}
2223 
2224 	return (ret);
2225 }
2226 
2227 uint64_t
2228 hvio_msg_setvalid(devhandle_t dev_hdl, pcie_msg_type_t msg_type,
2229     pcie_msg_valid_state_t msg_valid_state)
2230 {
2231 	uint64_t	ret = H_EOK;
2232 
2233 	switch (msg_valid_state) {
2234 	case PCIE_MSG_VALID:
2235 		switch (msg_type) {
2236 		case PCIE_PME_MSG:
2237 			CSR_BS((caddr_t)dev_hdl, PM_PME_MAPPING, V);
2238 			break;
2239 		case PCIE_PME_ACK_MSG:
2240 			CSR_BS((caddr_t)dev_hdl, PME_TO_ACK_MAPPING, V);
2241 			break;
2242 		case PCIE_CORR_MSG:
2243 			CSR_BS((caddr_t)dev_hdl, ERR_COR_MAPPING, V);
2244 			break;
2245 		case PCIE_NONFATAL_MSG:
2246 			CSR_BS((caddr_t)dev_hdl, ERR_NONFATAL_MAPPING, V);
2247 			break;
2248 		case PCIE_FATAL_MSG:
2249 			CSR_BS((caddr_t)dev_hdl, ERR_FATAL_MAPPING, V);
2250 			break;
2251 		default:
2252 			ret = H_EINVAL;
2253 			break;
2254 		}
2255 
2256 		break;
2257 	case PCIE_MSG_INVALID:
2258 		switch (msg_type) {
2259 		case PCIE_PME_MSG:
2260 			CSR_BC((caddr_t)dev_hdl, PM_PME_MAPPING, V);
2261 			break;
2262 		case PCIE_PME_ACK_MSG:
2263 			CSR_BC((caddr_t)dev_hdl, PME_TO_ACK_MAPPING, V);
2264 			break;
2265 		case PCIE_CORR_MSG:
2266 			CSR_BC((caddr_t)dev_hdl, ERR_COR_MAPPING, V);
2267 			break;
2268 		case PCIE_NONFATAL_MSG:
2269 			CSR_BC((caddr_t)dev_hdl, ERR_NONFATAL_MAPPING, V);
2270 			break;
2271 		case PCIE_FATAL_MSG:
2272 			CSR_BC((caddr_t)dev_hdl, ERR_FATAL_MAPPING, V);
2273 			break;
2274 		default:
2275 			ret = H_EINVAL;
2276 			break;
2277 		}
2278 		break;
2279 	default:
2280 		ret = H_EINVAL;
2281 	}
2282 
2283 	return (ret);
2284 }
2285 
2286 /*
2287  * Suspend/Resume Functions:
2288  *	(pec, mmu, ib)
2289  *	cb
2290  * Registers saved have all been touched in the XXX_init functions.
2291  */
2292 uint64_t
2293 hvio_suspend(devhandle_t dev_hdl, pxu_t *pxu_p)
2294 {
2295 	uint64_t	*config_state;
2296 	int		total_size;
2297 	int		i;
2298 
2299 	if (msiq_suspend(dev_hdl, pxu_p) != H_EOK)
2300 		return (H_EIO);
2301 
2302 	total_size = PEC_SIZE + MMU_SIZE + IB_SIZE + IB_MAP_SIZE;
2303 	config_state = kmem_zalloc(total_size, KM_NOSLEEP);
2304 
2305 	if (config_state == NULL) {
2306 		return (H_EIO);
2307 	}
2308 
2309 	/*
2310 	 * Soft state for suspend/resume  from pxu_t
2311 	 * uint64_t	*pec_config_state;
2312 	 * uint64_t	*mmu_config_state;
2313 	 * uint64_t	*ib_intr_map;
2314 	 * uint64_t	*ib_config_state;
2315 	 * uint64_t	*xcb_config_state;
2316 	 */
2317 
2318 	/* Save the PEC configuration states */
2319 	pxu_p->pec_config_state = config_state;
2320 	for (i = 0; i < PEC_KEYS; i++) {
2321 		pxu_p->pec_config_state[i] =
2322 		    CSR_XR((caddr_t)dev_hdl, pec_config_state_regs[i]);
2323 	}
2324 
2325 	/* Save the MMU configuration states */
2326 	pxu_p->mmu_config_state = pxu_p->pec_config_state + PEC_KEYS;
2327 	for (i = 0; i < MMU_KEYS; i++) {
2328 		pxu_p->mmu_config_state[i] =
2329 		    CSR_XR((caddr_t)dev_hdl, mmu_config_state_regs[i]);
2330 	}
2331 
2332 	/* Save the interrupt mapping registers */
2333 	pxu_p->ib_intr_map = pxu_p->mmu_config_state + MMU_KEYS;
2334 	for (i = 0; i < INTERRUPT_MAPPING_ENTRIES; i++) {
2335 		pxu_p->ib_intr_map[i] =
2336 		    CSRA_XR((caddr_t)dev_hdl, INTERRUPT_MAPPING, i);
2337 	}
2338 
2339 	/* Save the IB configuration states */
2340 	pxu_p->ib_config_state = pxu_p->ib_intr_map + INTERRUPT_MAPPING_ENTRIES;
2341 	for (i = 0; i < IB_KEYS; i++) {
2342 		pxu_p->ib_config_state[i] =
2343 		    CSR_XR((caddr_t)dev_hdl, ib_config_state_regs[i]);
2344 	}
2345 
2346 	return (H_EOK);
2347 }
2348 
2349 void
2350 hvio_resume(devhandle_t dev_hdl, devino_t devino, pxu_t *pxu_p)
2351 {
2352 	int		total_size;
2353 	sysino_t	sysino;
2354 	int		i;
2355 
2356 	/* Make sure that suspend actually did occur */
2357 	if (!pxu_p->pec_config_state) {
2358 		return;
2359 	}
2360 
2361 	/* Restore IB configuration states */
2362 	for (i = 0; i < IB_KEYS; i++) {
2363 		CSR_XS((caddr_t)dev_hdl, ib_config_state_regs[i],
2364 		    pxu_p->ib_config_state[i]);
2365 	}
2366 
2367 	/*
2368 	 * Restore the interrupt mapping registers
2369 	 * And make sure the intrs are idle.
2370 	 */
2371 	for (i = 0; i < INTERRUPT_MAPPING_ENTRIES; i++) {
2372 		CSRA_FS((caddr_t)dev_hdl, INTERRUPT_CLEAR, i,
2373 		    ENTRIES_INT_STATE, INTERRUPT_IDLE_STATE);
2374 		CSRA_XS((caddr_t)dev_hdl, INTERRUPT_MAPPING, i,
2375 		    pxu_p->ib_intr_map[i]);
2376 	}
2377 
2378 	/* Restore MMU configuration states */
2379 	/* Clear the cache. */
2380 	CSR_XS((caddr_t)dev_hdl, MMU_TTE_CACHE_INVALIDATE, -1ull);
2381 
2382 	for (i = 0; i < MMU_KEYS; i++) {
2383 		CSR_XS((caddr_t)dev_hdl, mmu_config_state_regs[i],
2384 		    pxu_p->mmu_config_state[i]);
2385 	}
2386 
2387 	/* Restore PEC configuration states */
2388 	/* Make sure all reset bits are low until error is detected */
2389 	CSR_XS((caddr_t)dev_hdl, LPU_RESET, 0ull);
2390 
2391 	for (i = 0; i < PEC_KEYS; i++) {
2392 		CSR_XS((caddr_t)dev_hdl, pec_config_state_regs[i],
2393 		    pxu_p->pec_config_state[i]);
2394 	}
2395 
2396 	/* Enable PCI-E interrupt */
2397 	(void) hvio_intr_devino_to_sysino(dev_hdl, pxu_p, devino, &sysino);
2398 
2399 	(void) hvio_intr_setstate(dev_hdl, sysino, INTR_IDLE_STATE);
2400 
2401 	total_size = PEC_SIZE + MMU_SIZE + IB_SIZE + IB_MAP_SIZE;
2402 	kmem_free(pxu_p->pec_config_state, total_size);
2403 
2404 	pxu_p->pec_config_state = NULL;
2405 	pxu_p->mmu_config_state = NULL;
2406 	pxu_p->ib_config_state = NULL;
2407 	pxu_p->ib_intr_map = NULL;
2408 
2409 	msiq_resume(dev_hdl, pxu_p);
2410 }
2411 
2412 uint64_t
2413 hvio_cb_suspend(devhandle_t dev_hdl, pxu_t *pxu_p)
2414 {
2415 	uint64_t	*config_state;
2416 	int		i;
2417 
2418 	config_state = kmem_zalloc(CB_SIZE, KM_NOSLEEP);
2419 
2420 	if (config_state == NULL) {
2421 		return (H_EIO);
2422 	}
2423 
2424 	/* Save the configuration states */
2425 	pxu_p->xcb_config_state = config_state;
2426 	for (i = 0; i < CB_KEYS; i++) {
2427 		pxu_p->xcb_config_state[i] =
2428 		    CSR_XR((caddr_t)dev_hdl, cb_config_state_regs[i]);
2429 	}
2430 
2431 	return (H_EOK);
2432 }
2433 
2434 void
2435 hvio_cb_resume(devhandle_t pci_dev_hdl, devhandle_t xbus_dev_hdl,
2436     devino_t devino, pxu_t *pxu_p)
2437 {
2438 	sysino_t	sysino;
2439 	int		i;
2440 
2441 	/*
2442 	 * No reason to have any reset bits high until an error is
2443 	 * detected on the link.
2444 	 */
2445 	CSR_XS((caddr_t)xbus_dev_hdl, JBC_ERROR_STATUS_CLEAR, -1ull);
2446 
2447 	ASSERT(pxu_p->xcb_config_state);
2448 
2449 	/* Restore the configuration states */
2450 	for (i = 0; i < CB_KEYS; i++) {
2451 		CSR_XS((caddr_t)xbus_dev_hdl, cb_config_state_regs[i],
2452 		    pxu_p->xcb_config_state[i]);
2453 	}
2454 
2455 	/* Enable XBC interrupt */
2456 	(void) hvio_intr_devino_to_sysino(pci_dev_hdl, pxu_p, devino, &sysino);
2457 
2458 	(void) hvio_intr_setstate(pci_dev_hdl, sysino, INTR_IDLE_STATE);
2459 
2460 	kmem_free(pxu_p->xcb_config_state, CB_SIZE);
2461 
2462 	pxu_p->xcb_config_state = NULL;
2463 }
2464 
2465 static uint64_t
2466 msiq_suspend(devhandle_t dev_hdl, pxu_t *pxu_p)
2467 {
2468 	size_t	bufsz;
2469 	volatile uint64_t *cur_p;
2470 	int i;
2471 
2472 	bufsz = MSIQ_STATE_SIZE + MSIQ_MAPPING_SIZE + MSIQ_OTHER_SIZE;
2473 	if ((pxu_p->msiq_config_state = kmem_zalloc(bufsz, KM_NOSLEEP)) ==
2474 	    NULL)
2475 		return (H_EIO);
2476 
2477 	cur_p = pxu_p->msiq_config_state;
2478 
2479 	/* Save each EQ state */
2480 	for (i = 0; i < EVENT_QUEUE_STATE_ENTRIES; i++, cur_p++)
2481 		*cur_p = CSRA_XR((caddr_t)dev_hdl, EVENT_QUEUE_STATE, i);
2482 
2483 	/* Save MSI mapping registers */
2484 	for (i = 0; i < MSI_MAPPING_ENTRIES; i++, cur_p++)
2485 		*cur_p = CSRA_XR((caddr_t)dev_hdl, MSI_MAPPING, i);
2486 
2487 	/* Save all other MSIQ registers */
2488 	for (i = 0; i < MSIQ_OTHER_KEYS; i++, cur_p++)
2489 		*cur_p = CSR_XR((caddr_t)dev_hdl, msiq_config_other_regs[i]);
2490 	return (H_EOK);
2491 }
2492 
2493 static void
2494 msiq_resume(devhandle_t dev_hdl, pxu_t *pxu_p)
2495 {
2496 	size_t	bufsz;
2497 	uint64_t *cur_p;
2498 	int i;
2499 
2500 	bufsz = MSIQ_STATE_SIZE + MSIQ_MAPPING_SIZE + MSIQ_OTHER_SIZE;
2501 	cur_p = pxu_p->msiq_config_state;
2502 	/*
2503 	 * Initialize EQ base address register and
2504 	 * Interrupt Mondo Data 0 register.
2505 	 */
2506 	(void) hvio_msiq_init(dev_hdl, pxu_p);
2507 
2508 	/* Restore EQ states */
2509 	for (i = 0; i < EVENT_QUEUE_STATE_ENTRIES; i++, cur_p++) {
2510 		if (((*cur_p) & EVENT_QUEUE_STATE_ENTRIES_STATE_MASK) ==
2511 		    EQ_ACTIVE_STATE) {
2512 			CSRA_BS((caddr_t)dev_hdl, EVENT_QUEUE_CONTROL_SET,
2513 			    i, ENTRIES_EN);
2514 		}
2515 	}
2516 
2517 	/* Restore MSI mapping */
2518 	for (i = 0; i < MSI_MAPPING_ENTRIES; i++, cur_p++)
2519 		CSRA_XS((caddr_t)dev_hdl, MSI_MAPPING, i, *cur_p);
2520 
2521 	/*
2522 	 * Restore all other registers. MSI 32 bit address and
2523 	 * MSI 64 bit address are restored as part of this.
2524 	 */
2525 	for (i = 0; i < MSIQ_OTHER_KEYS; i++, cur_p++)
2526 		CSR_XS((caddr_t)dev_hdl, msiq_config_other_regs[i], *cur_p);
2527 
2528 	kmem_free(pxu_p->msiq_config_state, bufsz);
2529 	pxu_p->msiq_config_state = NULL;
2530 }
2531 
2532 /*
2533  * sends PME_Turn_Off message to put the link in L2/L3 ready state.
2534  * called by px_goto_l23ready.
2535  * returns DDI_SUCCESS or DDI_FAILURE
2536  */
2537 int
2538 px_send_pme_turnoff(caddr_t csr_base)
2539 {
2540 	volatile uint64_t reg;
2541 
2542 	reg = CSR_XR(csr_base, TLU_PME_TURN_OFF_GENERATE);
2543 	/* If already pending, return failure */
2544 	if (reg & (1ull << TLU_PME_TURN_OFF_GENERATE_PTO)) {
2545 		DBG(DBG_PWR, NULL, "send_pme_turnoff: pending PTO bit "
2546 		    "tlu_pme_turn_off_generate = %x\n", reg);
2547 		return (DDI_FAILURE);
2548 	}
2549 
2550 	/* write to PME_Turn_off reg to boradcast */
2551 	reg |= (1ull << TLU_PME_TURN_OFF_GENERATE_PTO);
2552 	CSR_XS(csr_base,  TLU_PME_TURN_OFF_GENERATE, reg);
2553 
2554 	return (DDI_SUCCESS);
2555 }
2556 
2557 /*
2558  * Checks for link being in L1idle state.
2559  * Returns
2560  * DDI_SUCCESS - if the link is in L1idle
2561  * DDI_FAILURE - if the link is not in L1idle
2562  */
2563 int
2564 px_link_wait4l1idle(caddr_t csr_base)
2565 {
2566 	uint8_t ltssm_state;
2567 	int ntries = px_max_l1_tries;
2568 
2569 	while (ntries > 0) {
2570 		ltssm_state = CSR_FR(csr_base, LPU_LTSSM_STATUS1, LTSSM_STATE);
2571 		if (ltssm_state == LPU_LTSSM_L1_IDLE || (--ntries <= 0))
2572 			break;
2573 		delay(1);
2574 	}
2575 	DBG(DBG_PWR, NULL, "check_for_l1idle: ltssm_state %x\n", ltssm_state);
2576 	return ((ltssm_state == LPU_LTSSM_L1_IDLE) ? DDI_SUCCESS : DDI_FAILURE);
2577 }
2578 
2579 /*
2580  * Tranisition the link to L0, after it is down.
2581  */
2582 int
2583 px_link_retrain(caddr_t csr_base)
2584 {
2585 	volatile uint64_t reg;
2586 
2587 	reg = CSR_XR(csr_base, TLU_CONTROL);
2588 	if (!(reg & (1ull << TLU_REMAIN_DETECT_QUIET))) {
2589 		DBG(DBG_PWR, NULL, "retrain_link: detect.quiet bit not set\n");
2590 		return (DDI_FAILURE);
2591 	}
2592 
2593 	/* Clear link down bit in TLU Other Event Clear Status Register. */
2594 	CSR_BS(csr_base, TLU_OTHER_EVENT_STATUS_CLEAR, LDN_P);
2595 
2596 	/* Clear Drain bit in TLU Status Register */
2597 	CSR_BS(csr_base, TLU_STATUS, DRAIN);
2598 
2599 	/* Clear Remain in Detect.Quiet bit in TLU Control Register */
2600 	reg = CSR_XR(csr_base, TLU_CONTROL);
2601 	reg &= ~(1ull << TLU_REMAIN_DETECT_QUIET);
2602 	CSR_XS(csr_base, TLU_CONTROL, reg);
2603 
2604 	return (DDI_SUCCESS);
2605 }
2606 
2607 void
2608 px_enable_detect_quiet(caddr_t csr_base)
2609 {
2610 	volatile uint64_t tlu_ctrl;
2611 
2612 	tlu_ctrl = CSR_XR(csr_base, TLU_CONTROL);
2613 	tlu_ctrl |= (1ull << TLU_REMAIN_DETECT_QUIET);
2614 	CSR_XS(csr_base, TLU_CONTROL, tlu_ctrl);
2615 }
2616