xref: /titanic_51/usr/src/uts/sun4u/starfire/io/idn_proto.c (revision 3d5869cc80c26887ee2c78b0f35e58f574e97192)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 
22 /*
23  * Copyright 2009 Sun Microsystems, Inc.  All rights reserved.
24  * Use is subject to license terms.
25  */
26 
27 /*
28  * Inter-Domain Network
29  *
30  * IDN Protocol functions to support domain link/unlink/reconfig.
31  */
32 
33 #include <sys/types.h>
34 #include <sys/param.h>
35 #include <sys/machparam.h>
36 #include <sys/debug.h>
37 #include <sys/cpuvar.h>
38 #include <sys/kmem.h>
39 #include <sys/mutex.h>
40 #include <sys/rwlock.h>
41 #include <sys/systm.h>
42 #include <sys/stream.h>
43 #include <sys/strsun.h>
44 #include <sys/stropts.h>
45 #include <sys/sema_impl.h>
46 #include <sys/membar.h>
47 #include <sys/utsname.h>
48 #include <inet/common.h>
49 #include <inet/mi.h>
50 #include <netinet/ip6.h>
51 #include <inet/ip.h>
52 #include <netinet/in.h>
53 #include <sys/vm_machparam.h>
54 #include <sys/x_call.h>
55 #include <sys/ddi.h>
56 #include <sys/sunddi.h>
57 #include <sys/atomic.h>
58 #include <vm/as.h>		/* kas decl */
59 
60 #include <sys/idn.h>
61 #include <sys/idn_xf.h>
62 
63 #define	IDNBUG_CPUPERBOARD
64 
65 extern pri_t		maxclsyspri;
66 extern u_longlong_t	gettick();
67 
68 clock_t	idn_xmit_monitor_freq = 50;
69 
70 static int	idn_connect(int domid);
71 static int	idn_disconnect(int domid, idn_fin_t fintype,
72 				idn_finarg_t finarg, idn_finsync_t finsync);
73 static void	idn_deconfig(int domid);
74 static void	idn_unlink_domainset(domainset_t domset, idn_fin_t fintype,
75 				idn_finarg_t finarg, idn_finopt_t finopt,
76 				boardset_t idnset);
77 static void	idn_retry_execute(void *arg);
78 static void	idn_retry_submit(void (*func)(uint_t token, void *arg),
79 				void *arg, uint_t token, clock_t ticks);
80 static void	idn_shutdown_datapath(domainset_t domset, int force);
81 static mblk_t	*idn_fill_buffer(caddr_t bufp, int size, mblk_t *mp,
82 				uchar_t **data_rptrp);
83 static ushort_t	idn_cksum(register ushort_t *hdrp, register int count);
84 static int	idn_mark_awol(int domid, clock_t *atime);
85 
86 static void	idn_recv_proto(idn_protomsg_t *hp);
87 static void	idn_send_config(int domid, int phase);
88 static void	idn_recv_config(int domid, idn_msgtype_t *mtp,
89 				idn_xdcargs_t xargs);
90 static int	idn_send_master_config(int domid, int phase);
91 static int	idn_send_slave_config(int domid, int phase);
92 static uint_t	idn_check_master_config(int domid, uint_t *exp, uint_t *act);
93 static uint_t	idn_check_slave_config(int domid, uint_t *exp, uint_t *act);
94 static int	idn_recv_config_done(int domid);
95 static void	idn_nego_cleanup_check(int domid, int new_masterid,
96 				int new_cpuid);
97 static void	idn_recv_cmd(int domid, idn_msgtype_t *mtp,
98 				idn_xdcargs_t xargs);
99 static int	idn_recv_data(int domid, idn_msgtype_t *mtp,
100 				idn_xdcargs_t xargs);
101 static int	idn_send_data_loopback(idn_netaddr_t dst_netaddr,
102 				queue_t *wq, mblk_t *mp);
103 static void	idn_send_dataresp(int domid, idn_nack_t nacktype);
104 static int	idn_send_mboxdata(int domid, struct idn *sip, int channel,
105 				caddr_t bufp);
106 static int	idn_recv_mboxdata(int channel, caddr_t bufp);
107 static int	idn_program_hardware(int domid);
108 static int	idn_deprogram_hardware(int domid);
109 
110 static void	idn_send_cmd_nackresp(int domid, idn_msgtype_t *mtp,
111 			idn_cmd_t cmdtype, idn_nack_t nacktype);
112 static void	idn_local_cmd(idn_cmd_t cmdtype, uint_t arg1,
113 				uint_t arg2, uint_t arg3);
114 static void	idn_terminate_cmd(int domid, int serrno);
115 static void	idn_mboxarea_init(idn_mboxtbl_t *mtp, register int ntbls);
116 static void	idn_mainmbox_activate(int domid);
117 static void	idn_mainmbox_deactivate(ushort_t domset);
118 static void	idn_mainmbox_chan_register(int domid,
119 				idn_mainmbox_t *send_mmp,
120 				idn_mainmbox_t *recv_mmp, int channel);
121 static int	idn_mainmbox_chan_unregister(ushort_t domset, int channel);
122 static int	idn_mainmbox_flush(int domid, idn_mainmbox_t *mmp);
123 static void	idn_mainmbox_reset(int domid, idn_mainmbox_t *cmp);
124 static int	idn_activate_channel(idn_chanset_t chanset,
125 				idn_chanop_t chanop);
126 static void	idn_deactivate_channel(idn_chanset_t chanset,
127 				idn_chanop_t chanop);
128 static int	idn_deactivate_channel_services(int channel,
129 				idn_chanop_t chanop);
130 static int	idn_activate_channel_services(int channel);
131 static void	idn_chan_server(idn_chansvr_t **cspp);
132 #if 0
133 static void	idn_chan_flush(idn_chansvr_t *csp);
134 #endif /* 0 */
135 static void	idn_chan_action(int channel, idn_chanaction_t chanaction,
136 				int wait);
137 static void	idn_chan_addmbox(int channel, ushort_t domset);
138 static void	idn_chan_delmbox(int channel, ushort_t domset);
139 static void	idn_submit_chanactivate_job(int channel);
140 static void	idn_exec_chanactivate(void *chn);
141 
142 static void	idn_link_established(void *arg);
143 static void	idn_prealloc_slab(int nslabs);
144 static void	idn_recv_slaballoc_req(int domid, idn_msgtype_t *mtp,
145 				uint_t slab_size);
146 static void	idn_send_slaballoc_resp(int domid, idn_msgtype_t *mtp,
147 				uint_t slab_offset, uint_t slab_size,
148 				int serrno);
149 static void	idn_recv_slaballoc_resp(int domid, smr_offset_t slab_offset,
150 				uint_t slab_size, int serrno);
151 static void	idn_recv_slabreap_req(int domid, idn_msgtype_t *mtp,
152 				int nslabs);
153 static void	idn_recv_slabreap_resp(int domid, int nslabs, int serrno);
154 static void	idn_send_slabreap_resp(int domid, idn_msgtype_t *mtp,
155 				int nslabs, int serrno);
156 static void	idn_recv_slabfree_req(int domid, idn_msgtype_t *mtp,
157 				smr_offset_t slab_offset, uint_t slab_size);
158 static void	idn_recv_slabfree_resp(int domid, uint_t slab_offset,
159 				uint_t slab_size, int serrno);
160 static void	idn_send_slabfree_resp(int domid, idn_msgtype_t *mtp,
161 				uint_t slab_offset, uint_t slab_size,
162 				int serrno);
163 static void	idn_retry_nodename_req(void *arg);
164 static void	idn_send_nodename_req(int domid);
165 static void	idn_send_nodename_resp(int domid, idn_msgtype_t *mtp,
166 				uint_t bufoffset, int serrno);
167 static void	idn_recv_nodename_req(int domid, idn_msgtype_t *mtp,
168 				uint_t bufoffset);
169 static void	idn_recv_nodename_resp(int domid, uint_t bufoffset,
170 				int serrno);
171 
172 static void	idn_protocol_server(int *id);
173 static void	idn_protocol_server_killall();
174 static void	idn_protojob_free(idn_protojob_t *jp);
175 
176 static int	idn_xstate_transfunc(int domid, void *transarg);
177 static int	idn_xphase_transition(int domid, idn_msgtype_t *mtp,
178 				idn_xdcargs_t xargs);
179 static void	idn_sync_enter(int domid, idn_synccmd_t cmd,
180 				domainset_t xset, domainset_t rset,
181 				int (*transfunc)(), void *transarg);
182 static domainset_t
183 		idn_sync_register(int domid, idn_synccmd_t cmd,
184 				domainset_t ready_set, idn_syncreg_t regtype);
185 static void	idn_sync_register_awol(int domid);
186 static int	idn_verify_config_mbox(int domid);
187 static int	idn_select_master(int domid, int rmasterid, int rcpuid);
188 
189 static int	valid_mtu(uint_t mtu);
190 static int	valid_bufsize(uint_t bufsize);
191 static int	valid_slabsize(int slabsize);
192 static int	valid_nwrsize(int nwrsize);
193 
194 static int	idn_master_init();
195 static void	idn_master_deinit();
196 
197 static void	idn_send_acknack(int domid, idn_msgtype_t *mtp,
198 				idn_xdcargs_t xargs);
199 
200 static int	idn_send_nego(int domid, idn_msgtype_t *mtp,
201 				domainset_t conset);
202 static void	idn_retry_nego(uint_t token, void *arg);
203 static int	idn_check_nego(int domid, idn_msgtype_t *mtp,
204 				idn_xdcargs_t xargs);
205 static void	idn_action_nego_pend(int domid, idn_msgtype_t *mtp,
206 				idn_xdcargs_t xargs);
207 static void	idn_error_nego(int domid, idn_msgtype_t *mtp,
208 				idn_xdcargs_t xargs);
209 static void	idn_action_nego_sent(int domid, idn_msgtype_t *mtp,
210 				idn_xdcargs_t xargs);
211 static void	idn_action_nego_rcvd(int domid, idn_msgtype_t *mtp,
212 				idn_xdcargs_t xargs);
213 static void	idn_final_nego(int domid);
214 static void	idn_exit_nego(int domid, uint_t msgtype);
215 
216 static int	idn_send_con(int domid, idn_msgtype_t *mtp,
217 				idn_con_t contype, domainset_t conset);
218 static void	idn_retry_con(uint_t token, void *arg);
219 static int	idn_check_con(int domid, idn_msgtype_t *mtp,
220 				idn_xdcargs_t xargs);
221 static void	idn_action_con_pend(int domid, idn_msgtype_t *mtp,
222 				idn_xdcargs_t xargs);
223 static void	idn_error_con(int domid, idn_msgtype_t *mtp,
224 				idn_xdcargs_t xargs);
225 static void	idn_action_con_sent(int domid, idn_msgtype_t *mtp,
226 				idn_xdcargs_t xargs);
227 static void	idn_action_con_rcvd(int domid, idn_msgtype_t *mtp,
228 				idn_xdcargs_t xargs);
229 static void	idn_final_con(int domid);
230 static void	idn_exit_con(int domid, uint_t msgtype);
231 
232 static int	idn_send_fin(int domid, idn_msgtype_t *mtp, idn_fin_t fintype,
233 				idn_finarg_t finarg, idn_finopt_t finopt,
234 				domainset_t finset, uint_t finmaster);
235 static void	idn_retry_fin(uint_t token, void *arg);
236 static int	idn_check_fin_pend(int domid, idn_msgtype_t *mtp,
237 				idn_xdcargs_t xargs);
238 static void	idn_action_fin_pend(int domid, idn_msgtype_t *mtp,
239 				idn_xdcargs_t xargs);
240 static void	idn_error_fin_pend(int domid, idn_msgtype_t *mtp,
241 				idn_xdcargs_t xargs);
242 static int	idn_check_fin_sent(int domid, idn_msgtype_t *mtp,
243 				idn_xdcargs_t xargs);
244 static void	idn_action_fin_sent(int domid, idn_msgtype_t *mtp,
245 				idn_xdcargs_t xargs);
246 static void	idn_error_fin_sent(int domid, idn_msgtype_t *mtp,
247 				idn_xdcargs_t xargs);
248 static void	idn_action_fin_rcvd(int domid, idn_msgtype_t *mtp,
249 				idn_xdcargs_t xargs);
250 static void	idn_final_fin(int domid);
251 static void	idn_exit_fin(int domid, uint_t msgtype);
252 
253 /*
254  * We keep a small cache of protojob structures just
255  * in case allocation within idn_handler comes back
256  * with nothing from the land of kmem.
257  */
258 idn_protojob_t	idn_protojob_cache[IDN_DMV_PENDING_MAX];
259 idn_protojob_t	*idn_protojob_cache_list;
260 kmutex_t	idn_protojob_cache_lock;
261 
262 /*
263  *	- receive message.
264  *	- call check-function for current state.
265  *	- if (check-function == ok) then
266  *		call action-function for current state.
267  *	  else
268  *		call error-function for current state.
269  *	- transition state based on check results.
270  *	- if (next state == final state) then
271  *		call final-function.
272  */
273 static idn_xphase_t xphase_nego = {
274 	IDNP_NEGO,
275 	{
276 		{ IDNDS_NEGO_PEND,
277 			idn_check_nego,
278 			idn_action_nego_pend,
279 			idn_error_nego},
280 		{ IDNDS_NEGO_SENT,
281 			idn_check_nego,
282 			idn_action_nego_sent,
283 			idn_error_nego},
284 		{ IDNDS_NEGO_RCVD,
285 			NULL,
286 			idn_action_nego_rcvd,
287 			NULL },
288 		{ IDNDS_CONFIG, NULL, NULL, NULL },
289 	},
290 	idn_final_nego,
291 	idn_exit_nego
292 };
293 
294 static idn_xphase_t xphase_con = {
295 	IDNP_CON,
296 	{
297 		{ IDNDS_CON_PEND,
298 			idn_check_con,
299 			idn_action_con_pend,
300 			idn_error_con},
301 		{ IDNDS_CON_SENT,
302 			idn_check_con,
303 			idn_action_con_sent,
304 			idn_error_con},
305 		{ IDNDS_CON_RCVD,
306 			NULL,
307 			idn_action_con_rcvd,
308 			NULL },
309 		{ IDNDS_CON_READY, NULL, NULL, NULL },
310 	},
311 	idn_final_con,
312 	idn_exit_con
313 };
314 
315 static idn_xphase_t xphase_fin = {
316 	IDNP_FIN,
317 	{
318 		{ IDNDS_FIN_PEND,
319 			idn_check_fin_pend,
320 			idn_action_fin_pend,
321 			idn_error_fin_pend },
322 		{ IDNDS_FIN_SENT,
323 			idn_check_fin_sent,
324 			idn_action_fin_sent,
325 			idn_error_fin_sent },
326 		{ IDNDS_FIN_RCVD,
327 			NULL,
328 			idn_action_fin_rcvd,
329 			NULL },
330 		{ IDNDS_DMAP, NULL, NULL, NULL },
331 	},
332 	idn_final_fin,
333 	idn_exit_fin
334 };
335 
336 static int idnxs_state_table[4][5][2] = {
337 	{			/* IDNXS_PEND */
338 		{ IDNXS_SENT,	IDNXS_PEND },	/* 0 */
339 		{ IDNXS_RCVD,	IDNXS_PEND },	/* msg */
340 		{ IDNXS_NIL,	IDNXS_PEND },	/* msg+ack */
341 		{ IDNXS_PEND,	IDNXS_NIL },	/* ack */
342 		{ IDNXS_PEND,	IDNXS_NIL },	/* nack */
343 	},
344 	{			/* IDNXS_SENT */
345 		{ IDNXS_NIL,	IDNXS_NIL },	/* 0 */
346 		{ IDNXS_RCVD,	IDNXS_PEND },	/* msg */
347 		{ IDNXS_FINAL,	IDNXS_PEND },	/* msg+ack */
348 		{ IDNXS_NIL,	IDNXS_NIL },	/* ack */
349 		{ IDNXS_PEND,	IDNXS_NIL },	/* nack */
350 	},
351 	{			/* IDNXS_RCVD */
352 		{ IDNXS_NIL,	IDNXS_NIL },	/* 0 */
353 		{ IDNXS_NIL,	IDNXS_NIL },	/* msg */
354 		{ IDNXS_FINAL,	IDNXS_NIL },	/* msg+ack */
355 		{ IDNXS_FINAL,	IDNXS_NIL },	/* ack */
356 		{ IDNXS_PEND,	IDNXS_NIL },	/* nack */
357 	},
358 	{			/* IDNXS_FINAL */
359 		{ IDNXS_NIL,	IDNXS_NIL },	/* 0 */
360 		{ IDNXS_NIL,	IDNXS_NIL },	/* msg */
361 		{ IDNXS_NIL,	IDNXS_NIL },	/* msg+ack */
362 		{ IDNXS_NIL,	IDNXS_NIL },	/* ack */
363 		{ IDNXS_NIL,	IDNXS_NIL },	/* nack */
364 	}
365 };
366 
367 /*
368  * NONE		Respective domain does not have a master.
369  * OTHER	Respective domain has a master different
370  *		than either local or remote.
371  * LOCAL	Respective domain has chosen local as master.
372  * REMOTE	Respective domain has chosen remote as master.
373  *
374  * Actions:
375  *	VOTE		Compare votes and select one.
376  *	VOTE_RCFG	Compare votes and Reconfigure
377  *			if necessary, i.e. remote won.
378  *	CONNECT		Connect to remote's OTHER if different
379  *			than our local master.
380  *	LOCAL		Local domain is winner.
381  *	REMOTE		Remote domain is winner.
382  *	WAIT		Wait for remote to connect to our
383  *			master if his is different.
384  *	ERROR		An impossible condition.
385  *
386  * Index:
387  *	0 = Local
388  *	1 = Remote
389  */
390 static idn_master_select_t master_select_table[4][4] = {
391 	{				/* local	remote	*/
392 		MASTER_SELECT_VOTE,	/* NONE		NONE	*/
393 		MASTER_SELECT_CONNECT,	/* NONE		OTHER	*/
394 		MASTER_SELECT_LOCAL,	/* NONE		LOCAL	*/
395 		MASTER_SELECT_REMOTE	/* NONE		REMOTE	*/
396 	},
397 	{
398 		MASTER_SELECT_WAIT,	/* OTHER	NONE	*/
399 		MASTER_SELECT_CONNECT,	/* OTHER	OTHER	*/
400 		MASTER_SELECT_WAIT,	/* OTHER	LOCAL	*/
401 		MASTER_SELECT_WAIT	/* OTHER	REMOTE	*/
402 	},
403 	{
404 		MASTER_SELECT_LOCAL,	/* LOCAL	NONE	*/
405 		MASTER_SELECT_CONNECT,	/* LOCAL	OTHER	*/
406 		MASTER_SELECT_LOCAL,	/* LOCAL	LOCAL	*/
407 		MASTER_SELECT_VOTE_RCFG	/* LOCAL	REMOTE	*/
408 	},
409 	{
410 		MASTER_SELECT_REMOTE,	/* REMOTE	NONE	*/
411 		MASTER_SELECT_CONNECT,	/* REMOTE	OTHER	*/
412 		MASTER_SELECT_ERROR,	/* REMOTE	LOCAL	*/
413 		MASTER_SELECT_REMOTE	/* REMOTE	REMOTE	*/
414 	}
415 };
416 
417 void
418 idn_assign_cookie(int domid)
419 {
420 	static ushort_t	num = 0;
421 	ushort_t	cookie;
422 	procname_t	proc = "idn_assign_cookie";
423 
424 	if ((cookie = idn_domain[domid].dcookie_recv) != 0)
425 		return;
426 
427 	cookie = (ushort_t)(((uint64_t)&idn_domain[domid] >> 8) & 0xff);
428 	while ((cookie ^= num++ & 0xff) == 0)
429 		;
430 
431 	PR_PROTO("%s:%d: assigned RECV cookie 0x%x\n", proc, domid, cookie);
432 
433 	idn_domain[domid].dcookie_recv = cookie;
434 }
435 
436 void
437 idn_update_priority(int domid, int pri)
438 {
439 	idn_domain_t	*dp;
440 	procname_t	proc = "idn_update_priority";
441 
442 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
443 
444 	dp = &idn_domain[domid];
445 
446 	if (pri >= IDNVOTE_MINPRI) {
447 		dp->dvote.v.priority = pri & IDNVOTE_PRI_MASK;
448 
449 		PR_PROTO("%s:%d: SETTING PRIORITY to req(%d) "
450 		    "(localpri = 0x%x)\n",
451 		    proc, domid, pri, IDNVOTE_PRIVALUE(dp->dvote));
452 	} else {
453 		PR_PROTO("%s:%d: PRIORITIES UNCHANGED (pri = 0x%x)\n",
454 		    proc, domid, IDNVOTE_PRIVALUE(dp->dvote));
455 	}
456 }
457 
458 /*
459  * Initiate a link between the local domain and the remote domain
460  * containing the given cpuid.
461  */
462 int
463 idn_link(int domid, int cpuid, int pri, int waittime, idnsb_error_t *sep)
464 {
465 	int		rv;
466 	idn_domain_t	*dp;
467 	void		*opcookie;
468 	procname_t	proc = "idn_link";
469 
470 	if ((cpuid < 0) || (cpuid >= NCPU)) {
471 		cmn_err(CE_WARN,
472 		    "IDN: 201: (LINK) invalid CPU ID (%d)", cpuid);
473 		return (EINVAL);
474 	}
475 	if (waittime < 0) {
476 		cmn_err(CE_WARN,
477 		    "IDN: 202: (LINK) invalid time-out value (%d)",
478 		    waittime);
479 		return (EINVAL);
480 	}
481 	if (!VALID_DOMAINID(domid)) {
482 		cmn_err(CE_WARN,
483 		    "IDN: 203: (LINK) invalid domain ID (%d)",
484 		    domid);
485 		return (EINVAL);
486 	}
487 	if (domid == idn.localid)
488 		return (0);
489 
490 	IDN_SYNC_LOCK();
491 	IDN_DLOCK_EXCL(domid);
492 
493 	dp = &idn_domain[domid];
494 
495 	switch (dp->dstate) {
496 	case IDNDS_CLOSED:
497 		break;
498 
499 	case IDNDS_CONNECTED:
500 #ifdef DEBUG
501 		cmn_err(CE_NOTE,
502 		    "!IDN: domain %d (CPU ID %d) already connected",
503 		    domid, cpuid);
504 #endif /* DEBUG */
505 		IDN_DUNLOCK(domid);
506 		IDN_SYNC_UNLOCK();
507 		return (0);
508 
509 	default:
510 		cmn_err(CE_WARN,
511 		    "IDN: 204: domain %d state (%s) inappropriate",
512 		    domid, idnds_str[dp->dstate]);
513 		IDN_DUNLOCK(domid);
514 		IDN_SYNC_UNLOCK();
515 		return (EINVAL);
516 	}
517 
518 	rv = idn_open_domain(domid, cpuid, 0);
519 	if (rv != 0) {
520 		cmn_err(CE_WARN,
521 		    "IDN: 205: (%s) failed to open-domain(%d,%d)",
522 		    proc, domid, cpuid);
523 		IDN_DUNLOCK(domid);
524 		IDN_SYNC_UNLOCK();
525 		return (EIO);
526 	}
527 
528 
529 	IDN_DLOCK_EXCL(idn.localid);
530 	idn_update_priority(idn.localid, pri);
531 	IDN_DUNLOCK(idn.localid);
532 
533 	if (waittime > 0)
534 		opcookie = idn_init_op(IDNOP_CONNECTED, DOMAINSET(domid), sep);
535 
536 	(void) idn_connect(domid);
537 
538 	IDN_DUNLOCK(domid);
539 	IDN_SYNC_UNLOCK();
540 
541 	PR_PROTO("%s:%d: ALLOCATED idn_link(%d)\n", proc, domid, cpuid);
542 
543 	if (waittime > 0) {
544 		boardset_t	domset = 0;
545 		/*
546 		 * Well we've successfully allocated a domain id,
547 		 * but the link may not be fully established yet.
548 		 * Need to wait since it happens asynchronously.
549 		 */
550 		PR_PROTO("%s:%d: WAITING for op(%s) for (domset 0%x)...\n",
551 		    proc, domid, idnop_str[IDNOP_CONNECTED],
552 		    DOMAINSET(domid));
553 
554 		rv = idn_wait_op(opcookie, &domset, waittime);
555 	}
556 
557 #ifdef DEBUG
558 	if (rv == 0) {
559 		if (waittime > 0) {
560 			PR_PROTO("%s:%d: connect SUCCEEDED (cpu %d)\n",
561 			    proc, domid, cpuid);
562 		} else {
563 			PR_PROTO("%s:%d: connect KICKED OFF (cpu %d)\n",
564 			    proc, domid, cpuid);
565 		}
566 	} else {
567 		PR_PROTO("%s:%d: connect FAILED (cpu %d)\n",
568 		    proc, domid, cpuid);
569 	}
570 #endif /* DEBUG */
571 
572 	return (rv);
573 }
574 
575 /*
576  * Unlink the given domain from any domain cluster of
577  * which it might be a member.  Force indicates that domain
578  * should not go AWOL and if it's currently AWOL to close
579  * and remove it.
580  * IMPORTANT: If the (hard) force flag is set, the caller is
581  *	      assumed to GUARANTEE that the given domain will
582  *	      not attempt to communicate with the local domain
583  *	      in any manner.
584  */
585 int
586 idn_unlink(int domid, boardset_t idnset, idn_fin_t fintype,
587 		idn_finopt_t finopt, int waittime, idnsb_error_t *sep)
588 {
589 	int		rv = 0;
590 	domainset_t	domset;
591 	void		*opcookie;
592 	procname_t	proc = "idn_unlink";
593 
594 
595 	if (waittime < 0) {
596 		cmn_err(CE_WARN,
597 		    "IDN: 202: (UNLINK) invalid time-out value (%d)",
598 		    waittime);
599 		SET_IDNKERR_IDNERR(sep, IDNKERR_INVALID_WTIME);
600 		SET_IDNKERR_PARAM0(sep, waittime);
601 		return (EINVAL);
602 	}
603 	if (!VALID_DOMAINID(domid)) {
604 		cmn_err(CE_WARN,
605 		    "IDN: 203: (UNLINK) invalid domain ID (%d)",
606 		    domid);
607 		SET_IDNKERR_IDNERR(sep, IDNKERR_INVALID_DOMAIN);
608 		SET_IDNKERR_PARAM0(sep, domid);
609 		SET_IDNKERR_PARAM1(sep, -1);
610 		return (EINVAL);
611 	}
612 	if (idn.localid == IDN_NIL_DOMID) {
613 #ifdef DEBUG
614 		cmn_err(CE_NOTE,
615 		    "!IDN: %s: local domain not connected to an IDNnet",
616 		    proc);
617 #endif /* DEBUG */
618 		return (0);
619 	}
620 
621 	/*
622 	 * Lock ordering protocols requires that we grab the
623 	 * global lock _before_ the local domain's lock.
624 	 * However, non-local domains must have their lock
625 	 * grabbed _before_ the global lock.
626 	 */
627 	IDN_SYNC_LOCK();
628 	IDN_GLOCK_EXCL();
629 	domset = idn.domset.ds_trans_on | idn.domset.ds_trans_off;
630 	if ((idn.state == IDNGS_OFFLINE) && !domset) {
631 #ifdef DEBUG
632 		cmn_err(CE_WARN,
633 		    "!IDN: %s: local domain not connected to an IDNnet",
634 		    proc);
635 #endif /* DEBUG */
636 		IDN_GUNLOCK();
637 		IDN_SYNC_UNLOCK();
638 		return (0);
639 	}
640 
641 	if ((domid == IDN_NIL_DOMID) || (domid == idn.localid)) {
642 		domid = idn.localid;
643 		IDN_GSTATE_TRANSITION(IDNGS_DISCONNECT);
644 		IDN_SET_NEW_MASTERID(IDN_NIL_DOMID);
645 		domset = DOMAINSET_ALL;
646 		DOMAINSET_DEL(domset, idn.localid);
647 	} else {
648 		domset = DOMAINSET(domid);
649 	}
650 	IDN_GUNLOCK();
651 
652 	if (waittime > 0)
653 		opcookie = idn_init_op(IDNOP_DISCONNECTED, domset, sep);
654 
655 	idn_unlink_domainset(domset, fintype, IDNFIN_ARG_NONE, finopt, idnset);
656 
657 	IDN_SYNC_UNLOCK();
658 
659 	if (waittime > 0) {
660 		/*
661 		 * Well the unlink has successfully kicked off.
662 		 * Since process is asynchronous we need to wait
663 		 * for it to complete.
664 		 */
665 		PR_PROTO("%s:%d: WAITING for op(%s) for (domset 0%x)...\n",
666 		    proc, domid, idnop_str[IDNOP_DISCONNECTED],
667 		    domset);
668 
669 		rv = idn_wait_op(opcookie, &domset, waittime);
670 	}
671 
672 	if (rv == 0) {
673 		if (waittime > 0) {
674 			PR_PROTO("%s:%d: disconnect SUCCEEDED\n",
675 			    proc, domid);
676 		} else {
677 			PR_PROTO("%s:%d: disconnect KICKED OFF\n",
678 			    proc, domid);
679 		}
680 	} else {
681 		PR_PROTO("%s:%d: disconnect FAILED\n", proc, domid);
682 	}
683 
684 	return (rv);
685 }
686 
687 static void
688 idn_unlink_domainset(domainset_t domset, idn_fin_t fintype,
689 			idn_finarg_t finarg, idn_finopt_t finopt,
690 			boardset_t idnset)
691 {
692 	int		d;
693 	domainset_t	offset;
694 	procname_t	proc = "idn_unlink_domainset";
695 
696 	ASSERT(IDN_SYNC_IS_LOCKED());
697 
698 	/*
699 	 * Determine subset for which we have
700 	 * no active connections.
701 	 */
702 	offset = domset & ~(idn.domset.ds_trans_on |
703 	    idn.domset.ds_connected |
704 	    idn.domset.ds_trans_off |
705 	    idn.domset.ds_relink);
706 	/*
707 	 * Determine subset that are really candidates.
708 	 * Note that we include those already down the path
709 	 * since it's possible a request came in to upgrade
710 	 * their fintype (e.g. NORMAL->FORCE_SOFT).
711 	 */
712 	domset &= ~offset;
713 
714 	if (offset)
715 		idn_update_op(IDNOP_DISCONNECTED, offset, NULL);
716 
717 	IDN_GLOCK_EXCL();
718 	if ((finopt == IDNFIN_OPT_RELINK) && (idn.state != IDNGS_DISCONNECT)) {
719 		/*
720 		 * Don't add domains already transitioning off.
721 		 * If they caught on an earlier Reconfig wave then
722 		 * they'll already be in ds_relink anyway.  Otherwise,
723 		 * once a domain is transition off we can't upgrade
724 		 * him to a RELINK.
725 		 */
726 #ifdef DEBUG
727 		if (idn.domset.ds_hitlist & domset) {
728 			PR_HITLIST("%s: domset=%x, hitlist=%x, trans_off=%x "
729 			    "-> relink = %x -> %x\n",
730 			    proc, domset, idn.domset.ds_hitlist,
731 			    idn.domset.ds_relink, idn.domset.ds_trans_off,
732 			    idn.domset.ds_relink |
733 			    (domset & ~idn.domset.ds_trans_off));
734 		}
735 #endif /* DEBUG */
736 
737 		domset &= ~idn.domset.ds_trans_off;
738 		idn.domset.ds_relink |= domset;
739 	} else {
740 		idn.domset.ds_relink &= ~domset;
741 	}
742 	/*
743 	 * Update the ds_trans_on/off so we don't waste
744 	 * time talking to these folks.
745 	 */
746 	idn.domset.ds_trans_on  &= ~domset;
747 	idn.domset.ds_trans_off |= domset;
748 
749 	if (domset == 0) {
750 		if ((idn.domset.ds_trans_on |
751 		    idn.domset.ds_connected |
752 		    idn.domset.ds_trans_off |
753 		    idn.domset.ds_relink) == 0) {
754 			PR_HITLIST("%s:%x: HITLIST %x -> 0\n",
755 			    proc, domset, idn.domset.ds_hitlist);
756 			idn.domset.ds_hitlist = 0;
757 			IDN_GSTATE_TRANSITION(IDNGS_OFFLINE);
758 		}
759 		IDN_GUNLOCK();
760 		return;
761 	}
762 	IDN_GUNLOCK();
763 
764 	for (d = 0; d < MAX_DOMAINS; d++) {
765 		idn_domain_t	*dp;
766 		idn_fin_t	ftype;
767 
768 		if (!DOMAIN_IN_SET(domset, d))
769 			continue;
770 
771 		dp = &idn_domain[d];
772 		IDN_DLOCK_EXCL(d);
773 		IDN_HISTORY_LOG(IDNH_RELINK, d, dp->dstate,
774 		    idn.domset.ds_relink);
775 		ftype = fintype;
776 		if ((dp->dcpu != IDN_NIL_DCPU) && dp->dhw.dh_boardset) {
777 			/*
778 			 * If domain is not in the IDNSET passed
779 			 * down then we need to upgrade this to
780 			 * hard-force in order to prevent possible
781 			 * system failures (arbstop).  This is simply
782 			 * extra protection beyond that checked by
783 			 * the SSP.  IDNSET contains the set of boards
784 			 * that have a "link" to the local domain,
785 			 * including the SMD regs.
786 			 */
787 			if ((idnset & dp->dhw.dh_boardset) == 0) {
788 				PR_PROTO("%s:%d: boardset 0x%x "
789 				    "NOT in IDNSET 0x%x\n",
790 				    proc, d, dp->dhw.dh_boardset,
791 				    idnset);
792 				if (ftype != IDNFIN_FORCE_HARD)
793 					cmn_err(CE_NOTE,
794 					    "!IDN: 222: no IDN linkage "
795 					    "found (b=0x%x, i=0x%x) "
796 					    "upgrading unlink %s to %s",
797 					    dp->dhw.dh_boardset,
798 					    idnset, idnfin_str[ftype],
799 					    idnfin_str[IDNFIN_FORCE_HARD]);
800 
801 				ftype = IDNFIN_FORCE_HARD;
802 			} else {
803 				PR_PROTO("%s:%d: boardset 0x%x "
804 				    "FOUND in IDNSET 0x%x\n",
805 				    proc, d, dp->dhw.dh_boardset,
806 				    idnset);
807 			}
808 		}
809 		(void) idn_disconnect(d, ftype, finarg, IDNDS_SYNC_TYPE(dp));
810 		IDN_DUNLOCK(d);
811 	}
812 }
813 
814 /*
815  * Return w/locks held.
816  */
817 static int
818 idn_connect(int domid)
819 {
820 	idn_xdcargs_t	xargs;
821 	idn_domain_t	*dp = &idn_domain[domid];
822 	procname_t	proc = "idn_connect";
823 
824 	ASSERT(IDN_SYNC_IS_LOCKED());
825 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
826 
827 	ASSERT(dp->dcpu != IDN_NIL_DCPU);
828 
829 	if (dp->dstate != IDNDS_CLOSED) {
830 		if (DOMAIN_IN_SET(idn.domset.ds_trans_on |
831 		    idn.domset.ds_connected, domid)) {
832 			PR_PROTO("%s:%d: already connected or "
833 			    "in-progress\n", proc, domid);
834 		} else {
835 			PR_PROTO("%s:%d: current state (%s) != "
836 			    "CLOSED\n", proc, domid,
837 			    idnds_str[dp->dstate]);
838 		}
839 		return (-1);
840 	}
841 
842 	ASSERT(!DOMAIN_IN_SET(idn.domset.ds_connected, domid));
843 	ASSERT(!DOMAIN_IN_SET(idn.domset.ds_trans_off, domid));
844 
845 	dp->dxp = &xphase_nego;
846 	IDN_XSTATE_TRANSITION(dp, IDNXS_PEND);
847 
848 	(void) idn_xphase_transition(domid, NULL, xargs);
849 
850 	return (0);
851 }
852 
853 /*
854  * Return w/locks held.
855  */
856 static int
857 idn_disconnect(int domid, idn_fin_t fintype, idn_finarg_t finarg,
858     idn_finsync_t finsync)
859 {
860 	int		new_masterid, new_cpuid = IDN_NIL_DCPU;
861 	uint_t		token;
862 	uint_t		finmaster;
863 	idn_xdcargs_t	xargs;
864 	idn_finopt_t	finopt;
865 	idn_domain_t	*dp = &idn_domain[domid];
866 	procname_t	proc = "idn_disconnect";
867 
868 	ASSERT(IDN_SYNC_IS_LOCKED());
869 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
870 
871 	if (dp->dstate == IDNDS_CLOSED) {
872 		PR_PROTO("%s:%d: already CLOSED\n", proc, domid);
873 		idn_update_op(IDNOP_DISCONNECTED, DOMAINSET(domid), NULL);
874 		return (-1);
875 	}
876 
877 	/*
878 	 * Terminate any outstanding commands that were
879 	 * targeted towards this domain.
880 	 */
881 	idn_terminate_cmd(domid, ECANCELED);
882 
883 	/*
884 	 * Terminate any and all retries that may have
885 	 * outstanding for this domain.
886 	 */
887 	token = IDN_RETRY_TOKEN(domid, IDN_RETRY_TYPEALL);
888 	(void) idn_retry_terminate(token);
889 
890 	/*
891 	 * Stop all outstanding message timers for
892 	 * this guy.
893 	 */
894 	IDN_MSGTIMER_STOP(domid, 0, 0);
895 
896 	dp->dxp = &xphase_fin;
897 	IDN_XSTATE_TRANSITION(dp, IDNXS_PEND);
898 	if ((int)dp->dfin < (int)fintype) {
899 		/*
900 		 * You can only upgrade a fin type.
901 		 * We don't allow it to be downgraded
902 		 * as it's too dangerous since some
903 		 * state may have been blown away while
904 		 * we were fin'ing at a higher level.
905 		 */
906 		IDN_FSTATE_TRANSITION(dp, fintype);
907 	}
908 
909 	dp->dfin_sync = finsync;
910 	PR_PROTO("%s:%d: disconnect synchronously = %s\n",
911 	    proc, domid, (finsync == IDNFIN_SYNC_OFF) ? "OFF" :
912 	    (finsync == IDNFIN_SYNC_NO) ? "NO" : "YES");
913 
914 	IDN_GLOCK_SHARED();
915 	if (DOMAIN_IN_SET(idn.domset.ds_relink, domid) &&
916 	    (idn.state != IDNGS_DISCONNECT)) {
917 		finopt = IDNFIN_OPT_RELINK;
918 	} else {
919 		finopt = IDNFIN_OPT_UNLINK;
920 		PR_HITLIST("%s:%d: HITLIST %x -> %x\n",
921 		    proc, domid, idn.domset.ds_hitlist,
922 		    idn.domset.ds_hitlist | DOMAINSET(domid));
923 		DOMAINSET_ADD(idn.domset.ds_hitlist, domid);
924 	}
925 
926 	CLR_XARGS(xargs);
927 	SET_XARGS_FIN_TYPE(xargs, dp->dfin);
928 	SET_XARGS_FIN_ARG(xargs, finarg);
929 	SET_XARGS_FIN_OPT(xargs, finopt);
930 	SET_XARGS_FIN_DOMSET(xargs, 0);		/* unused when msg = 0 */
931 	new_masterid = IDN_GET_NEW_MASTERID();
932 	IDN_GUNLOCK();
933 	if (new_masterid != IDN_NIL_DOMID)
934 		new_cpuid = idn_domain[new_masterid].dcpu;
935 	finmaster = MAKE_FIN_MASTER(new_masterid, new_cpuid);
936 	SET_XARGS_FIN_MASTER(xargs, finmaster);
937 
938 	(void) idn_xphase_transition(domid, NULL, xargs);
939 
940 	return (0);
941 }
942 
943 static int
944 idn_next_xstate(idn_xstate_t o_xstate, int err, uint_t msg)
945 {
946 	int		index;
947 	procname_t	proc = "idn_next_xstate";
948 
949 	ASSERT(((int)o_xstate >= 0) && ((int)o_xstate <= 4));
950 
951 	if (!msg)
952 		index = 0;
953 	else if ((msg & IDNP_MSGTYPE_MASK) == 0)
954 		index = (msg & IDNP_ACK) ? 3 : (msg & IDNP_NACK) ? 4 : -1;
955 	else
956 		index = (msg & IDNP_ACK) ? 2 :
957 		    !(msg & IDNP_ACKNACK_MASK) ? 1 : -1;
958 
959 	if (index == -1) {
960 		STRING(str);
961 
962 		INUM2STR(msg, str);
963 		PR_PROTO("%s: (msg = 0x%x(%s))\n", proc, msg, str);
964 		return (IDNXS_NIL);
965 	}
966 
967 	if (err == -1) {
968 		int	n_xstate;
969 		/*
970 		 * Caller is just interested in querying is this
971 		 * is a valid message to receive in the current
972 		 * xstate.  A return value of IDNXS_NIL indicates
973 		 * that it's not.  A return value of non-IDNXS_NIL
974 		 * indicates it's cool.  An invalid message is
975 		 * determined by both err & !err states being IDNXS_NIL.
976 		 */
977 		n_xstate = idnxs_state_table[(int)o_xstate][index][0];
978 		if (n_xstate != IDNXS_NIL)
979 			return (n_xstate);
980 		else
981 			return (idnxs_state_table[(int)o_xstate][index][1]);
982 	} else {
983 		return (idnxs_state_table[(int)o_xstate][index][err ? 1 : 0]);
984 	}
985 }
986 
987 static int
988 idn_select_candidate(domainset_t master_set)
989 {
990 	int		d, best_id = IDN_NIL_DOMID;
991 	uint_t		best_vote = 0;
992 	idn_domain_t	*dp;
993 	procname_t	proc = "idn_select_candidate";
994 
995 	ASSERT(IDN_SYNC_IS_LOCKED());
996 
997 	if (master_set == 0) {
998 		PR_PROTO("%s: %x -> %d\n", proc, master_set, IDN_NIL_DOMID);
999 		return (IDN_NIL_DOMID);
1000 	}
1001 
1002 	for (d = 0; d < MAX_DOMAINS; d++) {
1003 		uint_t		vote;
1004 		idn_vote_t	v;
1005 
1006 		if (!DOMAIN_IN_SET(master_set, d))
1007 			continue;
1008 
1009 		dp = &idn_domain[d];
1010 
1011 		if ((dp->domid == IDN_NIL_DOMID) ||
1012 		    (dp->dcpu == IDN_NIL_DCPU) ||
1013 		    ((v.ticket = dp->dvote.ticket) == 0))
1014 			continue;
1015 
1016 		vote = IDNVOTE_ELECT(v);
1017 
1018 		if (vote > best_vote) {
1019 			best_vote = vote;
1020 			best_id = d;
1021 		}
1022 	}
1023 
1024 	PR_PROTO("%s: %x -> %d\n", proc, master_set, best_id);
1025 
1026 	return (best_id);
1027 }
1028 
1029 /*
1030  * If a non-zero value is returned then GLOCK will have been dropped.
1031  * Otherwise, routine returns with all incoming locks still held.
1032  */
1033 static int
1034 idn_select_master(int domid, int rmasterid, int rcpuid)
1035 {
1036 	char		*sel;
1037 	int		lmasterid, masterid;
1038 	int		do_reconfig = 0;
1039 	int		lindex, rindex;
1040 	idn_domain_t	*ldp, *rdp;
1041 	uint_t		rvote, lvote;
1042 	idn_master_select_t	select;
1043 	procname_t	proc = "idn_select_master";
1044 
1045 	ASSERT(IDN_SYNC_IS_LOCKED());
1046 	ASSERT(IDN_GLOCK_IS_EXCL());
1047 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
1048 
1049 	PR_PROTO("%s:%d: lmasterid = %d, rmasterid = %d, rcpuid = %d\n",
1050 	    proc, domid, IDN_GET_MASTERID(), rmasterid, rcpuid);
1051 
1052 	IDN_DLOCK_EXCL(idn.localid);
1053 
1054 	ldp = &idn_domain[idn.localid];
1055 	rdp = &idn_domain[domid];
1056 
1057 	/*
1058 	 * Clear master bits since mastership is derived from
1059 	 * other information (local/remote idn.masterid/idn.new_masterid)
1060 	 * and we don't want the vote master bit to confuse matters.
1061 	 */
1062 	lvote = IDNVOTE_ELECT(ldp->dvote);
1063 	rvote = IDNVOTE_ELECT(rdp->dvote);
1064 
1065 	lmasterid = IDN_GET_MASTERID();
1066 
1067 	lindex = (lmasterid == IDN_NIL_DOMID) ? MASTER_IS_NONE :
1068 	    (lmasterid == idn.localid) ? MASTER_IS_LOCAL :
1069 	    (lmasterid == domid) ? MASTER_IS_REMOTE :
1070 	    MASTER_IS_OTHER;
1071 
1072 	rindex = (rmasterid == IDN_NIL_DOMID) ? MASTER_IS_NONE :
1073 	    (rmasterid == domid) ? MASTER_IS_REMOTE :
1074 	    (rmasterid == idn.localid) ? MASTER_IS_LOCAL :
1075 	    MASTER_IS_OTHER;
1076 
1077 	select = master_select_table[lindex][rindex];
1078 
1079 	masterid = IDN_NIL_DOMID;
1080 
1081 	/*
1082 	 * Each case is responsible for dropping DLOCK(localid)
1083 	 * and GLOCK if it doesn't select a master, unless a
1084 	 * reconfig is necessary.
1085 	 */
1086 	switch (select) {
1087 	case MASTER_SELECT_VOTE_RCFG:
1088 		sel = "VOTE_RECONFIG";
1089 		if (lvote > rvote) {
1090 			/*
1091 			 * If the local domain is the winner then remote
1092 			 * domain will have to Reconfig.  We'll continue
1093 			 * through the connection process anyway.  The
1094 			 * remote domains will tell us to back-off while
1095 			 * Reconfigs, but that's okay as we'll keep retrying.
1096 			 */
1097 			masterid = idn.localid;
1098 		} else if (lvote < rvote) {
1099 			do_reconfig = 1;
1100 			/*
1101 			 * GLOCK will get dropped once reconfig
1102 			 * is kicked off.
1103 			 */
1104 		} else {
1105 			cmn_err(CE_WARN,
1106 			    "IDN: 206: cannot link domains "
1107 			    "with equal votes (L(%d),R(%d),0x%x)",
1108 			    idn.localid, domid, rvote);
1109 			IDN_GUNLOCK();
1110 		}
1111 		IDN_DUNLOCK(idn.localid);
1112 		break;
1113 
1114 	case MASTER_SELECT_VOTE:
1115 		sel = "VOTE";
1116 		if (lvote > rvote) {
1117 			masterid = idn.localid;
1118 			ldp->dvote.v.master = 1;
1119 			rdp->dvote.v.master = 0;
1120 		} else if (lvote < rvote) {
1121 			masterid = domid;
1122 			ldp->dvote.v.master = 0;
1123 			rdp->dvote.v.master = 1;
1124 		} else {
1125 			cmn_err(CE_WARN,
1126 			    "IDN: 206: cannot link domains "
1127 			    "with equal votes (L(%d),R(%d),0x%x)",
1128 			    idn.localid, domid, rvote);
1129 		}
1130 		ASSERT(IDN_GET_MASTERID() == IDN_NIL_DOMID);
1131 		if (masterid != IDN_NIL_DOMID) {
1132 			IDN_SET_MASTERID(masterid);
1133 			IDN_SET_NEW_MASTERID(IDN_NIL_DOMID);
1134 		} else {
1135 			IDN_GUNLOCK();
1136 		}
1137 		IDN_DUNLOCK(idn.localid);
1138 		break;
1139 
1140 	case MASTER_SELECT_REMOTE:
1141 		sel = "REMOTE";
1142 		masterid = domid;
1143 		if (IDN_GET_MASTERID() == IDN_NIL_DOMID) {
1144 			IDN_SET_MASTERID(masterid);
1145 			IDN_SET_NEW_MASTERID(IDN_NIL_DOMID);
1146 			ldp->dvote.v.master = 0;
1147 			rdp->dvote.v.master = 1;
1148 		}
1149 		ASSERT(IDN_GET_MASTERID() == domid);
1150 		IDN_DUNLOCK(idn.localid);
1151 		break;
1152 
1153 	case MASTER_SELECT_LOCAL:
1154 		sel = "LOCAL";
1155 		masterid = idn.localid;
1156 		if (IDN_GET_MASTERID() == IDN_NIL_DOMID) {
1157 			IDN_SET_MASTERID(masterid);
1158 			IDN_SET_NEW_MASTERID(IDN_NIL_DOMID);
1159 			ldp->dvote.v.master = 1;
1160 			rdp->dvote.v.master = 0;
1161 		}
1162 		ASSERT(IDN_GET_MASTERID() == idn.localid);
1163 		IDN_DUNLOCK(idn.localid);
1164 		break;
1165 
1166 	case MASTER_SELECT_CONNECT:
1167 		sel = "CONNECT";
1168 		if (rmasterid == lmasterid) {
1169 			/*
1170 			 * Local and remote have same master,
1171 			 * let him come onboard.
1172 			 */
1173 			masterid = lmasterid;
1174 			IDN_DUNLOCK(idn.localid);
1175 
1176 		} else {
1177 			int	rv;
1178 
1179 			IDN_DUNLOCK(idn.localid);
1180 			IDN_GUNLOCK();
1181 			IDN_DLOCK_EXCL(rmasterid);
1182 			PR_PROTO("%s:%d: attempting connect w/remote "
1183 			    "master %d\n",
1184 			    proc, domid, rmasterid);
1185 			rv = idn_open_domain(rmasterid, rcpuid, 0);
1186 			if (rv == 0) {
1187 				(void) idn_connect(rmasterid);
1188 			} else if (rv < 0) {
1189 				cmn_err(CE_WARN,
1190 				    "IDN: 205: (%s) failed to "
1191 				    "open-domain(%d,%d)",
1192 				    proc, rmasterid, rcpuid);
1193 			} else {
1194 				/*
1195 				 * Must already have a connection going.
1196 				 */
1197 				PR_PROTO("%s:%d: failed "
1198 				    "idn_open_domain(%d,%d,0) "
1199 				    "(rv = %d)\n",
1200 				    proc, domid, rmasterid,
1201 				    rcpuid, rv);
1202 			}
1203 			IDN_DUNLOCK(rmasterid);
1204 		}
1205 		break;
1206 
1207 	case MASTER_SELECT_WAIT:
1208 		sel = "WAIT";
1209 		/*
1210 		 * If the remote domain has the same master as the local
1211 		 * domain then there's no need to wait.
1212 		 */
1213 		if (rmasterid == lmasterid) {
1214 			masterid = lmasterid;
1215 		} else {
1216 			IDN_GUNLOCK();
1217 		}
1218 		IDN_DUNLOCK(idn.localid);
1219 		break;
1220 
1221 	case MASTER_SELECT_ERROR:
1222 		sel = "ERROR";
1223 		/*
1224 		 * Hit impossible condition.
1225 		 */
1226 		cmn_err(CE_WARN,
1227 		    "IDN: 207: local/remote master-id conflict "
1228 		    "(%d.lmasterid = %d, %d.rmasterid = %d)",
1229 		    idn.localid, lmasterid, domid, rmasterid);
1230 		IDN_GUNLOCK();
1231 		IDN_DUNLOCK(idn.localid);
1232 		break;
1233 
1234 	default:
1235 		cmn_err(CE_WARN,
1236 		    "IDN: 208: %s: unknown case (%d)",
1237 		    proc, (int)select);
1238 		IDN_GUNLOCK();
1239 		IDN_DUNLOCK(idn.localid);
1240 		ASSERT(0);
1241 		break;
1242 	}
1243 
1244 	if (masterid == IDN_NIL_DOMID) {
1245 		PR_PROTO("%s:%d: NO MASTER SELECTED (rmstr=%d) sel=%s\n",
1246 		    proc, domid, rmasterid, sel);
1247 	} else {
1248 		PR_PROTO("%s:%d: MASTER SELECTED = %d (%s)\n",
1249 		    proc, domid, masterid,
1250 		    (masterid == idn.localid) ? "LOCAL" :
1251 		    (masterid == domid) ? "REMOTE" : "OTHER");
1252 	}
1253 
1254 	if (do_reconfig) {
1255 		domainset_t	dis_set;
1256 
1257 		/*
1258 		 * Local domain already has a master.
1259 		 * Need to dismantle all connections
1260 		 * and reestablish one with new master.
1261 		 */
1262 		IDN_GKSTAT_GLOBAL_EVENT(gk_reconfigs, gk_reconfig_last);
1263 
1264 		PR_PROTO("%s:%d: RECONFIG new masterid = %d\n",
1265 		    proc, domid, domid);
1266 
1267 		IDN_GSTATE_TRANSITION(IDNGS_RECONFIG);
1268 		IDN_SET_NEW_MASTERID(domid);
1269 		IDN_GUNLOCK();
1270 
1271 		dis_set = idn.domset.ds_trans_on | idn.domset.ds_connected;
1272 		DOMAINSET_DEL(dis_set, domid);
1273 
1274 		idn_unlink_domainset(dis_set, IDNFIN_NORMAL, IDNFIN_ARG_NONE,
1275 		    IDNFIN_OPT_RELINK, BOARDSET_ALL);
1276 	}
1277 
1278 	return ((masterid == IDN_NIL_DOMID) ? -1 : 0);
1279 }
1280 
1281 /*ARGSUSED1*/
1282 static void
1283 idn_retry_query(uint_t token, void *arg)
1284 {
1285 	idn_retry_t	rtype = IDN_RETRY_TOKEN2TYPE(token);
1286 	int		d, domid = IDN_RETRY_TOKEN2DOMID(token);
1287 	idn_domain_t	*dp = &idn_domain[domid];
1288 	idn_synccmd_t	sync_cmd;
1289 	domainset_t	query_set, my_ready_set;
1290 	procname_t	proc = "idn_retry_query";
1291 
1292 	IDN_SYNC_LOCK();
1293 	IDN_DLOCK_EXCL(domid);
1294 
1295 	switch (rtype) {
1296 	case IDNRETRY_CONQ:
1297 		sync_cmd = IDNSYNC_CONNECT;
1298 		my_ready_set = idn.domset.ds_ready_on | idn.domset.ds_connected;
1299 		my_ready_set &= ~idn.domset.ds_trans_off;
1300 		DOMAINSET_ADD(my_ready_set, idn.localid);
1301 		break;
1302 
1303 	case IDNRETRY_FINQ:
1304 		sync_cmd = IDNSYNC_DISCONNECT;
1305 		my_ready_set = idn.domset.ds_ready_off |
1306 		    ~idn.domset.ds_connected;
1307 		break;
1308 
1309 	default:
1310 		IDN_DUNLOCK(domid);
1311 		IDN_SYNC_UNLOCK();
1312 		return;
1313 	}
1314 
1315 	if (dp->dsync.s_cmd == sync_cmd)
1316 		my_ready_set |= dp->dsync.s_set_rdy;
1317 
1318 	query_set = idn_sync_register(domid, sync_cmd, 0, IDNSYNC_REG_QUERY);
1319 
1320 	PR_PROTO("%s:%d: query_set = 0x%x\n", proc, domid, query_set);
1321 
1322 	if (query_set == 0) {
1323 		IDN_DUNLOCK(domid);
1324 		IDN_SYNC_UNLOCK();
1325 		return;
1326 	}
1327 
1328 	for (d = 0; d < MAX_DOMAINS; d++) {
1329 		if (!DOMAIN_IN_SET(query_set, d))
1330 			continue;
1331 
1332 		dp = &idn_domain[d];
1333 		if (d != domid)
1334 			IDN_DLOCK_EXCL(d);
1335 
1336 		if ((dp->dsync.s_cmd == sync_cmd) ||
1337 		    (!dp->dcookie_send &&
1338 		    (rtype == IDNRETRY_CONQ))) {
1339 			if (d != domid)
1340 				IDN_DUNLOCK(d);
1341 			continue;
1342 		}
1343 
1344 		IDN_SYNC_QUERY_UPDATE(domid, d);
1345 
1346 		if (rtype == IDNRETRY_CONQ)
1347 			(void) idn_send_con(d, NULL, IDNCON_QUERY,
1348 			    my_ready_set);
1349 		else
1350 			(void) idn_send_fin(d, NULL, IDNFIN_QUERY,
1351 			    IDNFIN_ARG_NONE, IDNFIN_OPT_NONE, my_ready_set,
1352 			    NIL_FIN_MASTER);
1353 		if (d != domid)
1354 			IDN_DUNLOCK(d);
1355 	}
1356 
1357 	IDN_DUNLOCK(domid);
1358 	IDN_SYNC_UNLOCK();
1359 }
1360 
1361 static int
1362 idn_send_nego(int domid, idn_msgtype_t *mtp, domainset_t conset)
1363 {
1364 	idn_domain_t	*ldp, *dp;
1365 	int		d, masterid;
1366 	uint_t		dmask;
1367 	uint_t		acknack;
1368 	uint_t		ticket;
1369 	idnneg_dset_t	dset;
1370 	idn_msgtype_t	mt;
1371 	procname_t	proc = "idn_send_nego";
1372 
1373 	ASSERT(IDN_SYNC_IS_LOCKED());
1374 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
1375 
1376 	if (mtp) {
1377 		acknack = mtp->mt_mtype & IDNP_ACKNACK_MASK;
1378 		mt.mt_mtype = mtp->mt_mtype;
1379 		mt.mt_atype = mtp->mt_atype;
1380 		mt.mt_cookie = mtp->mt_cookie;
1381 	} else {
1382 		acknack = 0;
1383 		mt.mt_mtype = IDNP_NEGO;
1384 		mt.mt_atype = 0;
1385 		mt.mt_cookie = IDN_TIMER_PUBLIC_COOKIE;
1386 	}
1387 
1388 	IDN_GLOCK_SHARED();
1389 
1390 	dp = &idn_domain[domid];
1391 	ldp = &idn_domain[idn.localid];
1392 
1393 	if ((idn.state == IDNGS_RECONFIG) ||
1394 	    ((masterid = IDN_GET_MASTERID()) == IDN_NIL_DOMID)) {
1395 		masterid = IDN_GET_NEW_MASTERID();
1396 		if ((masterid == idn.localid) || (masterid == domid)) {
1397 			/*
1398 			 * We only send the new-master "hint" to
1399 			 * "other" domains.  If the new-master is
1400 			 * ourself or we're talking to the new-master
1401 			 * then we need to be accurate about our
1402 			 * real master so that the correct master
1403 			 * is selected.
1404 			 */
1405 			masterid = IDN_NIL_DOMID;
1406 		}
1407 	}
1408 
1409 	DOMAINSET_DEL(conset, idn.localid);
1410 	DOMAINSET_DEL(conset, domid);
1411 	/*
1412 	 * Exclude domains from conset that are on
1413 	 * remote domain's hitlist.  He's not interested
1414 	 * in hearing about them.  SSP is probably requesting
1415 	 * such domains be unlinked - will eventually get to
1416 	 * local domain.
1417 	 */
1418 	conset &= ~idn.domset.ds_hitlist;
1419 	if ((masterid != IDN_NIL_DOMID) &&
1420 	    DOMAIN_IN_SET(idn.domset.ds_hitlist, masterid)) {
1421 		PR_PROTO("%s:%d: masterid(%d) on hitlist(0x%x) -> -1\n",
1422 		    proc, domid, masterid, idn.domset.ds_hitlist);
1423 		/*
1424 		 * Yikes, our chosen master is on the hitlist!
1425 		 */
1426 		masterid = IDN_NIL_DOMID;
1427 	}
1428 
1429 	dmask = IDNNEG_DSET_MYMASK();
1430 	IDNNEG_DSET_INIT(dset, dmask);
1431 	for (d = 0; d < MAX_DOMAINS; d++) {
1432 		int	cpuid;
1433 
1434 		if (!DOMAIN_IN_SET(conset, d))
1435 			continue;
1436 
1437 		if ((cpuid = idn_domain[d].dcpu) == IDN_NIL_DCPU) {
1438 			ASSERT(d != masterid);
1439 			continue;
1440 		}
1441 
1442 		IDNNEG_DSET_SET(dset, d, cpuid, dmask);
1443 	}
1444 	IDNNEG_DSET_SET_MASTER(dset, domid, masterid);
1445 	ASSERT((masterid != IDN_NIL_DOMID) ?
1446 	    (idn_domain[masterid].dcpu != IDN_NIL_DCPU) : 1);
1447 	IDN_GUNLOCK();
1448 
1449 	IDN_DLOCK_SHARED(idn.localid);
1450 	ticket = IDNVOTE_BASICS(ldp->dvote);
1451 	/*
1452 	 * We just want to send basic vote components without an
1453 	 * indication of mastership (master bit) since that's primarily
1454 	 * for local domain's usage.  There is more correct master
1455 	 * indications in the DSET.  Recall that if we were in a
1456 	 * Reconfig we would have transmitted the "new_masterid"
1457 	 * which might conflict with the local domain's vote.v.master
1458 	 * bit if he was originally the master prior to the Reconfig.
1459 	 */
1460 
1461 	PR_PROTO("%s:%d: sending nego%sto (cpu %d) "
1462 	    "[v=0x%x, cs=0x%x, mstr=%d]\n",
1463 	    proc, domid,
1464 	    (acknack & IDNP_ACK) ? "+ack " :
1465 	    (acknack & IDNP_NACK) ? "+nack " : " ",
1466 	    dp->dcpu, ticket, conset, masterid);
1467 
1468 	IDN_MSGTIMER_START(domid, IDNP_NEGO, 0,
1469 	    idn_msg_waittime[IDNP_NEGO], &mt.mt_cookie);
1470 
1471 	IDNXDC(domid, &mt, ticket, dset[0], dset[1], dset[2]);
1472 
1473 	IDN_DUNLOCK(idn.localid);
1474 
1475 	return (0);
1476 }
1477 
1478 static int
1479 idn_recv_nego(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs,
1480     ushort_t dcookie)
1481 {
1482 	uint_t		msg = mtp->mt_mtype;
1483 	idn_msgtype_t	mt;
1484 	idn_domain_t	*dp = &idn_domain[domid];
1485 	idn_xdcargs_t	nargs;
1486 	procname_t	proc = "idn_recv_nego";
1487 
1488 	ASSERT(IDN_SYNC_IS_LOCKED());
1489 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
1490 
1491 	mt.mt_cookie = mtp->mt_cookie;
1492 
1493 #ifdef DEBUG
1494 	if (DOMAIN_IN_SET(idn.domset.ds_hitlist, domid)) {
1495 		PR_HITLIST("%s:%d: dcpu=%d, dstate=%s, msg=%x, "
1496 		    "hitlist=%x\n",
1497 		    proc, domid, dp->dcpu, idnds_str[dp->dstate],
1498 		    msg, idn.domset.ds_hitlist);
1499 	}
1500 #endif /* DEBUG */
1501 
1502 	if (dp->dcpu == IDN_NIL_DCPU) {
1503 		int		cpuid;
1504 		uint_t		ticket;
1505 		/*
1506 		 * Brandnew link.  Need to open a new domain entry.
1507 		 */
1508 		ticket = GET_XARGS_NEGO_TICKET(xargs);
1509 		cpuid = dp->dcpu_last;
1510 		ASSERT(VALID_CPUID(cpuid));
1511 
1512 		if (idn_open_domain(domid, cpuid, ticket) != 0) {
1513 			PR_PROTO("%s:%d: FAILED to open doamin "
1514 			    "(ticket = 0x%x)\n",
1515 			    proc, domid, ticket);
1516 			return (-1);
1517 		}
1518 	}
1519 
1520 	if ((msg & IDNP_MSGTYPE_MASK) == IDNP_NEGO) {
1521 		PR_PROTO("%s:%d: assigned SEND cookie 0x%x\n",
1522 		    proc, domid, dcookie);
1523 		dp->dcookie_send = dcookie;
1524 	}
1525 
1526 	if ((dp->dxp == NULL) && IDNDS_IS_CLOSED(dp)) {
1527 		dp->dxp = &xphase_nego;
1528 		IDN_XSTATE_TRANSITION(dp, IDNXS_PEND);
1529 	} else if (dp->dxp != &xphase_nego) {
1530 		if (msg & IDNP_MSGTYPE_MASK) {
1531 			/*
1532 			 * If we already have a connection to somebody
1533 			 * trying to initiate a connection to us, then
1534 			 * possibly we've awaken from a coma or he did.
1535 			 * In any case, dismantle current connection
1536 			 * and attempt to establish a new one.
1537 			 */
1538 			if (dp->dstate == IDNDS_CONNECTED) {
1539 				DOMAINSET_ADD(idn.domset.ds_relink, domid);
1540 				IDN_HISTORY_LOG(IDNH_RELINK, domid,
1541 				    dp->dstate, idn.domset.ds_relink);
1542 				(void) idn_disconnect(domid, IDNFIN_NORMAL,
1543 				    IDNFIN_ARG_NONE, IDNFIN_SYNC_YES);
1544 			} else {
1545 				mt.mt_mtype = IDNP_NACK;
1546 				mt.mt_atype = msg;
1547 
1548 				CLR_XARGS(nargs);
1549 
1550 				if (DOMAIN_IN_SET(idn.domset.ds_hitlist,
1551 				    domid)) {
1552 					SET_XARGS_NACK_TYPE(nargs,
1553 					    IDNNACK_EXIT);
1554 				} else {
1555 					int	new_masterid;
1556 					int	new_cpuid = IDN_NIL_DCPU;
1557 
1558 					SET_XARGS_NACK_TYPE(nargs,
1559 					    IDNNACK_RETRY);
1560 					IDN_GLOCK_SHARED();
1561 					new_masterid = IDN_GET_NEW_MASTERID();
1562 					if (new_masterid == IDN_NIL_DOMID)
1563 						new_masterid =
1564 						    IDN_GET_MASTERID();
1565 					if (new_masterid != IDN_NIL_DOMID) {
1566 						idn_domain_t	*mdp;
1567 
1568 						mdp = &idn_domain[new_masterid];
1569 						new_cpuid = mdp->dcpu;
1570 					}
1571 					SET_XARGS_NACK_ARG1(nargs,
1572 					    new_masterid);
1573 					SET_XARGS_NACK_ARG2(nargs, new_cpuid);
1574 					IDN_GUNLOCK();
1575 				}
1576 				idn_send_acknack(domid, &mt, nargs);
1577 			}
1578 		}
1579 		return (0);
1580 	}
1581 
1582 	(void) idn_xphase_transition(domid, mtp, xargs);
1583 
1584 	return (0);
1585 }
1586 
1587 /*ARGSUSED1*/
1588 static void
1589 idn_retry_nego(uint_t token, void *arg)
1590 {
1591 	int		domid = IDN_RETRY_TOKEN2DOMID(token);
1592 	int		new_masterid;
1593 	idn_domain_t	*dp = &idn_domain[domid];
1594 	idn_xdcargs_t	xargs;
1595 	procname_t	proc = "idn_retry_nego";
1596 
1597 	ASSERT(IDN_RETRY_TOKEN2TYPE(token) == IDNRETRY_NEGO);
1598 
1599 	IDN_SYNC_LOCK();
1600 	IDN_DLOCK_EXCL(domid);
1601 
1602 	if (dp->dxp != &xphase_nego) {
1603 		STRING(str);
1604 
1605 #ifdef DEBUG
1606 		if (dp->dxp) {
1607 			INUM2STR(dp->dxp->xt_msgtype, str);
1608 		}
1609 #endif /* DEBUG */
1610 
1611 		PR_PROTO("%s:%d: dxp(%s) != NEGO...bailing...\n",
1612 		    proc, domid, dp->dxp ? str : "NULL");
1613 		IDN_DUNLOCK(domid);
1614 		IDN_SYNC_UNLOCK();
1615 		return;
1616 	}
1617 
1618 	if (dp->dxstate != IDNXS_PEND) {
1619 		PR_PROTO("%s:%d: xstate(%s) != %s...bailing\n",
1620 		    proc, domid, idnxs_str[dp->dxstate],
1621 		    idnxs_str[IDNXS_PEND]);
1622 		IDN_DUNLOCK(domid);
1623 		IDN_SYNC_UNLOCK();
1624 		return;
1625 	}
1626 
1627 	IDN_GLOCK_SHARED();
1628 	if (idn.state == IDNGS_RECONFIG) {
1629 		/*
1630 		 * Have to try again later after
1631 		 * reconfig has completed.
1632 		 */
1633 		PR_PROTO("%s:%d: reconfig in-progress...try later\n",
1634 		    proc, domid);
1635 		idn_retry_submit(idn_retry_nego, NULL, token,
1636 		    idn_msg_retrytime[IDNP_NEGO]);
1637 		IDN_GUNLOCK();
1638 		IDN_DUNLOCK(domid);
1639 		IDN_SYNC_UNLOCK();
1640 		return;
1641 	}
1642 	new_masterid = IDN_GET_NEW_MASTERID();
1643 	if ((idn.state == IDNGS_CONNECT) &&
1644 	    (new_masterid != IDN_NIL_DOMID) &&
1645 	    (domid != new_masterid) &&
1646 	    (idn.localid != new_masterid)) {
1647 		/*
1648 		 * We have a new master pending and this
1649 		 * guy isn't it.  Wait until the local domain
1650 		 * has a chance to connect with the new
1651 		 * master before going forward with this
1652 		 * guy.
1653 		 */
1654 		PR_PROTO("%s:%d: waiting for connect to new master %d\n",
1655 		    proc, domid, IDN_GET_NEW_MASTERID());
1656 		idn_retry_submit(idn_retry_nego, NULL, token,
1657 		    idn_msg_retrytime[IDNP_NEGO]);
1658 		IDN_GUNLOCK();
1659 		IDN_DUNLOCK(domid);
1660 		IDN_SYNC_UNLOCK();
1661 		return;
1662 	}
1663 	IDN_GUNLOCK();
1664 
1665 	(void) idn_xphase_transition(domid, NULL, xargs);
1666 
1667 	IDN_DUNLOCK(domid);
1668 	IDN_SYNC_UNLOCK();
1669 }
1670 
1671 static int
1672 idn_check_nego(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
1673 {
1674 	int		d, new_masterid, masterid;
1675 	int		cpuid, m_cpuid = -1;
1676 	uint_t		dmask;
1677 	uint_t		msg = mtp ? mtp->mt_mtype : 0;
1678 	idn_domain_t	*dp, *ldp;
1679 	domainset_t	con_set, pending_set;
1680 	idnneg_dset_t	dset;
1681 	procname_t	proc = "idn_check_nego";
1682 
1683 	ASSERT(IDN_SYNC_IS_LOCKED());
1684 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
1685 
1686 	dp = &idn_domain[domid];
1687 	ldp = &idn_domain[idn.localid];
1688 
1689 	if (msg & IDNP_NACK) {
1690 		if (GET_XARGS_NACK_TYPE(xargs) == IDNNACK_EXIT) {
1691 			PR_HITLIST("%s:%d(%s): (msg=%x) EXIT received, "
1692 			    "adding to hitlist %x -> %x\n",
1693 			    proc, domid, idnds_str[dp->dstate], msg,
1694 			    idn.domset.ds_hitlist,
1695 			    idn.domset.ds_hitlist | DOMAINSET(domid));
1696 
1697 			DOMAINSET_ADD(idn.domset.ds_hitlist, domid);
1698 			return (-1);
1699 		} else {
1700 			return (0);
1701 		}
1702 	}
1703 
1704 	if (DOMAIN_IN_SET(idn.domset.ds_hitlist, domid)) {
1705 		PR_HITLIST("%s:%d(%s): (msg=%x) domain in hitlist (%x) - "
1706 		    "exiting phase\n",
1707 		    proc, domid, idnds_str[dp->dstate], msg,
1708 		    idn.domset.ds_hitlist);
1709 		return (-1);
1710 	}
1711 
1712 	if ((dp->dstate == IDNDS_NEGO_PEND) && (msg & IDNP_MSGTYPE_MASK) &&
1713 	    (msg & IDNP_ACK))		/* nego+ack */
1714 		return (1);
1715 
1716 	dmask = (uint_t)-1;
1717 
1718 	IDN_GLOCK_EXCL();
1719 	if (idn.state == IDNGS_DISCONNECT) {
1720 		PR_PROTO("%s:%d: DISCONNECT in-progress >>> EXIT\n",
1721 		    proc, domid);
1722 		IDN_GUNLOCK();
1723 		return (-1);
1724 	} else if (idn.state == IDNGS_OFFLINE) {
1725 		IDN_GSTATE_TRANSITION(IDNGS_CONNECT);
1726 		IDN_PREP_HWINIT();
1727 		IDN_DLOCK_EXCL(idn.localid);
1728 		ldp->dvote.v.connected = 0;
1729 		IDN_DUNLOCK(idn.localid);
1730 	}
1731 
1732 	if (!DOMAIN_IN_SET(idn.domset.ds_trans_on, domid)) {
1733 		DOMAINSET_ADD(idn.domset.ds_trans_on, domid);
1734 		IDN_HISTORY_LOG(IDNH_NEGO, domid,
1735 		    idn.domset.ds_trans_on,
1736 		    idn.domset.ds_connected);
1737 	}
1738 
1739 	switch (idn.state) {
1740 	case IDNGS_RECONFIG:
1741 		PR_PROTO("%s:%d: RECONFIG in-progress >>> RETRY\n",
1742 		    proc, domid);
1743 		IDN_GUNLOCK();
1744 		return (1);
1745 
1746 	case IDNGS_CONNECT:
1747 		new_masterid = IDN_GET_NEW_MASTERID();
1748 		if ((new_masterid != IDN_NIL_DOMID) &&
1749 		    (domid != new_masterid) &&
1750 		    (idn.localid != new_masterid)) {
1751 			PR_PROTO("%s:%d: waiting for connect to "
1752 			    "new master %d\n",
1753 			    proc, domid, IDN_GET_NEW_MASTERID());
1754 			IDN_GUNLOCK();
1755 			return (1);
1756 		}
1757 		break;
1758 
1759 	default:
1760 		break;
1761 	}
1762 
1763 	ASSERT((idn.state == IDNGS_CONNECT) || (idn.state == IDNGS_ONLINE));
1764 
1765 	con_set = 0;
1766 
1767 	if (msg) {
1768 		idn_domain_t	*mdp;
1769 		idn_vote_t	vote;
1770 
1771 		vote.ticket = GET_XARGS_NEGO_TICKET(xargs);
1772 		/*
1773 		 * Sender should note have set master bit,
1774 		 * but just in case clear it so local domain
1775 		 * doesn't get confused.
1776 		 */
1777 		vote.v.master = 0;
1778 		dp->dvote.ticket = vote.ticket;
1779 		GET_XARGS_NEGO_DSET(xargs, dset);
1780 		/*LINTED*/
1781 		IDNNEG_DSET_GET_MASK(dset, domid, dmask);
1782 		IDNNEG_DSET_GET_MASTER(dset, new_masterid);
1783 		if (new_masterid == IDNNEG_NO_MASTER) {
1784 			new_masterid = IDN_NIL_DOMID;
1785 		} else {
1786 			/*
1787 			 * Remote domain has a master.  Find
1788 			 * his cpuid in the dset.  We may need
1789 			 * it to initiate a connection.
1790 			 */
1791 			if (new_masterid == domid) {
1792 				m_cpuid = dp->dcpu;
1793 			} else {
1794 				IDNNEG_DSET_GET(dset, new_masterid, m_cpuid,
1795 				    dmask);
1796 				if (m_cpuid == -1) {
1797 					/*
1798 					 * Something is bogus if remote domain
1799 					 * is reporting a valid masterid, but
1800 					 * doesn't have the cpuid for it.
1801 					 */
1802 					cmn_err(CE_WARN,
1803 					    "IDN: 209: remote domain (ID "
1804 					    "%d, CPU %d) reporting master "
1805 					    "(ID %d) without CPU ID",
1806 					    domid, dp->dcpu, new_masterid);
1807 					DOMAINSET_ADD(idn.domset.ds_hitlist,
1808 					    domid);
1809 					IDN_GUNLOCK();
1810 					return (-1);
1811 				}
1812 			}
1813 		}
1814 
1815 		for (d = 0; d < MAX_DOMAINS; d++) {
1816 			if ((d == idn.localid) || (d == domid))
1817 				continue;
1818 			IDNNEG_DSET_GET(dset, d, cpuid, dmask);
1819 			if (cpuid != -1) {
1820 				DOMAINSET_ADD(con_set, d);
1821 			}
1822 		}
1823 
1824 #ifdef DEBUG
1825 		if (idn.domset.ds_hitlist) {
1826 			PR_HITLIST("%s:%d: con_set %x -> %x (hitlist = %x)\n",
1827 			    proc, domid, con_set,
1828 			    con_set & ~idn.domset.ds_hitlist,
1829 			    idn.domset.ds_hitlist);
1830 		}
1831 #endif /* DEBUG */
1832 
1833 		con_set &= ~idn.domset.ds_hitlist;
1834 
1835 		ASSERT(!DOMAIN_IN_SET(con_set, idn.localid));
1836 		ASSERT(!DOMAIN_IN_SET(con_set, domid));
1837 
1838 		if ((new_masterid != IDN_NIL_DOMID) &&
1839 		    DOMAIN_IN_SET(idn.domset.ds_hitlist, new_masterid)) {
1840 			PR_HITLIST("%s:%d: new_mstr %d -> -1 (hitlist = %x)\n",
1841 			    proc, domid, new_masterid,
1842 			    idn.domset.ds_hitlist);
1843 			IDN_GUNLOCK();
1844 			return (1);
1845 		}
1846 
1847 		if (idn_select_master(domid, new_masterid, m_cpuid) < 0) {
1848 			/*
1849 			 * Returns w/GLOCK dropped if error.
1850 			 */
1851 			return (1);
1852 		}
1853 
1854 		masterid = IDN_GET_MASTERID();
1855 		ASSERT(masterid != IDN_NIL_DOMID);
1856 
1857 		if (idn.state == IDNGS_CONNECT) {
1858 			/*
1859 			 * This is the initial connection for
1860 			 * the local domain.
1861 			 */
1862 			IDN_DLOCK_EXCL(idn.localid);
1863 
1864 			if (masterid == idn.localid) {
1865 				if (idn_master_init() < 0) {
1866 					cmn_err(CE_WARN,
1867 					    "IDN: 210: failed to init "
1868 					    "MASTER context");
1869 					ldp->dvote.v.master = 0;
1870 					IDN_DUNLOCK(idn.localid);
1871 					IDN_GSTATE_TRANSITION(IDNGS_DISCONNECT);
1872 					IDN_SET_MASTERID(IDN_NIL_DOMID);
1873 					IDN_GUNLOCK();
1874 					return (-1);
1875 				}
1876 				DSLAB_LOCK_EXCL(idn.localid);
1877 				ldp->dslab_state = DSLAB_STATE_LOCAL;
1878 				DSLAB_UNLOCK(idn.localid);
1879 				ldp->dvote.v.connected = 1;
1880 			} else {
1881 				/*
1882 				 * Either the remote domain is the
1883 				 * master or its a new slave trying
1884 				 * to connect to us.  We can't allow
1885 				 * further progress until we've
1886 				 * sync'd up with the master.
1887 				 */
1888 				if (masterid != domid) {
1889 					IDN_DUNLOCK(idn.localid);
1890 					IDN_GUNLOCK();
1891 					return (1);
1892 				}
1893 				DSLAB_LOCK_EXCL(idn.localid);
1894 				ldp->dslab_state = DSLAB_STATE_REMOTE;
1895 				DSLAB_UNLOCK(idn.localid);
1896 			}
1897 			IDN_DUNLOCK(idn.localid);
1898 			/*
1899 			 * We've sync'd up with the new master.
1900 			 */
1901 			IDN_GSTATE_TRANSITION(IDNGS_ONLINE);
1902 		}
1903 
1904 		mdp = &idn_domain[masterid];
1905 
1906 		if ((masterid != domid) && !IDNDS_CONFIG_DONE(mdp)) {
1907 			/*
1908 			 * We can't progress any further with
1909 			 * other domains until we've exchanged all
1910 			 * the necessary CFG info with the master,
1911 			 * i.e. until we have a mailbox area from
1912 			 * which we can allocate mailboxes to
1913 			 * other domains.
1914 			 */
1915 			PR_PROTO("%s:%d: still exchanging CFG "
1916 			    "w/master(%d)\n", proc, domid, masterid);
1917 			IDN_GUNLOCK();
1918 			return (1);
1919 		}
1920 
1921 		DSLAB_LOCK_EXCL(domid);
1922 		dp->dslab_state = ldp->dslab_state;
1923 		DSLAB_UNLOCK(domid);
1924 		if (idn.state != IDNGS_ONLINE) {
1925 			IDN_GSTATE_TRANSITION(IDNGS_ONLINE);
1926 		}
1927 	}
1928 
1929 	IDN_GUNLOCK();
1930 
1931 	pending_set = con_set;
1932 	pending_set &= ~(idn.domset.ds_trans_on | idn.domset.ds_connected);
1933 	idn.domset.ds_trans_on |= pending_set;
1934 
1935 	con_set |= idn.domset.ds_trans_on | idn.domset.ds_connected;
1936 	con_set &= ~idn.domset.ds_trans_off;
1937 	DOMAINSET_ADD(con_set, idn.localid);
1938 
1939 	if (dp->dsync.s_cmd != IDNSYNC_CONNECT) {
1940 		idn_sync_exit(domid, IDNSYNC_DISCONNECT);
1941 		idn_sync_enter(domid, IDNSYNC_CONNECT,
1942 		    con_set, DOMAINSET(idn.localid), idn_xstate_transfunc,
1943 		    (void *)IDNP_CON);
1944 	}
1945 
1946 	/*
1947 	 * Get this domain registered as an expected domain on
1948 	 * the remaining domains in the CONNECT synchronization.
1949 	 */
1950 	(void) idn_sync_register(domid, IDNSYNC_CONNECT, 0, IDNSYNC_REG_NEW);
1951 
1952 	/*
1953 	 * Note that if (msg == 0), i.e. then there will be
1954 	 * no dset and also pending_set will be 0.
1955 	 * So, the following loop will never attempt to
1956 	 * look at the dset unless (msg != 0), implying
1957 	 * that we've been through the initial code above
1958 	 * and have initialized dmask.
1959 	 */
1960 	ASSERT(pending_set ? (dmask != (uint_t)-1) : 1);
1961 
1962 	for (d = 0; d < MAX_DOMAINS; d++) {
1963 		int	rv;
1964 
1965 		if (!DOMAIN_IN_SET(pending_set, d))
1966 			continue;
1967 
1968 		ASSERT((d != idn.localid) && (d != domid));
1969 
1970 		dp = &idn_domain[d];
1971 
1972 		IDNNEG_DSET_GET(dset, d, cpuid, dmask);
1973 		if (cpuid == -1) {
1974 			PR_PROTO("%s:%d: failed to get cpuid from dset "
1975 			    "for domain %d (pset = 0x%x)\n",
1976 			    proc, domid, d, pending_set);
1977 			DOMAINSET_DEL(idn.domset.ds_trans_on, d);
1978 			continue;
1979 		}
1980 
1981 		IDN_DLOCK_EXCL(d);
1982 		if ((rv = idn_open_domain(d, cpuid, 0)) != 0) {
1983 			PR_PROTO("%s:%d: failed "
1984 			    "idn_open_domain(%d,%d,0) (rv = %d)\n",
1985 			    proc, domid, d, cpuid, rv);
1986 			if (rv < 0) {
1987 				cmn_err(CE_WARN,
1988 				    "IDN: 205: (%s) failed to "
1989 				    "open-domain(%d,%d)",
1990 				    proc, d, cpuid);
1991 				DOMAINSET_DEL(idn.domset.ds_trans_on, d);
1992 			} else if (DOMAIN_IN_SET(idn.domset.ds_trans_off, d)) {
1993 				/*
1994 				 * We've requested to connect to a domain
1995 				 * from which we're disconnecting.  We
1996 				 * better mark this guy for relinking.
1997 				 */
1998 				DOMAINSET_ADD(idn.domset.ds_relink, d);
1999 				IDN_HISTORY_LOG(IDNH_RELINK, d, dp->dstate,
2000 				    idn.domset.ds_relink);
2001 			}
2002 			IDN_DUNLOCK(d);
2003 			continue;
2004 		}
2005 
2006 		(void) idn_connect(d);
2007 
2008 		IDN_DUNLOCK(d);
2009 	}
2010 
2011 	return (0);
2012 }
2013 
2014 /*ARGSUSED*/
2015 static void
2016 idn_action_nego_pend(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
2017 {
2018 	uint_t		msg = mtp ? mtp->mt_mtype : 0;
2019 	idn_msgtype_t	mt;
2020 	domainset_t	con_set;
2021 
2022 	ASSERT(IDN_SYNC_IS_LOCKED());
2023 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
2024 
2025 	con_set = idn.domset.ds_trans_on | idn.domset.ds_connected;
2026 	con_set &= ~idn.domset.ds_trans_off;
2027 
2028 	if (!msg) {
2029 		(void) idn_send_nego(domid, NULL, con_set);
2030 	} else {
2031 		mt.mt_mtype = IDNP_NEGO | IDNP_ACK;
2032 		mt.mt_atype = 0;
2033 		mt.mt_cookie = mtp->mt_cookie;
2034 		(void) idn_send_nego(domid, &mt, con_set);
2035 	}
2036 }
2037 
2038 /*ARGSUSED*/
2039 static void
2040 idn_error_nego(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
2041 {
2042 	int	new_masterid, new_cpuid;
2043 	int	retry = 1;
2044 	uint_t	msg = mtp ? mtp->mt_mtype : 0;
2045 	uint_t	token;
2046 
2047 	ASSERT(IDN_SYNC_IS_LOCKED());
2048 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
2049 
2050 	if (msg & IDNP_NACK) {
2051 		idn_nack_t	nack;
2052 
2053 		nack = GET_XARGS_NACK_TYPE(xargs);
2054 		switch (nack) {
2055 		case IDNNACK_RETRY:
2056 			new_masterid = (int)GET_XARGS_NACK_ARG1(xargs);
2057 			new_cpuid    = (int)GET_XARGS_NACK_ARG2(xargs);
2058 			break;
2059 
2060 		case IDNNACK_EXIT:
2061 			retry = 0;
2062 			/*FALLTHROUGH*/
2063 
2064 		default:
2065 			new_masterid = IDN_NIL_DOMID;
2066 			new_cpuid    = IDN_NIL_DCPU;
2067 			break;
2068 		}
2069 		idn_nego_cleanup_check(domid, new_masterid, new_cpuid);
2070 	}
2071 
2072 	if (msg & IDNP_MSGTYPE_MASK) {
2073 		idn_msgtype_t	mt;
2074 		idn_xdcargs_t	nargs;
2075 
2076 		mt.mt_mtype = IDNP_NACK;
2077 		mt.mt_atype = msg;
2078 		mt.mt_cookie = mtp->mt_cookie;
2079 		CLR_XARGS(nargs);
2080 		SET_XARGS_NACK_TYPE(nargs, IDNNACK_RETRY);
2081 		IDN_GLOCK_SHARED();
2082 		new_masterid = IDN_GET_NEW_MASTERID();
2083 		if (new_masterid == IDN_NIL_DOMID)
2084 			new_masterid = IDN_GET_MASTERID();
2085 		if (new_masterid != IDN_NIL_DOMID)
2086 			new_cpuid = idn_domain[new_masterid].dcpu;
2087 		else
2088 			new_cpuid = IDN_NIL_DCPU;
2089 		SET_XARGS_NACK_ARG1(nargs, new_masterid);
2090 		SET_XARGS_NACK_ARG2(nargs, new_cpuid);
2091 		IDN_GUNLOCK();
2092 		idn_send_acknack(domid, &mt, nargs);
2093 	}
2094 
2095 	if (retry) {
2096 		token = IDN_RETRY_TOKEN(domid, IDNRETRY_NEGO);
2097 		idn_retry_submit(idn_retry_nego, NULL, token,
2098 		    idn_msg_retrytime[(int)IDNRETRY_NEGO]);
2099 	} else {
2100 		DOMAINSET_DEL(idn.domset.ds_relink, domid);
2101 		IDN_RESET_COOKIES(domid);
2102 		(void) idn_disconnect(domid, IDNFIN_NORMAL, IDNFIN_ARG_NONE,
2103 		    IDNDS_SYNC_TYPE(&idn_domain[domid]));
2104 	}
2105 }
2106 
2107 /*ARGSUSED*/
2108 static void
2109 idn_action_nego_sent(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
2110 {
2111 	uint_t		msg = mtp ? mtp->mt_mtype : 0;
2112 	domainset_t	conset;
2113 	idn_msgtype_t	mt;
2114 
2115 	ASSERT(IDN_SYNC_IS_LOCKED());
2116 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
2117 
2118 	mt.mt_cookie = mtp ? mtp->mt_cookie : 0;
2119 
2120 	conset = idn.domset.ds_trans_on | idn.domset.ds_connected;
2121 	conset &= ~idn.domset.ds_trans_off;
2122 
2123 	if ((msg & IDNP_ACKNACK_MASK) == 0) {
2124 		/*
2125 		 * nego
2126 		 */
2127 		mt.mt_mtype = IDNP_NEGO | IDNP_ACK;
2128 		mt.mt_atype = 0;
2129 		(void) idn_send_nego(domid, &mt, conset);
2130 	} else if (msg & IDNP_MSGTYPE_MASK) {
2131 		int		d;
2132 		idn_xdcargs_t	nargs;
2133 		idnneg_dset_t	dset;
2134 		uint_t		dmask;
2135 		idn_vote_t	vote;
2136 
2137 		mt.mt_mtype = IDNP_ACK;
2138 		mt.mt_atype = msg;
2139 		DOMAINSET_DEL(conset, idn.localid);
2140 		DOMAINSET_DEL(conset, domid);
2141 
2142 		dmask = IDNNEG_DSET_MYMASK();
2143 		IDNNEG_DSET_INIT(dset, dmask);
2144 		for (d = 0; d < MAX_DOMAINS; d++) {
2145 			int	cpuid;
2146 
2147 			if (!DOMAIN_IN_SET(conset, d))
2148 				continue;
2149 
2150 			if ((cpuid = idn_domain[d].dcpu) == IDN_NIL_DCPU)
2151 				continue;
2152 
2153 			IDNNEG_DSET_SET(dset, d, cpuid, dmask);
2154 		}
2155 		IDNNEG_DSET_SET_MASTER(dset, domid, IDN_GET_MASTERID());
2156 		ASSERT((IDN_GET_MASTERID() != IDN_NIL_DOMID) ?
2157 		    (idn_domain[IDN_GET_MASTERID()].dcpu != IDN_NIL_DCPU) : 1);
2158 		vote.ticket = idn_domain[idn.localid].dvote.ticket;
2159 		vote.v.master = 0;
2160 		CLR_XARGS(nargs);
2161 		SET_XARGS_NEGO_TICKET(nargs, vote.ticket);
2162 		SET_XARGS_NEGO_DSET(nargs, dset);
2163 		/*
2164 		 * nego+ack
2165 		 */
2166 		idn_send_acknack(domid, &mt, nargs);
2167 	} else {
2168 		uint_t		token;
2169 		int		new_masterid, new_cpuid;
2170 		int		retry = 1;
2171 		idn_nack_t	nack;
2172 		/*
2173 		 * nack - retry
2174 		 *
2175 		 * It's possible if we've made it this far that
2176 		 * we may have already chosen a master and this
2177 		 * dude might be it!  If it is we need to clean up.
2178 		 */
2179 		nack = GET_XARGS_NACK_TYPE(xargs);
2180 		switch (nack) {
2181 		case IDNNACK_RETRY:
2182 			new_masterid = (int)GET_XARGS_NACK_ARG1(xargs);
2183 			new_cpuid = (int)GET_XARGS_NACK_ARG2(xargs);
2184 			break;
2185 
2186 		case IDNNACK_EXIT:
2187 			retry = 0;
2188 			/*FALLTHROUGH*/
2189 
2190 		default:
2191 			new_masterid = IDN_NIL_DOMID;
2192 			new_cpuid = IDN_NIL_DCPU;
2193 			break;
2194 		}
2195 
2196 		idn_nego_cleanup_check(domid, new_masterid, new_cpuid);
2197 
2198 		if (retry) {
2199 			token = IDN_RETRY_TOKEN(domid, IDNRETRY_NEGO);
2200 			idn_retry_submit(idn_retry_nego, NULL, token,
2201 			    idn_msg_retrytime[(int)IDNRETRY_NEGO]);
2202 		} else {
2203 			DOMAINSET_DEL(idn.domset.ds_relink, domid);
2204 			IDN_RESET_COOKIES(domid);
2205 			(void) idn_disconnect(domid, IDNFIN_NORMAL,
2206 			    IDNFIN_ARG_NONE,
2207 			    IDNDS_SYNC_TYPE(&idn_domain[domid]));
2208 		}
2209 	}
2210 }
2211 
2212 /*ARGSUSED*/
2213 static void
2214 idn_action_nego_rcvd(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
2215 {
2216 	uint_t	msg = mtp ? mtp->mt_mtype : 0;
2217 
2218 	ASSERT(IDN_SYNC_IS_LOCKED());
2219 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
2220 
2221 	if (msg & IDNP_NACK) {
2222 		uint_t		token;
2223 		int		new_masterid, new_cpuid;
2224 		int		retry = 1;
2225 		idn_nack_t	nack;
2226 		/*
2227 		 * nack - retry.
2228 		 *
2229 		 * At this stage of receiving a nack we need to
2230 		 * check whether we need to start over again with
2231 		 * selecting a new master.
2232 		 */
2233 		nack = GET_XARGS_NACK_TYPE(xargs);
2234 		switch (nack) {
2235 		case IDNNACK_RETRY:
2236 			new_masterid = (int)GET_XARGS_NACK_ARG1(xargs);
2237 			new_cpuid = (int)GET_XARGS_NACK_ARG2(xargs);
2238 			break;
2239 
2240 		case IDNNACK_EXIT:
2241 			retry = 0;
2242 			/*FALLTHROUGH*/
2243 
2244 		default:
2245 			new_masterid = IDN_NIL_DOMID;
2246 			new_cpuid = IDN_NIL_DCPU;
2247 			break;
2248 		}
2249 
2250 		idn_nego_cleanup_check(domid, new_masterid, new_cpuid);
2251 
2252 		if (retry) {
2253 			token = IDN_RETRY_TOKEN(domid, IDNRETRY_NEGO);
2254 			idn_retry_submit(idn_retry_nego, NULL, token,
2255 			    idn_msg_retrytime[(int)IDNRETRY_NEGO]);
2256 		} else {
2257 			DOMAINSET_DEL(idn.domset.ds_relink, domid);
2258 			IDN_RESET_COOKIES(domid);
2259 			(void) idn_disconnect(domid, IDNFIN_NORMAL,
2260 			    IDNFIN_ARG_NONE,
2261 			    IDNDS_SYNC_TYPE(&idn_domain[domid]));
2262 		}
2263 	}
2264 }
2265 
2266 static void
2267 idn_final_nego(int domid)
2268 {
2269 	idn_domain_t	*dp = &idn_domain[domid];
2270 
2271 	ASSERT(IDN_SYNC_IS_LOCKED());
2272 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
2273 
2274 	(void) idn_retry_terminate(IDN_RETRY_TOKEN(domid, IDNRETRY_NEGO));
2275 
2276 	ASSERT(dp->dstate == IDNDS_CONFIG);
2277 
2278 	dp->dxp = NULL;
2279 	IDN_XSTATE_TRANSITION(dp, IDNXS_NIL);
2280 
2281 	idn_send_config(domid, 1);
2282 }
2283 
2284 /*
2285  */
2286 /*ARGSUSED1*/
2287 static void
2288 idn_exit_nego(int domid, uint_t msgtype)
2289 {
2290 	idn_domain_t	*dp;
2291 	idn_fin_t	fintype;
2292 
2293 	ASSERT(IDN_SYNC_IS_LOCKED());
2294 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
2295 
2296 	dp = &idn_domain[domid];
2297 
2298 	fintype = msgtype ? IDNFIN_NORMAL : IDNFIN_FORCE_HARD;
2299 
2300 	(void) idn_retry_terminate(IDN_RETRY_TOKEN(domid, IDNRETRY_NEGO));
2301 
2302 	ASSERT(!DOMAIN_IN_SET(idn.domset.ds_connected, domid));
2303 	ASSERT(!DOMAIN_IN_SET(idn.domset.ds_ready_on, domid));
2304 	ASSERT(dp->dxp == &xphase_nego);
2305 
2306 	idn_nego_cleanup_check(domid, IDN_NIL_DOMID, IDN_NIL_DCPU);
2307 
2308 	IDN_GLOCK_SHARED();
2309 	if ((idn.state != IDNGS_DISCONNECT) &&
2310 	    !DOMAIN_IN_SET(idn.domset.ds_hitlist, domid)) {
2311 		DOMAINSET_ADD(idn.domset.ds_relink, domid);
2312 		IDN_HISTORY_LOG(IDNH_RELINK, domid, dp->dstate,
2313 		    idn.domset.ds_relink);
2314 	} else {
2315 		idn_update_op(IDNOP_ERROR, DOMAINSET(domid), NULL);
2316 		DOMAINSET_DEL(idn.domset.ds_relink, domid);
2317 	}
2318 	IDN_GUNLOCK();
2319 	/*
2320 	 * Reset send cookie to 0 so that receiver does not validate
2321 	 * cookie.  This is necessary since at this early stage it's
2322 	 * possible we may not have exchanged appropriate cookies.
2323 	 */
2324 	IDN_RESET_COOKIES(domid);
2325 	(void) idn_disconnect(domid, fintype, IDNFIN_ARG_NONE,
2326 	    IDNDS_SYNC_TYPE(dp));
2327 }
2328 
2329 static void
2330 idn_nego_cleanup_check(int domid, int new_masterid, int new_cpuid)
2331 {
2332 	idn_domain_t	*ldp, *dp;
2333 	procname_t	proc = "idn_nego_cleanup_check";
2334 
2335 	ASSERT(IDN_SYNC_IS_LOCKED());
2336 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
2337 
2338 	dp = &idn_domain[domid];
2339 	ldp = &idn_domain[idn.localid];
2340 
2341 	IDN_GLOCK_EXCL();
2342 
2343 	if (((idn.state == IDNGS_ONLINE) && !idn.domset.ds_connected) ||
2344 	    (idn.state == IDNGS_CONNECT)) {
2345 		domainset_t	trans_on;
2346 		int		masterid;
2347 		int		retry_domid = IDN_NIL_DOMID;
2348 		int		rv;
2349 
2350 		IDN_DLOCK_EXCL(idn.localid);
2351 		masterid = (idn.state == IDNGS_ONLINE) ?
2352 		    IDN_GET_MASTERID() : IDN_GET_NEW_MASTERID();
2353 		trans_on = idn.domset.ds_trans_on;
2354 		DOMAINSET_DEL(trans_on, domid);
2355 		if (trans_on == 0) {
2356 			int		d;
2357 			domainset_t	relink = idn.domset.ds_relink;
2358 			/*
2359 			 * This was the only guy we were trying
2360 			 * to connect with.
2361 			 */
2362 			ASSERT((idn.state == IDNGS_ONLINE) ?
2363 			    ((idn.localid == masterid) ||
2364 			    (domid == masterid)) : 1);
2365 			if (idn.localid == masterid)
2366 				idn_master_deinit();
2367 			ldp->dvote.v.connected = 0;
2368 			ldp->dvote.v.master = 0;
2369 			dp->dvote.v.master = 0;
2370 			IDN_SET_MASTERID(IDN_NIL_DOMID);
2371 			IDN_SET_NEW_MASTERID(new_masterid);
2372 			IDN_GSTATE_TRANSITION(IDNGS_CONNECT);
2373 			IDN_PREP_HWINIT();
2374 			IDN_DUNLOCK(idn.localid);
2375 			IDN_GUNLOCK();
2376 			/*
2377 			 * If there's a new master available then
2378 			 * just try and relink with him unless
2379 			 * it's ourself.
2380 			 */
2381 			if ((new_masterid != IDN_NIL_DOMID) &&
2382 			    (new_masterid != idn.localid) &&
2383 			    (new_masterid != domid)) {
2384 				IDN_DLOCK_EXCL(new_masterid);
2385 				rv = idn_open_domain(new_masterid,
2386 				    new_cpuid, 0);
2387 				if (rv < 0) {
2388 					cmn_err(CE_WARN,
2389 					    "IDN: 205: (%s) failed to "
2390 					    "open-domain(%d,%d)",
2391 					    proc, new_masterid, new_cpuid);
2392 					IDN_GLOCK_EXCL();
2393 					IDN_SET_NEW_MASTERID(IDN_NIL_DOMID);
2394 					IDN_GUNLOCK();
2395 				} else {
2396 					relink = DOMAINSET(new_masterid);
2397 				}
2398 				IDN_DUNLOCK(new_masterid);
2399 			}
2400 			DOMAINSET_DEL(relink, domid);
2401 			if (relink)
2402 				for (d = 0; d < MAX_DOMAINS; d++) {
2403 					if (!DOMAIN_IN_SET(relink, d))
2404 						continue;
2405 					retry_domid = d;
2406 					break;
2407 				}
2408 		} else if (domid == masterid) {
2409 			/*
2410 			 * There are other domains we were trying
2411 			 * to connect to.  As long as the chosen
2412 			 * master was somebody other then this
2413 			 * domain that nack'd us, life is cool, but
2414 			 * if it was this remote domain we'll need
2415 			 * to start over.
2416 			 */
2417 			IDN_DUNLOCK(idn.localid);
2418 			dp->dvote.v.master = 0;
2419 			IDN_SET_MASTERID(IDN_NIL_DOMID);
2420 			IDN_SET_NEW_MASTERID(new_masterid);
2421 
2422 			if (idn.state == IDNGS_ONLINE) {
2423 				IDN_GKSTAT_GLOBAL_EVENT(gk_reconfigs,
2424 				    gk_reconfig_last);
2425 				IDN_GSTATE_TRANSITION(IDNGS_RECONFIG);
2426 				IDN_GUNLOCK();
2427 				idn_unlink_domainset(trans_on, IDNFIN_NORMAL,
2428 				    IDNFIN_ARG_NONE,
2429 				    IDNFIN_OPT_RELINK,
2430 				    BOARDSET_ALL);
2431 			} else if ((new_masterid != IDN_NIL_DOMID) &&
2432 			    (new_masterid != idn.localid) &&
2433 			    (new_masterid != domid) &&
2434 			    !DOMAIN_IN_SET(trans_on, new_masterid)) {
2435 				IDN_GUNLOCK();
2436 				IDN_DLOCK_EXCL(new_masterid);
2437 				rv = idn_open_domain(new_masterid,
2438 				    new_cpuid, 0);
2439 				IDN_GLOCK_EXCL();
2440 				IDN_DUNLOCK(new_masterid);
2441 				if (rv < 0) {
2442 					cmn_err(CE_WARN,
2443 					    "IDN: 205: (%s) failed to "
2444 					    "open-domain(%d,%d)",
2445 					    proc, new_masterid,
2446 					    new_cpuid);
2447 					IDN_SET_NEW_MASTERID(IDN_NIL_DOMID);
2448 					new_masterid = IDN_NIL_DOMID;
2449 				} else {
2450 					retry_domid = new_masterid;
2451 				}
2452 				IDN_GUNLOCK();
2453 			} else {
2454 				IDN_GUNLOCK();
2455 			}
2456 		} else {
2457 			IDN_DUNLOCK(idn.localid);
2458 			IDN_GUNLOCK();
2459 		}
2460 		if (retry_domid != IDN_NIL_DOMID) {
2461 			uint_t		token;
2462 			idn_domain_t	*rdp = &idn_domain[retry_domid];
2463 
2464 			IDN_DLOCK_EXCL(retry_domid);
2465 			rdp->dxp = &xphase_nego;
2466 			IDN_XSTATE_TRANSITION(rdp, IDNXS_PEND);
2467 			IDN_DUNLOCK(retry_domid);
2468 			token = IDN_RETRY_TOKEN(retry_domid, IDNRETRY_NEGO);
2469 			idn_retry_submit(idn_retry_nego, NULL, token,
2470 			    idn_msg_retrytime[(int)IDNRETRY_NEGO]);
2471 		}
2472 	} else {
2473 		IDN_GUNLOCK();
2474 	}
2475 }
2476 
2477 static int
2478 idn_send_con(int domid, idn_msgtype_t *mtp, idn_con_t contype, domainset_t
2479     conset)
2480 {
2481 	idn_msgtype_t	mt;
2482 	uint_t		acknack;
2483 	procname_t	proc = "idn_send_con";
2484 
2485 	ASSERT(IDN_SYNC_IS_LOCKED());
2486 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
2487 
2488 	if (mtp) {
2489 		acknack = mtp->mt_mtype & IDNP_ACKNACK_MASK;
2490 		mt.mt_mtype = mtp->mt_mtype;
2491 		mt.mt_atype = mtp->mt_atype;
2492 		mt.mt_cookie = mtp->mt_cookie;
2493 	} else {
2494 		acknack = 0;
2495 		mt.mt_mtype = IDNP_CON;
2496 		mt.mt_atype = 0;
2497 		/*
2498 		 * For simple CON queries we want a unique
2499 		 * timer assigned.  For others, they
2500 		 * effectively share one.
2501 		 */
2502 		if (contype == IDNCON_QUERY)
2503 			mt.mt_cookie = 0;
2504 		else
2505 			mt.mt_cookie = IDN_TIMER_PUBLIC_COOKIE;
2506 	}
2507 
2508 	ASSERT((contype == IDNCON_QUERY) ? idn_domain[domid].dcookie_send : 1);
2509 
2510 	PR_PROTO("%s:%d: sending con%sto (cpu %d) [ct=%s, cs=0x%x]\n",
2511 	    proc, domid,
2512 	    (acknack & IDNP_ACK) ? "+ack " :
2513 	    (acknack & IDNP_NACK) ? "+nack " : " ",
2514 	    idn_domain[domid].dcpu,
2515 	    idncon_str[contype], conset);
2516 
2517 	IDN_MSGTIMER_START(domid, IDNP_CON, (ushort_t)contype,
2518 	    idn_msg_waittime[IDNP_CON], &mt.mt_cookie);
2519 
2520 	IDNXDC(domid, &mt, (uint_t)contype, (uint_t)conset, 0, 0);
2521 
2522 	return (0);
2523 }
2524 
2525 /*
2526  * Must leave w/DLOCK dropped and SYNC_LOCK held.
2527  */
2528 static int
2529 idn_recv_con(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
2530 {
2531 	uint_t		msg = mtp ? mtp->mt_mtype : 0;
2532 	uint_t		msgarg = mtp ? mtp->mt_atype : 0;
2533 	idn_con_t	contype;
2534 	domainset_t	my_ready_set, ready_set;
2535 	idn_msgtype_t	mt;
2536 	idn_domain_t	*dp = &idn_domain[domid];
2537 	idn_xdcargs_t	aargs;
2538 	procname_t	proc = "idn_recv_con";
2539 
2540 	ASSERT(IDN_SYNC_IS_LOCKED());
2541 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
2542 
2543 	mt.mt_cookie = mtp ? mtp->mt_cookie : 0;
2544 
2545 	contype   = GET_XARGS_CON_TYPE(xargs);
2546 	ready_set = GET_XARGS_CON_DOMSET(xargs);
2547 
2548 	CLR_XARGS(aargs);
2549 
2550 	if (!(msg & IDNP_NACK) && (contype == IDNCON_QUERY)) {
2551 		domainset_t	query_set;
2552 
2553 		query_set = idn_sync_register(domid, IDNSYNC_CONNECT,
2554 		    ready_set, IDNSYNC_REG_REG);
2555 
2556 		my_ready_set = idn.domset.ds_connected | idn.domset.ds_ready_on;
2557 		my_ready_set &= ~idn.domset.ds_trans_off;
2558 		DOMAINSET_ADD(my_ready_set, idn.localid);
2559 
2560 		if (msg & IDNP_MSGTYPE_MASK) {
2561 			mt.mt_mtype = IDNP_ACK;
2562 			mt.mt_atype = IDNP_CON;
2563 			SET_XARGS_CON_TYPE(aargs, contype);
2564 			SET_XARGS_CON_DOMSET(aargs, my_ready_set);
2565 			idn_send_acknack(domid, &mt, aargs);
2566 		}
2567 
2568 		if (query_set) {
2569 			uint_t	token;
2570 
2571 			token = IDN_RETRY_TOKEN(domid, IDNRETRY_CONQ);
2572 			idn_retry_submit(idn_retry_query, NULL, token,
2573 			    idn_msg_retrytime[(int)IDNRETRY_CONQ]);
2574 		}
2575 
2576 		return (0);
2577 	}
2578 
2579 	if (dp->dxp == NULL) {
2580 		STRING(mstr);
2581 		STRING(lstr);
2582 		/*
2583 		 * Must have received an inappropriate error
2584 		 * message as we should already be registered
2585 		 * by the time we reach here.
2586 		 */
2587 		INUM2STR(msg, mstr);
2588 		INUM2STR(msgarg, lstr);
2589 
2590 		PR_PROTO("%s:%d: ERROR: NOT YET REGISTERED (%s/%s)\n",
2591 		    proc, domid, mstr, lstr);
2592 
2593 		if (msg & IDNP_MSGTYPE_MASK) {
2594 			mt.mt_mtype = IDNP_NACK;
2595 			mt.mt_atype = msg;
2596 			SET_XARGS_NACK_TYPE(aargs, IDNNACK_RETRY);
2597 			idn_send_acknack(domid, &mt, aargs);
2598 		}
2599 
2600 		return (-1);
2601 	}
2602 
2603 	(void) idn_xphase_transition(domid, mtp, xargs);
2604 
2605 	return (0);
2606 }
2607 
2608 /*ARGSUSED1*/
2609 static void
2610 idn_retry_con(uint_t token, void *arg)
2611 {
2612 	int		domid = IDN_RETRY_TOKEN2DOMID(token);
2613 	idn_domain_t	*dp = &idn_domain[domid];
2614 	idn_xdcargs_t	xargs;
2615 	procname_t	proc = "idn_retry_con";
2616 
2617 	ASSERT(IDN_RETRY_TOKEN2TYPE(token) == IDNRETRY_CON);
2618 
2619 	IDN_SYNC_LOCK();
2620 	IDN_DLOCK_EXCL(domid);
2621 
2622 	if (dp->dxp != &xphase_con) {
2623 		STRING(str);
2624 
2625 #ifdef DEBUG
2626 		if (dp->dxp) {
2627 			INUM2STR(dp->dxp->xt_msgtype, str);
2628 		}
2629 #endif /* DEBUG */
2630 
2631 		PR_PROTO("%s:%d: dxp(%s) != CON...bailing...\n",
2632 		    proc, domid, dp->dxp ? str : "NULL");
2633 		IDN_DUNLOCK(domid);
2634 		IDN_SYNC_UNLOCK();
2635 		return;
2636 	}
2637 
2638 	if ((dp->dsync.s_cmd != IDNSYNC_CONNECT) ||
2639 	    (dp->dxstate != IDNXS_PEND)) {
2640 		PR_PROTO("%s:%d: cmd (%s) and/or xstate (%s) not "
2641 		    "expected (%s/%s)\n",
2642 		    proc, domid, idnsync_str[dp->dsync.s_cmd],
2643 		    idnxs_str[dp->dxstate], idnsync_str[IDNSYNC_CONNECT],
2644 		    idnxs_str[IDNXS_PEND]);
2645 		IDN_DUNLOCK(domid);
2646 		IDN_SYNC_UNLOCK();
2647 		return;
2648 	}
2649 
2650 	(void) idn_xphase_transition(domid, NULL, xargs);
2651 
2652 	IDN_DUNLOCK(domid);
2653 	IDN_SYNC_UNLOCK();
2654 }
2655 
2656 static int
2657 idn_check_con(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
2658 {
2659 	int		ready;
2660 	uint_t		msg = mtp ? mtp->mt_mtype : 0;
2661 	idn_domain_t	*dp = &idn_domain[domid];
2662 	domainset_t	ready_set, my_ready_set, query_set;
2663 
2664 	ASSERT(IDN_SYNC_IS_LOCKED());
2665 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
2666 
2667 	if (msg & IDNP_NACK)
2668 		return (0);
2669 
2670 	if ((dp->dstate == IDNDS_CON_PEND) &&
2671 	    (msg & IDNP_MSGTYPE_MASK) && (msg & IDNP_ACK))	/* con+ack */
2672 		return (1);
2673 
2674 	if (msg == 0) {
2675 		ready_set = idn.domset.ds_connected &
2676 		    ~idn.domset.ds_trans_off;
2677 	} else {
2678 		ready_set = GET_XARGS_CON_DOMSET(xargs);
2679 		DOMAINSET_ADD(idn.domset.ds_ready_on, domid);
2680 	}
2681 
2682 	DOMAINSET_ADD(ready_set, idn.localid);
2683 
2684 	query_set = idn_sync_register(domid, IDNSYNC_CONNECT,
2685 	    ready_set, IDNSYNC_REG_REG);
2686 	/*
2687 	 * No need to query this domain as he's already
2688 	 * in the CON sequence.
2689 	 */
2690 	DOMAINSET_DEL(query_set, domid);
2691 
2692 	ready = (dp->dsync.s_set_exp == dp->dsync.s_set_rdy) ? 1 : 0;
2693 	if (ready) {
2694 		DOMAINSET_DEL(idn.domset.ds_ready_on, domid);
2695 		DOMAINSET_ADD(idn.domset.ds_connected, domid);
2696 	}
2697 
2698 	if (query_set) {
2699 		int	d;
2700 
2701 		my_ready_set = idn.domset.ds_ready_on |
2702 		    idn.domset.ds_connected;
2703 		my_ready_set &= ~idn.domset.ds_trans_off;
2704 		DOMAINSET_ADD(my_ready_set, idn.localid);
2705 
2706 		for (d = 0; d < MAX_DOMAINS; d++) {
2707 			if (!DOMAIN_IN_SET(query_set, d))
2708 				continue;
2709 
2710 			dp = &idn_domain[d];
2711 
2712 			IDN_DLOCK_EXCL(d);
2713 			if ((dp->dsync.s_cmd == IDNSYNC_CONNECT) ||
2714 			    !dp->dcookie_send) {
2715 				IDN_DUNLOCK(d);
2716 				continue;
2717 			}
2718 
2719 			IDN_SYNC_QUERY_UPDATE(domid, d);
2720 
2721 			(void) idn_send_con(d, NULL, IDNCON_QUERY,
2722 			    my_ready_set);
2723 			IDN_DUNLOCK(d);
2724 		}
2725 	}
2726 
2727 	return (!msg ? 0 : (ready ? 0 : 1));
2728 }
2729 
2730 /*ARGSUSED2*/
2731 static void
2732 idn_error_con(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
2733 {
2734 	uint_t	token;
2735 	uint_t	msg = mtp ? mtp->mt_mtype : 0;
2736 
2737 	ASSERT(IDN_SYNC_IS_LOCKED());
2738 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
2739 
2740 	if (msg & IDNP_MSGTYPE_MASK) {
2741 		idn_msgtype_t	mt;
2742 		idn_xdcargs_t	nargs;
2743 
2744 		mt.mt_mtype = IDNP_NACK;
2745 		mt.mt_atype = msg;
2746 		mt.mt_cookie = mtp->mt_cookie;
2747 		CLR_XARGS(nargs);
2748 		SET_XARGS_NACK_TYPE(nargs, IDNNACK_RETRY);
2749 		idn_send_acknack(domid, &mt, nargs);
2750 	}
2751 
2752 	token = IDN_RETRY_TOKEN(domid, IDNRETRY_CON);
2753 	idn_retry_submit(idn_retry_con, NULL, token,
2754 	    idn_msg_retrytime[(int)IDNRETRY_CON]);
2755 }
2756 
2757 /*ARGSUSED*/
2758 static void
2759 idn_action_con_pend(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
2760 {
2761 	uint_t		msg = mtp ? mtp->mt_mtype : 0;
2762 	idn_domain_t	*dp = &idn_domain[domid];
2763 	idn_msgtype_t	mt;
2764 	domainset_t	my_ready_set;
2765 
2766 	ASSERT(IDN_SYNC_IS_LOCKED());
2767 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
2768 
2769 	my_ready_set = dp->dsync.s_set_rdy | idn.domset.ds_ready_on |
2770 	    idn.domset.ds_connected;
2771 	my_ready_set &= ~idn.domset.ds_trans_off;
2772 	DOMAINSET_ADD(my_ready_set, idn.localid);
2773 
2774 	if (!msg) {
2775 		(void) idn_send_con(domid, NULL, IDNCON_NORMAL, my_ready_set);
2776 	} else {
2777 		mt.mt_mtype = IDNP_CON | IDNP_ACK;
2778 		mt.mt_atype = 0;
2779 		mt.mt_cookie = mtp->mt_cookie;
2780 		(void) idn_send_con(domid, &mt, IDNCON_NORMAL, my_ready_set);
2781 	}
2782 }
2783 
2784 static void
2785 idn_action_con_sent(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
2786 {
2787 	uint_t		msg = mtp ? mtp->mt_mtype : 0;
2788 	idn_domain_t	*dp = &idn_domain[domid];
2789 	idn_con_t	contype;
2790 	domainset_t	my_ready_set;
2791 	idn_msgtype_t	mt;
2792 
2793 	ASSERT(IDN_SYNC_IS_LOCKED());
2794 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
2795 
2796 	mt.mt_cookie = mtp ? mtp->mt_cookie : 0;
2797 
2798 	my_ready_set = dp->dsync.s_set_rdy | idn.domset.ds_ready_on |
2799 	    idn.domset.ds_connected;
2800 	my_ready_set &= ~idn.domset.ds_trans_off;
2801 	DOMAINSET_ADD(my_ready_set, idn.localid);
2802 
2803 	contype = GET_XARGS_CON_TYPE(xargs);
2804 
2805 	if ((msg & IDNP_ACKNACK_MASK) == 0) {
2806 		/*
2807 		 * con
2808 		 */
2809 		mt.mt_mtype = IDNP_CON | IDNP_ACK;
2810 		mt.mt_atype = 0;
2811 		(void) idn_send_con(domid, &mt, contype, my_ready_set);
2812 	} else if (msg & IDNP_MSGTYPE_MASK) {
2813 		idn_xdcargs_t	cargs;
2814 
2815 		mt.mt_mtype = IDNP_ACK;
2816 		mt.mt_atype = msg;
2817 		CLR_XARGS(cargs);
2818 		SET_XARGS_CON_TYPE(cargs, contype);
2819 		SET_XARGS_CON_DOMSET(cargs, my_ready_set);
2820 		/*
2821 		 * con+ack
2822 		 */
2823 		idn_send_acknack(domid, &mt, cargs);
2824 	} else {
2825 		uint_t	token;
2826 		/*
2827 		 * nack - retry
2828 		 */
2829 		token = IDN_RETRY_TOKEN(domid, IDNRETRY_CON);
2830 		idn_retry_submit(idn_retry_con, NULL, token,
2831 		    idn_msg_retrytime[(int)IDNRETRY_CON]);
2832 	}
2833 }
2834 
2835 /*ARGSUSED*/
2836 static void
2837 idn_action_con_rcvd(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
2838 {
2839 	uint_t	msg = mtp ? mtp->mt_mtype : 0;
2840 
2841 	ASSERT(IDN_SYNC_IS_LOCKED());
2842 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
2843 
2844 	if (msg & IDNP_NACK) {
2845 		uint_t	token;
2846 		/*
2847 		 * nack - retry
2848 		 */
2849 		token = IDN_RETRY_TOKEN(domid, IDNRETRY_CON);
2850 		idn_retry_submit(idn_retry_con, NULL, token,
2851 		    idn_msg_retrytime[(int)IDNRETRY_CON]);
2852 	}
2853 }
2854 
2855 static void
2856 idn_final_con(int domid)
2857 {
2858 	uint_t		targ;
2859 	uint_t		token = IDN_RETRY_TOKEN(domid, IDNRETRY_CON);
2860 	idn_domain_t	*dp = &idn_domain[domid];
2861 	procname_t	proc = "idn_final_con";
2862 
2863 	ASSERT(IDN_SYNC_IS_LOCKED());
2864 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
2865 
2866 	(void) idn_retry_terminate(token);
2867 
2868 	dp->dxp = NULL;
2869 	IDN_XSTATE_TRANSITION(dp, IDNXS_NIL);
2870 
2871 	idn_sync_exit(domid, IDNSYNC_CONNECT);
2872 
2873 	CHECKPOINT_OPENED(IDNSB_CHKPT_LINK, dp->dhw.dh_boardset, 1);
2874 
2875 	DOMAINSET_DEL(idn.domset.ds_trans_on, domid);
2876 	DOMAINSET_DEL(idn.domset.ds_relink, domid);
2877 	IDN_FSTATE_TRANSITION(dp, IDNFIN_OFF);
2878 
2879 	PR_PROTO("%s:%d: CONNECTED\n", proc, domid);
2880 
2881 	if (idn.domset.ds_trans_on == 0) {
2882 		if ((idn.domset.ds_trans_off | idn.domset.ds_relink) == 0) {
2883 			PR_HITLIST("%s:%d: HITLIST %x -> 0\n",
2884 			    proc, domid, idn.domset.ds_hitlist);
2885 			idn.domset.ds_hitlist = 0;
2886 		}
2887 		PR_PROTO("%s:%d: ALL CONNECTED ************ "
2888 		    "(0x%x + 0x%x) = 0x%x\n", proc, domid,
2889 		    DOMAINSET(idn.localid), idn.domset.ds_connected,
2890 		    DOMAINSET(idn.localid) | idn.domset.ds_connected);
2891 	} else {
2892 		PR_PROTO("%s:%d: >>> ds_trans_on = 0x%x, ds_ready_on = 0x%x\n",
2893 		    proc, domid,
2894 		    idn.domset.ds_trans_on, idn.domset.ds_ready_on);
2895 	}
2896 
2897 	if (idn_verify_config_mbox(domid)) {
2898 		idnsb_error_t	idnerr;
2899 		/*
2900 		 * Mailbox is not cool. Need to disconnect.
2901 		 */
2902 		INIT_IDNKERR(&idnerr);
2903 		SET_IDNKERR_ERRNO(&idnerr, EPROTO);
2904 		SET_IDNKERR_IDNERR(&idnerr, IDNKERR_SMR_CORRUPTED);
2905 		SET_IDNKERR_PARAM0(&idnerr, domid);
2906 		idn_update_op(IDNOP_ERROR, DOMAINSET(domid), &idnerr);
2907 		/*
2908 		 * We cannot disconnect from an individual domain
2909 		 * unless all domains are attempting to disconnect
2910 		 * from him also, especially now since we touched
2911 		 * the SMR and now we have a potential cache conflicts
2912 		 * with the other domains with respect to this
2913 		 * domain.  Disconnect attempt will effectively
2914 		 * shutdown connection with respective domain
2915 		 * which is the effect we really want anyway.
2916 		 */
2917 		(void) idn_disconnect(domid, IDNFIN_NORMAL, IDNFIN_ARG_SMRBAD,
2918 		    IDNFIN_SYNC_YES);
2919 
2920 		return;
2921 	}
2922 
2923 	if (lock_try(&idn.first_swlink)) {
2924 		/*
2925 		 * This is our first connection.  Need to
2926 		 * kick some stuff into gear.
2927 		 */
2928 		idndl_dlpi_init();
2929 		(void) idn_activate_channel(CHANSET_ALL, IDNCHAN_ONLINE);
2930 
2931 		targ = 0xf0;
2932 	} else {
2933 		targ = 0;
2934 	}
2935 
2936 	idn_mainmbox_activate(domid);
2937 
2938 	idn_update_op(IDNOP_CONNECTED, DOMAINSET(domid), NULL);
2939 
2940 	IDN_GKSTAT_GLOBAL_EVENT(gk_links, gk_link_last);
2941 
2942 	membar_stst_ldst();
2943 
2944 	IDN_DSTATE_TRANSITION(dp, IDNDS_CONNECTED);
2945 	/*
2946 	 * Need to kick off initial commands in background.
2947 	 * We do not want to do them within the context of
2948 	 * a protocol server because they may sleep and thus
2949 	 * cause the protocol server to incur a soft-deadlock,
2950 	 * i.e. he's sleeping waiting in the slab-waiting area
2951 	 * for a response that will arrive on his protojob
2952 	 * queue, but which he obviously can't process since
2953 	 * he's not waiting on his protojob queue.
2954 	 */
2955 	targ |= domid & 0x0f;
2956 	(void) timeout(idn_link_established, (void *)(uintptr_t)targ, 50);
2957 
2958 	cmn_err(CE_NOTE,
2959 	    "!IDN: 200: link (domain %d, CPU %d) connected",
2960 	    dp->domid, dp->dcpu);
2961 }
2962 
2963 static void
2964 idn_exit_con(int domid, uint_t msgtype)
2965 {
2966 	idn_domain_t	*dp = &idn_domain[domid];
2967 	idn_fin_t	fintype;
2968 	procname_t	proc = "idn_exit_con";
2969 	STRING(str);
2970 
2971 	ASSERT(IDN_SYNC_IS_LOCKED());
2972 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
2973 
2974 	INUM2STR(msgtype, str);
2975 	PR_PROTO("%s:%d: msgtype = 0x%x(%s)\n", proc, domid, msgtype, str);
2976 
2977 	fintype = msgtype ? IDNFIN_NORMAL : IDNFIN_FORCE_HARD;
2978 
2979 	IDN_GLOCK_SHARED();
2980 	if (idn.state != IDNGS_DISCONNECT) {
2981 		DOMAINSET_ADD(idn.domset.ds_relink, domid);
2982 		IDN_HISTORY_LOG(IDNH_RELINK, domid, dp->dstate,
2983 		    idn.domset.ds_relink);
2984 	} else {
2985 		DOMAINSET_DEL(idn.domset.ds_relink, domid);
2986 	}
2987 	IDN_GUNLOCK();
2988 
2989 	(void) idn_disconnect(domid, fintype, IDNFIN_ARG_NONE,
2990 	    IDNDS_SYNC_TYPE(dp));
2991 }
2992 
2993 static int
2994 idn_send_fin(int domid, idn_msgtype_t *mtp, idn_fin_t fintype, idn_finarg_t
2995     finarg, idn_finopt_t finopt, domainset_t finset, uint_t finmaster)
2996 {
2997 	int		need_timer = 1;
2998 	uint_t		acknack;
2999 	uint_t		fintypearg = 0;
3000 	idn_msgtype_t	mt;
3001 	idn_domain_t	*dp = &idn_domain[domid];
3002 	procname_t	proc = "idn_send_fin";
3003 
3004 	ASSERT(IDN_SYNC_IS_LOCKED());
3005 	ASSERT(IDN_DLOCK_IS_HELD(domid));
3006 
3007 	ASSERT((fintype != IDNFIN_QUERY) ? (finopt != IDNFIN_OPT_NONE) : 1);
3008 
3009 	if (mtp) {
3010 		acknack = mtp->mt_mtype & IDNP_ACKNACK_MASK;
3011 		mt.mt_mtype = mtp->mt_mtype;
3012 		mt.mt_atype = mtp->mt_atype;
3013 		mt.mt_cookie = mtp->mt_cookie;
3014 	} else {
3015 		acknack = 0;
3016 		mt.mt_mtype = IDNP_FIN;
3017 		mt.mt_atype = 0;
3018 		/*
3019 		 * For simple FIN queries we want a unique
3020 		 * timer assigned.  For others, they
3021 		 * effectively share one.
3022 		 */
3023 		if (fintype == IDNFIN_QUERY)
3024 			mt.mt_cookie = 0;
3025 		else
3026 			mt.mt_cookie = IDN_TIMER_PUBLIC_COOKIE;
3027 	}
3028 
3029 	PR_PROTO("%s:%d: sending fin%sto (cpu %d) "
3030 	    "[ft=%s, fa=%s, fs=0x%x, fo=%s, fm=(%d,%d)]\n",
3031 	    proc, domid,
3032 	    (acknack & IDNP_ACK) ? "+ack " :
3033 	    (acknack & IDNP_NACK) ? "+nack " : " ",
3034 	    dp->dcpu, idnfin_str[fintype], idnfinarg_str[finarg],
3035 	    (int)finset, idnfinopt_str[finopt],
3036 	    FIN_MASTER_DOMID(finmaster), FIN_MASTER_CPUID(finmaster));
3037 
3038 	if (need_timer) {
3039 		IDN_MSGTIMER_START(domid, IDNP_FIN, (ushort_t)fintype,
3040 		    idn_msg_waittime[IDNP_FIN], &mt.mt_cookie);
3041 	}
3042 
3043 	SET_FIN_TYPE(fintypearg, fintype);
3044 	SET_FIN_ARG(fintypearg, finarg);
3045 
3046 	IDNXDC(domid, &mt, fintypearg, (uint_t)finset, (uint_t)finopt,
3047 	    finmaster);
3048 
3049 	return (0);
3050 }
3051 
3052 /*
3053  * Must leave w/DLOCK dropped and SYNC_LOCK held.
3054  */
3055 static int
3056 idn_recv_fin(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
3057 {
3058 	uint_t		msg = mtp ? mtp->mt_mtype : 0;
3059 	idn_fin_t	fintype;
3060 	idn_finarg_t	finarg;
3061 	idn_finopt_t	finopt;
3062 	domainset_t	my_ready_set, ready_set;
3063 	idn_msgtype_t	mt;
3064 	idn_domain_t	*dp = &idn_domain[domid];
3065 	idn_xdcargs_t	aargs;
3066 	procname_t	proc = "idn_recv_fin";
3067 
3068 	ASSERT(IDN_SYNC_IS_LOCKED());
3069 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
3070 
3071 	mt.mt_cookie = mtp ? mtp->mt_cookie : 0;
3072 
3073 	fintype   = GET_XARGS_FIN_TYPE(xargs);
3074 	finarg    = GET_XARGS_FIN_ARG(xargs);
3075 	ready_set = GET_XARGS_FIN_DOMSET(xargs);
3076 	finopt    = GET_XARGS_FIN_OPT(xargs);
3077 
3078 	CLR_XARGS(aargs);
3079 
3080 	if (msg & IDNP_NACK) {
3081 		PR_PROTO("%s:%d: received NACK (type = %s)\n",
3082 		    proc, domid, idnnack_str[xargs[0]]);
3083 	} else {
3084 		PR_PROTO("%s:%d: fintype = %s, finopt = %s, "
3085 		    "finarg = %s, ready_set = 0x%x\n",
3086 		    proc, domid, idnfin_str[fintype],
3087 		    idnfinopt_str[finopt],
3088 		    idnfinarg_str[finarg], ready_set);
3089 	}
3090 
3091 	if (!(msg & IDNP_NACK) && (fintype == IDNFIN_QUERY)) {
3092 		domainset_t	query_set;
3093 
3094 		query_set = idn_sync_register(domid, IDNSYNC_DISCONNECT,
3095 		    ready_set, IDNSYNC_REG_REG);
3096 
3097 		my_ready_set = ~idn.domset.ds_connected |
3098 		    idn.domset.ds_ready_off;
3099 
3100 		if (msg & IDNP_MSGTYPE_MASK) {
3101 			mt.mt_mtype = IDNP_ACK;
3102 			mt.mt_atype = IDNP_FIN;
3103 			SET_XARGS_FIN_TYPE(aargs, fintype);
3104 			SET_XARGS_FIN_ARG(aargs, finarg);
3105 			SET_XARGS_FIN_DOMSET(aargs, my_ready_set);
3106 			SET_XARGS_FIN_OPT(aargs, IDNFIN_OPT_NONE);
3107 			SET_XARGS_FIN_MASTER(aargs, NIL_FIN_MASTER);
3108 			idn_send_acknack(domid, &mt, aargs);
3109 		}
3110 
3111 		if (query_set) {
3112 			uint_t	token;
3113 
3114 			token = IDN_RETRY_TOKEN(domid, IDNRETRY_FINQ);
3115 			idn_retry_submit(idn_retry_query, NULL, token,
3116 			    idn_msg_retrytime[(int)IDNRETRY_FINQ]);
3117 		}
3118 
3119 		return (0);
3120 	}
3121 
3122 	if (dp->dxp != &xphase_fin) {
3123 		uint_t	token;
3124 
3125 		if (IDNDS_IS_CLOSED(dp)) {
3126 			PR_PROTO("%s:%d: domain already closed (%s)\n",
3127 			    proc, domid, idnds_str[dp->dstate]);
3128 			if (msg & IDNP_MSGTYPE_MASK) {
3129 				/*
3130 				 * fin or fin+ack.
3131 				 */
3132 				mt.mt_mtype = IDNP_NACK;
3133 				mt.mt_atype = msg;
3134 				SET_XARGS_NACK_TYPE(aargs, IDNNACK_NOCONN);
3135 				idn_send_acknack(domid, &mt, aargs);
3136 			}
3137 			return (0);
3138 		}
3139 		dp->dfin_sync = IDNDS_SYNC_TYPE(dp);
3140 
3141 		/*
3142 		 * Need to do some clean-up ala idn_disconnect().
3143 		 *
3144 		 * Terminate any outstanding commands that were
3145 		 * targeted towards this domain.
3146 		 */
3147 		idn_terminate_cmd(domid, ECANCELED);
3148 
3149 		/*
3150 		 * Terminate any and all retries that may have
3151 		 * outstanding for this domain.
3152 		 */
3153 		token = IDN_RETRY_TOKEN(domid, IDN_RETRY_TYPEALL);
3154 		(void) idn_retry_terminate(token);
3155 
3156 		/*
3157 		 * Stop all outstanding message timers for
3158 		 * this guy.
3159 		 */
3160 		IDN_MSGTIMER_STOP(domid, 0, 0);
3161 
3162 		dp->dxp = &xphase_fin;
3163 		IDN_XSTATE_TRANSITION(dp, IDNXS_PEND);
3164 	}
3165 
3166 	if (msg & IDNP_NACK) {
3167 		idn_nack_t	nack;
3168 
3169 		nack = GET_XARGS_NACK_TYPE(xargs);
3170 		if (nack == IDNNACK_NOCONN) {
3171 			/*
3172 			 * We're trying to FIN with somebody we're
3173 			 * already disconnected from.  Need to
3174 			 * speed this guy through.
3175 			 */
3176 			DOMAINSET_ADD(idn.domset.ds_ready_off, domid);
3177 			(void) idn_sync_register(domid, IDNSYNC_DISCONNECT,
3178 			    DOMAINSET_ALL, IDNSYNC_REG_REG);
3179 			ready_set = (uint_t)DOMAINSET_ALL;
3180 			/*
3181 			 * Need to transform message to allow us to
3182 			 * pass this guy right through and not waste time
3183 			 * talking to him.
3184 			 */
3185 			IDN_FSTATE_TRANSITION(dp, IDNFIN_FORCE_HARD);
3186 
3187 			switch (dp->dstate) {
3188 			case IDNDS_FIN_PEND:
3189 				mtp->mt_mtype = 0;
3190 				mtp->mt_atype = 0;
3191 				break;
3192 
3193 			case IDNDS_FIN_SENT:
3194 				mtp->mt_mtype = IDNP_FIN | IDNP_ACK;
3195 				mtp->mt_atype = 0;
3196 				break;
3197 
3198 			case IDNDS_FIN_RCVD:
3199 				mtp->mt_mtype = IDNP_ACK;
3200 				mtp->mt_atype = IDNP_FIN | IDNP_ACK;
3201 				break;
3202 
3203 			default:
3204 #ifdef DEBUG
3205 				cmn_err(CE_PANIC,
3206 				    "%s:%d: UNEXPECTED state = %s",
3207 				    proc, domid,
3208 				    idnds_str[dp->dstate]);
3209 #endif /* DEBUG */
3210 				break;
3211 			}
3212 		}
3213 		fintype = (uint_t)dp->dfin;
3214 		finopt = DOMAIN_IN_SET(idn.domset.ds_relink, domid) ?
3215 		    IDNFIN_OPT_RELINK : IDNFIN_OPT_UNLINK;
3216 
3217 		CLR_XARGS(xargs);
3218 		SET_XARGS_FIN_TYPE(xargs, fintype);
3219 		SET_XARGS_FIN_ARG(xargs, finarg);
3220 		SET_XARGS_FIN_DOMSET(xargs, ready_set);
3221 		SET_XARGS_FIN_OPT(xargs, finopt);
3222 		SET_XARGS_FIN_MASTER(xargs, NIL_FIN_MASTER);
3223 	}
3224 
3225 	(void) idn_xphase_transition(domid, mtp, xargs);
3226 
3227 	return (0);
3228 }
3229 
3230 /*ARGSUSED1*/
3231 static void
3232 idn_retry_fin(uint_t token, void *arg)
3233 {
3234 	int		domid = IDN_RETRY_TOKEN2DOMID(token);
3235 	int		new_masterid, new_cpuid = IDN_NIL_DCPU;
3236 	uint_t		finmaster;
3237 	idn_domain_t	*dp = &idn_domain[domid];
3238 	idn_xdcargs_t	xargs;
3239 	idn_finopt_t	finopt;
3240 	procname_t	proc = "idn_retry_fin";
3241 
3242 	ASSERT(IDN_RETRY_TOKEN2TYPE(token) == IDNRETRY_FIN);
3243 
3244 	IDN_SYNC_LOCK();
3245 	IDN_DLOCK_EXCL(domid);
3246 
3247 	if (dp->dxp != &xphase_fin) {
3248 		PR_PROTO("%s:%d: dxp(0x%p) != xstate_fin(0x%p)...bailing\n",
3249 		    proc, domid, (void *)dp->dxp, (void *)&xphase_fin);
3250 		IDN_DUNLOCK(domid);
3251 		IDN_SYNC_UNLOCK();
3252 		return;
3253 	}
3254 
3255 	if (dp->dxstate != IDNXS_PEND) {
3256 		PR_PROTO("%s:%d: xstate(%s) != %s...bailing\n",
3257 		    proc, domid, idnxs_str[dp->dxstate],
3258 		    idnxs_str[IDNXS_PEND]);
3259 		IDN_DUNLOCK(domid);
3260 		IDN_SYNC_UNLOCK();
3261 		return;
3262 	}
3263 
3264 	finopt = DOMAIN_IN_SET(idn.domset.ds_relink, domid) ?
3265 	    IDNFIN_OPT_RELINK : IDNFIN_OPT_UNLINK;
3266 
3267 	CLR_XARGS(xargs);
3268 	SET_XARGS_FIN_TYPE(xargs, dp->dfin);
3269 	/*LINTED*/
3270 	SET_XARGS_FIN_ARG(xargs, IDNFIN_ARG_NONE);
3271 	SET_XARGS_FIN_OPT(xargs, finopt);
3272 	SET_XARGS_FIN_DOMSET(xargs, 0);		/* unused when msg == 0 */
3273 	IDN_GLOCK_SHARED();
3274 	new_masterid = IDN_GET_NEW_MASTERID();
3275 	IDN_GUNLOCK();
3276 	if (new_masterid != IDN_NIL_DOMID)
3277 		new_cpuid = idn_domain[new_masterid].dcpu;
3278 	finmaster = MAKE_FIN_MASTER(new_masterid, new_cpuid);
3279 	SET_XARGS_FIN_MASTER(xargs, finmaster);
3280 
3281 	(void) idn_xphase_transition(domid, NULL, xargs);
3282 
3283 	IDN_DUNLOCK(domid);
3284 	IDN_SYNC_UNLOCK();
3285 }
3286 
3287 static int
3288 idn_check_fin_pend(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
3289 {
3290 	idn_domain_t	*dp = &idn_domain[domid];
3291 	idn_fin_t	fintype;
3292 	idn_finopt_t	finopt;
3293 	idn_finarg_t	finarg;
3294 	int		ready;
3295 	int		finmasterid;
3296 	int		fincpuid;
3297 	uint_t		finmaster;
3298 	uint_t		msg = mtp ? mtp->mt_mtype : 0;
3299 	domainset_t	query_set, ready_set, conn_set;
3300 	domainset_t	my_ready_set, shutdown_set;
3301 	procname_t	proc = "idn_check_fin_pend";
3302 
3303 	ASSERT(IDN_SYNC_IS_LOCKED());
3304 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
3305 
3306 	if (msg & IDNP_NACK)
3307 		return (0);
3308 
3309 	if ((dp->dstate == IDNDS_FIN_PEND) && (msg & IDNP_MSGTYPE_MASK) &&
3310 	    (msg & IDNP_ACK))		/* fin+ack */
3311 		return (1);
3312 
3313 	query_set = 0;
3314 
3315 	if (!DOMAIN_IN_SET(idn.domset.ds_trans_off, domid)) {
3316 		/*
3317 		 * Can't remove domain from ds_connected yet,
3318 		 * since he's still officially connected until
3319 		 * we get an ACK from him.
3320 		 */
3321 		DOMAINSET_DEL(idn.domset.ds_trans_on, domid);
3322 		DOMAINSET_ADD(idn.domset.ds_trans_off, domid);
3323 	}
3324 
3325 	IDN_GLOCK_SHARED();
3326 	conn_set = (idn.domset.ds_connected | idn.domset.ds_trans_on) &
3327 	    ~idn.domset.ds_trans_off;
3328 	if ((idn.state == IDNGS_DISCONNECT) ||
3329 	    (idn.state == IDNGS_RECONFIG) ||
3330 	    (domid == IDN_GET_MASTERID()) || !conn_set) {
3331 		/*
3332 		 * If we're disconnecting, reconfiguring,
3333 		 * unlinking from the master, or unlinking
3334 		 * the last of our connections, then we need
3335 		 * to shutdown all the channels.
3336 		 */
3337 		shutdown_set = DOMAINSET_ALL;
3338 	} else {
3339 		shutdown_set = DOMAINSET(domid);
3340 	}
3341 	IDN_GUNLOCK();
3342 
3343 	idn_shutdown_datapath(shutdown_set, (dp->dfin == IDNFIN_FORCE_HARD));
3344 
3345 	IDN_GLOCK_EXCL();
3346 	/*
3347 	 * Remap the SMR back to our local space if the remote
3348 	 * domain going down is the master.  We do this now before
3349 	 * flushing caches.  This will help guarantee that any
3350 	 * accidental accesses to the SMR after the cache flush
3351 	 * will only go to local memory.
3352 	 */
3353 	if ((domid == IDN_GET_MASTERID()) && (idn.smr.rempfn != PFN_INVALID)) {
3354 		PR_PROTO("%s:%d: deconfiging CURRENT MASTER - SMR remap\n",
3355 		    proc, domid);
3356 		IDN_DLOCK_EXCL(idn.localid);
3357 		/*
3358 		 * We're going to remap the SMR,
3359 		 * so gotta blow away our local
3360 		 * pointer to the mbox table.
3361 		 */
3362 		idn_domain[idn.localid].dmbox.m_tbl = NULL;
3363 		IDN_DUNLOCK(idn.localid);
3364 
3365 		idn.smr.rempfn = PFN_INVALID;
3366 		idn.smr.rempfnlim = PFN_INVALID;
3367 
3368 		smr_remap(&kas, idn.smr.vaddr, idn.smr.locpfn, IDN_SMR_SIZE);
3369 	}
3370 	IDN_GUNLOCK();
3371 
3372 	if (DOMAIN_IN_SET(idn.domset.ds_flush, domid)) {
3373 		idnxf_flushall_ecache();
3374 		CHECKPOINT_CLOSED(IDNSB_CHKPT_CACHE, dp->dhw.dh_boardset, 2);
3375 		DOMAINSET_DEL(idn.domset.ds_flush, domid);
3376 	}
3377 
3378 	fintype   = GET_XARGS_FIN_TYPE(xargs);
3379 	finarg    = GET_XARGS_FIN_ARG(xargs);
3380 	ready_set = GET_XARGS_FIN_DOMSET(xargs);
3381 	finopt    = GET_XARGS_FIN_OPT(xargs);
3382 
3383 	ASSERT(fintype != IDNFIN_QUERY);
3384 	if (!VALID_FIN(fintype)) {
3385 		/*
3386 		 * If for some reason remote domain
3387 		 * sent us an invalid FIN type,
3388 		 * override it to a  NORMAL fin.
3389 		 */
3390 		PR_PROTO("%s:%d: WARNING invalid fintype (%d) -> %s(%d)\n",
3391 		    proc, domid, (int)fintype,
3392 		    idnfin_str[IDNFIN_NORMAL], (int)IDNFIN_NORMAL);
3393 		fintype = IDNFIN_NORMAL;
3394 	}
3395 
3396 	if (!VALID_FINOPT(finopt)) {
3397 		PR_PROTO("%s:%d: WARNING invalid finopt (%d) -> %s(%d)\n",
3398 		    proc, domid, (int)finopt,
3399 		    idnfinopt_str[IDNFIN_OPT_UNLINK],
3400 		    (int)IDNFIN_OPT_UNLINK);
3401 		finopt = IDNFIN_OPT_UNLINK;
3402 	}
3403 
3404 	finmaster = GET_XARGS_FIN_MASTER(xargs);
3405 	finmasterid = FIN_MASTER_DOMID(finmaster);
3406 	fincpuid = FIN_MASTER_CPUID(finmaster);
3407 
3408 	if ((finarg != IDNFIN_ARG_NONE) &&
3409 	    !DOMAIN_IN_SET(idn.domset.ds_hitlist, domid)) {
3410 		idnsb_error_t	idnerr;
3411 
3412 		INIT_IDNKERR(&idnerr);
3413 		SET_IDNKERR_ERRNO(&idnerr, EPROTO);
3414 		SET_IDNKERR_IDNERR(&idnerr, FINARG2IDNKERR(finarg));
3415 		SET_IDNKERR_PARAM0(&idnerr, domid);
3416 
3417 		if (IDNFIN_ARG_IS_FATAL(finarg)) {
3418 			finopt = IDNFIN_OPT_UNLINK;
3419 			DOMAINSET_DEL(idn.domset.ds_relink, domid);
3420 			DOMAINSET_ADD(idn.domset.ds_hitlist, domid);
3421 
3422 			if (idn.domset.ds_connected == 0) {
3423 				domainset_t	domset;
3424 
3425 				IDN_GLOCK_EXCL();
3426 				domset = ~idn.domset.ds_relink;
3427 				if (idn.domset.ds_relink == 0) {
3428 					IDN_GSTATE_TRANSITION(IDNGS_DISCONNECT);
3429 				}
3430 				domset &= ~idn.domset.ds_hitlist;
3431 				/*
3432 				 * The primary domain we were trying to
3433 				 * connect to fin'd us with a fatal argument.
3434 				 * Something isn't cool in our IDN environment,
3435 				 * e.g. corrupted SMR or non-compatible CONFIG
3436 				 * parameters.  In any case we need to dismantle
3437 				 * ourselves completely.
3438 				 */
3439 				IDN_SET_NEW_MASTERID(IDN_NIL_DOMID);
3440 				IDN_GUNLOCK();
3441 				IDN_DUNLOCK(domid);
3442 
3443 				DOMAINSET_DEL(domset, idn.localid);
3444 				DOMAINSET_DEL(domset, domid);
3445 
3446 				idn_update_op(IDNOP_ERROR, DOMAINSET_ALL,
3447 				    &idnerr);
3448 
3449 				PR_HITLIST("%s:%d: unlink_domainset(%x) "
3450 				    "due to CFG error (relink=%x, "
3451 				    "hitlist=%x)\n", proc, domid, domset,
3452 				    idn.domset.ds_relink,
3453 				    idn.domset.ds_hitlist);
3454 
3455 				idn_unlink_domainset(domset, IDNFIN_NORMAL,
3456 				    finarg, IDNFIN_OPT_UNLINK, BOARDSET_ALL);
3457 				IDN_DLOCK_EXCL(domid);
3458 			}
3459 			PR_HITLIST("%s:%d: CFG error, (conn=%x, relink=%x, "
3460 			    "hitlist=%x)\n",
3461 			    proc, domid, idn.domset.ds_connected,
3462 			    idn.domset.ds_relink, idn.domset.ds_hitlist);
3463 		}
3464 		idn_update_op(IDNOP_ERROR, DOMAINSET(domid), &idnerr);
3465 	}
3466 
3467 	if ((finmasterid != IDN_NIL_DOMID) && (!VALID_DOMAINID(finmasterid) ||
3468 	    DOMAIN_IN_SET(idn.domset.ds_hitlist, domid))) {
3469 		PR_HITLIST("%s:%d: finmasterid = %d -> -1, relink=%x, "
3470 		    "hitlist=%x\n",
3471 		    proc, domid, finmasterid, idn.domset.ds_relink,
3472 		    idn.domset.ds_hitlist);
3473 		PR_PROTO("%s:%d: WARNING invalid finmasterid (%d) -> -1\n",
3474 		    proc, domid, finmasterid);
3475 		finmasterid = IDN_NIL_DOMID;
3476 	}
3477 
3478 	IDN_GLOCK_EXCL();
3479 
3480 	if ((finopt == IDNFIN_OPT_RELINK) && (idn.state != IDNGS_DISCONNECT)) {
3481 		DOMAINSET_ADD(idn.domset.ds_relink, domid);
3482 		IDN_HISTORY_LOG(IDNH_RELINK, domid, dp->dstate,
3483 		    idn.domset.ds_relink);
3484 	} else {
3485 		DOMAINSET_DEL(idn.domset.ds_relink, domid);
3486 		DOMAINSET_ADD(idn.domset.ds_hitlist, domid);
3487 	}
3488 
3489 	if ((domid == IDN_GET_NEW_MASTERID()) &&
3490 	    !DOMAIN_IN_SET(idn.domset.ds_relink, domid)) {
3491 		IDN_SET_NEW_MASTERID(IDN_NIL_DOMID);
3492 	}
3493 
3494 	if ((idn.state != IDNGS_DISCONNECT) && (idn.state != IDNGS_RECONFIG) &&
3495 	    (domid == IDN_GET_MASTERID())) {
3496 		domainset_t	dis_set, master_candidates;
3497 
3498 		IDN_GKSTAT_GLOBAL_EVENT(gk_reconfigs, gk_reconfig_last);
3499 
3500 		IDN_GSTATE_TRANSITION(IDNGS_RECONFIG);
3501 		IDN_GUNLOCK();
3502 
3503 		if ((finmasterid != IDN_NIL_DOMID) &&
3504 		    (finmasterid != idn.localid)) {
3505 			if (finmasterid != domid)
3506 				IDN_DLOCK_EXCL(finmasterid);
3507 			if (idn_open_domain(finmasterid, fincpuid, 0) < 0) {
3508 				cmn_err(CE_WARN,
3509 				    "IDN: 205: (%s) failed to "
3510 				    "open-domain(%d,%d)",
3511 				    proc, finmasterid, fincpuid);
3512 				if (finmasterid != domid)
3513 					IDN_DUNLOCK(finmasterid);
3514 				finmasterid = IDN_NIL_DOMID;
3515 			}
3516 			if (finmasterid != domid)
3517 				IDN_DUNLOCK(finmasterid);
3518 		}
3519 
3520 		IDN_GLOCK_EXCL();
3521 		if (finmasterid == IDN_NIL_DOMID) {
3522 			int	m;
3523 
3524 			master_candidates = idn.domset.ds_trans_on |
3525 			    idn.domset.ds_connected |
3526 			    idn.domset.ds_relink;
3527 			master_candidates &= ~(idn.domset.ds_trans_off &
3528 			    ~idn.domset.ds_relink);
3529 			DOMAINSET_DEL(master_candidates, domid);
3530 			/*
3531 			 * Local domain gets to participate also.
3532 			 */
3533 			DOMAINSET_ADD(master_candidates, idn.localid);
3534 
3535 			m = idn_select_candidate(master_candidates);
3536 			IDN_SET_NEW_MASTERID(m);
3537 		} else {
3538 			IDN_SET_NEW_MASTERID(finmasterid);
3539 		}
3540 		IDN_GUNLOCK();
3541 
3542 		dis_set = idn.domset.ds_trans_on | idn.domset.ds_connected;
3543 		DOMAINSET_DEL(dis_set, domid);
3544 
3545 		idn_unlink_domainset(dis_set, IDNFIN_NORMAL, IDNFIN_ARG_NONE,
3546 		    IDNFIN_OPT_RELINK, BOARDSET_ALL);
3547 	} else {
3548 		IDN_GUNLOCK();
3549 	}
3550 
3551 	/*
3552 	 * My local ready-set are those domains from which I
3553 	 * have confirmed no datapaths exist.
3554 	 */
3555 	my_ready_set = ~idn.domset.ds_connected;
3556 
3557 	switch (dp->dfin) {
3558 	case IDNFIN_NORMAL:
3559 	case IDNFIN_FORCE_SOFT:
3560 	case IDNFIN_FORCE_HARD:
3561 		if (fintype < dp->dfin) {
3562 			/*
3563 			 * Remote domain has requested a
3564 			 * FIN of lower priority than what
3565 			 * we're currently running.  Just
3566 			 * leave the priority where it is.
3567 			 */
3568 			break;
3569 		}
3570 		/*FALLTHROUGH*/
3571 
3572 	default:
3573 		IDN_FSTATE_TRANSITION(dp, fintype);
3574 		break;
3575 	}
3576 
3577 	ASSERT(dp->dfin_sync != IDNFIN_SYNC_OFF);
3578 
3579 	if (msg == 0) {
3580 		/*
3581 		 * Local domain is initiating a FIN sequence
3582 		 * to remote domid.  Note that remote domain
3583 		 * remains in ds_connected even though he's
3584 		 * in thet ready-set from the local domain's
3585 		 * perspective.  We can't remove him from
3586 		 * ds_connected until we get a confirmed message
3587 		 * from him indicating he has ceased communication.
3588 		 */
3589 		ready_set = my_ready_set;
3590 	} else {
3591 		/*
3592 		 * Remote domain initiated a FIN sequence
3593 		 * to local domain.  This implies that he
3594 		 * has shutdown his datapath to us.  Since
3595 		 * we shutdown our datapath to him, we're
3596 		 * effectively now in his ready-set.
3597 		 */
3598 		DOMAINSET_ADD(ready_set, idn.localid);
3599 		/*
3600 		 * Since we know both sides of the connection
3601 		 * have ceased, this remote domain is effectively
3602 		 * considered disconnected.
3603 		 */
3604 		DOMAINSET_ADD(idn.domset.ds_ready_off, domid);
3605 	}
3606 
3607 	if (dp->dfin == IDNFIN_FORCE_HARD) {
3608 		/*
3609 		 * If we're doing a hard disconnect
3610 		 * of this domain then we want to
3611 		 * blow straight through and not
3612 		 * waste time trying to talk to the
3613 		 * remote domain nor to domains we
3614 		 * believe are AWOL.  Although we will
3615 		 * try and do it cleanly with
3616 		 * everybody else.
3617 		 */
3618 		DOMAINSET_ADD(my_ready_set, domid);
3619 		my_ready_set |= idn.domset.ds_awol;
3620 		ready_set = DOMAINSET_ALL;
3621 
3622 	} else if (dp->dfin_sync == IDNFIN_SYNC_NO) {
3623 		/*
3624 		 * If we're not fin'ing this domain
3625 		 * synchronously then the only
3626 		 * expected domain set is himself.
3627 		 */
3628 		ready_set |= ~DOMAINSET(domid);
3629 		my_ready_set |= ~DOMAINSET(domid);
3630 	}
3631 
3632 	if (dp->dsync.s_cmd != IDNSYNC_DISCONNECT) {
3633 		idn_sync_exit(domid, IDNSYNC_CONNECT);
3634 		idn_sync_enter(domid, IDNSYNC_DISCONNECT, DOMAINSET_ALL,
3635 		    my_ready_set, idn_xstate_transfunc,	(void *)IDNP_FIN);
3636 	}
3637 
3638 	query_set = idn_sync_register(domid, IDNSYNC_DISCONNECT, ready_set,
3639 	    IDNSYNC_REG_REG);
3640 
3641 	/*
3642 	 * No need to query this domain as he's already
3643 	 * in the FIN sequence.
3644 	 */
3645 	DOMAINSET_DEL(query_set, domid);
3646 
3647 	ready = (dp->dsync.s_set_exp == dp->dsync.s_set_rdy) ? 1 : 0;
3648 	if (ready) {
3649 		DOMAINSET_DEL(idn.domset.ds_ready_off, domid);
3650 		DOMAINSET_DEL(idn.domset.ds_connected, domid);
3651 	}
3652 
3653 	if (query_set) {
3654 		int	d;
3655 
3656 		my_ready_set = idn.domset.ds_ready_off |
3657 		    ~idn.domset.ds_connected;
3658 
3659 		for (d = 0; d < MAX_DOMAINS; d++) {
3660 			if (!DOMAIN_IN_SET(query_set, d))
3661 				continue;
3662 
3663 			dp = &idn_domain[d];
3664 
3665 			IDN_DLOCK_EXCL(d);
3666 
3667 			if (dp->dsync.s_cmd == IDNSYNC_DISCONNECT) {
3668 				IDN_DUNLOCK(d);
3669 				continue;
3670 			}
3671 
3672 			IDN_SYNC_QUERY_UPDATE(domid, d);
3673 
3674 			(void) idn_send_fin(d, NULL, IDNFIN_QUERY,
3675 			    IDNFIN_ARG_NONE, IDNFIN_OPT_NONE, my_ready_set,
3676 			    NIL_FIN_MASTER);
3677 			IDN_DUNLOCK(d);
3678 		}
3679 	}
3680 
3681 	return (!msg ? 0 : (ready ? 0 : 1));
3682 }
3683 
3684 /*ARGSUSED*/
3685 static void
3686 idn_error_fin_pend(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
3687 {
3688 	uint_t	msg = mtp ? mtp->mt_mtype : 0;
3689 	uint_t	token;
3690 
3691 	ASSERT(IDN_SYNC_IS_LOCKED());
3692 	ASSERT(IDN_DLOCK_IS_HELD(domid));
3693 
3694 	/*
3695 	 * Don't communicate with domains that
3696 	 * we're forcing a hard disconnect.
3697 	 */
3698 	if ((idn_domain[domid].dfin != IDNFIN_FORCE_HARD) &&
3699 	    (msg & IDNP_MSGTYPE_MASK)) {
3700 		idn_msgtype_t	mt;
3701 		idn_xdcargs_t	nargs;
3702 
3703 		mt.mt_mtype = IDNP_NACK;
3704 		mt.mt_atype = msg;
3705 		mt.mt_cookie = mtp->mt_cookie;
3706 		CLR_XARGS(nargs);
3707 		SET_XARGS_NACK_TYPE(nargs, IDNNACK_RETRY);
3708 		idn_send_acknack(domid, &mt, nargs);
3709 	}
3710 
3711 	token = IDN_RETRY_TOKEN(domid, IDNRETRY_FIN);
3712 	idn_retry_submit(idn_retry_fin, NULL, token,
3713 	    idn_msg_retrytime[(int)IDNRETRY_FIN]);
3714 }
3715 
3716 static void
3717 idn_action_fin_pend(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
3718 {
3719 	idn_domain_t	*dp = &idn_domain[domid];
3720 	domainset_t	my_ready_set;
3721 	idn_finopt_t	finopt;
3722 	idn_finarg_t	finarg;
3723 	uint_t		finmaster;
3724 	int		new_masterid, new_cpuid = IDN_NIL_DCPU;
3725 	uint_t		msg = mtp ? mtp->mt_mtype : 0;
3726 	idn_msgtype_t	mt;
3727 
3728 	ASSERT(IDN_SYNC_IS_LOCKED());
3729 	ASSERT(IDN_DLOCK_IS_HELD(domid));
3730 
3731 	my_ready_set = dp->dsync.s_set_rdy | idn.domset.ds_ready_off |
3732 	    ~idn.domset.ds_connected;
3733 
3734 	ASSERT(xargs[0] != (uint_t)IDNFIN_QUERY);
3735 
3736 	finarg = GET_XARGS_FIN_ARG(xargs);
3737 	finopt = DOMAIN_IN_SET(idn.domset.ds_relink, domid) ?
3738 	    IDNFIN_OPT_RELINK : IDNFIN_OPT_UNLINK;
3739 
3740 	mt.mt_cookie = mtp ? mtp->mt_cookie : 0;
3741 
3742 	IDN_GLOCK_SHARED();
3743 	new_masterid = IDN_GET_NEW_MASTERID();
3744 	IDN_GUNLOCK();
3745 	if (new_masterid != IDN_NIL_DOMID)
3746 		new_cpuid = idn_domain[new_masterid].dcpu;
3747 	finmaster = MAKE_FIN_MASTER(new_masterid, new_cpuid);
3748 
3749 	if (dp->dfin == IDNFIN_FORCE_HARD) {
3750 		ASSERT(IDN_DLOCK_IS_EXCL(domid));
3751 
3752 		if (!msg) {
3753 			mt.mt_mtype = IDNP_FIN | IDNP_ACK;
3754 			mt.mt_atype = 0;
3755 		} else {
3756 			mt.mt_mtype = IDNP_ACK;
3757 			mt.mt_atype = IDNP_FIN | IDNP_ACK;
3758 		}
3759 		(void) idn_xphase_transition(domid, &mt, xargs);
3760 	} else if (!msg) {
3761 		(void) idn_send_fin(domid, NULL, dp->dfin, finarg,
3762 		    finopt, my_ready_set, finmaster);
3763 	} else if ((msg & IDNP_ACKNACK_MASK) == 0) {
3764 		/*
3765 		 * fin
3766 		 */
3767 		mt.mt_mtype = IDNP_FIN | IDNP_ACK;
3768 		mt.mt_atype = 0;
3769 		(void) idn_send_fin(domid, &mt, dp->dfin, finarg,
3770 		    finopt, my_ready_set, finmaster);
3771 	} else {
3772 		uint_t	token;
3773 		/*
3774 		 * nack - retry
3775 		 */
3776 		token = IDN_RETRY_TOKEN(domid, IDNRETRY_FIN);
3777 		idn_retry_submit(idn_retry_fin, NULL, token,
3778 		    idn_msg_retrytime[(int)IDNRETRY_FIN]);
3779 	}
3780 }
3781 
3782 static int
3783 idn_check_fin_sent(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
3784 {
3785 	int		ready;
3786 	uint_t		msg = mtp ? mtp->mt_mtype : 0;
3787 	idn_fin_t	fintype;
3788 	idn_finopt_t	finopt;
3789 	idn_domain_t	*dp = &idn_domain[domid];
3790 	domainset_t	query_set, ready_set;
3791 
3792 	ASSERT(IDN_SYNC_IS_LOCKED());
3793 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
3794 
3795 	if (msg & IDNP_NACK)
3796 		return (0);
3797 
3798 	fintype   = GET_XARGS_FIN_TYPE(xargs);
3799 	ready_set = GET_XARGS_FIN_DOMSET(xargs);
3800 	finopt    = GET_XARGS_FIN_OPT(xargs);
3801 
3802 	ASSERT(fintype != IDNFIN_QUERY);
3803 	if (!VALID_FIN(fintype)) {
3804 		/*
3805 		 * If for some reason remote domain
3806 		 * sent us an invalid FIN type,
3807 		 * override it to a  NORMAL fin.
3808 		 */
3809 		fintype = IDNFIN_NORMAL;
3810 	}
3811 
3812 	if (!VALID_FINOPT(finopt)) {
3813 		finopt = IDNFIN_OPT_UNLINK;
3814 	}
3815 	IDN_GLOCK_SHARED();
3816 	if ((finopt == IDNFIN_OPT_RELINK) && (idn.state != IDNGS_DISCONNECT)) {
3817 		DOMAINSET_ADD(idn.domset.ds_relink, domid);
3818 		IDN_HISTORY_LOG(IDNH_RELINK, domid, dp->dstate,
3819 		    idn.domset.ds_relink);
3820 	} else {
3821 		DOMAINSET_DEL(idn.domset.ds_relink, domid);
3822 	}
3823 	IDN_GUNLOCK();
3824 
3825 	switch (dp->dfin) {
3826 	case IDNFIN_NORMAL:
3827 	case IDNFIN_FORCE_SOFT:
3828 	case IDNFIN_FORCE_HARD:
3829 		if (fintype < dp->dfin) {
3830 			/*
3831 			 * Remote domain has requested a
3832 			 * FIN of lower priority than what
3833 			 * we're current running.  Just
3834 			 * leave the priority where it is.
3835 			 */
3836 			break;
3837 		}
3838 		/*FALLTHROUGH*/
3839 
3840 	default:
3841 		IDN_FSTATE_TRANSITION(dp, fintype);
3842 		break;
3843 	}
3844 
3845 	if (dp->dfin == IDNFIN_FORCE_HARD) {
3846 		/*
3847 		 * If we're doing a hard disconnect
3848 		 * of this domain then we want to
3849 		 * blow straight through and not
3850 		 * waste time trying to talk to the
3851 		 * remote domain.  By registering him
3852 		 * as ready with respect to all
3853 		 * possible domains he'll transition
3854 		 * immediately.  Note that we'll still
3855 		 * try and do it coherently with
3856 		 * other domains to which we're connected.
3857 		 */
3858 		ready_set = DOMAINSET_ALL;
3859 	} else {
3860 		DOMAINSET_ADD(ready_set, idn.localid);
3861 	}
3862 
3863 	DOMAINSET_ADD(idn.domset.ds_ready_off, domid);
3864 
3865 	query_set = idn_sync_register(domid, IDNSYNC_DISCONNECT,
3866 	    ready_set, IDNSYNC_REG_REG);
3867 	/*
3868 	 * No need to query this domain as he's already
3869 	 * in the FIN sequence.
3870 	 */
3871 	DOMAINSET_DEL(query_set, domid);
3872 
3873 	ready = (dp->dsync.s_set_exp == dp->dsync.s_set_rdy) ? 1 : 0;
3874 	if (ready) {
3875 		DOMAINSET_DEL(idn.domset.ds_ready_off, domid);
3876 		DOMAINSET_DEL(idn.domset.ds_connected, domid);
3877 	}
3878 
3879 	if (query_set) {
3880 		int		d;
3881 		domainset_t	my_ready_set;
3882 
3883 		my_ready_set = idn.domset.ds_ready_off |
3884 		    ~idn.domset.ds_connected;
3885 
3886 		for (d = 0; d < MAX_DOMAINS; d++) {
3887 			if (!DOMAIN_IN_SET(query_set, d))
3888 				continue;
3889 
3890 			dp = &idn_domain[d];
3891 
3892 			IDN_DLOCK_EXCL(d);
3893 
3894 			if (dp->dsync.s_cmd == IDNSYNC_DISCONNECT) {
3895 				IDN_DUNLOCK(d);
3896 				continue;
3897 			}
3898 
3899 			IDN_SYNC_QUERY_UPDATE(domid, d);
3900 
3901 			(void) idn_send_fin(d, NULL, IDNFIN_QUERY,
3902 			    IDNFIN_ARG_NONE, IDNFIN_OPT_NONE, my_ready_set,
3903 			    NIL_FIN_MASTER);
3904 			IDN_DUNLOCK(d);
3905 		}
3906 	}
3907 
3908 	return ((ready > 0) ? 0 : 1);
3909 }
3910 
3911 /*ARGSUSED*/
3912 static void
3913 idn_error_fin_sent(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
3914 {
3915 	uint_t	msg = mtp ? mtp->mt_mtype : 0;
3916 	uint_t	token;
3917 
3918 	ASSERT(IDN_SYNC_IS_LOCKED());
3919 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
3920 
3921 	/*
3922 	 * Don't communicate with domains that
3923 	 * we're forcing a hard disconnect.
3924 	 */
3925 	if ((idn_domain[domid].dfin != IDNFIN_FORCE_HARD) &&
3926 	    (msg & IDNP_MSGTYPE_MASK)) {
3927 		idn_msgtype_t	mt;
3928 		idn_xdcargs_t	nargs;
3929 
3930 		mt.mt_mtype = IDNP_NACK;
3931 		mt.mt_atype = msg;
3932 		mt.mt_cookie = mtp->mt_cookie;
3933 		CLR_XARGS(nargs);
3934 		SET_XARGS_NACK_TYPE(nargs, IDNNACK_RETRY);
3935 		idn_send_acknack(domid, &mt, nargs);
3936 	}
3937 
3938 	token = IDN_RETRY_TOKEN(domid, IDNRETRY_FIN);
3939 	idn_retry_submit(idn_retry_fin, NULL, token,
3940 	    idn_msg_retrytime[(int)IDNRETRY_FIN]);
3941 }
3942 
3943 static void
3944 idn_action_fin_sent(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
3945 {
3946 	uint_t		msg = mtp ? mtp->mt_mtype : 0;
3947 	int		new_masterid, new_cpuid = IDN_NIL_DCPU;
3948 	uint_t		finmaster;
3949 	idn_msgtype_t	mt;
3950 	idn_finopt_t	finopt;
3951 	idn_finarg_t	finarg;
3952 	domainset_t	my_ready_set;
3953 	idn_domain_t	*dp = &idn_domain[domid];
3954 
3955 	ASSERT(IDN_SYNC_IS_LOCKED());
3956 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
3957 
3958 	mt.mt_cookie = mtp ? mtp->mt_cookie : 0;
3959 
3960 	finopt = DOMAIN_IN_SET(idn.domset.ds_relink, domid) ?
3961 	    IDNFIN_OPT_RELINK : IDNFIN_OPT_UNLINK;
3962 
3963 	finarg = GET_XARGS_FIN_ARG(xargs);
3964 
3965 	my_ready_set = dp->dsync.s_set_rdy | idn.domset.ds_ready_off |
3966 	    ~idn.domset.ds_connected;
3967 
3968 	IDN_GLOCK_SHARED();
3969 	new_masterid = IDN_GET_NEW_MASTERID();
3970 	IDN_GUNLOCK();
3971 	if (new_masterid != IDN_NIL_DOMID)
3972 		new_cpuid = idn_domain[new_masterid].dcpu;
3973 	finmaster = MAKE_FIN_MASTER(new_masterid, new_cpuid);
3974 
3975 	if ((msg & IDNP_ACKNACK_MASK) == 0) {
3976 		/*
3977 		 * fin
3978 		 */
3979 		if (dp->dfin == IDNFIN_FORCE_HARD) {
3980 			mt.mt_mtype = IDNP_ACK;
3981 			mt.mt_atype = IDNP_FIN | IDNP_ACK;
3982 			(void) idn_xphase_transition(domid, &mt, xargs);
3983 		} else {
3984 			mt.mt_mtype = IDNP_FIN | IDNP_ACK;
3985 			mt.mt_atype = 0;
3986 			(void) idn_send_fin(domid, &mt, dp->dfin, finarg,
3987 			    finopt, my_ready_set, finmaster);
3988 		}
3989 	} else if (msg & IDNP_MSGTYPE_MASK) {
3990 		/*
3991 		 * fin+ack
3992 		 */
3993 		if (dp->dfin != IDNFIN_FORCE_HARD) {
3994 			idn_xdcargs_t	fargs;
3995 
3996 			mt.mt_mtype = IDNP_ACK;
3997 			mt.mt_atype = msg;
3998 			CLR_XARGS(fargs);
3999 			SET_XARGS_FIN_TYPE(fargs, dp->dfin);
4000 			SET_XARGS_FIN_ARG(fargs, finarg);
4001 			SET_XARGS_FIN_DOMSET(fargs, my_ready_set);
4002 			SET_XARGS_FIN_OPT(fargs, finopt);
4003 			SET_XARGS_FIN_MASTER(fargs, finmaster);
4004 			idn_send_acknack(domid, &mt, fargs);
4005 		}
4006 	} else {
4007 		uint_t	token;
4008 		/*
4009 		 * nack - retry
4010 		 */
4011 		token = IDN_RETRY_TOKEN(domid, IDNRETRY_FIN);
4012 		idn_retry_submit(idn_retry_fin, NULL, token,
4013 		    idn_msg_retrytime[(int)IDNRETRY_FIN]);
4014 	}
4015 }
4016 
4017 /*ARGSUSED*/
4018 static void
4019 idn_action_fin_rcvd(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
4020 {
4021 	uint_t	msg = mtp ? mtp->mt_mtype : 0;
4022 
4023 	ASSERT(IDN_SYNC_IS_LOCKED());
4024 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
4025 
4026 	if (msg & IDNP_NACK) {
4027 		uint_t	token;
4028 		/*
4029 		 * nack - retry.
4030 		 */
4031 		token = IDN_RETRY_TOKEN(domid, IDNRETRY_FIN);
4032 		idn_retry_submit(idn_retry_fin, NULL, token,
4033 		    idn_msg_retrytime[(int)IDNRETRY_FIN]);
4034 	}
4035 }
4036 
4037 static void
4038 idn_final_fin(int domid)
4039 {
4040 	int		do_relink;
4041 	int		rv, d, new_masterid = IDN_NIL_DOMID;
4042 	idn_gstate_t	next_gstate;
4043 	domainset_t	relinkset;
4044 	uint_t		token = IDN_RETRY_TOKEN(domid, IDNRETRY_FIN);
4045 	idn_domain_t	*ldp, *dp = &idn_domain[domid];
4046 	procname_t	proc = "idn_final_fin";
4047 
4048 	ASSERT(IDN_SYNC_IS_LOCKED());
4049 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
4050 	ASSERT(dp->dstate == IDNDS_DMAP);
4051 
4052 	(void) idn_retry_terminate(token);
4053 
4054 	dp->dxp = NULL;
4055 	IDN_XSTATE_TRANSITION(dp, IDNXS_NIL);
4056 
4057 	idn_sync_exit(domid, IDNSYNC_DISCONNECT);
4058 
4059 	DOMAINSET_DEL(idn.domset.ds_trans_off, domid);
4060 
4061 	do_relink = DOMAIN_IN_SET(idn.domset.ds_relink, domid) ? 1 : 0;
4062 
4063 	/*
4064 	 * idn_deconfig will idn_close_domain.
4065 	 */
4066 	idn_deconfig(domid);
4067 
4068 	PR_PROTO("%s:%d: DISCONNECTED\n", proc, domid);
4069 
4070 	IDN_GLOCK_EXCL();
4071 	/*
4072 	 * It's important that this update-op occur within
4073 	 * the context of holding the glock(EXCL).  There is
4074 	 * still some additional state stuff to cleanup which
4075 	 * will be completed once the glock is dropped in
4076 	 * this flow.  Which means anybody that's doing a
4077 	 * SSI_INFO and waiting on glock will not actually
4078 	 * run until the clean-up is completed, which is what
4079 	 * we want.  Recall that a separate thread processes
4080 	 * the SSI_LINK/UNLINK calls and when they complete
4081 	 * (i.e. are awakened) they will immediately SSI_INFO
4082 	 * and we don't want them to prematurely pick up stale
4083 	 * information.
4084 	 */
4085 	idn_update_op(IDNOP_DISCONNECTED, DOMAINSET(domid), NULL);
4086 
4087 	ASSERT(idn.state != IDNGS_OFFLINE);
4088 	ASSERT(!DOMAIN_IN_SET(idn.domset.ds_trans_on, domid));
4089 
4090 	if (domid == IDN_GET_MASTERID()) {
4091 		IDN_SET_MASTERID(IDN_NIL_DOMID);
4092 		dp->dvote.v.master = 0;
4093 	}
4094 
4095 	if ((domid == IDN_GET_NEW_MASTERID()) && !do_relink) {
4096 		IDN_SET_NEW_MASTERID(IDN_NIL_DOMID);
4097 	}
4098 
4099 	if (idn.state == IDNGS_RECONFIG)
4100 		new_masterid = IDN_GET_NEW_MASTERID();
4101 
4102 	if ((idn.domset.ds_trans_on | idn.domset.ds_trans_off |
4103 	    idn.domset.ds_relink) == 0) {
4104 		PR_HITLIST("%s:%d: HITLIST %x -> 0\n",
4105 		    proc, domid, idn.domset.ds_hitlist);
4106 		idn.domset.ds_hitlist = 0;
4107 	}
4108 
4109 	if (idn.domset.ds_connected || idn.domset.ds_trans_off) {
4110 		PR_PROTO("%s:%d: ds_connected = 0x%x, ds_trans_off = 0x%x\n",
4111 		    proc, domid, idn.domset.ds_connected,
4112 		    idn.domset.ds_trans_off);
4113 		IDN_GUNLOCK();
4114 		goto fin_done;
4115 	}
4116 
4117 	IDN_DLOCK_EXCL(idn.localid);
4118 	ldp = &idn_domain[idn.localid];
4119 
4120 	if (idn.domset.ds_trans_on != 0) {
4121 		ASSERT((idn.state != IDNGS_DISCONNECT) &&
4122 		    (idn.state != IDNGS_OFFLINE));
4123 
4124 		switch (idn.state) {
4125 		case IDNGS_CONNECT:
4126 			if (idn.localid == IDN_GET_MASTERID()) {
4127 				idn_master_deinit();
4128 				IDN_SET_MASTERID(IDN_NIL_DOMID);
4129 				ldp->dvote.v.master = 0;
4130 			}
4131 			/*FALLTHROUGH*/
4132 		case IDNGS_ONLINE:
4133 			next_gstate = idn.state;
4134 			break;
4135 
4136 		case IDNGS_RECONFIG:
4137 			if (idn.localid == IDN_GET_MASTERID()) {
4138 				idn_master_deinit();
4139 				IDN_SET_MASTERID(IDN_NIL_DOMID);
4140 				ldp->dvote.v.master = 0;
4141 			}
4142 			ASSERT(IDN_GET_MASTERID() == IDN_NIL_DOMID);
4143 			next_gstate = IDNGS_CONNECT;
4144 			ldp->dvote.v.connected = 0;
4145 			/*
4146 			 * Need to do HWINIT since we won't
4147 			 * be transitioning through OFFLINE
4148 			 * which would normally be caught in
4149 			 * idn_check_nego() when we
4150 			 * initially go to CONNECT.
4151 			 */
4152 			IDN_PREP_HWINIT();
4153 			break;
4154 
4155 		case IDNGS_DISCONNECT:
4156 		case IDNGS_OFFLINE:
4157 			cmn_err(CE_WARN,
4158 			    "IDN: 211: disconnect domain %d, "
4159 			    "unexpected Gstate (%s)",
4160 			    domid, idngs_str[idn.state]);
4161 			IDN_DUNLOCK(idn.localid);
4162 			IDN_GUNLOCK();
4163 			goto fin_done;
4164 
4165 		default:
4166 			/*
4167 			 * XXX
4168 			 * Go into FATAL state?
4169 			 */
4170 			cmn_err(CE_PANIC,
4171 			    "IDN: 212: disconnect domain %d, "
4172 			    "bad Gstate (%d)",
4173 			    domid, idn.state);
4174 			/* not reached */
4175 			break;
4176 		}
4177 	} else {
4178 		if (idn.localid == IDN_GET_MASTERID()) {
4179 			idn_master_deinit();
4180 			IDN_SET_MASTERID(IDN_NIL_DOMID);
4181 			ldp->dvote.v.master = 0;
4182 		}
4183 		next_gstate = IDNGS_OFFLINE;
4184 		if (idn.domset.ds_relink == 0) {
4185 			IDN_SET_NEW_MASTERID(IDN_NIL_DOMID);
4186 		}
4187 	}
4188 	IDN_DUNLOCK(idn.localid);
4189 
4190 	/*
4191 	 * If we reach here we've effectively disconnected all
4192 	 * existing links, however new ones may be pending.
4193 	 */
4194 	PR_PROTO("%s:%d: ALL DISCONNECTED *****************\n", proc, domid);
4195 
4196 	IDN_GSTATE_TRANSITION(next_gstate);
4197 
4198 	ASSERT((idn.state == IDNGS_OFFLINE) ?
4199 	    (IDN_GET_MASTERID() == IDN_NIL_DOMID) : 1);
4200 
4201 	IDN_GUNLOCK();
4202 
4203 	/*
4204 	 * If we have no new masterid and yet there are relinkers
4205 	 * out there, then force us to attempt to link with one
4206 	 * of them.
4207 	 */
4208 	if ((new_masterid == IDN_NIL_DOMID) && idn.domset.ds_relink)
4209 		new_masterid = idn.localid;
4210 
4211 	if (new_masterid != IDN_NIL_DOMID) {
4212 		/*
4213 		 * If the local domain is the selected
4214 		 * master then we'll want to initiate
4215 		 * a link with one of the other candidates.
4216 		 * If not, then we want to initiate a link
4217 		 * with the master only.
4218 		 */
4219 		relinkset = (new_masterid == idn.localid) ?
4220 		    idn.domset.ds_relink : DOMAINSET(new_masterid);
4221 
4222 		DOMAINSET_DEL(relinkset, idn.localid);
4223 
4224 		for (d = 0; d < MAX_DOMAINS; d++) {
4225 			int	lock_held;
4226 
4227 			if (!DOMAIN_IN_SET(relinkset, d))
4228 				continue;
4229 
4230 			if (d == domid) {
4231 				do_relink = 0;
4232 				lock_held = 0;
4233 			} else {
4234 				IDN_DLOCK_EXCL(d);
4235 				lock_held = 1;
4236 			}
4237 
4238 			rv = idn_open_domain(d, -1, 0);
4239 			if (rv == 0) {
4240 				rv = idn_connect(d);
4241 				if (lock_held)
4242 					IDN_DUNLOCK(d);
4243 				/*
4244 				 * If we're able to kick off at
4245 				 * least one connect then that's
4246 				 * good enough for now.  The others
4247 				 * will fall into place normally.
4248 				 */
4249 				if (rv == 0)
4250 					break;
4251 			} else if (rv < 0) {
4252 				if (lock_held)
4253 					IDN_DUNLOCK(d);
4254 				cmn_err(CE_WARN,
4255 				    "IDN: 205: (%s.1) failed to "
4256 				    "open-domain(%d,%d)",
4257 				    proc, domid, -1);
4258 				DOMAINSET_DEL(idn.domset.ds_relink, d);
4259 			} else {
4260 				if (lock_held)
4261 					IDN_DUNLOCK(d);
4262 				PR_PROTO("%s:%d: failed to "
4263 				    "re-open domain %d "
4264 				    "(cpu %d) [rv = %d]\n",
4265 				    proc, domid, d, idn_domain[d].dcpu,
4266 				    rv);
4267 			}
4268 		}
4269 	}
4270 
4271 fin_done:
4272 	if (do_relink) {
4273 		ASSERT(IDN_DLOCK_IS_EXCL(domid));
4274 
4275 		rv = idn_open_domain(domid, -1, 0);
4276 		if (rv == 0) {
4277 			(void) idn_connect(domid);
4278 		} else if (rv < 0) {
4279 			cmn_err(CE_WARN,
4280 			    "IDN: 205: (%s.2) failed to "
4281 			    "open-domain(%d,%d)",
4282 			    proc, domid, -1);
4283 			DOMAINSET_DEL(idn.domset.ds_relink, domid);
4284 		}
4285 	}
4286 }
4287 
4288 static void
4289 idn_exit_fin(int domid, uint_t msgtype)
4290 {
4291 	idn_domain_t	*dp = &idn_domain[domid];
4292 	uint_t		token;
4293 	procname_t	proc = "idn_exit_fin";
4294 	STRING(str);
4295 
4296 	ASSERT(IDN_SYNC_IS_LOCKED());
4297 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
4298 
4299 	INUM2STR(msgtype, str);
4300 	PR_PROTO("%s:%d: msgtype = 0x%x(%s)\n", proc, domid, msgtype, str);
4301 
4302 	token = IDN_RETRY_TOKEN(domid, IDNRETRY_FIN);
4303 	(void) idn_retry_terminate(token);
4304 
4305 	DOMAINSET_DEL(idn.domset.ds_ready_off, domid);
4306 
4307 	dp->dxp = &xphase_fin;
4308 	IDN_XSTATE_TRANSITION(dp, IDNXS_PEND);
4309 
4310 	idn_retry_submit(idn_retry_fin, NULL, token,
4311 	    idn_msg_retrytime[(int)IDNRETRY_FIN]);
4312 }
4313 
4314 /*
4315  * Must return w/locks held.
4316  */
4317 static int
4318 idn_xphase_transition(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
4319 {
4320 	uint_t		msg = mtp ? mtp->mt_mtype : 0;
4321 	uint_t		msgarg = mtp ? mtp->mt_atype : 0;
4322 	idn_xphase_t	*xp;
4323 	idn_domain_t	*dp;
4324 	int		(*cfunc)(int, idn_msgtype_t *, idn_xdcargs_t);
4325 	void		(*ffunc)(int);
4326 	void		(*afunc)(int, idn_msgtype_t *, idn_xdcargs_t);
4327 	void		(*efunc)(int, idn_msgtype_t *, idn_xdcargs_t);
4328 	void		(*xfunc)(int, uint_t);
4329 	int		err = 0;
4330 	uint_t		msgtype;
4331 	idn_xstate_t	o_xstate, n_xstate;
4332 	procname_t	proc = "idn_xphase_transition";
4333 	STRING(mstr);
4334 	STRING(astr);
4335 
4336 	ASSERT(IDN_SYNC_IS_LOCKED());
4337 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
4338 
4339 	INUM2STR(msg, mstr);
4340 	INUM2STR(msgarg, astr);
4341 
4342 	dp = &idn_domain[domid];
4343 	if ((xp = dp->dxp) == NULL) {
4344 		PR_PROTO("%s:%d: WARNING: domain xsp is NULL (msg = %s, "
4345 		    "msgarg = %s) <<<<<<<<<<<<\n",
4346 		    proc, domid, mstr, astr);
4347 		return (-1);
4348 	}
4349 	o_xstate = dp->dxstate;
4350 
4351 	xfunc = xp->xt_exit;
4352 
4353 	if ((msgtype = (msg & IDNP_MSGTYPE_MASK)) == 0)
4354 		msgtype = msgarg & IDNP_MSGTYPE_MASK;
4355 
4356 	if ((o_xstate == IDNXS_PEND) && msg &&
4357 	    ((msg & IDNP_ACKNACK_MASK) == msg)) {
4358 		PR_PROTO("%s:%d: unwanted acknack received (o_xstate = %s, "
4359 		    "msg = %s/%s - dropping message\n",
4360 		    proc, domid, idnxs_str[(int)o_xstate], mstr, astr);
4361 		return (0);
4362 	}
4363 
4364 	/*
4365 	 * Validate that message received is following
4366 	 * the expected protocol for the current state.
4367 	 */
4368 	if (idn_next_xstate(o_xstate, -1, msg) == IDNXS_NIL) {
4369 		PR_PROTO("%s:%d: WARNING: o_xstate = %s, msg = %s -> NIL "
4370 		    "<<<<<<<<<\n",
4371 		    proc, domid, idnxs_str[(int)o_xstate], mstr);
4372 		if (xfunc)
4373 			(*xfunc)(domid, msgtype);
4374 		return (-1);
4375 	}
4376 
4377 	if (msg || msgarg) {
4378 		/*
4379 		 * Verify that message type is correct for
4380 		 * the given xstate.
4381 		 */
4382 		if (msgtype != xp->xt_msgtype) {
4383 			STRING(xstr);
4384 			STRING(tstr);
4385 
4386 			INUM2STR(xp->xt_msgtype, xstr);
4387 			INUM2STR(msgtype, tstr);
4388 			PR_PROTO("%s:%d: WARNING: msg expected %s(0x%x), "
4389 			    "actual %s(0x%x) [msg=%s(0x%x), "
4390 			    "msgarg=%s(0x%x)]\n",
4391 			    proc, domid, xstr, xp->xt_msgtype,
4392 			    tstr, msgtype, mstr, msg, astr, msgarg);
4393 			if (xfunc)
4394 				(*xfunc)(domid, msgtype);
4395 			return (-1);
4396 		}
4397 	}
4398 
4399 	cfunc = xp->xt_trans[(int)o_xstate].t_check;
4400 
4401 	if (cfunc && ((err = (*cfunc)(domid, mtp, xargs)) < 0)) {
4402 		if (o_xstate != IDNXS_PEND) {
4403 			IDN_XSTATE_TRANSITION(dp, IDNXS_PEND);
4404 		}
4405 		if (xfunc)
4406 			(*xfunc)(domid, msgtype);
4407 		return (-1);
4408 	}
4409 
4410 	n_xstate = idn_next_xstate(o_xstate, err, msg);
4411 
4412 	if (n_xstate == IDNXS_NIL) {
4413 		PR_PROTO("%s:%d: WARNING: n_xstate = %s, msg = %s -> NIL "
4414 		    "<<<<<<<<<\n",
4415 		    proc, domid, idnxs_str[(int)n_xstate], mstr);
4416 		if (xfunc)
4417 			(*xfunc)(domid, msgtype);
4418 		return (-1);
4419 	}
4420 
4421 	if (n_xstate != o_xstate) {
4422 		IDN_XSTATE_TRANSITION(dp, n_xstate);
4423 	}
4424 
4425 	if (err) {
4426 		if ((efunc = xp->xt_trans[(int)o_xstate].t_error) != NULL)
4427 			(*efunc)(domid, mtp, xargs);
4428 	} else if ((afunc = xp->xt_trans[(int)o_xstate].t_action) != NULL) {
4429 		(*afunc)(domid, mtp, xargs);
4430 	}
4431 
4432 	if ((n_xstate == IDNXS_FINAL) && ((ffunc = xp->xt_final) != NULL))
4433 		(*ffunc)(domid);
4434 
4435 	return (0);
4436 }
4437 
4438 /*
4439  * Entered and returns w/DLOCK & SYNC_LOCK held.
4440  */
4441 static int
4442 idn_xstate_transfunc(int domid, void *transarg)
4443 {
4444 	uint_t		msg = (uint_t)(uintptr_t)transarg;
4445 	uint_t		token;
4446 	procname_t	proc = "idn_xstate_transfunc";
4447 
4448 	ASSERT(IDN_SYNC_IS_LOCKED());
4449 
4450 	switch (msg) {
4451 	case IDNP_CON:
4452 		DOMAINSET_ADD(idn.domset.ds_connected, domid);
4453 		break;
4454 
4455 	case IDNP_FIN:
4456 		DOMAINSET_DEL(idn.domset.ds_connected, domid);
4457 		break;
4458 
4459 	default:
4460 		PR_PROTO("%s:%d: ERROR: unknown msg (0x%x) <<<<<<<<\n",
4461 		    proc, domid, msg);
4462 		return (0);
4463 	}
4464 
4465 	token = IDN_RETRY_TOKEN(domid, (msg == IDNP_CON) ?
4466 	    IDNRETRY_CON : IDNRETRY_FIN);
4467 	if (msg == IDNP_CON)
4468 		idn_retry_submit(idn_retry_con, NULL, token,
4469 		    idn_msg_retrytime[(int)IDNRETRY_CON]);
4470 	else
4471 		idn_retry_submit(idn_retry_fin, NULL, token,
4472 		    idn_msg_retrytime[(int)IDNRETRY_FIN]);
4473 
4474 	return (1);
4475 }
4476 
4477 /*
4478  * Entered and returns w/DLOCK & SYNC_LOCK held.
4479  */
4480 static void
4481 idn_sync_enter(int domid, idn_synccmd_t cmd, domainset_t xset,
4482     domainset_t rset, int (*transfunc)(), void *transarg)
4483 {
4484 	int		z;
4485 	idn_syncop_t	*sp;
4486 	idn_synczone_t	*zp;
4487 	procname_t	proc = "idn_sync_enter";
4488 
4489 	ASSERT(IDN_SYNC_IS_LOCKED());
4490 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
4491 
4492 	z = IDN_SYNC_GETZONE(cmd);
4493 	ASSERT(z >= 0);
4494 	zp = &idn.sync.sz_zone[z];
4495 
4496 	PR_SYNC("%s:%d: cmd=%s(%d), z=%d, xs=0x%x, rx=0x%x, cnt=%d\n",
4497 	    proc, domid, idnsync_str[cmd], cmd, z, xset, rset, zp->sc_cnt);
4498 
4499 	sp = &idn_domain[domid].dsync;
4500 
4501 	sp->s_domid = domid;
4502 	sp->s_cmd = cmd;
4503 	sp->s_msg = 0;
4504 	sp->s_set_exp = xset;
4505 	sp->s_set_rdy = rset;
4506 	sp->s_transfunc = transfunc;
4507 	sp->s_transarg = transarg;
4508 	IDN_SYNC_QUERY_INIT(domid);
4509 
4510 	sp->s_next = zp->sc_op;
4511 	zp->sc_op = sp;
4512 	zp->sc_cnt++;
4513 }
4514 
4515 /*
4516  * Entered and returns w/DLOCK & SYNC_LOCK held.
4517  */
4518 void
4519 idn_sync_exit(int domid, idn_synccmd_t cmd)
4520 {
4521 	int		d, z, zone, tot_queries, tot_domains;
4522 	idn_syncop_t	*sp;
4523 	idn_synczone_t	*zp = NULL;
4524 	procname_t	proc = "idn_sync_exit";
4525 
4526 	ASSERT(IDN_SYNC_IS_LOCKED());
4527 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
4528 
4529 	sp = &idn_domain[domid].dsync;
4530 
4531 	z = IDN_SYNC_GETZONE(sp->s_cmd);
4532 
4533 	zone = IDN_SYNC_GETZONE(cmd);
4534 
4535 	PR_SYNC("%s:%d: cmd=%s(%d) (z=%d, zone=%d)\n",
4536 	    proc, domid, idnsync_str[cmd], cmd, z, zone);
4537 
4538 #ifdef DEBUG
4539 	if (z != -1) {
4540 		tot_queries = tot_domains = 0;
4541 
4542 		for (d = 0; d < MAX_DOMAINS; d++) {
4543 			int	qv;
4544 
4545 			if ((qv = sp->s_query[d]) > 0) {
4546 				tot_queries += qv;
4547 				tot_domains++;
4548 				PR_SYNC("%s:%d: query_count = %d\n",
4549 				    proc, domid, qv);
4550 			}
4551 		}
4552 		PR_SYNC("%s:%d: tot_queries = %d, tot_domaines = %d\n",
4553 		    proc, domid, tot_queries, tot_domains);
4554 	}
4555 #endif /* DEBUG */
4556 
4557 	zp = (z != -1) ? &idn.sync.sz_zone[z] : NULL;
4558 
4559 	if (zp) {
4560 		idn_syncop_t	**spp;
4561 
4562 		for (spp = &zp->sc_op; *spp; spp = &((*spp)->s_next)) {
4563 			if (*spp == sp) {
4564 				*spp = sp->s_next;
4565 				sp->s_next = NULL;
4566 				zp->sc_cnt--;
4567 				break;
4568 			}
4569 		}
4570 	}
4571 
4572 	sp->s_cmd = IDNSYNC_NIL;
4573 
4574 	for (z = 0; z < IDN_SYNC_NUMZONE; z++) {
4575 		idn_syncop_t	**spp, **nspp;
4576 
4577 		if ((zone != -1) && (z != zone))
4578 			continue;
4579 
4580 		zp = &idn.sync.sz_zone[z];
4581 
4582 		for (spp = &zp->sc_op; *spp; spp = nspp) {
4583 			sp = *spp;
4584 			nspp = &sp->s_next;
4585 
4586 			if (!DOMAIN_IN_SET(sp->s_set_exp, domid))
4587 				continue;
4588 
4589 			DOMAINSET_DEL(sp->s_set_exp, domid);
4590 			DOMAINSET_DEL(sp->s_set_rdy, domid);
4591 
4592 			if ((sp->s_set_exp == sp->s_set_rdy) &&
4593 			    sp->s_transfunc) {
4594 				int	delok;
4595 
4596 				ASSERT(sp->s_domid != domid);
4597 
4598 				PR_SYNC("%s:%d invoking transfunc "
4599 				    "for domain %d\n",
4600 				    proc, domid, sp->s_domid);
4601 				delok = (*sp->s_transfunc)(sp->s_domid,
4602 				    sp->s_transarg);
4603 				if (delok) {
4604 					*spp = sp->s_next;
4605 					sp->s_next = NULL;
4606 					zp->sc_cnt--;
4607 					nspp = spp;
4608 				}
4609 			}
4610 		}
4611 	}
4612 }
4613 
4614 /*
4615  * Entered and returns w/DLOCK & SYNC_LOCK held.
4616  */
4617 static domainset_t
4618 idn_sync_register(int domid, idn_synccmd_t cmd, domainset_t ready_set,
4619     idn_syncreg_t regtype)
4620 {
4621 	int		z;
4622 	idn_synczone_t	*zp;
4623 	idn_syncop_t	*sp, **spp, **nspp;
4624 	domainset_t	query_set = 0, trans_set;
4625 	procname_t	proc = "idn_sync_register";
4626 
4627 	ASSERT(IDN_SYNC_IS_LOCKED());
4628 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
4629 
4630 	if ((z = IDN_SYNC_GETZONE(cmd)) == -1) {
4631 		PR_SYNC("%s:%d: ERROR: unexpected sync cmd(%d)\n",
4632 		    proc, domid, cmd);
4633 		return (0);
4634 	}
4635 
4636 	/*
4637 	 * Find out what domains are in transition with respect
4638 	 * to given command.  There will be no need to query
4639 	 * these folks.
4640 	 */
4641 	trans_set = IDN_SYNC_GETTRANS(cmd);
4642 
4643 	zp = &idn.sync.sz_zone[z];
4644 
4645 	PR_SYNC("%s:%d: cmd=%s(%d), z=%d, rset=0x%x, "
4646 	    "regtype=%s(%d), sc_op=%s\n",
4647 	    proc, domid, idnsync_str[cmd], cmd, z, ready_set,
4648 	    idnreg_str[regtype], regtype,
4649 	    zp->sc_op ? idnsync_str[zp->sc_op->s_cmd] : "NULL");
4650 
4651 	for (spp = &zp->sc_op; *spp; spp = nspp) {
4652 		sp = *spp;
4653 		nspp = &sp->s_next;
4654 
4655 		if (regtype == IDNSYNC_REG_NEW) {
4656 			DOMAINSET_ADD(sp->s_set_exp, domid);
4657 			PR_SYNC("%s:%d: adding new to %d (exp=0x%x)\n",
4658 			    proc, domid, sp->s_domid, sp->s_set_exp);
4659 		} else if (regtype == IDNSYNC_REG_QUERY) {
4660 			query_set |= ~sp->s_set_rdy & sp->s_set_exp;
4661 			continue;
4662 		}
4663 
4664 		if (!DOMAIN_IN_SET(sp->s_set_exp, domid))
4665 			continue;
4666 
4667 		if (!DOMAIN_IN_SET(ready_set, sp->s_domid)) {
4668 			/*
4669 			 * Given domid doesn't have a desired
4670 			 * domain in his ready-set.  We'll need
4671 			 * to query him again.
4672 			 */
4673 			DOMAINSET_ADD(query_set, domid);
4674 			continue;
4675 		}
4676 
4677 		/*
4678 		 * If we reach here, then an expected domain
4679 		 * has marked its respective datapath to
4680 		 * sp->s_domid as down (i.e. in his ready_set).
4681 		 */
4682 		DOMAINSET_ADD(sp->s_set_rdy, domid);
4683 
4684 		PR_SYNC("%s:%d: mark READY for domain %d "
4685 		    "(r=0x%x, x=0x%x)\n",
4686 		    proc, domid, sp->s_domid,
4687 		    sp->s_set_rdy, sp->s_set_exp);
4688 
4689 		query_set |= ~sp->s_set_rdy & sp->s_set_exp;
4690 
4691 		if (sp->s_set_exp == sp->s_set_rdy) {
4692 #ifdef DEBUG
4693 			if (sp->s_msg == 0) {
4694 				sp->s_msg = 1;
4695 				PR_SYNC("%s:%d: >>>>>>>>>>> DOMAIN %d "
4696 				    "ALL CHECKED IN (0x%x)\n",
4697 				    proc, domid, sp->s_domid,
4698 				    sp->s_set_exp);
4699 			}
4700 #endif /* DEBUG */
4701 
4702 			if ((sp->s_domid != domid) && sp->s_transfunc) {
4703 				int	delok;
4704 
4705 				PR_SYNC("%s:%d invoking transfunc "
4706 				    "for domain %d\n",
4707 				    proc, domid, sp->s_domid);
4708 				delok = (*sp->s_transfunc)(sp->s_domid,
4709 				    sp->s_transarg);
4710 				if (delok) {
4711 					*spp = sp->s_next;
4712 					sp->s_next = NULL;
4713 					zp->sc_cnt--;
4714 					nspp = spp;
4715 				}
4716 			}
4717 		}
4718 	}
4719 
4720 	PR_SYNC("%s:%d: trans_set = 0x%x, query_set = 0x%x -> 0x%x\n",
4721 	    proc, domid, trans_set, query_set, query_set & ~trans_set);
4722 
4723 	query_set &= ~trans_set;
4724 
4725 	return (query_set);
4726 }
4727 
4728 static void
4729 idn_sync_register_awol(int domid)
4730 {
4731 	int		z;
4732 	idn_synccmd_t	cmd = IDNSYNC_DISCONNECT;
4733 	idn_synczone_t	*zp;
4734 	idn_syncop_t	*sp;
4735 	procname_t	proc = "idn_sync_register_awol";
4736 
4737 	ASSERT(IDN_SYNC_IS_LOCKED());
4738 
4739 	if ((z = IDN_SYNC_GETZONE(cmd)) == -1) {
4740 		PR_SYNC("%s:%d: ERROR: unexpected sync cmd(%d)\n",
4741 		    proc, domid, cmd);
4742 		return;
4743 	}
4744 
4745 	zp = &idn.sync.sz_zone[z];
4746 
4747 	PR_SYNC("%s:%d: cmd=%s(%d), z=%d (domain %d = AWOL)\n",
4748 	    proc, domid, idnsync_str[cmd], cmd, z, domid);
4749 
4750 	for (sp = zp->sc_op; sp; sp = sp->s_next) {
4751 		idn_domain_t	*dp;
4752 
4753 		dp = &idn_domain[sp->s_domid];
4754 		if (dp->dfin == IDNFIN_FORCE_HARD) {
4755 			DOMAINSET_ADD(sp->s_set_rdy, domid);
4756 			PR_SYNC("%s:%d: adding new to %d (rdy=0x%x)\n",
4757 			    proc, domid, sp->s_domid, sp->s_set_rdy);
4758 		}
4759 	}
4760 }
4761 
4762 static void
4763 idn_link_established(void *arg)
4764 {
4765 	int	first_link;
4766 	int	domid, masterid;
4767 	uint_t	info = (uint_t)(uintptr_t)arg;
4768 
4769 	first_link = (int)(info & 0xf0);
4770 	domid = (int)(info & 0x0f);
4771 
4772 	IDN_GLOCK_SHARED();
4773 	masterid = IDN_GET_MASTERID();
4774 	if ((masterid == IDN_NIL_DOMID) ||
4775 	    (idn_domain[masterid].dstate != IDNDS_CONNECTED)) {
4776 		/*
4777 		 * No point in doing this unless we're connected
4778 		 * to the master.
4779 		 */
4780 		if ((masterid != IDN_NIL_DOMID) &&
4781 		    (idn.state == IDNGS_ONLINE)) {
4782 			/*
4783 			 * As long as we're still online keep
4784 			 * trying.
4785 			 */
4786 			(void) timeout(idn_link_established, arg, 50);
4787 		}
4788 		IDN_GUNLOCK();
4789 		return;
4790 	}
4791 	IDN_GUNLOCK();
4792 
4793 	if (first_link && IDN_SLAB_PREALLOC)
4794 		idn_prealloc_slab(IDN_SLAB_PREALLOC);
4795 
4796 	/*
4797 	 * No guarantee, but it might save a little
4798 	 * time.
4799 	 */
4800 	if (idn_domain[domid].dstate == IDNDS_CONNECTED) {
4801 		/*
4802 		 * Get the remote domain's dname.
4803 		 */
4804 		idn_send_nodename_req(domid);
4805 	}
4806 
4807 	/*
4808 	 * May have had some streams backed up waiting for
4809 	 * this connection.  Prod them.
4810 	 */
4811 	rw_enter(&idn.struprwlock, RW_READER);
4812 	mutex_enter(&idn.sipwenlock);
4813 	idndl_wenable(NULL);
4814 	mutex_exit(&idn.sipwenlock);
4815 	rw_exit(&idn.struprwlock);
4816 }
4817 
4818 /*
4819  * Send the following chunk of data received from above onto
4820  * the IDN wire.  This is raw data as far as the IDN driver
4821  * is concerned.
4822  * Returns:
4823  *	IDNXMIT_LOOP	- Msg handled in loopback and thus
4824  *			  still active (i.e. don't free).
4825  *	IDNXMIT_OKAY	- Data handled (freemsg).
4826  *	IDNXMIT_DROP	- Packet should be dropped.
4827  *	IDNXMIT_RETRY	- Packet should be requeued and retried.
4828  *	IDNXMIT_REQUEUE	- Packet should be requeued, but not
4829  *			  immediatetly retried.
4830  */
4831 int
4832 idn_send_data(int dst_domid, idn_netaddr_t dst_netaddr, queue_t *wq, mblk_t *mp)
4833 {
4834 	int		pktcnt = 0;
4835 	int		msglen;
4836 	int		rv = IDNXMIT_OKAY;
4837 	int		xfersize = 0;
4838 	caddr_t		iobufp, iodatap;
4839 	uchar_t		*data_rptr;
4840 	int		cpuindex;
4841 	int		serrno;
4842 	int		channel;
4843 	int		retry_reclaim;
4844 	idn_chansvr_t	*csp = NULL;
4845 	uint_t		netports = 0;
4846 	struct idnstr	*stp;
4847 	struct idn	*sip;
4848 	idn_domain_t	*dp;
4849 	struct ether_header	*ehp;
4850 	smr_pkthdr_t	*hdrp;
4851 	idn_msgtype_t	mt;
4852 	procname_t	proc = "idn_send_data";
4853 #ifdef DEBUG
4854 	size_t		orig_msglen = msgsize(mp);
4855 #endif /* DEBUG */
4856 
4857 	ASSERT(DB_TYPE(mp) == M_DATA);
4858 
4859 	mt.mt_mtype = IDNP_DATA;
4860 	mt.mt_atype = 0;
4861 	mt.mt_cookie = 0;
4862 
4863 	channel = (int)dst_netaddr.net.chan;
4864 
4865 	msglen = msgdsize(mp);
4866 	PR_DATA("%s:%d: (netaddr 0x%x) msgsize=%ld, msgdsize=%d\n",
4867 	    proc, dst_domid, dst_netaddr.netaddr, msgsize(mp), msglen);
4868 
4869 	ASSERT(wq->q_ptr);
4870 
4871 	stp = (struct idnstr *)wq->q_ptr;
4872 	sip = stp->ss_sip;
4873 	ASSERT(sip);
4874 
4875 	if (msglen < 0) {
4876 		/*
4877 		 * No data to send.  That was easy!
4878 		 */
4879 		PR_DATA("%s:%d: BAD msg length (%d) (netaddr 0x%x)\n",
4880 		    proc, dst_domid, msglen, dst_netaddr.netaddr);
4881 		return (IDNXMIT_DROP);
4882 	}
4883 
4884 	ASSERT(RW_READ_HELD(&stp->ss_rwlock));
4885 
4886 	if (dst_domid == IDN_NIL_DOMID) {
4887 		cmn_err(CE_WARN,
4888 		    "IDN: 213: no destination specified "
4889 		    "(d=%d, c=%d, n=0x%x)",
4890 		    dst_domid, dst_netaddr.net.chan,
4891 		    dst_netaddr.net.netid);
4892 		IDN_KSTAT_INC(sip, si_nolink);
4893 		IDN_KSTAT_INC(sip, si_macxmt_errors);
4894 		rv = IDNXMIT_DROP;
4895 		goto nocando;
4896 	}
4897 
4898 	ehp = (struct ether_header *)mp->b_rptr;
4899 	PR_DATA("%s:%d: destination channel = %d\n", proc, dst_domid, channel);
4900 
4901 #ifdef DEBUG
4902 	{
4903 		uchar_t	echn;
4904 
4905 		echn = (uchar_t)
4906 		    ehp->ether_shost.ether_addr_octet[IDNETHER_CHANNEL];
4907 		ASSERT((uchar_t)channel == echn);
4908 	}
4909 #endif /* DEBUG */
4910 	ASSERT(msglen <= IDN_DATA_SIZE);
4911 
4912 	dp = &idn_domain[dst_domid];
4913 	/*
4914 	 * Get reader lock.  We hold for the duration
4915 	 * of the transfer so that our state doesn't
4916 	 * change during this activity.  Note that since
4917 	 * we grab the reader lock, we can still permit
4918 	 * simultaneous tranfers from different threads
4919 	 * to the same domain.
4920 	 * Before we waste a bunch of time gathering locks, etc.
4921 	 * do a an unprotected check to make sure things are
4922 	 * semi-copesetic.  If these values are in flux,
4923 	 * that's okay.
4924 	 */
4925 	if ((dp->dstate != IDNDS_CONNECTED) || (idn.state != IDNGS_ONLINE)) {
4926 		IDN_KSTAT_INC(sip, si_linkdown);
4927 		if (idn.state != IDNGS_ONLINE) {
4928 			rv = IDNXMIT_REQUEUE;
4929 		} else {
4930 			IDN_KSTAT_INC(sip, si_macxmt_errors);
4931 			rv = IDNXMIT_DROP;
4932 		}
4933 		goto nocando;
4934 	}
4935 
4936 	if (idn.chan_servers[channel].ch_send.c_checkin) {
4937 		/*
4938 		 * Gotta bail, somethin' s'up.
4939 		 */
4940 		rv = IDNXMIT_REQUEUE;
4941 		goto nocando;
4942 	}
4943 
4944 	csp = &idn.chan_servers[channel];
4945 	IDN_CHAN_LOCK_SEND(csp);
4946 
4947 	if (dst_netaddr.net.netid == IDN_BROADCAST_ALLNETID) {
4948 		/*
4949 		 * We're doing a broadcast.  Need to set
4950 		 * up IDN netaddr's one at a time.
4951 		 * We set the ethernet destination to the same
4952 		 * instance as the sending address.  The instance
4953 		 * numbers effectively represent subnets.
4954 		 */
4955 		dst_netaddr.net.netid = dp->dnetid;
4956 
4957 		(void) idndl_domain_etheraddr(dst_domid, channel,
4958 		    &ehp->ether_dhost);
4959 
4960 		if (dst_domid == idn.localid) {
4961 			mblk_t	*nmp;
4962 			/*
4963 			 * If this is a broadcast and going to
4964 			 * the local domain, then we need to make
4965 			 * a private copy of the message since
4966 			 * the current one will be reused when
4967 			 * transmitting to other domains.
4968 			 */
4969 			PR_DATA("%s:%d: dup broadcast msg for local domain\n",
4970 			    proc, dst_domid);
4971 			if ((nmp = copymsg(mp)) == NULL) {
4972 				/*
4973 				 * Couldn't get a duplicate copy.
4974 				 */
4975 				IDN_CHAN_UNLOCK_SEND(csp);
4976 				csp = NULL;
4977 				IDN_KSTAT_INC(sip, si_allocbfail);
4978 				IDN_KSTAT_INC(sip, si_noxmtbuf);
4979 				rv = IDNXMIT_DROP;
4980 				goto nocando;
4981 			}
4982 			mp = nmp;
4983 		}
4984 	}
4985 
4986 	if (dp->dnetid != dst_netaddr.net.netid) {
4987 		PR_DATA("%s:%d: dest netid (0x%x) != expected (0x%x)\n",
4988 		    proc, dst_domid, (uint_t)dst_netaddr.net.netid,
4989 		    (uint_t)dp->dnetid);
4990 		IDN_CHAN_UNLOCK_SEND(csp);
4991 		csp = NULL;
4992 		IDN_KSTAT_INC(sip, si_nolink);
4993 		IDN_KSTAT_INC(sip, si_macxmt_errors);
4994 		rv = IDNXMIT_DROP;
4995 		goto nocando;
4996 	}
4997 
4998 	if (dst_domid == idn.localid) {
4999 		int	lbrv;
5000 		/*
5001 		 * Sending to our local domain! Loopback.
5002 		 * Note that idn_send_data_loop returning 0
5003 		 * does not mean the message can now be freed.
5004 		 * We need to return (-1) so that caller doesn't
5005 		 * try to free mblk.
5006 		 */
5007 		IDN_CHAN_UNLOCK_SEND(csp);
5008 		rw_exit(&stp->ss_rwlock);
5009 		lbrv = idn_send_data_loopback(dst_netaddr, wq, mp);
5010 		rw_enter(&stp->ss_rwlock, RW_READER);
5011 		if (lbrv == 0) {
5012 			return (IDNXMIT_LOOP);
5013 		} else {
5014 			IDN_KSTAT_INC(sip, si_macxmt_errors);
5015 			return (IDNXMIT_DROP);
5016 		}
5017 	}
5018 
5019 	if (dp->dstate != IDNDS_CONNECTED) {
5020 		/*
5021 		 * Can't send data unless a link has already been
5022 		 * established with the target domain.  Normally,
5023 		 * a user cannot set the remote netaddr unless a
5024 		 * link has already been established, however it
5025 		 * is possible the connection may have become
5026 		 * disconnected since that time.
5027 		 */
5028 		IDN_CHAN_UNLOCK_SEND(csp);
5029 		csp = NULL;
5030 		IDN_KSTAT_INC(sip, si_linkdown);
5031 		IDN_KSTAT_INC(sip, si_macxmt_errors);
5032 		rv = IDNXMIT_DROP;
5033 		goto nocando;
5034 	}
5035 
5036 	/*
5037 	 * Need to make sure the channel is active and that the
5038 	 * domain to which we're sending is allowed to receive stuff.
5039 	 */
5040 	if (!IDN_CHANNEL_IS_SEND_ACTIVE(csp)) {
5041 		int	not_active;
5042 		/*
5043 		 * See if we can activate channel.
5044 		 */
5045 		IDN_CHAN_UNLOCK_SEND(csp);
5046 		not_active = idn_activate_channel(CHANSET(channel),
5047 		    IDNCHAN_OPEN);
5048 		if (!not_active) {
5049 			/*
5050 			 * Only grab the lock for a recheck if we were
5051 			 * able to activate the channel.
5052 			 */
5053 			IDN_CHAN_LOCK_SEND(csp);
5054 		}
5055 		/*
5056 		 * Verify channel still active now that we have the lock.
5057 		 */
5058 		if (not_active || !IDN_CHANNEL_IS_SEND_ACTIVE(csp)) {
5059 			if (!not_active) {
5060 				/*
5061 				 * Only need to drop the lock if it was
5062 				 * acquired while we thought we had
5063 				 * activated the channel.
5064 				 */
5065 				IDN_CHAN_UNLOCK_SEND(csp);
5066 			}
5067 			ASSERT(!IDN_CHAN_SEND_IS_LOCKED(csp));
5068 			/*
5069 			 * Damn!   Must have went inactive during the window
5070 			 * before we regrabbed the send lock.  Oh well, can't
5071 			 * spend all day doing this, bail out.  Set csp to
5072 			 * NULL to prevent inprogress update at bottom.
5073 			 */
5074 			csp = NULL;
5075 			/*
5076 			 * Channel is not active, should not be used.
5077 			 */
5078 			PR_DATA("%s:%d: dest channel %d NOT ACTIVE\n",
5079 			    proc, dst_domid, channel);
5080 			IDN_KSTAT_INC(sip, si_linkdown);
5081 			rv = IDNXMIT_REQUEUE;
5082 			goto nocando;
5083 		}
5084 		ASSERT(IDN_CHAN_SEND_IS_LOCKED(csp));
5085 	}
5086 	/*
5087 	 * If we made it here then the channel is active
5088 	 * Make sure the target domain is registered to receive stuff,
5089 	 * i.e. we're still linked.
5090 	 */
5091 	if (!IDN_CHAN_DOMAIN_IS_REGISTERED(csp, dst_domid)) {
5092 		/*
5093 		 * If domain is not even registered with this channel
5094 		 * then we have no business being here.  Doesn't matter
5095 		 * whether it's active or not.
5096 		 */
5097 		PR_DATA("%s:%d: domain not registered with channel %d\n",
5098 		    proc, dst_domid, channel);
5099 		/*
5100 		 * Set csp to NULL to prevent in-progress update below.
5101 		 */
5102 		IDN_CHAN_UNLOCK_SEND(csp);
5103 		csp = NULL;
5104 		IDN_KSTAT_INC(sip, si_linkdown);
5105 		IDN_KSTAT_INC(sip, si_macxmt_errors);
5106 		rv = IDNXMIT_DROP;
5107 		goto nocando;
5108 	}
5109 
5110 	IDN_CHAN_SEND_INPROGRESS(csp);
5111 	IDN_CHAN_UNLOCK_SEND(csp);
5112 
5113 	/*
5114 	 * Find a target cpu to send interrupt to if
5115 	 * it becomes necessary (i.e. remote channel
5116 	 * server is idle).
5117 	 */
5118 	cpuindex = dp->dcpuindex;
5119 
5120 	/*
5121 	 * dcpuindex is atomically incremented, but other than
5122 	 * that is not well protected and that's okay.  The
5123 	 * intention is to simply spread around the interrupts
5124 	 * at the destination domain, however we don't have to
5125 	 * anal about it.  If we hit the same cpu multiple times
5126 	 * in a row that's okay, it will only be for a very short
5127 	 * period anyway before the cpuindex is incremented
5128 	 * to the next cpu.
5129 	 */
5130 	if (cpuindex < NCPU) {
5131 		ATOMIC_INC(dp->dcpuindex);
5132 	}
5133 	if (dp->dcpuindex >= NCPU)
5134 		dp->dcpuindex = 0;
5135 
5136 	IDN_ASSIGN_DCPU(dp, cpuindex);
5137 
5138 #ifdef XXX_DLPI_UNFRIENDLY
5139 	{
5140 		ushort_t	dstport = (ushort_t)dp->dcpu;
5141 
5142 		/*
5143 		 * XXX
5144 		 * This is not DLPI friendly, but we need some way
5145 		 * of distributing our XDC interrupts to the cpus
5146 		 * on the remote domain in a relatively random fashion
5147 		 * while trying to remain constant for an individual
5148 		 * network connection.  Don't want the target network
5149 		 * appl pinging around cpus thrashing the caches.
5150 		 * So, we'll pick target cpus based on the destination
5151 		 * TCP/IP port (socket).  The (simple) alternative to
5152 		 * this is to simply send all messages destined for
5153 		 * particular domain to the same cpu (dcpu), but
5154 		 * will lower our bandwidth and introduce a lot of
5155 		 * contention on that target cpu.
5156 		 */
5157 		if (ehp->ether_type == ETHERTYPE_IP) {
5158 			ipha_t	*ipha;
5159 			uchar_t	*dstporta;
5160 			int	hdr_length;
5161 			mblk_t	*nmp = mp;
5162 			uchar_t	*rptr = mp->b_rptr +
5163 			    sizeof (struct ether_header);
5164 			if (nmp->b_wptr <= rptr) {
5165 				/*
5166 				 * Only the ethernet header was contained
5167 				 * in the first block.  Check for the
5168 				 * next packet.
5169 				 */
5170 				if ((nmp = mp->b_cont) != NULL)
5171 					rptr = nmp->b_rptr;
5172 			}
5173 			/*
5174 			 * If we still haven't found the IP header packet
5175 			 * then don't bother.  Can't search forever.
5176 			 */
5177 			if (nmp &&
5178 			    ((nmp->b_wptr - rptr) >= IP_SIMPLE_HDR_LENGTH)) {
5179 				ipha = (ipha_t *)ALIGN32(rptr);
5180 
5181 				ASSERT(DB_TYPE(mp) == M_DATA);
5182 				hdr_length = IPH_HDR_LENGTH(ipha);
5183 
5184 				switch (ipha->ipha_protocol) {
5185 				case IPPROTO_UDP:
5186 				case IPPROTO_TCP:
5187 					/*
5188 					 * TCP/UDP Protocol Header (1st word)
5189 					 * 0	    15,16	31
5190 					 * -----------------------
5191 					 * | src port | dst port |
5192 					 * -----------------------
5193 					 */
5194 					dstporta = (uchar_t *)ipha + hdr_length;
5195 					netports = *(uint_t *)dstporta;
5196 					dstporta += 2;
5197 					dstport  = *(ushort_t *)dstporta;
5198 					break;
5199 				default:
5200 					break;
5201 				}
5202 			}
5203 
5204 		}
5205 		IDN_ASSIGN_DCPU(dp, dstport);
5206 
5207 		PR_DATA("%s:%d: (dstport %d) assigned %d\n",
5208 		    proc, dst_domid, (int)dstport, dp->dcpu);
5209 	}
5210 #endif /* XXX_DLPI_UNFRIENDLY */
5211 
5212 	data_rptr = mp->b_rptr;
5213 
5214 	ASSERT(dp->dcpu != IDN_NIL_DCPU);
5215 
5216 	ASSERT(idn_domain[dst_domid].dmbox.m_send);
5217 
5218 	retry_reclaim = 1;
5219 retry:
5220 	if ((dp->dio >= IDN_RECLAIM_MIN) || dp->diowanted) {
5221 		int	reclaim_req;
5222 		/*
5223 		 * Reclaim however many outstanding buffers
5224 		 * there are up to IDN_RECLAIM_MAX if it's set.
5225 		 */
5226 		reclaim_req = dp->diowanted ? -1 : IDN_RECLAIM_MAX ?
5227 		    MIN(dp->dio, IDN_RECLAIM_MAX) : dp->dio;
5228 		(void) idn_reclaim_mboxdata(dst_domid, channel,
5229 		    reclaim_req);
5230 	}
5231 
5232 	if (dp->dio >= IDN_WINDOW_EMAX) {
5233 
5234 		if (lock_try(&dp->diocheck)) {
5235 			IDN_MSGTIMER_START(dst_domid, IDNP_DATA, 0,
5236 			    idn_msg_waittime[IDNP_DATA],
5237 			    &mt.mt_cookie);
5238 			/*
5239 			 * We have exceeded the minimum window for
5240 			 * outstanding I/O buffers to this domain.
5241 			 * Need to start the MSG timer to check for
5242 			 * possible response from remote domain.
5243 			 * The remote domain may be hung.  Send a
5244 			 * wakeup!  Specify all channels for given
5245 			 * domain since we don't know precisely which
5246 			 * is backed up (dio is global).
5247 			 */
5248 			IDNXDC(dst_domid, &mt,
5249 			    (uint_t)dst_netaddr.net.chan, 0, 0, 0);
5250 		}
5251 
5252 		/*
5253 		 * Yikes!  We have exceeded the maximum window
5254 		 * which means no more packets going to remote
5255 		 * domain until he frees some up.
5256 		 */
5257 		IDN_KSTAT_INC(sip, si_txmax);
5258 		IDN_KSTAT_INC(sip, si_macxmt_errors);
5259 		rv = IDNXMIT_DROP;
5260 		goto nocando;
5261 	}
5262 
5263 	/*
5264 	 * Allocate a SMR I/O buffer and send it.
5265 	 */
5266 	if (msglen == 0) {
5267 		/*
5268 		 * A zero length messages is effectively a signal
5269 		 * to just send an interrupt to the remote domain.
5270 		 */
5271 		IDN_MSGTIMER_START(dst_domid, IDNP_DATA, 0,
5272 		    idn_msg_waittime[IDNP_DATA],
5273 		    &mt.mt_cookie);
5274 		IDNXDC(dst_domid, &mt,
5275 		    (uint_t)dst_netaddr.net.chan, 0, 0, 0);
5276 	}
5277 	for (; (msglen > 0) && mp; msglen -= xfersize) {
5278 		int		xrv;
5279 		smr_offset_t	bufoffset;
5280 #ifdef DEBUG
5281 		int		n_xfersize;
5282 #endif /* DEBUG */
5283 
5284 		ASSERT(msglen <= IDN_DATA_SIZE);
5285 		xfersize = msglen;
5286 
5287 		serrno = smr_buf_alloc(dst_domid, xfersize, &iobufp);
5288 		if (serrno) {
5289 			PR_DATA("%s:%d: failed to alloc SMR I/O buffer "
5290 			    "(serrno = %d)\n",
5291 			    proc, dst_domid, serrno);
5292 			/*
5293 			 * Failure is either due to a timeout waiting
5294 			 * for the master to give us a slab, OR the
5295 			 * local domain exhausted its slab quota!
5296 			 * In either case we'll have to bail from
5297 			 * here and let higher layers decide what
5298 			 * to do.
5299 			 * We also could have had locking problems.
5300 			 * A negative serrno indicates we lost the lock
5301 			 * on dst_domid, so no need in dropping lock.
5302 			 */
5303 
5304 			if (lock_try(&dp->diowanted) && retry_reclaim) {
5305 				/*
5306 				 * We were the first to acquire the
5307 				 * lock indicating that it wasn't
5308 				 * set on entry to idn_send_data.
5309 				 * So, let's go back and see if we
5310 				 * can't reclaim some buffers and
5311 				 * try again.
5312 				 * It's very likely diowanted will be
5313 				 * enough to prevent us from looping
5314 				 * on retrying here, however to protect
5315 				 * against the small window where a
5316 				 * race condition might exist, we use
5317 				 * the retry_reclaim flag so that we
5318 				 * don't retry more than once.
5319 				 */
5320 				retry_reclaim = 0;
5321 				goto retry;
5322 			}
5323 
5324 			rv = (serrno > 0) ? serrno : -serrno;
5325 			IDN_KSTAT_INC(sip, si_notbufs);
5326 			IDN_KSTAT_INC(sip, si_noxmtbuf);	/* MIB II */
5327 			switch (rv) {
5328 			case ENOMEM:
5329 			case EBUSY:
5330 			case ENOLCK:
5331 			case ETIMEDOUT:
5332 			case EDQUOT:
5333 				/*
5334 				 * These are all transient conditions
5335 				 * which should be recoverable over
5336 				 * time.
5337 				 */
5338 				rv = IDNXMIT_REQUEUE;
5339 				break;
5340 
5341 			default:
5342 				rv = IDNXMIT_DROP;
5343 				break;
5344 			}
5345 			goto nocando;
5346 		}
5347 
5348 		lock_clear(&dp->diowanted);
5349 
5350 		hdrp = IDN_BUF2HDR(iobufp);
5351 		bufoffset = (smr_offset_t)IDN_ALIGNPTR(sizeof (smr_pkthdr_t),
5352 		    data_rptr);
5353 		/*
5354 		 * If the alignment of bufoffset took us pass the
5355 		 * length of a smr_pkthdr_t then we need to possibly
5356 		 * lower xfersize since it was calulated based on
5357 		 * a perfect alignment.  However, if we're in DLPI
5358 		 * mode then shouldn't be necessary since the length
5359 		 * of the incoming packet (mblk) should have already
5360 		 * taken into consideration this possible adjustment.
5361 		 */
5362 #ifdef DEBUG
5363 		if (bufoffset != sizeof (smr_pkthdr_t))
5364 			PR_DATA("%s:%d: offset ALIGNMENT (%lu -> %u) "
5365 			    "(data_rptr = %p)\n",
5366 			    proc, dst_domid, sizeof (smr_pkthdr_t),
5367 			    bufoffset, (void *)data_rptr);
5368 
5369 		n_xfersize = MIN(xfersize, (IDN_SMR_BUFSIZE - bufoffset));
5370 		if (xfersize != n_xfersize) {
5371 			PR_DATA("%s:%d: xfersize ADJUST (%d -> %d)\n",
5372 			    proc, dst_domid, xfersize, n_xfersize);
5373 			cmn_err(CE_WARN, "%s: ERROR (xfersize = %d, > "
5374 			    "bufsize(%d)-bufoffset(%d) = %d)",
5375 			    proc, xfersize, IDN_SMR_BUFSIZE,
5376 			    bufoffset,
5377 			    IDN_SMR_BUFSIZE - bufoffset);
5378 		}
5379 #endif /* DEBUG */
5380 		xfersize = MIN(xfersize, (int)(IDN_SMR_BUFSIZE - bufoffset));
5381 
5382 		iodatap = IDN_BUF2DATA(iobufp, bufoffset);
5383 		mp = idn_fill_buffer(iodatap, xfersize, mp, &data_rptr);
5384 
5385 		hdrp->b_netaddr  = dst_netaddr.netaddr;
5386 		hdrp->b_netports = netports;
5387 		hdrp->b_offset   = bufoffset;
5388 		hdrp->b_length   = xfersize;
5389 		hdrp->b_next	 = IDN_NIL_SMROFFSET;
5390 		hdrp->b_rawio	 = 0;
5391 		hdrp->b_cksum    = IDN_CKSUM_PKT(hdrp);
5392 
5393 		xrv = idn_send_mboxdata(dst_domid, sip, channel, iobufp);
5394 		if (xrv) {
5395 			/*
5396 			 * Reclaim packet.
5397 			 * Return error on this packet so it can be retried
5398 			 * (putbq).  Note that it should be safe to assume
5399 			 * that this for-loop is only executed once when in
5400 			 * DLPI mode and so no need to worry about fractured
5401 			 * mblk packet.
5402 			 */
5403 			PR_DATA("%s:%d: DATA XFER to chan %d FAILED "
5404 			    "(ret=%d)\n",
5405 			    proc, dst_domid, channel, xrv);
5406 			(void) smr_buf_free(dst_domid, iobufp, xfersize);
5407 
5408 			PR_DATA("%s:%d: (line %d) dec(dio) -> %d\n",
5409 			    proc, dst_domid, __LINE__, dp->dio);
5410 
5411 			rv = IDNXMIT_DROP;
5412 			IDN_KSTAT_INC(sip, si_macxmt_errors);
5413 			goto nocando;
5414 		} else {
5415 			pktcnt++;
5416 			/*
5417 			 * Packet will get freed on a subsequent send
5418 			 * when we reclaim buffers that the receivers
5419 			 * has finished consuming.
5420 			 */
5421 		}
5422 	}
5423 
5424 #ifdef DEBUG
5425 	if (pktcnt > 1)
5426 		cmn_err(CE_WARN,
5427 		    "%s: ERROR: sent multi-pkts (%d), len = %ld",
5428 		    proc, pktcnt, orig_msglen);
5429 #endif /* DEBUG */
5430 
5431 	PR_DATA("%s:%d: SENT %d packets (%d @ 0x%x)\n",
5432 	    proc, dst_domid, pktcnt, dst_netaddr.net.chan,
5433 	    dst_netaddr.net.netid);
5434 
5435 	IDN_CHAN_LOCK_SEND(csp);
5436 	IDN_CHAN_SEND_DONE(csp);
5437 	IDN_CHAN_UNLOCK_SEND(csp);
5438 
5439 	return (IDNXMIT_OKAY);
5440 
5441 nocando:
5442 
5443 	if (csp) {
5444 		IDN_CHAN_LOCK_SEND(csp);
5445 		IDN_CHAN_SEND_DONE(csp);
5446 		IDN_CHAN_UNLOCK_SEND(csp);
5447 	}
5448 
5449 	if (rv == IDNXMIT_REQUEUE) {
5450 		/*
5451 		 * Better kick off monitor to check when
5452 		 * it's ready to reenable the queues for
5453 		 * this channel.
5454 		 */
5455 		idn_xmit_monitor_kickoff(channel);
5456 	}
5457 
5458 	return (rv);
5459 }
5460 
5461 /*
5462  * Function to support local loopback testing of IDN driver.
5463  * Primarily geared towards measuring stream-head and IDN driver
5464  * overhead with respect to data messages.  Setting idn_strhead_only
5465  * allows routine to focus on stream-head overhead by simply putting
5466  * the message straight to the 'next' queue of the destination
5467  * read-queue.  Current implementation puts the message directly to
5468  * the read-queue thus sending the message right back to the IDN driver
5469  * as though the data came in off the wire.  No need to worry about
5470  * any IDN layers attempting to ack data as that's normally handled
5471  * by idnh_recv_data.
5472  *
5473  * dst_netaddr = destination port-n-addr on local domain.
5474  * wq          = write queue from whence message came.
5475  * mp          = the (data-only) message.
5476  *
5477  * Returns 0		Indicates data handled.
5478  *	   errno	EAGAIN indicates data can be retried.
5479  *			Other errno's indicate failure to handle.
5480  */
5481 static int
5482 idn_send_data_loopback(idn_netaddr_t dst_netaddr, queue_t *wq, mblk_t *mp)
5483 {
5484 	register struct idnstr	*stp;
5485 	struct idn	*sip;
5486 	int		rv = 0;
5487 	procname_t	proc = "idn_send_data_loopback";
5488 
5489 	if (dst_netaddr.net.netid != idn_domain[idn.localid].dnetid) {
5490 		PR_DATA("%s: dst_netaddr.net.netid 0x%x != local 0x%x\n",
5491 		    proc, dst_netaddr.net.netid,
5492 		    idn_domain[idn.localid].dnetid);
5493 		rv = EADDRNOTAVAIL;
5494 		goto done;
5495 	}
5496 	stp = (struct idnstr *)wq->q_ptr;
5497 	if (!stp || !stp->ss_rq) {
5498 		rv = EDESTADDRREQ;
5499 		goto done;
5500 	}
5501 	sip = stp->ss_sip;
5502 
5503 	idndl_read(sip, mp);
5504 	rv = 0;
5505 
5506 done:
5507 	return (rv);
5508 }
5509 
5510 /*
5511  * Fill bufp with as much data as possible from the message pointed
5512  * to by mp up to size bytes.
5513  * Save our current read pointer in the variable parameter (data_rptrp)
5514  * so we know where to start on the next go around.  Don't want to
5515  * bump the actual b_rptr in the mblk because the mblk may need to
5516  * be reused, e.g. broadcast.
5517  * Return the mblk pointer to the position we had to stop.
5518  */
5519 static mblk_t *
5520 idn_fill_buffer(caddr_t bufp, int size, mblk_t *mp, uchar_t **data_rptrp)
5521 {
5522 	int	copysize;
5523 
5524 	ASSERT(bufp && size);
5525 
5526 	if (mp == NULL)
5527 		return (NULL);
5528 
5529 	while ((size > 0) && mp) {
5530 
5531 		copysize = MIN(mp->b_wptr - (*data_rptrp), size);
5532 
5533 		if (copysize > 0) {
5534 			/*
5535 			 * If there's data to copy, do it.
5536 			 */
5537 			bcopy((*data_rptrp), bufp, copysize);
5538 			(*data_rptrp) += copysize;
5539 			bufp += copysize;
5540 			size -= copysize;
5541 		}
5542 		if (mp->b_wptr <= (*data_rptrp)) {
5543 			/*
5544 			 * If we emptied the mblk, then
5545 			 * move on to the next one.
5546 			 */
5547 			for (mp = mp->b_cont;
5548 			    mp && (mp->b_datap->db_type != M_DATA);
5549 			    mp = mp->b_cont)
5550 				;
5551 			if (mp)
5552 				*data_rptrp = mp->b_rptr;
5553 		}
5554 	}
5555 	return (mp);
5556 }
5557 
5558 /*
5559  * Messages received here do NOT arrive on a stream, but are
5560  * instead handled via the idn_protocol_servers.  This routine
5561  * is effectively the job processor for the protocol servers.
5562  */
5563 static void
5564 idn_recv_proto(idn_protomsg_t *hp)
5565 {
5566 	int		domid, cpuid;
5567 	int		sync_lock = 0;
5568 	idn_domain_t	*dp;
5569 	register uint_t	mtype;
5570 	register uint_t	msgtype, acktype;
5571 	idn_msgtype_t	mt;
5572 	ushort_t	dcookie, tcookie;
5573 	procname_t	proc = "idn_recv_proto";
5574 
5575 
5576 	if (idn.state == IDNGS_IGNORE) {
5577 		/*
5578 		 * Fault injection to simulate non-responsive domain.
5579 		 */
5580 		return;
5581 	}
5582 
5583 	domid   = hp->m_domid;
5584 	cpuid   = hp->m_cpuid;
5585 	msgtype = hp->m_msgtype;
5586 	acktype = hp->m_acktype;
5587 	dcookie = IDN_DCOOKIE(hp->m_cookie);
5588 	tcookie = IDN_TCOOKIE(hp->m_cookie);
5589 	/*
5590 	 * msgtype =	Is the type of message we received,
5591 	 *		e.g. nego, ack, nego+ack, etc.
5592 	 *
5593 	 * acktype =	If we received a pure ack or nack
5594 	 *		then this variable is set to the
5595 	 *		type of message that was ack/nack'd.
5596 	 */
5597 	if ((mtype = msgtype & IDNP_MSGTYPE_MASK) == 0) {
5598 		/*
5599 		 * Received a pure ack/nack.
5600 		 */
5601 		mtype = acktype & IDNP_MSGTYPE_MASK;
5602 	}
5603 
5604 	if (!VALID_MSGTYPE(mtype)) {
5605 		PR_PROTO("%s:%d: ERROR: invalid message type (0x%x)\n",
5606 		    proc, domid, mtype);
5607 		return;
5608 	}
5609 	if (!VALID_CPUID(cpuid)) {
5610 		PR_PROTO("%s:%d: ERROR: invalid cpuid (%d)\n",
5611 		    proc, domid, cpuid);
5612 		return;
5613 	}
5614 
5615 	/*
5616 	 * No pure data packets should reach this level.
5617 	 * Data+ack messages will reach here, but only
5618 	 * for the purpose of stopping the timer which
5619 	 * happens by default when this routine is called.
5620 	 */
5621 	ASSERT(msgtype != IDNP_DATA);
5622 
5623 	/*
5624 	 * We should never receive a request from ourself,
5625 	 * except for commands in the case of broadcasts!
5626 	 */
5627 	if ((domid == idn.localid) && (mtype != IDNP_CMD)) {
5628 		char	str[15];
5629 
5630 		inum2str(hp->m_msgtype, str);
5631 
5632 		cmn_err(CE_WARN,
5633 		    "IDN: 214: received message (%s[0x%x]) from self "
5634 		    "(domid %d)",
5635 		    str, hp->m_msgtype, domid);
5636 		return;
5637 	}
5638 
5639 	IDN_SYNC_LOCK();
5640 	/*
5641 	 * Set a flag indicating whether we really need
5642 	 * SYNC-LOCK.  We'll drop it in a little bit if
5643 	 * we really don't need it.
5644 	 */
5645 	switch (mtype) {
5646 	case IDNP_CON:
5647 	case IDNP_FIN:
5648 	case IDNP_NEGO:
5649 		sync_lock = 1;
5650 		break;
5651 
5652 	default:
5653 		break;
5654 	}
5655 
5656 	dp = &idn_domain[domid];
5657 	IDN_DLOCK_EXCL(domid);
5658 
5659 	/*
5660 	 * The only messages we do _not_ check the cookie are:
5661 	 *	nego
5662 	 *	nego+ack
5663 	 *	fin	 - if received cookie is 0.
5664 	 *	fin+ack	 - if received cookie is 0.
5665 	 *	ack/fin	 - if received cookie is 0.
5666 	 *	nack/fin - if received cookie is 0.
5667 	 */
5668 	if (((msgtype & IDNP_MSGTYPE_MASK) != IDNP_NEGO) &&
5669 	    ((mtype != IDNP_FIN) || (dcookie && dp->dcookie_recv))) {
5670 		if (dp->dcookie_recv != dcookie) {
5671 			dp->dcookie_errcnt++;
5672 			if (dp->dcookie_err == 0) {
5673 				/*
5674 				 * Set cookie error to prevent a
5675 				 * possible flood of bogus cookies
5676 				 * and thus error messages.
5677 				 */
5678 				dp->dcookie_err = 1;
5679 				cmn_err(CE_WARN,
5680 				    "IDN: 215: invalid cookie (0x%x) "
5681 				    "for message (0x%x) from domain %d",
5682 				    dcookie, hp->m_msgtype, domid);
5683 
5684 				PR_PROTO("%s:%d: received cookie (0x%x), "
5685 				    "expected (0x%x) [errcnt = %d]\n",
5686 				    proc, domid, dcookie,
5687 				    dp->dcookie_recv, dp->dcookie_errcnt);
5688 			}
5689 			IDN_DUNLOCK(domid);
5690 			IDN_SYNC_UNLOCK();
5691 			return;
5692 		}
5693 	}
5694 	dp->dcookie_err = 0;
5695 	IDN_GLOCK_EXCL();
5696 
5697 	idn_clear_awol(domid);
5698 
5699 	IDN_GUNLOCK();
5700 	if (!sync_lock)		/* really don't need SYNC-LOCK past here */
5701 		IDN_SYNC_UNLOCK();
5702 
5703 	/*
5704 	 * Stop any timers that may have been outstanding for
5705 	 * this domain, for this particular message type.
5706 	 * Note that CFG timers are directly managed by
5707 	 * config recv/send code.
5708 	 */
5709 	if ((mtype != IDNP_CFG) && (msgtype & IDNP_ACKNACK_MASK) && tcookie) {
5710 		IDN_MSGTIMER_STOP(domid, mtype, tcookie);
5711 	}
5712 
5713 	/*
5714 	 * Keep track of the last cpu to send us a message.
5715 	 * If the domain has not yet been assigned, we'll need
5716 	 * this cpuid in order to send back a respond.
5717 	 */
5718 	dp->dcpu_last = cpuid;
5719 
5720 	mt.mt_mtype = (ushort_t)msgtype;
5721 	mt.mt_atype = (ushort_t)acktype;
5722 	mt.mt_cookie = tcookie;
5723 
5724 	switch (mtype) {
5725 	case IDNP_NEGO:
5726 		(void) idn_recv_nego(domid, &mt, hp->m_xargs, dcookie);
5727 		break;
5728 
5729 	case IDNP_CFG:
5730 		idn_recv_config(domid, &mt, hp->m_xargs);
5731 		break;
5732 
5733 	case IDNP_CON:
5734 		(void) idn_recv_con(domid, &mt, hp->m_xargs);
5735 		break;
5736 
5737 	case IDNP_FIN:
5738 		(void) idn_recv_fin(domid, &mt, hp->m_xargs);
5739 		break;
5740 
5741 	case IDNP_CMD:
5742 		idn_recv_cmd(domid, &mt, hp->m_xargs);
5743 		break;
5744 
5745 	case IDNP_DATA:
5746 		ASSERT(msgtype & IDNP_ACKNACK_MASK);
5747 		/*
5748 		 * When doing the fast track we simply process
5749 		 * possible nack error conditions.  The actual
5750 		 * processing of the SMR data buffer is taken
5751 		 * care of in idnh_recv_dataack.  When NOT doing
5752 		 * the fast track, we do all the processing here
5753 		 * in the protocol server.
5754 		 */
5755 		(void) idn_recv_data(domid, &mt, hp->m_xargs);
5756 		break;
5757 
5758 	default:
5759 		/*
5760 		 * Should be receiving 0 inum and 0 acknack.
5761 		 */
5762 #ifdef DEBUG
5763 		cmn_err(CE_PANIC,
5764 #else /* DEBUG */
5765 		    cmn_err(CE_WARN,
5766 #endif /* DEBUG */
5767 			/* CSTYLED */
5768 			"IDN: 216: (0x%x)msgtype/(0x%x)acktype rcvd from "
5769 			/* CSTYLED */
5770 			"domain %d", msgtype, acktype, domid);
5771 		break;
5772 	}
5773 
5774 	IDN_DUNLOCK(domid);
5775 	/*
5776 	 * All receiving routines are responsible for dropping drwlock.
5777 	 */
5778 
5779 	if (sync_lock)
5780 		IDN_SYNC_UNLOCK();
5781 }
5782 
5783 /*
5784  * Once the CONFIG state is hit we immediately blast out all
5785  * of our config info.  This guarantees that the CONFIG state
5786  * effectively signifies that the sender has sent _all_ of
5787  * their config info.
5788  */
5789 static void
5790 idn_send_config(int domid, int phase)
5791 {
5792 	idn_domain_t	*dp;
5793 	int		rv;
5794 	clock_t		cfg_waittime = idn_msg_waittime[IDNP_CFG];
5795 	procname_t	proc = "idn_send_config";
5796 
5797 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
5798 
5799 	dp = &idn_domain[domid];
5800 
5801 	ASSERT(dp->dstate == IDNDS_CONFIG);
5802 
5803 	if (phase == 1) {
5804 		/*
5805 		 * Reset stuff in dtmp to 0:
5806 		 *	dcfgphase
5807 		 *	dcksum
5808 		 *	dncfgitems
5809 		 *	dmaxnets
5810 		 *	dmboxpernet
5811 		 */
5812 		dp->dtmp = 0;
5813 	}
5814 
5815 	if (dp->dcfgsnddone) {
5816 		if (!dp->dcfgrcvdone) {
5817 			IDN_MSGTIMER_START(domid, IDNP_CFG, 0,
5818 			    cfg_waittime, NULL);
5819 		}
5820 		return;
5821 	}
5822 
5823 	IDN_DLOCK_SHARED(idn.localid);
5824 
5825 	PR_PROTO("%s:%d: sending %s config (phase %d)\n",
5826 	    proc, domid,
5827 	    idn_domain[idn.localid].dvote.v.master ? "MASTER" : "SLAVE",
5828 	    phase);
5829 
5830 	if (idn_domain[idn.localid].dvote.v.master)
5831 		rv = idn_send_master_config(domid, phase);
5832 	else
5833 		rv = idn_send_slave_config(domid, phase);
5834 
5835 	IDN_DUNLOCK(idn.localid);
5836 
5837 	if (rv >= 0) {
5838 
5839 		if (rv == 1) {
5840 			dp->dcfgsnddone = 1;
5841 			PR_PROTO("%s:%d: SEND config DONE\n", proc, domid);
5842 			if (!dp->dcfgrcvdone) {
5843 				IDN_MSGTIMER_START(domid, IDNP_CFG, 0,
5844 				    cfg_waittime, NULL);
5845 			}
5846 		} else {
5847 			IDN_MSGTIMER_START(domid, IDNP_CFG, 0,
5848 			    cfg_waittime, NULL);
5849 		}
5850 	}
5851 }
5852 
5853 /*
5854  * Clear out the mailbox table.
5855  * NOTE: This routine touches the SMR.
5856  */
5857 static void
5858 idn_reset_mboxtbl(idn_mboxtbl_t *mtp)
5859 {
5860 	int		qi;
5861 	idn_mboxmsg_t	*mp = &mtp->mt_queue[0];
5862 
5863 	qi = 0;
5864 	do {
5865 		mp[qi].ms_bframe = 0;
5866 		mp[qi].ms_owner = 0;
5867 		mp[qi].ms_flag = 0;
5868 		IDN_MMBOXINDEX_INC(qi);
5869 	} while (qi);
5870 }
5871 
5872 static int
5873 idn_get_mbox_config(int domid, int *mindex, smr_offset_t *mtable,
5874     smr_offset_t *mdomain)
5875 {
5876 	idn_domain_t	*dp, *ldp;
5877 
5878 	dp = &idn_domain[domid];
5879 	ldp = &idn_domain[idn.localid];
5880 
5881 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
5882 	ASSERT(IDN_DLOCK_IS_SHARED(idn.localid));
5883 	ASSERT(IDN_GET_MASTERID() != IDN_NIL_DOMID);
5884 
5885 	/*
5886 	 * Get SMR offset of receive mailbox assigned
5887 	 * to respective domain.  If I'm a slave then
5888 	 * my dmbox.m_tbl will not have been assigned yet.
5889 	 * Instead of sending the actual offset I send
5890 	 * the master his assigned index.  Since the
5891 	 * master knows what offset it will assign to
5892 	 * me he can determine his assigned (recv) mailbox
5893 	 * based on the offset and given index.  The local
5894 	 * domain can also use this information once the
5895 	 * dmbox.m_tbl is received to properly assign the
5896 	 * correct mbox offset to the master.
5897 	 */
5898 	if (ldp->dmbox.m_tbl == NULL) {
5899 		/*
5900 		 * Local domain has not yet been assigned a
5901 		 * (recv) mailbox table.  This must be the
5902 		 * initial connection of this domain.
5903 		 */
5904 		ASSERT(dp->dvote.v.master && !ldp->dvote.v.master);
5905 		ASSERT(mindex);
5906 		*mindex = domid;
5907 	} else {
5908 		idn_mboxtbl_t	*mtp;
5909 
5910 		mtp = IDN_MBOXTBL_PTR(ldp->dmbox.m_tbl, domid);
5911 
5912 		ASSERT(mdomain);
5913 		*mdomain = IDN_ADDR2OFFSET(mtp);
5914 
5915 		if (ldp->dvote.v.master) {
5916 			/*
5917 			 * Need to calculate mailbox table to
5918 			 * assign to the given domain.  Since
5919 			 * I'm the master his mailbox is in
5920 			 * the (all-domains) mailbox table.
5921 			 */
5922 			mtp = IDN_MBOXAREA_BASE(idn.mboxarea, domid);
5923 			ASSERT(mtable);
5924 			*mtable = IDN_ADDR2OFFSET(mtp);
5925 
5926 			dp->dmbox.m_tbl = mtp;
5927 		}
5928 	}
5929 
5930 	return (0);
5931 }
5932 
5933 /*
5934  * RETURNS:
5935  *	1	Unexpected/unnecessary phase.
5936  *	0	Successfully handled, timer needed.
5937  */
5938 static int
5939 idn_send_master_config(int domid, int phase)
5940 {
5941 	idn_cfgsubtype_t	cfg_subtype;
5942 	int		rv = 0;
5943 	idn_domain_t	*dp, *ldp;
5944 	idn_msgtype_t	mt;
5945 	int		nmcadr;
5946 	uint_t		barpfn, larpfn;
5947 	uint_t		cpus_u32, cpus_l32;
5948 	uint_t		mcadr[3];
5949 	smr_offset_t	mbox_table, mbox_domain;
5950 	register int	b, p, m;
5951 	procname_t	proc = "idn_send_master_config";
5952 
5953 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
5954 	ASSERT(IDN_DLOCK_IS_SHARED(idn.localid));
5955 
5956 	dp = &idn_domain[domid];
5957 	ldp = &idn_domain[idn.localid];
5958 
5959 	ASSERT(dp->dstate == IDNDS_CONFIG);
5960 	ASSERT(dp->dvote.v.master == 0);
5961 	ASSERT(ldp->dvote.v.master == 1);
5962 
5963 	mt.mt_mtype = IDNP_CFG;
5964 	mt.mt_atype = 0;
5965 	mt.mt_cookie = 0;
5966 	m = 0;
5967 	mcadr[0] = mcadr[1] = mcadr[2] = 0;
5968 	cfg_subtype.val = 0;
5969 
5970 	switch (phase) {
5971 
5972 	case 1:
5973 		mbox_table = mbox_domain = IDN_NIL_SMROFFSET;
5974 		(void) idn_get_mbox_config(domid, NULL, &mbox_table,
5975 		    &mbox_domain);
5976 		/*
5977 		 * ----------------------------------------------------
5978 		 * Send: SLABSIZE, DATAMBOX.DOMAIN, DATAMBOX.TABLE
5979 		 * ----------------------------------------------------
5980 		 */
5981 		cfg_subtype.param.p[0] = IDN_CFGPARAM(IDNCFG_SIZE,
5982 		    IDNCFGARG_SIZE_SLAB);
5983 		cfg_subtype.param.p[1] = IDN_CFGPARAM(IDNCFG_DATAMBOX,
5984 		    IDNCFGARG_DATAMBOX_DOMAIN);
5985 		cfg_subtype.param.p[2] = IDN_CFGPARAM(IDNCFG_DATAMBOX,
5986 		    IDNCFGARG_DATAMBOX_TABLE);
5987 		cfg_subtype.info.num = 3;
5988 		cfg_subtype.info.phase = phase;
5989 		dp->dcfgphase = phase;
5990 
5991 		ASSERT(mbox_domain != IDN_NIL_SMROFFSET);
5992 		ASSERT(mbox_table != IDN_NIL_SMROFFSET);
5993 
5994 		PR_PROTO("%s:%d:%d: sending SLABSIZE (%d), "
5995 		    "DATAMBOX.DOMAIN (0x%x), DATAMBOX.TABLE (0x%x)\n",
5996 		    proc, domid, phase, IDN_SLAB_BUFCOUNT, mbox_domain,
5997 		    mbox_table);
5998 
5999 		IDNXDC(domid, &mt, cfg_subtype.val, IDN_SLAB_BUFCOUNT,
6000 		    mbox_domain, mbox_table);
6001 		break;
6002 
6003 	case 2:
6004 		barpfn = idn.smr.locpfn;
6005 		larpfn = barpfn + (uint_t)btop(MB2B(IDN_SMR_SIZE));
6006 		/*
6007 		 * ----------------------------------------------------
6008 		 * Send: NETID, BARLAR
6009 		 * ----------------------------------------------------
6010 		 */
6011 		cfg_subtype.param.p[0] = IDN_CFGPARAM(IDNCFG_NETID, 0);
6012 		cfg_subtype.param.p[1] = IDN_CFGPARAM(IDNCFG_BARLAR,
6013 		    IDNCFGARG_BARLAR_BAR);
6014 		cfg_subtype.param.p[2] = IDN_CFGPARAM(IDNCFG_BARLAR,
6015 		    IDNCFGARG_BARLAR_LAR);
6016 		cfg_subtype.info.num = 3;
6017 		cfg_subtype.info.phase = phase;
6018 		dp->dcfgphase = phase;
6019 
6020 		PR_PROTO("%s:%d:%d: sending NETID (%d), "
6021 		    "BARPFN/LARPFN (0x%x/0x%x)\n",
6022 		    proc, domid, phase, ldp->dnetid, barpfn, larpfn);
6023 
6024 		IDNXDC(domid, &mt, cfg_subtype.val,
6025 		    (uint_t)ldp->dnetid, barpfn, larpfn);
6026 		break;
6027 
6028 	case 3:
6029 		nmcadr = ldp->dhw.dh_nmcadr;
6030 		cpus_u32 = UPPER32_CPUMASK(ldp->dcpuset);
6031 		cpus_l32 = LOWER32_CPUMASK(ldp->dcpuset);
6032 		/*
6033 		 * ----------------------------------------------------
6034 		 * Send: CPUSET, NMCADR
6035 		 * ----------------------------------------------------
6036 		 */
6037 		cfg_subtype.param.p[0] = IDN_CFGPARAM(IDNCFG_CPUSET,
6038 		    IDNCFGARG_CPUSET_UPPER);
6039 		cfg_subtype.param.p[1] = IDN_CFGPARAM(IDNCFG_CPUSET,
6040 		    IDNCFGARG_CPUSET_LOWER);
6041 		cfg_subtype.param.p[2] = IDN_CFGPARAM(IDNCFG_NMCADR, 0);
6042 		cfg_subtype.info.num = 3;
6043 		cfg_subtype.info.phase = phase;
6044 		dp->dcfgphase = phase;
6045 
6046 		PR_PROTO("%s:%d:%d: sending CPUSET (0x%x.%x), NMCADR (%d)\n",
6047 		    proc, domid, phase, cpus_u32, cpus_l32, nmcadr);
6048 
6049 		IDNXDC(domid, &mt, cfg_subtype.val,
6050 		    cpus_u32, cpus_l32, nmcadr);
6051 		break;
6052 
6053 	case 4:
6054 		/*
6055 		 * ----------------------------------------------------
6056 		 * Send: BOARDSET, MTU, BUFSIZE
6057 		 * ----------------------------------------------------
6058 		 */
6059 		cfg_subtype.param.p[0] = IDN_CFGPARAM(IDNCFG_BOARDSET, 0);
6060 		cfg_subtype.param.p[1] = IDN_CFGPARAM(IDNCFG_SIZE,
6061 		    IDNCFGARG_SIZE_MTU);
6062 		cfg_subtype.param.p[2] = IDN_CFGPARAM(IDNCFG_SIZE,
6063 		    IDNCFGARG_SIZE_BUF);
6064 		cfg_subtype.info.num = 3;
6065 		cfg_subtype.info.phase = phase;
6066 		dp->dcfgphase = phase;
6067 
6068 		PR_PROTO("%s:%d:%d: sending BOARDSET (0x%x), MTU (0x%lx), "
6069 		    "BUFSIZE (0x%x)\n", proc, domid, phase,
6070 		    ldp->dhw.dh_boardset, IDN_MTU, IDN_SMR_BUFSIZE);
6071 
6072 		IDNXDC(domid, &mt, cfg_subtype.val,
6073 		    ldp->dhw.dh_boardset, IDN_MTU, IDN_SMR_BUFSIZE);
6074 		break;
6075 
6076 	case 5:
6077 		/*
6078 		 * ----------------------------------------------------
6079 		 * Send: MAXNETS, MBOXPERNET, CKSUM
6080 		 * ----------------------------------------------------
6081 		 */
6082 		cfg_subtype.param.p[0] = IDN_CFGPARAM(IDNCFG_DATASVR,
6083 		    IDNCFGARG_DATASVR_MAXNETS);
6084 		cfg_subtype.param.p[1] = IDN_CFGPARAM(IDNCFG_DATASVR,
6085 		    IDNCFGARG_DATASVR_MBXPERNET);
6086 		cfg_subtype.param.p[2] = IDN_CFGPARAM(IDNCFG_OPTIONS,
6087 		    IDNCFGARG_CHECKSUM);
6088 		cfg_subtype.info.num = 3;
6089 		cfg_subtype.info.phase = phase;
6090 		dp->dcfgphase = phase;
6091 
6092 		PR_PROTO("%s:%d:%d: sending MAXNETS (%d), "
6093 		    "MBOXPERNET (%d), CKSUM (%d)\n",
6094 		    proc, domid, phase,
6095 		    IDN_MAX_NETS, IDN_MBOX_PER_NET,
6096 		    IDN_CHECKSUM);
6097 
6098 		IDNXDC(domid, &mt, cfg_subtype.val,
6099 		    IDN_MAX_NETS, IDN_MBOX_PER_NET, IDN_CHECKSUM);
6100 		break;
6101 
6102 	case 6:
6103 		/*
6104 		 * ----------------------------------------------------
6105 		 * Send: NWRSIZE (piggyback on MCADRs)
6106 		 * ----------------------------------------------------
6107 		 */
6108 		cfg_subtype.param.p[0] = IDN_CFGPARAM(IDNCFG_SIZE,
6109 		    IDNCFGARG_SIZE_NWR);
6110 		mcadr[0] = IDN_NWR_SIZE;
6111 		m = 1;
6112 
6113 		/*FALLTHROUGH*/
6114 
6115 	default:	/* case 7 and above */
6116 		/*
6117 		 * ----------------------------------------------------
6118 		 * Send: MCADR's
6119 		 * ----------------------------------------------------
6120 		 * First need to figure how many we've already sent
6121 		 * based on what phase of CONFIG we're in.
6122 		 * ----------------------------------------------------
6123 		 */
6124 		if (phase > 6) {
6125 			p = ((phase - 7) * 3) + 2;
6126 			for (b = 0; (b < MAX_BOARDS) && (p > 0); b++)
6127 				if (ldp->dhw.dh_mcadr[b])
6128 					p--;
6129 		} else {
6130 			b = 0;
6131 		}
6132 
6133 		for (; (b < MAX_BOARDS) && (m < 3); b++) {
6134 			if (ldp->dhw.dh_mcadr[b] == 0)
6135 				continue;
6136 			mcadr[m] = ldp->dhw.dh_mcadr[b];
6137 			cfg_subtype.param.p[m] = IDN_CFGPARAM(IDNCFG_MCADR, b);
6138 			m++;
6139 		}
6140 		if (m > 0) {
6141 			if (phase == 6) {
6142 				PR_PROTO("%s:%d:%d: sending NWRSIZE (%d), "
6143 				    "MCADRs (0x%x, 0x%x)\n",
6144 				    proc, domid, phase,
6145 				    mcadr[0], mcadr[1], mcadr[2]);
6146 			} else {
6147 				PR_PROTO("%s:%d:%d: sending MCADRs "
6148 				    "(0x%x, 0x%x, 0x%x)\n",
6149 				    proc, domid, phase,
6150 				    mcadr[0], mcadr[1], mcadr[2]);
6151 			}
6152 			cfg_subtype.info.num = m;
6153 			cfg_subtype.info.phase = phase;
6154 			dp->dcfgphase = phase;
6155 
6156 			IDNXDC(domid, &mt, cfg_subtype.val,
6157 			    mcadr[0], mcadr[1], mcadr[2]);
6158 		} else {
6159 			rv = 1;
6160 		}
6161 		break;
6162 	}
6163 
6164 	return (rv);
6165 }
6166 
6167 /*
6168  * RETURNS:
6169  *	1	Unexpected/unnecessary phase.
6170  *	0	Successfully handled.
6171  */
6172 static int
6173 idn_send_slave_config(int domid, int phase)
6174 {
6175 	idn_cfgsubtype_t	cfg_subtype;
6176 	int		rv = 0;
6177 	idn_domain_t	*dp, *ldp;
6178 	smr_offset_t	mbox_domain;
6179 	idn_msgtype_t	mt;
6180 	int		mbox_index;
6181 	uint_t		cpus_u32, cpus_l32;
6182 	procname_t	proc = "idn_send_slave_config";
6183 
6184 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
6185 	ASSERT(IDN_DLOCK_IS_SHARED(idn.localid));
6186 
6187 	mt.mt_mtype = IDNP_CFG;
6188 	mt.mt_atype = 0;
6189 	dp = &idn_domain[domid];
6190 	ldp = &idn_domain[idn.localid];
6191 
6192 	ASSERT(dp->dstate == IDNDS_CONFIG);
6193 	ASSERT(ldp->dvote.v.master == 0);
6194 
6195 	switch (phase) {
6196 
6197 	case 1:
6198 		mbox_index = IDN_NIL_DOMID;
6199 		mbox_domain = IDN_NIL_SMROFFSET;
6200 		(void) idn_get_mbox_config(domid, &mbox_index, NULL,
6201 		    &mbox_domain);
6202 		/*
6203 		 * ----------------------------------------------------
6204 		 * Send: DATAMBOX.DOMAIN or DATAMBOX.INDEX,
6205 		 *	 DATASVR.MAXNETS, DATASVR.MBXPERNET
6206 		 * ----------------------------------------------------
6207 		 */
6208 		cfg_subtype.val = 0;
6209 		if (mbox_index == IDN_NIL_DOMID) {
6210 			ASSERT(mbox_domain != IDN_NIL_SMROFFSET);
6211 			cfg_subtype.param.p[0] = IDN_CFGPARAM(IDNCFG_DATAMBOX,
6212 			    IDNCFGARG_DATAMBOX_DOMAIN);
6213 		} else {
6214 			/*
6215 			 * Should only be sending Index to
6216 			 * the master and not another slave.
6217 			 */
6218 			ASSERT(dp->dvote.v.master);
6219 			ASSERT(mbox_domain == IDN_NIL_SMROFFSET);
6220 			cfg_subtype.param.p[0] = IDN_CFGPARAM(IDNCFG_DATAMBOX,
6221 			    IDNCFGARG_DATAMBOX_INDEX);
6222 		}
6223 		cfg_subtype.param.p[1] = IDN_CFGPARAM(IDNCFG_DATASVR,
6224 		    IDNCFGARG_DATASVR_MAXNETS);
6225 		cfg_subtype.param.p[2] = IDN_CFGPARAM(IDNCFG_DATASVR,
6226 		    IDNCFGARG_DATASVR_MBXPERNET);
6227 		cfg_subtype.info.num = 3;
6228 		cfg_subtype.info.phase = phase;
6229 		dp->dcfgphase = phase;
6230 
6231 		PR_PROTO("%s:%d:%d: sending DATAMBOX.%s (0x%x), "
6232 		    "MAXNETS (%d), MBXPERNET (%d)\n",
6233 		    proc, domid, phase,
6234 		    (IDN_CFGPARAM_ARG(cfg_subtype.param.p[0])
6235 		    == IDNCFGARG_DATAMBOX_INDEX) ? "INDEX" : "DOMAIN",
6236 		    (mbox_index == IDN_NIL_DOMID) ? mbox_domain : mbox_index,
6237 		    IDN_MAX_NETS, IDN_MBOX_PER_NET);
6238 
6239 		IDNXDC(domid, &mt, cfg_subtype.val,
6240 		    ((mbox_index == IDN_NIL_DOMID) ? mbox_domain : mbox_index),
6241 		    IDN_MAX_NETS, IDN_MBOX_PER_NET);
6242 		break;
6243 
6244 	case 2:
6245 		cpus_u32 = UPPER32_CPUMASK(ldp->dcpuset);
6246 		cpus_l32 = LOWER32_CPUMASK(ldp->dcpuset);
6247 		/*
6248 		 * ----------------------------------------------------
6249 		 * Send: NETID, CPUSET
6250 		 * ----------------------------------------------------
6251 		 */
6252 		cfg_subtype.val = 0;
6253 		cfg_subtype.param.p[0] = IDN_CFGPARAM(IDNCFG_NETID, 0);
6254 		cfg_subtype.param.p[1] = IDN_CFGPARAM(IDNCFG_CPUSET,
6255 		    IDNCFGARG_CPUSET_UPPER);
6256 		cfg_subtype.param.p[2] = IDN_CFGPARAM(IDNCFG_CPUSET,
6257 		    IDNCFGARG_CPUSET_LOWER);
6258 		cfg_subtype.info.num = 3;
6259 		cfg_subtype.info.phase = phase;
6260 		dp->dcfgphase = phase;
6261 
6262 		PR_PROTO("%s:%d:%d: sending NETID (%d), "
6263 		    "CPUSET (0x%x.%x)\n", proc, domid, phase,
6264 		    ldp->dnetid, cpus_u32, cpus_l32);
6265 
6266 		IDNXDC(domid, &mt, cfg_subtype.val,
6267 		    (uint_t)ldp->dnetid, cpus_u32, cpus_l32);
6268 		break;
6269 
6270 	case 3:
6271 		/*
6272 		 * ----------------------------------------------------
6273 		 * Send: BOARDSET, MTU, BUFSIZE
6274 		 * ----------------------------------------------------
6275 		 */
6276 		cfg_subtype.val = 0;
6277 		cfg_subtype.param.p[0] = IDN_CFGPARAM(IDNCFG_BOARDSET, 0);
6278 		cfg_subtype.param.p[1] = IDN_CFGPARAM(IDNCFG_SIZE,
6279 		    IDNCFGARG_SIZE_MTU);
6280 		cfg_subtype.param.p[2] = IDN_CFGPARAM(IDNCFG_SIZE,
6281 		    IDNCFGARG_SIZE_BUF);
6282 		cfg_subtype.info.num = 3;
6283 		cfg_subtype.info.phase = phase;
6284 		dp->dcfgphase = phase;
6285 
6286 		PR_PROTO("%s:%d:%d: sending BOARDSET (0x%x), MTU (0x%lx), "
6287 		    "BUFSIZE (0x%x)\n",
6288 		    proc, domid, phase, ldp->dhw.dh_boardset, IDN_MTU,
6289 		    IDN_SMR_BUFSIZE);
6290 
6291 		IDNXDC(domid, &mt, cfg_subtype.val,
6292 		    ldp->dhw.dh_boardset, IDN_MTU, IDN_SMR_BUFSIZE);
6293 		break;
6294 
6295 	case 4:
6296 		/*
6297 		 * ----------------------------------------------------
6298 		 * Send: SLABSIZE, OPTIONS.CHECKSUM, NWR_SIZE
6299 		 * ----------------------------------------------------
6300 		 */
6301 		cfg_subtype.val = 0;
6302 		cfg_subtype.param.p[0] = IDN_CFGPARAM(IDNCFG_SIZE,
6303 		    IDNCFGARG_SIZE_SLAB);
6304 		cfg_subtype.param.p[1] = IDN_CFGPARAM(IDNCFG_OPTIONS,
6305 		    IDNCFGARG_CHECKSUM);
6306 		cfg_subtype.param.p[2] = IDN_CFGPARAM(IDNCFG_SIZE,
6307 		    IDNCFGARG_SIZE_NWR);
6308 		cfg_subtype.info.num = 3;
6309 		cfg_subtype.info.phase = phase;
6310 		dp->dcfgphase = phase;
6311 
6312 		PR_PROTO("%s:%d:%d: sending SLABSIZE (%d), CKSUM (%d), "
6313 		    "NWRSIZE (%d)\n",
6314 		    proc, domid, phase, IDN_SLAB_BUFCOUNT,
6315 		    IDN_CHECKSUM, IDN_NWR_SIZE);
6316 
6317 		IDNXDC(domid, &mt, cfg_subtype.val,
6318 		    IDN_SLAB_BUFCOUNT, IDN_CHECKSUM, IDN_NWR_SIZE);
6319 		break;
6320 
6321 	default:
6322 		rv = 1;
6323 		break;
6324 	}
6325 
6326 	return (rv);
6327 }
6328 
6329 #define	CFG_FATAL	((uint_t)-1)	/* reset link */
6330 #define	CFG_CONTINUE	0x0000		/* looking for more */
6331 #define	CFG_DONE	0x0001		/* got everything expected */
6332 #define	CFG_ERR_MTU	0x0002
6333 #define	CFG_ERR_BUF	0x0004
6334 #define	CFG_ERR_SLAB	0x0008
6335 #define	CFG_ERR_NWR	0x0010
6336 #define	CFG_ERR_NETS	0x0020
6337 #define	CFG_ERR_MBOX	0x0040
6338 #define	CFG_ERR_NMCADR	0x0080
6339 #define	CFG_ERR_MCADR	0x0100
6340 #define	CFG_ERR_CKSUM	0x0200
6341 #define	CFG_ERR_SMR	0x0400
6342 #define	CFG_MAX_ERRORS	16
6343 
6344 #define	CFGERR2IDNKERR(ce) \
6345 	(((ce) & CFG_ERR_MTU)	? IDNKERR_CONFIG_MTU 	: \
6346 	((ce) & CFG_ERR_BUF)	? IDNKERR_CONFIG_BUF 	: \
6347 	((ce) & CFG_ERR_SLAB)	? IDNKERR_CONFIG_SLAB 	: \
6348 	((ce) & CFG_ERR_NWR)	? IDNKERR_CONFIG_NWR 	: \
6349 	((ce) & CFG_ERR_NETS)	? IDNKERR_CONFIG_NETS 	: \
6350 	((ce) & CFG_ERR_MBOX)	? IDNKERR_CONFIG_MBOX 	: \
6351 	((ce) & CFG_ERR_NMCADR)	? IDNKERR_CONFIG_NMCADR	: \
6352 	((ce) & CFG_ERR_MCADR)	? IDNKERR_CONFIG_MCADR	: \
6353 	((ce) & CFG_ERR_CKSUM)	? IDNKERR_CONFIG_CKSUM	: \
6354 	((ce) & CFG_ERR_SMR)	? IDNKERR_CONFIG_SMR	: 0)
6355 
6356 #define	CFGERR2FINARG(ce) \
6357 	(((ce) & CFG_ERR_MTU)	? IDNFIN_ARG_CFGERR_MTU    : \
6358 	((ce) & CFG_ERR_BUF)	? IDNFIN_ARG_CFGERR_BUF    : \
6359 	((ce) & CFG_ERR_SLAB)	? IDNFIN_ARG_CFGERR_SLAB   : \
6360 	((ce) & CFG_ERR_NWR)	? IDNFIN_ARG_CFGERR_NWR    : \
6361 	((ce) & CFG_ERR_NETS)	? IDNFIN_ARG_CFGERR_NETS   : \
6362 	((ce) & CFG_ERR_MBOX)	? IDNFIN_ARG_CFGERR_MBOX   : \
6363 	((ce) & CFG_ERR_NMCADR)	? IDNFIN_ARG_CFGERR_NMCADR : \
6364 	((ce) & CFG_ERR_MCADR)	? IDNFIN_ARG_CFGERR_MCADR  : \
6365 	((ce) & CFG_ERR_CKSUM)	? IDNFIN_ARG_CFGERR_CKSUM  : \
6366 	((ce) & CFG_ERR_SMR)	? IDNFIN_ARG_CFGERR_SMR	   : IDNFIN_ARG_NONE)
6367 
6368 /*
6369  * Called when some CFG messages arrive.  We use dncfgitems to count the
6370  * total number of items received so far since we'll receive multiple CFG
6371  * messages during the CONFIG phase.  Note that dncfgitems is initialized
6372  * in idn_send_config.
6373  */
6374 static void
6375 idn_recv_config(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
6376 {
6377 	uint_t		msg = mtp->mt_mtype;
6378 	uint_t		rv, rv_expected, rv_actual;
6379 	int		pnum;
6380 	int		phase;
6381 	register int	p;
6382 	register int	c;
6383 	idn_mainmbox_t	*mmp;
6384 	register uint_t	subtype, subtype_arg;
6385 	idn_domain_t	*dp;
6386 	int		index;
6387 	idn_domain_t	*ldp = &idn_domain[idn.localid];
6388 	idn_mboxtbl_t	*mbtp;
6389 	idn_cfgsubtype_t	cfg_subtype;
6390 	idn_xdcargs_t	cfg_arg;
6391 	idn_msgtype_t	mt;
6392 	idnsb_error_t	idnerr;
6393 	procname_t	proc = "idn_recv_config";
6394 
6395 	ASSERT(domid != idn.localid);
6396 
6397 	GET_XARGS(xargs, &cfg_subtype.val, &cfg_arg[0], &cfg_arg[1],
6398 	    &cfg_arg[2]);
6399 	cfg_arg[3] = 0;
6400 
6401 	dp = &idn_domain[domid];
6402 
6403 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
6404 
6405 	if (dp->dstate != IDNDS_CONFIG) {
6406 		/*
6407 		 * Not ready to receive config info.
6408 		 * Drop whatever he sent us.  Let the
6409 		 * timer continue and timeout if needed.
6410 		 */
6411 		PR_PROTO("%s:%d: WARNING state(%s) != CONFIG\n",
6412 		    proc, domid, idnds_str[dp->dstate]);
6413 		return;
6414 	}
6415 
6416 	if ((msg & IDNP_ACKNACK_MASK) || dp->dcfgsnddone) {
6417 		IDN_MSGTIMER_STOP(domid, IDNP_CFG, 0);
6418 	}
6419 
6420 	if (msg & IDNP_ACKNACK_MASK) {
6421 		/*
6422 		 * ack/cfg
6423 		 */
6424 		phase = GET_XARGS_CFG_PHASE(xargs);
6425 
6426 		PR_PROTO("%s:%d: received ACK for CFG phase %d\n",
6427 		    proc, domid, phase);
6428 		if (phase != (int)dp->dcfgphase) {
6429 			/*
6430 			 * Phase is not what we were
6431 			 * expecting.  Something got lost
6432 			 * in the shuffle.  Restart the
6433 			 * timer and let it timeout if necessary
6434 			 * and reestablish the connection.
6435 			 */
6436 			IDN_MSGTIMER_START(domid, IDNP_CFG, dp->dcfgphase,
6437 			    idn_msg_waittime[IDNP_CFG], NULL);
6438 		} else {
6439 			idn_send_config(domid, phase + 1);
6440 
6441 			if (dp->dcfgsnddone && dp->dcfgrcvdone) {
6442 				IDN_DUNLOCK(domid);
6443 				IDN_SYNC_LOCK();
6444 				IDN_DLOCK_EXCL(domid);
6445 				if (dp->dstate == IDNDS_CONFIG) {
6446 					dp->dxp = &xphase_con;
6447 					IDN_XSTATE_TRANSITION(dp, IDNXS_PEND);
6448 					bzero(xargs, sizeof (xargs));
6449 
6450 					(void) idn_xphase_transition(domid,
6451 					    NULL, xargs);
6452 				}
6453 				IDN_SYNC_UNLOCK();
6454 			}
6455 		}
6456 		return;
6457 	}
6458 
6459 	pnum = (int)cfg_subtype.info.num;
6460 	phase = (int)cfg_subtype.info.phase;
6461 
6462 	for (p = 0; p < pnum; p++) {
6463 		int	board;
6464 #ifdef DEBUG
6465 		uint_t	val;
6466 		char	*str;
6467 
6468 		val = 0;
6469 		str = NULL;
6470 #define	RCVCFG(s, v)	{ str = (s); val = (v); }
6471 #else
6472 #define	RCVCFG(s, v)	{}
6473 #endif /* DEBUG */
6474 
6475 		subtype	    = IDN_CFGPARAM_TYPE(cfg_subtype.param.p[p]);
6476 		subtype_arg = IDN_CFGPARAM_ARG(cfg_subtype.param.p[p]);
6477 
6478 		switch (subtype) {
6479 
6480 		case IDNCFG_BARLAR:
6481 			IDN_GLOCK_EXCL();
6482 			switch (subtype_arg) {
6483 
6484 			case IDNCFGARG_BARLAR_BAR:
6485 				if (idn.smr.rempfn == PFN_INVALID) {
6486 					idn.smr.rempfn = (pfn_t)cfg_arg[p];
6487 					dp->dncfgitems++;
6488 					RCVCFG("BARLAR_BAR", cfg_arg[p]);
6489 				}
6490 				break;
6491 
6492 			case IDNCFGARG_BARLAR_LAR:
6493 				if (idn.smr.rempfnlim == PFN_INVALID) {
6494 					idn.smr.rempfnlim = (pfn_t)cfg_arg[p];
6495 					dp->dncfgitems++;
6496 					RCVCFG("BARLAR_LAR", cfg_arg[p]);
6497 				}
6498 				break;
6499 
6500 			default:
6501 				cmn_err(CE_WARN,
6502 				    "IDN 217: unknown CFGARG type (%d) "
6503 				    "from domain %d",
6504 				    subtype_arg, domid);
6505 				break;
6506 			}
6507 			IDN_GUNLOCK();
6508 			break;
6509 
6510 		case IDNCFG_MCADR:
6511 			board = subtype_arg;
6512 			if ((board >= 0) && (board < MAX_BOARDS) &&
6513 			    (dp->dhw.dh_mcadr[board] == 0)) {
6514 				dp->dhw.dh_mcadr[board] = cfg_arg[p];
6515 				dp->dncfgitems++;
6516 				RCVCFG("MCADR", cfg_arg[p]);
6517 			}
6518 			break;
6519 
6520 		case IDNCFG_NMCADR:
6521 			if (dp->dhw.dh_nmcadr == 0) {
6522 				dp->dhw.dh_nmcadr = cfg_arg[p];
6523 				dp->dncfgitems++;
6524 				RCVCFG("NMCADR", cfg_arg[p]);
6525 			}
6526 			break;
6527 
6528 		case IDNCFG_CPUSET:
6529 			switch (subtype_arg) {
6530 
6531 			case IDNCFGARG_CPUSET_UPPER:
6532 			{
6533 				cpuset_t	tmpset;
6534 
6535 				MAKE64_CPUMASK(tmpset, cfg_arg[p], 0);
6536 				CPUSET_OR(dp->dcpuset, tmpset);
6537 				dp->dncfgitems++;
6538 				RCVCFG("CPUSET_UPPER", cfg_arg[p]);
6539 				break;
6540 			}
6541 			case IDNCFGARG_CPUSET_LOWER:
6542 			{
6543 				cpuset_t	tmpset;
6544 
6545 				MAKE64_CPUMASK(tmpset, 0, cfg_arg[p]);
6546 				CPUSET_OR(dp->dcpuset, tmpset);
6547 				dp->dncfgitems++;
6548 				RCVCFG("CPUSET_LOWER", cfg_arg[p]);
6549 				break;
6550 			}
6551 			default:
6552 				ASSERT(0);
6553 				break;
6554 			}
6555 			break;
6556 
6557 		case IDNCFG_NETID:
6558 			if (dp->dnetid == (ushort_t)-1) {
6559 				dp->dnetid = (ushort_t)cfg_arg[p];
6560 				dp->dncfgitems++;
6561 				RCVCFG("NETID", cfg_arg[p]);
6562 			}
6563 			break;
6564 
6565 		case IDNCFG_BOARDSET:
6566 			if ((dp->dhw.dh_boardset & cfg_arg[p])
6567 			    == dp->dhw.dh_boardset) {
6568 				/*
6569 				 * Boardset better include what we
6570 				 * already know about.
6571 				 */
6572 				dp->dhw.dh_boardset = cfg_arg[p];
6573 				dp->dncfgitems++;
6574 				RCVCFG("BOARDSET", cfg_arg[p]);
6575 			}
6576 			break;
6577 
6578 		case IDNCFG_SIZE:
6579 			switch (subtype_arg) {
6580 
6581 			case IDNCFGARG_SIZE_MTU:
6582 				if (dp->dmtu == 0) {
6583 					dp->dmtu = cfg_arg[p];
6584 					dp->dncfgitems++;
6585 					RCVCFG("MTU", cfg_arg[p]);
6586 				}
6587 				break;
6588 
6589 			case IDNCFGARG_SIZE_BUF:
6590 				if (dp->dbufsize == 0) {
6591 					dp->dbufsize = cfg_arg[p];
6592 					dp->dncfgitems++;
6593 					RCVCFG("BUFSIZE", cfg_arg[p]);
6594 				}
6595 				break;
6596 
6597 			case IDNCFGARG_SIZE_SLAB:
6598 				if (dp->dslabsize == 0) {
6599 					dp->dslabsize = (short)cfg_arg[p];
6600 					dp->dncfgitems++;
6601 					RCVCFG("SLABSIZE", cfg_arg[p]);
6602 				}
6603 				break;
6604 
6605 			case IDNCFGARG_SIZE_NWR:
6606 				if (dp->dnwrsize == 0) {
6607 					dp->dnwrsize = (short)cfg_arg[p];
6608 					dp->dncfgitems++;
6609 					RCVCFG("NWRSIZE", cfg_arg[p]);
6610 				}
6611 				break;
6612 
6613 			default:
6614 				ASSERT(0);
6615 				break;
6616 			}
6617 			break;
6618 
6619 		case IDNCFG_DATAMBOX:
6620 			switch (subtype_arg) {
6621 
6622 			case IDNCFGARG_DATAMBOX_TABLE:
6623 				if (ldp->dmbox.m_tbl ||
6624 				    !dp->dvote.v.master ||
6625 				    !VALID_NWROFFSET(cfg_arg[p], 4)) {
6626 					/*
6627 					 * Only a master should be
6628 					 * sending us a datambox table.
6629 					 */
6630 					break;
6631 				}
6632 				IDN_DLOCK_EXCL(idn.localid);
6633 				ldp->dmbox.m_tbl = (idn_mboxtbl_t *)
6634 				    IDN_OFFSET2ADDR(cfg_arg[p]);
6635 				IDN_DUNLOCK(idn.localid);
6636 				dp->dncfgitems++;
6637 				RCVCFG("DATAMBOX.TABLE", cfg_arg[p]);
6638 				break;
6639 
6640 			case IDNCFGARG_DATAMBOX_DOMAIN:
6641 				if (dp->dmbox.m_send->mm_smr_mboxp ||
6642 				    !VALID_NWROFFSET(cfg_arg[p], 4))
6643 					break;
6644 				mbtp = (idn_mboxtbl_t *)
6645 				    IDN_OFFSET2ADDR(cfg_arg[p]);
6646 				mmp = dp->dmbox.m_send;
6647 				for (c = 0; c < IDN_MAX_NETS; c++) {
6648 
6649 					mutex_enter(&mmp[c].mm_mutex);
6650 					mmp[c].mm_smr_mboxp = mbtp;
6651 					mutex_exit(&mmp[c].mm_mutex);
6652 
6653 					IDN_MBOXTBL_PTR_INC(mbtp);
6654 				}
6655 				if (c <= 0)
6656 					break;
6657 				dp->dncfgitems++;
6658 				RCVCFG("DATAMBOX.DOMAIN", cfg_arg[p]);
6659 				break;
6660 
6661 			case IDNCFGARG_DATAMBOX_INDEX:
6662 				if (!ldp->dvote.v.master ||
6663 				    dp->dmbox.m_send->mm_smr_mboxp) {
6664 					/*
6665 					 * If I'm not the master then
6666 					 * I can't handle processing a
6667 					 * mailbox index.
6668 					 * OR, if I already have the send
6669 					 * mailbox, I'm done with this
6670 					 * config item.
6671 					 */
6672 					break;
6673 				}
6674 				ASSERT(dp->dmbox.m_tbl);
6675 				index = (int)cfg_arg[p];
6676 				/*
6677 				 * The given index is the local domain's
6678 				 * index into the remote domain's mailbox
6679 				 * table that contains the mailbox that
6680 				 * remote domain wants the local domain to
6681 				 * use as the send mailbox for messages
6682 				 * destined for the remote domain.
6683 				 * I.e. from the remote domain's
6684 				 *	perspective, this is his receive
6685 				 *	mailbox.
6686 				 */
6687 				mbtp = IDN_MBOXTBL_PTR(dp->dmbox.m_tbl, index);
6688 				mmp = dp->dmbox.m_send;
6689 				for (c = 0; c < IDN_MAX_NETS; c++) {
6690 
6691 					mutex_enter(&mmp[c].mm_mutex);
6692 					mmp[c].mm_smr_mboxp = mbtp;
6693 					mutex_exit(&mmp[c].mm_mutex);
6694 
6695 					IDN_MBOXTBL_PTR_INC(mbtp);
6696 				}
6697 				if (c <= 0)
6698 					break;
6699 				dp->dncfgitems++;
6700 				RCVCFG("DATAMBOX.INDEX", cfg_arg[p]);
6701 				break;
6702 
6703 			default:
6704 				ASSERT(0);
6705 				break;
6706 			}
6707 			break;
6708 
6709 		case IDNCFG_DATASVR:
6710 			switch (subtype_arg) {
6711 
6712 			case IDNCFGARG_DATASVR_MAXNETS:
6713 				if (dp->dmaxnets)
6714 					break;
6715 				dp->dmaxnets = (uint_t)(cfg_arg[p] & 0x3f);
6716 				dp->dncfgitems++;
6717 				RCVCFG("DATASVR.MAXNETS", cfg_arg[p]);
6718 				break;
6719 
6720 			case IDNCFGARG_DATASVR_MBXPERNET:
6721 				if (dp->dmboxpernet)
6722 					break;
6723 				dp->dmboxpernet = (uint_t)(cfg_arg[p] & 0x1ff);
6724 				dp->dncfgitems++;
6725 				RCVCFG("DATASVR.MBXPERNET", cfg_arg[p]);
6726 				break;
6727 
6728 			default:
6729 				ASSERT(0);
6730 				break;
6731 			}
6732 			break;
6733 
6734 		case IDNCFG_OPTIONS:
6735 			switch (subtype_arg) {
6736 
6737 			case IDNCFGARG_CHECKSUM:
6738 				if (dp->dcksum)
6739 					break;
6740 				if ((cfg_arg[p] & 0xff) == 0)
6741 					dp->dcksum = 1;		/* off */
6742 				else
6743 					dp->dcksum = 2;		/* on */
6744 				dp->dncfgitems++;
6745 				RCVCFG("OPTIONS.CHECKSUM", cfg_arg[p]);
6746 				break;
6747 
6748 			default:
6749 				ASSERT(0);
6750 				break;
6751 			}
6752 
6753 		default:
6754 			break;
6755 		}
6756 #ifdef DEBUG
6757 		PR_PROTO("%s:%d: received %s (0x%x)\n",
6758 		    proc, domid, str ? str : "<empty>", val);
6759 #endif /* DEBUG */
6760 	}
6761 
6762 	mt.mt_mtype = IDNP_ACK;
6763 	mt.mt_atype = IDNP_CFG;
6764 	mt.mt_cookie = mtp->mt_cookie;
6765 	CLR_XARGS(cfg_arg);
6766 	SET_XARGS_CFG_PHASE(cfg_arg, phase);
6767 	idn_send_acknack(domid, &mt, cfg_arg);
6768 
6769 	rv_expected = rv_actual = 0;
6770 
6771 	if (dp->dvote.v.master == 0) {
6772 		/*
6773 		 * Remote domain is a slave, check if we've received
6774 		 * all that we were expecting, and if so transition to
6775 		 * the next state.
6776 		 */
6777 		rv = idn_check_slave_config(domid, &rv_expected, &rv_actual);
6778 	} else {
6779 		/*
6780 		 * Remote domain is a master, check if this slave has
6781 		 * received all that it was expecting, and if so
6782 		 * transition to the next state.
6783 		 */
6784 		rv = idn_check_master_config(domid, &rv_expected, &rv_actual);
6785 	}
6786 
6787 	switch (rv) {
6788 	case CFG_DONE:
6789 		/*
6790 		 * All config info received that was expected, wrap up.
6791 		 */
6792 		if (!idn_recv_config_done(domid) && dp->dvote.v.master) {
6793 			IDN_DLOCK_EXCL(idn.localid);
6794 			ldp->dvote.v.connected = 1;
6795 			IDN_DUNLOCK(idn.localid);
6796 		}
6797 		break;
6798 
6799 	case CFG_CONTINUE:
6800 		/*
6801 		 * If we're not done sending our own config, then
6802 		 * there's no need to set a timer since one will
6803 		 * automatically be set when we send a config
6804 		 * message waiting for an acknowledgement.
6805 		 */
6806 		if (dp->dcfgsnddone) {
6807 			/*
6808 			 * We haven't yet received all the config
6809 			 * information we were expecting.  Need to
6810 			 * restart CFG timer if we've sent everything..
6811 			 */
6812 			IDN_MSGTIMER_START(domid, IDNP_CFG, 0,
6813 			    idn_msg_waittime[IDNP_CFG], NULL);
6814 		}
6815 		break;
6816 
6817 	case CFG_FATAL:
6818 		/*
6819 		 * Fatal error occurred during config exchange.
6820 		 * We need to shutdown connection in this
6821 		 * case, so initiate a (non-relink) FIN.
6822 		 * so let's get the show on the road.
6823 		 */
6824 		IDN_DUNLOCK(domid);
6825 		IDN_SYNC_LOCK();
6826 		IDN_DLOCK_EXCL(domid);
6827 		/*
6828 		 * If the state has changed from CONFIG
6829 		 * then somebody else has taken over
6830 		 * control of this domain so we can just
6831 		 * bail out.
6832 		 */
6833 		if (dp->dstate == IDNDS_CONFIG) {
6834 			INIT_IDNKERR(&idnerr);
6835 			SET_IDNKERR_ERRNO(&idnerr, EPROTO);
6836 			SET_IDNKERR_IDNERR(&idnerr, IDNKERR_CONFIG_FATAL);
6837 			SET_IDNKERR_PARAM0(&idnerr, domid);
6838 			idn_update_op(IDNOP_ERROR, DOMAINSET(domid), &idnerr);
6839 			/*
6840 			 * Keep this guy around so we can try again.
6841 			 */
6842 			DOMAINSET_ADD(idn.domset.ds_relink, domid);
6843 			IDN_HISTORY_LOG(IDNH_RELINK, domid, dp->dstate,
6844 			    idn.domset.ds_relink);
6845 			(void) idn_disconnect(domid, IDNFIN_NORMAL,
6846 			    IDNFIN_ARG_CFGERR_FATAL,
6847 			    IDNFIN_SYNC_NO);
6848 		}
6849 		IDN_SYNC_UNLOCK();
6850 		break;
6851 
6852 	default:	/* parameter conflict */
6853 		IDN_DUNLOCK(domid);
6854 		IDN_SYNC_LOCK();
6855 		IDN_DLOCK_EXCL(domid);
6856 		if (dp->dstate != IDNDS_CONFIG) {
6857 			/*
6858 			 * Hmmm...changed in the short period
6859 			 * we had dropped the lock, oh well.
6860 			 */
6861 			IDN_SYNC_UNLOCK();
6862 			break;
6863 		}
6864 		c = 0;
6865 		for (p = 0; p < CFG_MAX_ERRORS; p++)
6866 			if (rv & (1 << p))
6867 				c++;
6868 		INIT_IDNKERR(&idnerr);
6869 		SET_IDNKERR_ERRNO(&idnerr, EINVAL);
6870 		SET_IDNKERR_PARAM0(&idnerr, domid);
6871 		if (c > 1) {
6872 			SET_IDNKERR_IDNERR(&idnerr, IDNKERR_CONFIG_MULTIPLE);
6873 			SET_IDNKERR_PARAM1(&idnerr, c);
6874 		} else {
6875 			SET_IDNKERR_IDNERR(&idnerr, CFGERR2IDNKERR(rv));
6876 			SET_IDNKERR_PARAM1(&idnerr, rv_expected);
6877 			SET_IDNKERR_PARAM2(&idnerr, rv_actual);
6878 		}
6879 		/*
6880 		 * Any parameter conflicts are grounds for dismissal.
6881 		 */
6882 		if (idn.domset.ds_connected == 0) {
6883 			domainset_t	domset;
6884 			/*
6885 			 * We have no other connections yet.
6886 			 * We must blow out of here completely
6887 			 * unless we have relinkers left from
6888 			 * a RECONFIG.
6889 			 */
6890 			IDN_GLOCK_EXCL();
6891 			domset = ~idn.domset.ds_relink;
6892 			if (idn.domset.ds_relink == 0) {
6893 				IDN_GSTATE_TRANSITION(IDNGS_DISCONNECT);
6894 			}
6895 			domset &= ~idn.domset.ds_hitlist;
6896 			IDN_SET_NEW_MASTERID(IDN_NIL_DOMID);
6897 			IDN_GUNLOCK();
6898 			IDN_DUNLOCK(domid);
6899 
6900 			DOMAINSET_DEL(domset, idn.localid);
6901 
6902 			idn_update_op(IDNOP_ERROR, DOMAINSET_ALL, &idnerr);
6903 
6904 			PR_HITLIST("%s:%d: unlink_domainset(%x) due to "
6905 			    "CFG error (relink=%x, hitlist=%x)\n",
6906 			    proc, domid, domset, idn.domset.ds_relink,
6907 			    idn.domset.ds_hitlist);
6908 
6909 			idn_unlink_domainset(domset, IDNFIN_NORMAL,
6910 			    CFGERR2FINARG(rv),
6911 			    IDNFIN_OPT_UNLINK,
6912 			    BOARDSET_ALL);
6913 			IDN_SYNC_UNLOCK();
6914 			IDN_DLOCK_EXCL(domid);
6915 		} else {
6916 			PR_HITLIST("%s:%d: idn_disconnect(%d) due to CFG "
6917 			    "error (conn=%x, relink=%x, hitlist=%x)\n",
6918 			    proc, domid, domid, idn.domset.ds_connected,
6919 			    idn.domset.ds_relink, idn.domset.ds_hitlist);
6920 			/*
6921 			 * If we have other connections then
6922 			 * we're only going to blow away this
6923 			 * single connection.
6924 			 */
6925 			idn_update_op(IDNOP_ERROR, DOMAINSET(domid), &idnerr);
6926 
6927 			DOMAINSET_DEL(idn.domset.ds_relink, domid);
6928 			(void) idn_disconnect(domid, IDNFIN_NORMAL,
6929 			    CFGERR2FINARG(rv), IDNFIN_SYNC_NO);
6930 			IDN_SYNC_UNLOCK();
6931 		}
6932 		break;
6933 	}
6934 }
6935 
6936 /*
6937  * Called by master or slave which expects exactly the following
6938  * with respect to config info received from a SLAVE:
6939  * 	IDNCFG_CPUSET
6940  *	IDNCFG_NETID
6941  *	IDNCFG_BOARDSET
6942  *	IDNCFG_SIZE (MTU, BUF, SLAB, NWR)
6943  *	IDNCFG_DATAMBOX (DOMAIN or INDEX if caller is master)
6944  *	IDNCFG_DATASVR (MAXNETS, MBXPERNET)
6945  *	IDNCFG_OPTIONS (CHECKSUM)
6946  */
6947 static uint_t
6948 idn_check_slave_config(int domid, uint_t *exp, uint_t *act)
6949 {
6950 	uint_t		rv = 0;
6951 	idn_domain_t	*ldp, *dp;
6952 	procname_t	proc = "idn_check_slave_config";
6953 
6954 	dp = &idn_domain[domid];
6955 	ldp = &idn_domain[idn.localid];
6956 
6957 	ASSERT(domid != idn.localid);
6958 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
6959 	ASSERT(dp->dstate == IDNDS_CONFIG);
6960 
6961 	PR_PROTO("%s:%d: number received %d, number expected %d\n",
6962 	    proc, domid, (int)dp->dncfgitems, IDN_SLAVE_NCFGITEMS);
6963 
6964 	if ((int)dp->dncfgitems < IDN_SLAVE_NCFGITEMS)
6965 		return (CFG_CONTINUE);
6966 
6967 	if ((dp->dnetid == (ushort_t)-1) ||
6968 	    CPUSET_ISNULL(dp->dcpuset) ||
6969 	    (dp->dhw.dh_boardset == 0) ||
6970 	    (dp->dmbox.m_send->mm_smr_mboxp == NULL) ||
6971 	    (dp->dmaxnets == 0) ||
6972 	    (dp->dmboxpernet == 0) ||
6973 	    (dp->dcksum == 0) ||
6974 	    (dp->dmtu == 0) ||
6975 	    (dp->dbufsize == 0) ||
6976 	    (dp->dslabsize == 0) ||
6977 	    (dp->dnwrsize == 0)) {
6978 		/*
6979 		 * We received our IDN_SLAVE_NCFGITEMS config items,
6980 		 * but not all what we were expecting!  Gotta nack and
6981 		 * close connection.
6982 		 */
6983 		cmn_err(CE_WARN,
6984 		    "IDN: 218: missing some required config items from "
6985 		    "domain %d", domid);
6986 
6987 		rv = CFG_FATAL;
6988 		goto done;
6989 	}
6990 
6991 	if (!valid_mtu(dp->dmtu)) {
6992 		cmn_err(CE_WARN,
6993 		    "IDN: 219: remote domain %d MTU (%d) invalid "
6994 		    "(local.mtu = %d)", dp->domid, dp->dmtu, ldp->dmtu);
6995 
6996 		*exp = (uint_t)ldp->dmtu;
6997 		*act = (uint_t)dp->dmtu;
6998 		rv |= CFG_ERR_MTU;
6999 	}
7000 	if (!valid_bufsize(dp->dbufsize)) {
7001 		cmn_err(CE_WARN,
7002 		    "IDN: 220: remote domain %d BUFSIZE (%d) invalid "
7003 		    "(local.bufsize = %d)", dp->domid, dp->dbufsize,
7004 		    ldp->dbufsize);
7005 
7006 		*exp = (uint_t)ldp->dbufsize;
7007 		*act = (uint_t)dp->dbufsize;
7008 		rv |= CFG_ERR_BUF;
7009 	}
7010 	if (!valid_slabsize((int)dp->dslabsize)) {
7011 		cmn_err(CE_WARN,
7012 		    "IDN: 221: remote domain %d SLABSIZE (%d) invalid "
7013 		    "(local.slabsize = %d)",
7014 		    dp->domid, dp->dslabsize, ldp->dslabsize);
7015 
7016 		*exp = (uint_t)ldp->dslabsize;
7017 		*act = (uint_t)dp->dslabsize;
7018 		rv |= CFG_ERR_SLAB;
7019 	}
7020 	if (!valid_nwrsize((int)dp->dnwrsize)) {
7021 		cmn_err(CE_WARN,
7022 		    "IDN: 223: remote domain %d NWRSIZE (%d) invalid "
7023 		    "(local.nwrsize = %d)",
7024 		    dp->domid, dp->dnwrsize, ldp->dnwrsize);
7025 
7026 		*exp = (uint_t)ldp->dnwrsize;
7027 		*act = (uint_t)dp->dnwrsize;
7028 		rv |= CFG_ERR_NWR;
7029 	}
7030 	if ((int)dp->dmaxnets != IDN_MAX_NETS) {
7031 		cmn_err(CE_WARN,
7032 		    "IDN: 224: remote domain %d MAX_NETS (%d) invalid "
7033 		    "(local.maxnets = %d)",
7034 		    dp->domid, (int)dp->dmaxnets, IDN_MAX_NETS);
7035 
7036 		*exp = (uint_t)IDN_MAX_NETS;
7037 		*act = (uint_t)dp->dmaxnets;
7038 		rv |= CFG_ERR_NETS;
7039 	}
7040 	if ((int)dp->dmboxpernet != IDN_MBOX_PER_NET) {
7041 		cmn_err(CE_WARN,
7042 		    "IDN: 225: remote domain %d MBOX_PER_NET (%d) "
7043 		    "invalid (local.mboxpernet = %d)",
7044 		    dp->domid, (int)dp->dmboxpernet, IDN_MBOX_PER_NET);
7045 
7046 		*exp = (uint_t)IDN_MBOX_PER_NET;
7047 		*act = (uint_t)dp->dmboxpernet;
7048 		rv |= CFG_ERR_MBOX;
7049 	}
7050 	if ((dp->dcksum - 1) != (uchar_t)IDN_CHECKSUM) {
7051 		cmn_err(CE_WARN,
7052 		    "IDN: 226: remote domain %d CHECKSUM flag (%d) "
7053 		    "mismatches local domain's (%d)",
7054 		    dp->domid, (int)dp->dcksum - 1, IDN_CHECKSUM);
7055 
7056 		*exp = (uint_t)IDN_CHECKSUM;
7057 		*act = (uint_t)(dp->dcksum - 1);
7058 		rv |= CFG_ERR_CKSUM;
7059 	}
7060 
7061 done:
7062 
7063 	return (rv ? rv : CFG_DONE);
7064 }
7065 
7066 /*
7067  * Called by slave ONLY which expects exactly the following
7068  * config info from the MASTER:
7069  *	IDNCFG_BARLAR
7070  *	IDNCFG_MCADR
7071  *	IDNCFG_NMCADR
7072  * 	IDNCFG_CPUSET
7073  *	IDNCFG_NETID
7074  *	IDNCFG_BOARDSET
7075  *	IDNCFG_SIZE (MTU, BUF, SLAB, NWR)
7076  *	IDNCFG_DATAMBOX (TABLE, DOMAIN)
7077  *	IDNCFG_DATASVR (MAXNETS, MBXPERNET)
7078  *	IDNCFG_OPTIONS (CHECKSUM)
7079  */
7080 static uint_t
7081 idn_check_master_config(int domid, uint_t *exp, uint_t *act)
7082 {
7083 	uint_t		rv = 0;
7084 	int		nmcadr;
7085 	int		total_expitems;
7086 	int		p, m, err;
7087 	idn_domain_t	*dp;
7088 	idn_domain_t	*ldp = &idn_domain[idn.localid];
7089 	procname_t	proc = "idn_check_master_config";
7090 
7091 	dp = &idn_domain[domid];
7092 
7093 	ASSERT(IDN_GET_MASTERID() != idn.localid);
7094 	ASSERT(domid != idn.localid);
7095 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
7096 	ASSERT(dp->dstate == IDNDS_CONFIG);
7097 
7098 	PR_PROTO("%s:%d: number received %d, minimum number expected %d\n",
7099 	    proc, domid, (int)dp->dncfgitems, IDN_MASTER_NCFGITEMS);
7100 
7101 	if ((int)dp->dncfgitems < IDN_MASTER_NCFGITEMS)
7102 		return (CFG_CONTINUE);
7103 
7104 	/*
7105 	 * We have at least IDN_MASTER_NCFGITEMS items which
7106 	 * means we have at least one MCADR.  Need to make sure
7107 	 * we have all that we're expecting, NMCADR.
7108 	 */
7109 	total_expitems = IDN_MASTER_NCFGITEMS + dp->dhw.dh_nmcadr - 1;
7110 	if ((dp->dhw.dh_nmcadr == 0) ||
7111 	    ((int)dp->dncfgitems < total_expitems)) {
7112 		/*
7113 		 * We have not yet received all the MCADRs
7114 		 * we're expecting.
7115 		 */
7116 		PR_PROTO("%s:%d: haven't received all MCADRs yet.\n",
7117 		    proc, domid);
7118 		return (CFG_CONTINUE);
7119 	}
7120 
7121 	nmcadr = 0;
7122 	for (p = 0; p < MAX_BOARDS; p++)
7123 		if (dp->dhw.dh_mcadr[p] != 0)
7124 			nmcadr++;
7125 
7126 	IDN_GLOCK_SHARED();
7127 	if ((idn.smr.rempfn == PFN_INVALID) ||
7128 	    (idn.smr.rempfnlim == PFN_INVALID) ||
7129 	    (dp->dnetid == (ushort_t)-1) ||
7130 	    CPUSET_ISNULL(dp->dcpuset) ||
7131 	    (dp->dhw.dh_boardset == 0) ||
7132 	    (nmcadr != dp->dhw.dh_nmcadr) ||
7133 	    (dp->dmbox.m_send->mm_smr_mboxp == NULL) ||
7134 	    (ldp->dmbox.m_tbl == NULL) ||
7135 	    (dp->dmaxnets == 0) ||
7136 	    (dp->dmboxpernet == 0) ||
7137 	    (dp->dcksum == 0) ||
7138 	    (dp->dmtu == 0) ||
7139 	    (dp->dbufsize == 0) ||
7140 	    (dp->dnwrsize == 0)) {
7141 
7142 		IDN_GUNLOCK();
7143 		/*
7144 		 * We received all of our config items, but not
7145 		 * all what we were expecting!  Gotta reset and
7146 		 * close connection.
7147 		 */
7148 		cmn_err(CE_WARN,
7149 		    "IDN: 227: missing some required config items from "
7150 		    "domain %d", domid);
7151 
7152 		rv = CFG_FATAL;
7153 		goto done;
7154 	}
7155 	if ((idn.smr.rempfnlim - idn.smr.rempfn) > btop(MB2B(IDN_SMR_SIZE))) {
7156 		/*
7157 		 * The master's SMR region is larger than
7158 		 * mine!  This means that this domain may
7159 		 * receive I/O buffers which are out of the
7160 		 * range of this local domain's SMR virtual
7161 		 * address space.  The master SMR has to be
7162 		 * no larger than the local SMR in order to
7163 		 * guarantee enough local virtual addresses
7164 		 * to see all of the SMR space.
7165 		 * XXX - Possibly add negotiating SMR size.
7166 		 *	 Try to create a new virtual mapping.
7167 		 *	 Could let domains negotiate SMR size.
7168 		 *	 Winning size would have to be smallest
7169 		 *	 in DC.  If so, how to handle incoming
7170 		 *	 domains with even smaller SMRs?
7171 		 *	 - Could either disallow connection
7172 		 *	 - Could reconfigure to use smaller SMR.
7173 		 */
7174 		cmn_err(CE_WARN,
7175 		    "IDN: 228: master's SMR (%ld) larger than "
7176 		    "local's SMR (%ld)",
7177 		    idn.smr.rempfnlim - idn.smr.rempfn,
7178 		    btop(MB2B(IDN_SMR_SIZE)));
7179 
7180 		*exp = (uint_t)IDN_SMR_SIZE;
7181 		*act = (uint_t)B2MB(ptob(idn.smr.rempfnlim - idn.smr.rempfn));
7182 		rv |= CFG_ERR_SMR;
7183 	}
7184 	IDN_GUNLOCK();
7185 
7186 	if (!valid_mtu(dp->dmtu)) {
7187 		cmn_err(CE_WARN,
7188 		    "IDN: 219: remote domain %d MTU (%d) invalid "
7189 		    "(local.mtu = %d)", dp->domid, dp->dmtu, ldp->dmtu);
7190 
7191 		*exp = (uint_t)ldp->dmtu;
7192 		*act = (uint_t)dp->dmtu;
7193 		rv |= CFG_ERR_MTU;
7194 	}
7195 	if (!valid_bufsize(dp->dbufsize)) {
7196 		cmn_err(CE_WARN,
7197 		    "IDN: 220: remote domain %d BUFSIZE (%d) invalid "
7198 		    "(local.bufsize = %d)", dp->domid, dp->dbufsize,
7199 		    ldp->dbufsize);
7200 
7201 		*exp = (uint_t)ldp->dbufsize;
7202 		*act = (uint_t)dp->dbufsize;
7203 		rv |= CFG_ERR_BUF;
7204 	}
7205 	if (!valid_nwrsize((int)dp->dnwrsize)) {
7206 		cmn_err(CE_WARN,
7207 		    "IDN: 223: remote domain %d NWRSIZE (%d) invalid "
7208 		    "(local.nwrsize = %d)",
7209 		    dp->domid, dp->dnwrsize, ldp->dnwrsize);
7210 
7211 		*exp = (uint_t)ldp->dnwrsize;
7212 		*act = (uint_t)dp->dnwrsize;
7213 		rv |= CFG_ERR_NWR;
7214 	}
7215 	if ((int)dp->dmaxnets != IDN_MAX_NETS) {
7216 		cmn_err(CE_WARN,
7217 		    "IDN: 224: remote domain %d MAX_NETS (%d) invalid "
7218 		    "(local.maxnets = %d)",
7219 		    dp->domid, (int)dp->dmaxnets, IDN_MAX_NETS);
7220 
7221 		*exp = (uint_t)IDN_MAX_NETS;
7222 		*act = (uint_t)dp->dmaxnets;
7223 		rv |= CFG_ERR_NETS;
7224 	}
7225 	if ((int)dp->dmboxpernet != IDN_MBOX_PER_NET) {
7226 		cmn_err(CE_WARN,
7227 		    "IDN: 225: remote domain %d MBOX_PER_NET (%d) "
7228 		    "invalid (local.mboxpernet = %d)",
7229 		    dp->domid, (int)dp->dmboxpernet, IDN_MBOX_PER_NET);
7230 
7231 		*exp = (uint_t)IDN_MBOX_PER_NET;
7232 		*act = (uint_t)dp->dmboxpernet;
7233 		rv |= CFG_ERR_MBOX;
7234 	}
7235 	if ((dp->dcksum - 1) != (uchar_t)IDN_CHECKSUM) {
7236 		cmn_err(CE_WARN,
7237 		    "IDN: 226: remote domain %d CHECKSUM flag (%d) "
7238 		    "mismatches local domain's (%d)",
7239 		    dp->domid, (int)dp->dcksum - 1, IDN_CHECKSUM);
7240 
7241 		*exp = (uint_t)IDN_CHECKSUM;
7242 		*act = (uint_t)(dp->dcksum - 1);
7243 		rv |= CFG_ERR_CKSUM;
7244 	}
7245 	nmcadr = 0;
7246 	err = 0;
7247 	for (m = 0; m < MAX_BOARDS; m++) {
7248 		if (!BOARD_IN_SET(dp->dhw.dh_boardset, m) &&
7249 		    dp->dhw.dh_mcadr[m]) {
7250 			cmn_err(CE_WARN,
7251 			    "IDN: 229: remote domain %d boardset (0x%x) "
7252 			    "conflicts with MCADR(board %d) [0x%x]",
7253 			    dp->domid, (uint_t)dp->dhw.dh_boardset, m,
7254 			    dp->dhw.dh_mcadr[m]);
7255 			err++;
7256 		}
7257 		if (dp->dhw.dh_mcadr[m])
7258 			nmcadr++;
7259 	}
7260 	if (err) {
7261 		*exp = 0;
7262 		*act = err;
7263 		rv |= CFG_ERR_MCADR;
7264 	} else if (nmcadr != dp->dhw.dh_nmcadr) {
7265 		cmn_err(CE_WARN,
7266 		    "IDN: 230: remote domain %d reported number of "
7267 		    "MCADRs (%d) mismatches received (%d)",
7268 		    dp->domid, dp->dhw.dh_nmcadr, nmcadr);
7269 		*exp = (uint_t)dp->dhw.dh_nmcadr;
7270 		*act = (uint_t)nmcadr;
7271 		rv |= CFG_ERR_NMCADR;
7272 	}
7273 
7274 done:
7275 
7276 	return (rv ? rv : CFG_DONE);
7277 }
7278 
7279 static int
7280 idn_recv_config_done(int domid)
7281 {
7282 	boardset_t		b_conflicts;
7283 	cpuset_t		p_conflicts;
7284 	register int		p, i;
7285 	register idn_domain_t	*dp;
7286 	idnsb_error_t		idnerr;
7287 	procname_t		proc = "idn_recv_config_done";
7288 
7289 	ASSERT(domid != IDN_NIL_DOMID);
7290 	dp = &idn_domain[domid];
7291 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
7292 
7293 	/*
7294 	 * Well, we received all that we were expecting
7295 	 * so stop any CFG timers we had going.
7296 	 */
7297 	IDN_MSGTIMER_STOP(domid, IDNP_CFG, 0);
7298 
7299 	dp->dncpus = 0;
7300 	for (p = 0; p < NCPU; p++)
7301 		if (CPU_IN_SET(dp->dcpuset, p))
7302 			dp->dncpus++;
7303 	dp->dhw.dh_nboards = 0;
7304 	for (p = 0; p < MAX_BOARDS; p++)
7305 		if (BOARD_IN_SET(dp->dhw.dh_boardset, p))
7306 			dp->dhw.dh_nboards++;
7307 
7308 	IDN_GLOCK_EXCL();
7309 	/*
7310 	 * Verify dcpuset and dhw.dh_boardset don't
7311 	 * conflict with any existing DC member.
7312 	 */
7313 	b_conflicts = idn.dc_boardset & dp->dhw.dh_boardset;
7314 	CPUSET_ZERO(p_conflicts);
7315 	CPUSET_OR(p_conflicts, idn.dc_cpuset);
7316 	CPUSET_AND(p_conflicts, dp->dcpuset);
7317 
7318 	if (b_conflicts || !CPUSET_ISNULL(p_conflicts)) {
7319 		if (b_conflicts) {
7320 			cmn_err(CE_WARN,
7321 			    "IDN: 231: domain %d boardset "
7322 			    "(0x%x) conflicts with existing "
7323 			    "IDN boardset (0x%x)",
7324 			    domid, dp->dhw.dh_boardset,
7325 			    b_conflicts);
7326 		}
7327 		if (!CPUSET_ISNULL(p_conflicts)) {
7328 			cmn_err(CE_WARN,
7329 			    "IDN: 232: domain %d cpuset "
7330 			    "(0x%x.%0x) conflicts with existing "
7331 			    "IDN cpuset (0x%x.%0x)", domid,
7332 			    UPPER32_CPUMASK(dp->dcpuset),
7333 			    LOWER32_CPUMASK(dp->dcpuset),
7334 			    UPPER32_CPUMASK(p_conflicts),
7335 			    LOWER32_CPUMASK(p_conflicts));
7336 		}
7337 		IDN_GUNLOCK();
7338 		/*
7339 		 * Need to disconnect and not retry with this guy.
7340 		 */
7341 		IDN_DUNLOCK(domid);
7342 		IDN_SYNC_LOCK();
7343 		DOMAINSET_DEL(idn.domset.ds_relink, domid);
7344 		IDN_DLOCK_EXCL(domid);
7345 
7346 		INIT_IDNKERR(&idnerr);
7347 		SET_IDNKERR_ERRNO(&idnerr, EPROTO);
7348 		SET_IDNKERR_IDNERR(&idnerr, IDNKERR_CONFIG_FATAL);
7349 		SET_IDNKERR_PARAM0(&idnerr, domid);
7350 		idn_update_op(IDNOP_ERROR, DOMAINSET(domid), &idnerr);
7351 
7352 		(void) idn_disconnect(domid, IDNFIN_FORCE_HARD,
7353 		    IDNFIN_ARG_CFGERR_FATAL, IDNFIN_SYNC_NO);
7354 		IDN_SYNC_UNLOCK();
7355 
7356 		return (-1);
7357 	}
7358 
7359 	idn_mainmbox_reset(domid, dp->dmbox.m_send);
7360 	idn_mainmbox_reset(domid, dp->dmbox.m_recv);
7361 
7362 #ifdef IDNBUG_CPUPERBOARD
7363 	/*
7364 	 * We only allow connections to domains whose (mem) boards
7365 	 * all have at least one cpu.  This is necessary so that
7366 	 * we can program the CICs of that respective board.  This
7367 	 * is primarily only a requirement if the remote domain
7368 	 * is the master _and_ has the SMR in that particular board.
7369 	 * To simplify the checking we simply restrict connections to
7370 	 * domains that have at least one cpu on all boards that
7371 	 * contain memory.
7372 	 */
7373 	if (!idn_cpu_per_board((void *)NULL, dp->dcpuset, &dp->dhw)) {
7374 		cmn_err(CE_WARN,
7375 		    "IDN: 233: domain %d missing CPU per "
7376 		    "memory boardset (0x%x), CPU boardset (0x%x)",
7377 		    domid, dp->dhw.dh_boardset,
7378 		    cpuset2boardset(dp->dcpuset));
7379 
7380 		IDN_GUNLOCK();
7381 		/*
7382 		 * Need to disconnect and not retry with this guy.
7383 		 */
7384 		IDN_DUNLOCK(domid);
7385 		IDN_SYNC_LOCK();
7386 		DOMAINSET_DEL(idn.domset.ds_relink, domid);
7387 		IDN_DLOCK_EXCL(domid);
7388 
7389 		INIT_IDNKERR(&idnerr);
7390 		SET_IDNKERR_ERRNO(&idnerr, EINVAL);
7391 		SET_IDNKERR_IDNERR(&idnerr, IDNKERR_CPU_CONFIG);
7392 		SET_IDNKERR_PARAM0(&idnerr, domid);
7393 		idn_update_op(IDNOP_ERROR, DOMAINSET(domid), &idnerr);
7394 
7395 		(void) idn_disconnect(domid, IDNFIN_FORCE_HARD,
7396 		    IDNFIN_ARG_CPUCFG, IDNFIN_SYNC_NO);
7397 		IDN_SYNC_UNLOCK();
7398 
7399 		return (-1);
7400 	}
7401 #endif /* IDNBUG_CPUPERBOARD */
7402 
7403 	CPUSET_OR(idn.dc_cpuset, dp->dcpuset);
7404 	idn.dc_boardset |= dp->dhw.dh_boardset;
7405 
7406 	IDN_GUNLOCK();
7407 
7408 	/*
7409 	 * Set up the portmap for this domain.
7410 	 */
7411 	i = -1;
7412 	for (p = 0; p < NCPU; p++) {
7413 		BUMP_INDEX(dp->dcpuset, i);
7414 		dp->dcpumap[p] = (uchar_t)i;
7415 	}
7416 
7417 	/*
7418 	 * Got everything we need from the remote
7419 	 * domain, now we can program hardware as needed.
7420 	 */
7421 	if (idn_program_hardware(domid) != 0) {
7422 		domainset_t	domset;
7423 		/*
7424 		 * Yikes!  Failed to program hardware.
7425 		 * Gotta bail.
7426 		 */
7427 		cmn_err(CE_WARN,
7428 		    "IDN: 234: failed to program hardware for domain %d "
7429 		    "(boardset = 0x%x)",
7430 		    domid, dp->dhw.dh_boardset);
7431 
7432 		IDN_DUNLOCK(domid);
7433 		/*
7434 		 * If we're having problems programming our
7435 		 * hardware we better unlink completely from
7436 		 * the IDN before things get really bad.
7437 		 */
7438 		IDN_SYNC_LOCK();
7439 		IDN_GLOCK_EXCL();
7440 		IDN_GSTATE_TRANSITION(IDNGS_DISCONNECT);
7441 		domset = DOMAINSET_ALL;
7442 		DOMAINSET_DEL(domset, idn.localid);
7443 		IDN_SET_NEW_MASTERID(IDN_NIL_DOMID);
7444 		IDN_GUNLOCK();
7445 
7446 		INIT_IDNKERR(&idnerr);
7447 		SET_IDNKERR_ERRNO(&idnerr, EINVAL);
7448 		SET_IDNKERR_IDNERR(&idnerr, IDNKERR_HW_ERROR);
7449 		SET_IDNKERR_PARAM0(&idnerr, domid);
7450 		idn_update_op(IDNOP_ERROR, DOMAINSET_ALL, &idnerr);
7451 
7452 		idn_unlink_domainset(domset, IDNFIN_NORMAL, IDNFIN_ARG_HWERR,
7453 		    IDNFIN_OPT_UNLINK, BOARDSET_ALL);
7454 
7455 		IDN_SYNC_UNLOCK();
7456 		IDN_DLOCK_EXCL(domid);
7457 
7458 		return (-1);
7459 	}
7460 
7461 	/*
7462 	 * Now that hardware has been programmed we can
7463 	 * remap the SMR into our local space, if necessary.
7464 	 */
7465 	IDN_GLOCK_EXCL();
7466 	if (domid == IDN_GET_MASTERID()) {
7467 		/*
7468 		 * No need to worry about disabling the data
7469 		 * server since at this stage there is only
7470 		 * one and he doesn't go active until his
7471 		 * mailbox (dmbox.m_recv->mm_smr_mboxp) is set up.
7472 		 */
7473 		smr_remap(&kas, idn.smr.vaddr, idn.smr.rempfn, IDN_SMR_SIZE);
7474 	}
7475 	IDN_GUNLOCK();
7476 
7477 	/*
7478 	 * There is no need to ACK the CFG messages since remote
7479 	 * domain would not progress to the next state (CON_SENT)
7480 	 * unless he has received everything.
7481 	 */
7482 
7483 	dp->dcfgrcvdone = 1;
7484 	PR_PROTO("%s:%d: RECV config DONE\n", proc, domid);
7485 
7486 	if (dp->dcfgsnddone) {
7487 		idn_xdcargs_t	xargs;
7488 		/*
7489 		 * Well, we've received all that we were expecting,
7490 		 * but we don't know if the remote domain has
7491 		 * received all that it was expecting from us,
7492 		 * although we know we transferred everything
7493 		 * so let's get the show on the road.
7494 		 */
7495 		IDN_DUNLOCK(domid);
7496 		IDN_SYNC_LOCK();
7497 		IDN_DLOCK_EXCL(domid);
7498 		/*
7499 		 * If the state has changed from CONFIG
7500 		 * then somebody else has taken over
7501 		 * control of this domain so we can just
7502 		 * bail out.
7503 		 */
7504 		if (dp->dstate == IDNDS_CONFIG) {
7505 			dp->dxp = &xphase_con;
7506 			IDN_XSTATE_TRANSITION(dp, IDNXS_PEND);
7507 			bzero(xargs, sizeof (xargs));
7508 
7509 			(void) idn_xphase_transition(domid, NULL, xargs);
7510 		}
7511 		IDN_SYNC_UNLOCK();
7512 	}
7513 
7514 	return (0);
7515 }
7516 
7517 static int
7518 idn_verify_config_mbox(int domid)
7519 {
7520 	idn_domain_t	*ldp, *dp;
7521 	idn_mainmbox_t	*mmp;
7522 	idn_mboxtbl_t	*mtp;
7523 	int		c, rv = 0;
7524 	uint_t		activeptr, readyptr;
7525 	ushort_t	mbox_csum;
7526 
7527 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
7528 
7529 	dp = &idn_domain[domid];
7530 	ldp = &idn_domain[idn.localid];
7531 
7532 	/*
7533 	 * The master will have assigned us the dmbox.m_tbl
7534 	 * from which we assign our receive mailboxes.
7535 	 * The first (0) entry contains the cookie used
7536 	 * for verification.
7537 	 */
7538 	IDN_DLOCK_SHARED(idn.localid);
7539 	/*
7540 	 * Now that we have an assigned mboxtbl from the
7541 	 * master, we can determine which receive mailbox
7542 	 * we indirectly assigned to him at the time we
7543 	 * sent him his MBOX_INDEX.  Prep it, however note
7544 	 * that the master will have not been able to
7545 	 * validate it because of the chicken 'n egg
7546 	 * problem between a master and slave.  Thus we
7547 	 * need to reset the cookie after the prep.
7548 	 */
7549 	mmp = dp->dmbox.m_recv;
7550 	mtp = IDN_MBOXTBL_PTR(ldp->dmbox.m_tbl, domid);
7551 	for (c = 0; c < IDN_MAX_NETS; c++) {
7552 		mutex_enter(&mmp[c].mm_mutex);
7553 		ASSERT(!mmp[c].mm_smr_mboxp);
7554 
7555 		mmp[c].mm_smr_mboxp = mtp;
7556 		mbox_csum = IDN_CKSUM_MBOX(&mtp->mt_header);
7557 		if (!VALID_MBOXHDR(&mtp->mt_header, c, mbox_csum)) {
7558 			cmn_err(CE_WARN,
7559 			    "IDN: 235: [recv] mailbox (domain %d, "
7560 			    "channel %d) SMR CORRUPTED - RELINK",
7561 			    domid, c);
7562 			cmn_err(CE_CONT,
7563 			    "IDN: 235: [recv] expected (cookie 0x%x, "
7564 			    "cksum 0x%x) actual (cookie 0x%x, "
7565 			    "cksum 0x%x)\n",
7566 			    IDN_GET_MBOXHDR_COOKIE(&mtp->mt_header),
7567 			    (int)mtp->mt_header.mh_cksum,
7568 			    IDN_MAKE_MBOXHDR_COOKIE(0, 0, c),
7569 			    (int)mbox_csum);
7570 			mutex_exit(&mmp[c].mm_mutex);
7571 			rv = -1;
7572 			break;
7573 		}
7574 		activeptr = mtp->mt_header.mh_svr_active_ptr;
7575 		readyptr = mtp->mt_header.mh_svr_ready_ptr;
7576 		/*
7577 		 * Verify pointers are valid.
7578 		 */
7579 		if (!activeptr || !VALID_NWROFFSET(activeptr, 2) ||
7580 		    !readyptr || !VALID_NWROFFSET(readyptr, 2)) {
7581 			cmn_err(CE_WARN,
7582 			    "IDN: 235: [recv] mailbox (domain %d, "
7583 			    "channel %d) SMR CORRUPTED - RELINK",
7584 			    domid, c);
7585 			cmn_err(CE_CONT,
7586 			    "IDN: 235: [recv] activeptr (0x%x), "
7587 			    "readyptr (0x%x)\n",
7588 			    activeptr, readyptr);
7589 			mutex_exit(&mmp[c].mm_mutex);
7590 			rv = -1;
7591 			break;
7592 		}
7593 		mmp[c].mm_smr_activep =	(ushort_t *)IDN_OFFSET2ADDR(activeptr);
7594 		mmp[c].mm_smr_readyp =	(ushort_t *)IDN_OFFSET2ADDR(readyptr);
7595 		mutex_exit(&mmp[c].mm_mutex);
7596 		IDN_MBOXTBL_PTR_INC(mtp);
7597 	}
7598 
7599 	IDN_DUNLOCK(idn.localid);
7600 
7601 	if (rv)
7602 		return (rv);
7603 
7604 	/*
7605 	 * Now we need to translate SMR offsets for send mailboxes
7606 	 * to actual virtual addresses.
7607 	 */
7608 	mmp = dp->dmbox.m_send;
7609 	for (c = 0; c < IDN_MAX_NETS; mmp++, c++) {
7610 		mutex_enter(&mmp->mm_mutex);
7611 		if ((mtp = mmp->mm_smr_mboxp) == NULL) {
7612 			mutex_exit(&mmp->mm_mutex);
7613 			rv = -1;
7614 			break;
7615 		}
7616 
7617 		mbox_csum = IDN_CKSUM_MBOX(&mtp->mt_header);
7618 
7619 		if (!VALID_MBOXHDR(&mtp->mt_header, c, mbox_csum)) {
7620 			cmn_err(CE_WARN,
7621 			    "IDN: 235: [send] mailbox (domain %d, "
7622 			    "channel %d) SMR CORRUPTED - RELINK",
7623 			    domid, c);
7624 			cmn_err(CE_CONT,
7625 			    "IDN: 235: [send] expected (cookie 0x%x, "
7626 			    "cksum 0x%x) actual (cookie 0x%x, "
7627 			    "cksum 0x%x)\n",
7628 			    IDN_GET_MBOXHDR_COOKIE(&mtp->mt_header),
7629 			    (int)mtp->mt_header.mh_cksum,
7630 			    IDN_MAKE_MBOXHDR_COOKIE(0, 0, c),
7631 			    (int)mbox_csum);
7632 			mutex_exit(&mmp->mm_mutex);
7633 			rv = -1;
7634 			break;
7635 		}
7636 		activeptr = mtp->mt_header.mh_svr_active_ptr;
7637 		readyptr = mtp->mt_header.mh_svr_ready_ptr;
7638 		/*
7639 		 * Paranoid check.
7640 		 */
7641 		if (!activeptr || !VALID_NWROFFSET(activeptr, 2) ||
7642 		    !readyptr || !VALID_NWROFFSET(readyptr, 2)) {
7643 			cmn_err(CE_WARN,
7644 			    "IDN: 235: [send] mailbox (domain %d, "
7645 			    "channel %d) SMR CORRUPTED - RELINK",
7646 			    domid, c);
7647 			cmn_err(CE_CONT,
7648 			    "IDN: 235: [send] activeptr (0x%x), "
7649 			    "readyptr (0x%x)\n",
7650 			    activeptr, readyptr);
7651 			mutex_exit(&mmp->mm_mutex);
7652 			rv = -1;
7653 			break;
7654 		}
7655 		mmp->mm_smr_activep = (ushort_t *)IDN_OFFSET2ADDR(activeptr);
7656 		mmp->mm_smr_readyp = (ushort_t *)IDN_OFFSET2ADDR(readyptr);
7657 		idn_reset_mboxtbl(mtp);
7658 		mutex_exit(&mmp->mm_mutex);
7659 		IDN_MBOXTBL_PTR_INC(mtp);
7660 	}
7661 
7662 	return (rv);
7663 }
7664 
7665 /*
7666  * The BUFSIZEs between domains have to be equal so that slave buffers
7667  * and the master's slabpool are consistent.
7668  * The MTUs between domains have to be equal so they can transfer
7669  * packets consistently without possible data truncation.
7670  *
7671  * ZZZ - Perhaps these could be negotiated?
7672  */
7673 static int
7674 valid_mtu(uint_t mtu)
7675 {
7676 	return ((mtu == idn_domain[idn.localid].dmtu) && mtu);
7677 }
7678 
7679 static int
7680 valid_bufsize(uint_t bufsize)
7681 {
7682 	return ((bufsize == idn_domain[idn.localid].dbufsize) && bufsize);
7683 }
7684 
7685 static int
7686 valid_slabsize(int slabsize)
7687 {
7688 	return ((slabsize == idn_domain[idn.localid].dslabsize) && slabsize);
7689 }
7690 
7691 static int
7692 valid_nwrsize(int nwrsize)
7693 {
7694 	return ((nwrsize == idn_domain[idn.localid].dnwrsize) && nwrsize);
7695 }
7696 
7697 static int
7698 idn_program_hardware(int domid)
7699 {
7700 	int		rv, is_master;
7701 	idn_domain_t	*dp;
7702 	uint_t		*mcadrp;
7703 	pfn_t		rem_pfn, rem_pfnlimit;
7704 	procname_t	proc = "idn_program_hardware";
7705 
7706 	PR_PROTO("%s:%d: program hw in domain %d w.r.t remote domain %d\n",
7707 	    proc, domid, idn.localid, domid);
7708 
7709 	dp = &idn_domain[domid];
7710 
7711 	ASSERT(domid != idn.localid);
7712 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
7713 	ASSERT(dp->dstate == IDNDS_CONFIG);
7714 
7715 	IDN_GLOCK_EXCL();
7716 
7717 	if (DOMAIN_IN_SET(idn.domset.ds_hwlinked, domid)) {
7718 		IDN_GUNLOCK();
7719 		return (0);
7720 	}
7721 
7722 	DOMAINSET_ADD(idn.domset.ds_flush, domid);
7723 	CHECKPOINT_OPENED(IDNSB_CHKPT_CACHE, dp->dhw.dh_boardset, 1);
7724 
7725 	if (domid != IDN_GET_MASTERID()) {
7726 		/*
7727 		 * If the remote domain is a slave, then
7728 		 * all we have to program is the CIC sm_mask.
7729 		 */
7730 		is_master = 0;
7731 		if ((idn.localid == IDN_GET_MASTERID()) &&
7732 		    lock_try(&idn.first_hwlink)) {
7733 			/*
7734 			 * This is our first HW link and I'm the
7735 			 * master, which means we need to program
7736 			 * our local bar/lar.
7737 			 */
7738 			ASSERT(idn.first_hwmasterid == (short)IDN_NIL_DOMID);
7739 			idn.first_hwmasterid = (short)idn.localid;
7740 			rem_pfn = idn.smr.locpfn;
7741 			rem_pfnlimit = idn.smr.locpfn +
7742 			    btop(MB2B(IDN_SMR_SIZE));
7743 		} else {
7744 			/*
7745 			 * Otherwise, just a slave linking to
7746 			 * another slave.  No bar/lar updating
7747 			 * necessary.
7748 			 */
7749 			rem_pfn = rem_pfnlimit = PFN_INVALID;
7750 		}
7751 		mcadrp = NULL;
7752 	} else {
7753 		/*
7754 		 * If the remote domain is a master, then
7755 		 * we need to program the CIC sm_mask/sm_bar/sm_lar,
7756 		 * and PC's.
7757 		 */
7758 		is_master = 1;
7759 		rem_pfn = idn.smr.rempfn;
7760 		rem_pfnlimit = idn.smr.rempfnlim;
7761 		mcadrp = dp->dhw.dh_mcadr;
7762 		ASSERT(idn.first_hwmasterid == (short)IDN_NIL_DOMID);
7763 		idn.first_hwmasterid = (short)domid;
7764 	}
7765 
7766 	PR_PROTO("%s:%d: ADD bset (0x%x)\n", proc, domid, dp->dhw.dh_boardset);
7767 
7768 	rv = idnxf_shmem_add(is_master, dp->dhw.dh_boardset,
7769 	    rem_pfn, rem_pfnlimit, mcadrp);
7770 
7771 	if (rv == 0) {
7772 		DOMAINSET_ADD(idn.domset.ds_hwlinked, domid);
7773 	} else {
7774 		if (rem_pfn == idn.smr.locpfn)
7775 			lock_clear(&idn.first_hwlink);
7776 
7777 		if (idn.first_hwmasterid == (short)domid)
7778 			idn.first_hwmasterid = (short)IDN_NIL_DOMID;
7779 
7780 		(void) idnxf_shmem_sub(is_master, dp->dhw.dh_boardset);
7781 	}
7782 
7783 	IDN_GUNLOCK();
7784 
7785 	return (rv);
7786 }
7787 
7788 static int
7789 idn_deprogram_hardware(int domid)
7790 {
7791 	int		rv, is_master;
7792 	idn_domain_t	*dp;
7793 	procname_t	proc = "idn_deprogram_hardware";
7794 
7795 
7796 	dp = &idn_domain[domid];
7797 
7798 	ASSERT(domid != idn.localid);
7799 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
7800 
7801 	/*
7802 	 * Need to take into consideration what boards remote
7803 	 * domain was connected to.  If we don't have a connection to
7804 	 * them ourself, then we better remove them now , otherwise
7805 	 * they'll never be removed (unless we link to them at some point).
7806 	 */
7807 #if 0
7808 	DEBUG_USECDELAY(500000);
7809 #endif /* 0 */
7810 
7811 	IDN_GLOCK_EXCL();
7812 
7813 	if (!DOMAIN_IN_SET(idn.domset.ds_hwlinked, domid)) {
7814 		IDN_GUNLOCK();
7815 		return (0);
7816 	}
7817 
7818 	PR_PROTO("%s:%d: DEprogram hw in domain %d w.r.t remote domain %d\n",
7819 	    proc, domid, idn.localid, domid);
7820 
7821 	/*
7822 	 * It's possible to come through this flow for domains that
7823 	 * have not been programmed, i.e. not in idn.hwlinked_domset,
7824 	 * so don't bother asserting that they might be in there.
7825 	 * This can occur if we lose a domain during the config/syn
7826 	 * sequence.  If this occurs we won't know whether the remote
7827 	 * domain has programmed its hardware or not.  If it has then
7828 	 * it will have to go through the DMAP sequence and thus we
7829 	 * have to go through it also.  So, if we reach at least the
7830 	 * CONFIG state, we need to go through the DMAP handshake.
7831 	 */
7832 
7833 	PR_PROTO("%s:%d: SUB bset (0x%x)\n", proc, domid, dp->dhw.dh_boardset);
7834 
7835 	if (idn.first_hwmasterid == (short)domid) {
7836 		is_master = 1;
7837 		idn.first_hwmasterid = (short)IDN_NIL_DOMID;
7838 	} else {
7839 		is_master = 0;
7840 	}
7841 	rv = idnxf_shmem_sub(is_master, dp->dhw.dh_boardset);
7842 
7843 	if (rv == 0)
7844 		DOMAINSET_DEL(idn.domset.ds_hwlinked, domid);
7845 
7846 	IDN_GUNLOCK();
7847 
7848 	return (rv);
7849 }
7850 
7851 /*
7852  * Remember can't send slabs back to master at this point.
7853  * Entered with write-drwlock held.
7854  * Returns with drwlock dropped.
7855  */
7856 static void
7857 idn_deconfig(int domid)
7858 {
7859 	idn_domain_t	*dp, *ldp;
7860 	smr_slab_t	*sp;
7861 	int		c, masterid;
7862 	procname_t	proc = "idn_deconfig";
7863 
7864 	ASSERT(IDN_SYNC_IS_LOCKED());
7865 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
7866 	ASSERT(domid != idn.localid);
7867 
7868 	ldp = &idn_domain[idn.localid];
7869 	dp = &idn_domain[domid];
7870 
7871 	ASSERT(dp->dstate == IDNDS_DMAP);
7872 
7873 	PR_PROTO("%s:%d: (dio=%d, dioerr=%d, dnslabs=%d)\n",
7874 	    proc, domid, dp->dio, dp->dioerr, dp->dnslabs);
7875 
7876 	IDN_GLOCK_EXCL();
7877 	masterid = IDN_GET_MASTERID();
7878 
7879 	idn.dc_boardset &= ~dp->dhw.dh_boardset;
7880 	for (c = 0; c < NCPU; c++) {
7881 		if (CPU_IN_SET(dp->dcpuset, c)) {
7882 			CPUSET_DEL(idn.dc_cpuset, c);
7883 		}
7884 	}
7885 
7886 	IDN_GUNLOCK();
7887 
7888 	(void) smr_buf_free_all(domid);
7889 
7890 	if (idn.localid == masterid) {
7891 		/*
7892 		 * Since I'm the master there may
7893 		 * have been slabs in this domain's
7894 		 * idn_domain[] entry.
7895 		 */
7896 		DSLAB_LOCK_EXCL(domid);
7897 		if ((sp = dp->dslab) != NULL) {
7898 			PR_PROTO("%s:%d: freeing up %d dead slabs\n",
7899 			    proc, domid, dp->dnslabs);
7900 			smr_slab_free(domid, sp);
7901 			dp->dslab = NULL;
7902 			dp->dnslabs = 0;
7903 			dp->dslab_state = DSLAB_STATE_UNKNOWN;
7904 		}
7905 		DSLAB_UNLOCK(domid);
7906 	} else if (domid == masterid) {
7907 		/*
7908 		 * We're shutting down the master!
7909 		 * We need to blow away our local slab
7910 		 * data structures.
7911 		 * Since I'm not the master, there should
7912 		 * be no slab structures in the given
7913 		 * domain's idn_domain[] entry.  They should
7914 		 * only exist in the local domain's entry.
7915 		 */
7916 		DSLAB_LOCK_EXCL(idn.localid);
7917 		ASSERT(dp->dslab == NULL);
7918 #ifdef DEBUG
7919 		{
7920 			int	nbusy = 0;
7921 			uint_t	dommask = 0;
7922 			for (sp = ldp->dslab; sp; sp = sp->sl_next) {
7923 				smr_slabbuf_t *bp;
7924 
7925 				if (!smr_slab_busy(sp))
7926 					continue;
7927 				nbusy++;
7928 				for (bp = sp->sl_inuse; bp; bp = bp->sb_next)
7929 					if (bp->sb_domid != IDN_NIL_DOMID)
7930 						DOMAINSET_ADD(dommask,
7931 						    bp->sb_domid);
7932 			}
7933 			if (nbusy)
7934 				PR_PROTO("%s:%d: found %d busy slabs "
7935 				    "(dommask = 0x%x)\n",
7936 				    proc, domid, nbusy, dommask);
7937 		}
7938 #endif /* DEBUG */
7939 		if ((sp = ldp->dslab) != NULL) {
7940 			PR_PROTO("%s:%d: freeing up %d local slab "
7941 			    "structs\n", proc, domid, ldp->dnslabs);
7942 			smr_slab_garbage_collection(sp);
7943 			ldp->dslab = NULL;
7944 			ldp->dnslabs = 0;
7945 			ldp->dslab_state = DSLAB_STATE_UNKNOWN;
7946 		}
7947 		DSLAB_UNLOCK(idn.localid);
7948 	}
7949 	if (dp->dio) {
7950 		PR_PROTO("%s:%d: reset dio (%d) to 0\n", proc, domid, dp->dio);
7951 		dp->dio = 0;
7952 	}
7953 	dp->dioerr = 0;
7954 
7955 	PR_PROTO("%s:%d: reset diocheck (%x) to 0\n",
7956 	    proc, domid, dp->diocheck);
7957 	lock_clear(&dp->diocheck);
7958 
7959 	CHECKPOINT_CLOSED(IDNSB_CHKPT_LINK, dp->dhw.dh_boardset, 2);
7960 
7961 	/*
7962 	 * Should have already flush our memory before
7963 	 * reaching this stage.  The issue is that by the
7964 	 * time we reach here the remote domains may have
7965 	 * already reprogrammed their hardware and so flushing
7966 	 * out caches now could result in a arbstop/hang
7967 	 * if we have data that needs to go back to one
7968 	 * of the remote domains that has already reprogrammed
7969 	 * its hardware.
7970 	 */
7971 	ASSERT(!DOMAIN_IN_SET(idn.domset.ds_flush, domid));
7972 
7973 	(void) idn_deprogram_hardware(domid);
7974 	/*
7975 	 * XXX - what to do if we
7976 	 *	 fail to program hardware
7977 	 *	 probably should panic since
7978 	 *	 demise of system may be near?
7979 	 *	 Sufficient to just shutdown network?
7980 	 */
7981 
7982 	IDN_DSTATE_TRANSITION(dp, IDNDS_CLOSED);
7983 
7984 	idn_close_domain(domid);
7985 }
7986 
7987 /*
7988  * If we're sending a Reset we better make sure we don't have any
7989  * references or traffic headed in the direction of this guy, since
7990  * when he receives the reset, he'll start shutting down which means
7991  * we effectively have to shutdown _before_ sending the reset.
7992  * DO NOT HOLD ANY DOMAIN RWLOCKS ON ENTRY.  Could result in deadlock
7993  * due to channel server looping back through STREAMs and attempting
7994  * to acquire domain lock, i.e. channel server will never "stop".
7995  */
7996 static void
7997 idn_shutdown_datapath(domainset_t domset, int force)
7998 {
7999 	int		do_allchan;
8000 	idn_domain_t	*dp;
8001 	register int	d;
8002 	procname_t	proc = "idn_shutdown_datapath";
8003 
8004 
8005 	PR_CHAN("%s: domset = 0x%x\n", proc, (uint_t)domset);
8006 
8007 	do_allchan = (domset == DOMAINSET_ALL) ? 1 : 0;
8008 
8009 	DOMAINSET_DEL(domset, idn.localid);
8010 
8011 	if (do_allchan) {
8012 		/*
8013 		 * Need to stop all outgoing and
8014 		 * incoming SMR references.
8015 		 */
8016 		idn_deactivate_channel(CHANSET_ALL, IDNCHAN_OFFLINE);
8017 	}
8018 
8019 	/*
8020 	 * If force is set then we don't want to reference
8021 	 * the SMR at all, so deactivate the domains from
8022 	 * channels first.  This will result in the mainmbox-flush
8023 	 * routines to just clean up without referencing the
8024 	 * SMR space.
8025 	 */
8026 	if (force)
8027 		idn_mainmbox_deactivate(domset);
8028 
8029 	/*
8030 	 * Flush out mailboxes (clear smr reference).
8031 	 */
8032 	for (d = 0; d < MAX_DOMAINS; d++) {
8033 		if (!DOMAIN_IN_SET(domset, d))
8034 			continue;
8035 
8036 		dp = &idn_domain[d];
8037 		if ((dp->dmbox.m_send == NULL) && (dp->dmbox.m_recv == NULL))
8038 			continue;
8039 
8040 		IDN_MBOX_LOCK(d);
8041 		if (dp->dmbox.m_send)
8042 			(void) idn_mainmbox_flush(d, dp->dmbox.m_send);
8043 		if (dp->dmbox.m_recv)
8044 			(void) idn_mainmbox_flush(d, dp->dmbox.m_recv);
8045 		IDN_MBOX_UNLOCK(d);
8046 	}
8047 	/*
8048 	 * Deactivate all domain references also.
8049 	 * Only necessary if it wasn't already done above.
8050 	 */
8051 	if (!force)
8052 		idn_mainmbox_deactivate(domset);
8053 }
8054 
8055 void
8056 idn_send_cmd(int domid, idn_cmd_t cmdtype, uint_t arg1, uint_t arg2, uint_t
8057     arg3)
8058 {
8059 	idn_msgtype_t	mt;
8060 	procname_t	proc = "idn_send_cmd";
8061 
8062 	mt.mt_mtype = IDNP_CMD;
8063 	mt.mt_atype = 0;
8064 	mt.mt_cookie = 0;
8065 
8066 	ASSERT(IDN_DLOCK_IS_HELD(domid));
8067 
8068 	PR_PROTO("%s:%d: sending command %s\n", proc, domid,
8069 	    VALID_IDNCMD(cmdtype) ? idncmd_str[cmdtype] : "unknown");
8070 
8071 	IDN_MSGTIMER_START(domid, IDNP_CMD, (ushort_t)cmdtype,
8072 	    idn_msg_waittime[IDNP_CMD], &mt.mt_cookie);
8073 
8074 	IDNXDC(domid, &mt, (uint_t)cmdtype, arg1, arg2, arg3);
8075 }
8076 
8077 void
8078 idn_send_cmdresp(int domid, idn_msgtype_t *mtp, idn_cmd_t cmdtype, uint_t arg1,
8079     uint_t arg2, uint_t cerrno)
8080 {
8081 	idn_msgtype_t	mt;
8082 
8083 	ASSERT(IDN_DLOCK_IS_HELD(domid));
8084 
8085 	if (domid == idn.localid) {
8086 		/*
8087 		 * It's possible local domain received a command
8088 		 * from itself.  However, we cannot send a normal
8089 		 * "ack" response (XDC) to ourself.
8090 		 */
8091 		return;
8092 	}
8093 
8094 	mt.mt_mtype = IDNP_CMD | IDNP_ACK;
8095 	mt.mt_atype = 0;
8096 	mt.mt_cookie = mtp->mt_cookie;
8097 
8098 	IDNXDC(domid, &mt, (uint_t)cmdtype, arg1, arg2, cerrno);
8099 }
8100 
8101 static void
8102 idn_send_cmd_nackresp(int domid, idn_msgtype_t *mtp, idn_cmd_t cmdtype,
8103     idn_nack_t nacktype)
8104 {
8105 	idn_msgtype_t	mt;
8106 
8107 	if (domid == idn.localid)
8108 		return;
8109 
8110 	mt.mt_mtype = IDNP_CMD | IDNP_NACK;
8111 	mt.mt_atype = 0;
8112 	mt.mt_cookie = mtp->mt_cookie;
8113 
8114 	(void) IDNXDC(domid, &mt, (uint_t)cmdtype, (uint_t)nacktype, 0, 0);
8115 }
8116 
8117 void
8118 idn_broadcast_cmd(idn_cmd_t cmdtype, uint_t arg1, uint_t arg2, uint_t arg3)
8119 {
8120 	idn_msgtype_t	mt;
8121 	domainset_t	domset;
8122 	procname_t	proc = "idn_broadcast_cmd";
8123 
8124 	IDN_GLOCK_SHARED();
8125 
8126 	domset = idn.domset.ds_connected;
8127 	DOMAINSET_DEL(domset, idn.localid);
8128 
8129 	PR_PROTO("%s: broadcasting command (%s) to domainset 0x%x\n",
8130 	    proc, VALID_IDNCMD(cmdtype) ? idncmd_str[cmdtype] : "unknown",
8131 	    domset);
8132 
8133 	mt.mt_mtype = IDNP_CMD;
8134 	mt.mt_atype = 0;
8135 	mt.mt_cookie = 0;
8136 
8137 	IDNXDC_BROADCAST(domset, &mt, (uint_t)cmdtype, arg1, arg2, arg3);
8138 
8139 	IDN_GUNLOCK();
8140 	/*
8141 	 * This is a broadcast which means local domain needs
8142 	 * to process it also.  Since we can't XDC to ourselves
8143 	 * we simply call a local function.
8144 	 */
8145 	idn_local_cmd(cmdtype, arg1, arg2, arg3);
8146 }
8147 
8148 /*
8149  * Since xargs[0] contains the cmdtype, only xargs[1], xargs[2], xargs[3]
8150  * are valid possible response arguments.
8151  */
8152 static void
8153 idn_recv_cmd(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
8154 {
8155 	uint_t			msg = mtp->mt_mtype;
8156 	register idn_domain_t	*dp;
8157 	idn_cmd_t		cmdtype;
8158 	uint_t			acknack;
8159 	uint_t			cmdarg1, cmdarg2, cmdarg3;
8160 	int			islocal;
8161 	int			unsup_cmd_sent, unsup_cmd_recvd;
8162 	procname_t		proc = "idn_recv_cmd";
8163 
8164 	acknack = msg & IDNP_ACKNACK_MASK;
8165 	GET_XARGS(xargs, &cmdtype, &cmdarg1, &cmdarg2, &cmdarg3);
8166 
8167 	dp = &idn_domain[domid];
8168 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
8169 
8170 	IDN_GLOCK_SHARED();
8171 
8172 	islocal = (domid == idn.localid);
8173 
8174 	ASSERT(!acknack || (acknack & IDNP_ACKNACK_MASK));
8175 
8176 	PR_PROTO("%s:%d: (local=%d) acknack=0x%x, cmdtype=%s(%d), "
8177 	    "a1=0x%x, a2=0x%x, a3=0x%x\n",
8178 	    proc, domid, islocal, acknack,
8179 	    VALID_IDNCMD(cmdtype) ? idncmd_str[cmdtype] : "unknown",
8180 	    cmdtype, cmdarg1, cmdarg2, cmdarg3);
8181 
8182 	unsup_cmd_sent = unsup_cmd_recvd = 0;
8183 
8184 	if ((IDN_GET_MASTERID() == IDN_NIL_DOMID) ||
8185 	    (dp->dstate != IDNDS_CONNECTED)) {
8186 		/*
8187 		 * Commands cannot be handled without a valid
8188 		 * master.  If this is a request then nack him.
8189 		 */
8190 		PR_PROTO("%s:%d: cannot process CMD w/o master (%d, %s)\n",
8191 		    proc, domid, IDN_GET_MASTERID(),
8192 		    idnds_str[dp->dstate]);
8193 
8194 		if (!islocal && !(acknack & IDNP_ACKNACK_MASK))
8195 			idn_send_cmd_nackresp(domid, mtp, cmdtype,
8196 			    IDNNACK_NOCONN);
8197 		IDN_GUNLOCK();
8198 		return;
8199 	}
8200 	IDN_GUNLOCK();
8201 
8202 	if (acknack & IDNP_ACKNACK_MASK) {
8203 		idn_nack_t	nack;
8204 		/*
8205 		 * Receiving a cmd+ack or cmd+nack in response to some
8206 		 * earlier command we must have issued.
8207 		 * If the response is a nack, there are two possibilites:
8208 		 *
8209 		 *	1. Remote domain failed to allocate due
8210 		 *	   to limited resources.
8211 		 *
8212 		 *	2. Remote domain does not support this
8213 		 *	   particular command.
8214 		 *
8215 		 * In the case of #2, the argument immediately after
8216 		 * the cmdtype (xargs[1]) will be (-1).
8217 		 */
8218 		nack = (idn_nack_t)cmdarg1;
8219 		if ((acknack & IDNP_NACK) && (nack == IDNNACK_BADCMD))
8220 			unsup_cmd_sent++;
8221 
8222 		if (islocal) {
8223 			/*
8224 			 * Shouldn't be receiving local commands w/acks.
8225 			 */
8226 			cmdtype = (idn_cmd_t)0;
8227 		}
8228 
8229 		switch (cmdtype) {
8230 		case IDNCMD_SLABALLOC:
8231 			idn_recv_slaballoc_resp(domid, cmdarg1, cmdarg2,
8232 			    cmdarg3);
8233 			break;
8234 
8235 		case IDNCMD_SLABFREE:
8236 			idn_recv_slabfree_resp(domid, cmdarg1, cmdarg2,
8237 			    cmdarg3);
8238 			break;
8239 
8240 		case IDNCMD_SLABREAP:
8241 			/*
8242 			 * We only care if successful.
8243 			 */
8244 			if (acknack & IDNP_ACK)
8245 				idn_recv_slabreap_resp(domid, cmdarg1, cmdarg3);
8246 			break;
8247 
8248 		case IDNCMD_NODENAME:
8249 			if ((acknack & IDNP_NACK) == 0) {
8250 				idn_recv_nodename_resp(domid, cmdarg1, cmdarg3);
8251 				break;
8252 			}
8253 			switch (nack) {
8254 			case IDNNACK_NOCONN:
8255 			case IDNNACK_RETRY:
8256 				/*
8257 				 * Remote domain was not quite
8258 				 * ready, try again.
8259 				 */
8260 				PR_PROTO("%s:%d: remote not ready "
8261 				    "for %s - retrying "
8262 				    "[dstate=%s]\n",
8263 				    proc, domid,
8264 				    idncmd_str[IDNCMD_NODENAME],
8265 				    idnds_str[dp->dstate]);
8266 
8267 				if (dp->dstate == IDNDS_CONNECTED)
8268 					(void) timeout(idn_retry_nodename_req,
8269 					    (void *)(uintptr_t)domid, hz);
8270 			default:
8271 				break;
8272 			}
8273 			break;
8274 
8275 		default:
8276 			/*
8277 			 * Unsupported command.
8278 			 */
8279 			unsup_cmd_recvd++;
8280 			break;
8281 		}
8282 		if (unsup_cmd_sent) {
8283 			PR_PROTO("%s:%d: unsupported command "
8284 			    "requested (0x%x)\n",
8285 			    proc, domid, cmdtype);
8286 		}
8287 		if (unsup_cmd_recvd) {
8288 			PR_PROTO("%s:%d: unsupported command "
8289 			    "response (0x%x)\n",
8290 			    proc, domid, cmdtype);
8291 		}
8292 	} else {
8293 		/*
8294 		 * Receiving a regular cmd from a remote domain.
8295 		 */
8296 		switch (cmdtype) {
8297 		case IDNCMD_SLABALLOC:
8298 			idn_recv_slaballoc_req(domid, mtp, cmdarg1);
8299 			break;
8300 
8301 		case IDNCMD_SLABFREE:
8302 			idn_recv_slabfree_req(domid, mtp, cmdarg1, cmdarg2);
8303 			break;
8304 
8305 		case IDNCMD_SLABREAP:
8306 			idn_recv_slabreap_req(domid, mtp, cmdarg1);
8307 			break;
8308 
8309 		case IDNCMD_NODENAME:
8310 			idn_recv_nodename_req(domid, mtp, cmdarg1);
8311 			break;
8312 
8313 		default:
8314 			/*
8315 			 * Unsupported command.
8316 			 */
8317 			unsup_cmd_recvd++;
8318 			break;
8319 		}
8320 		if (!islocal && unsup_cmd_recvd) {
8321 			/*
8322 			 * Received an unsupported IDN command.
8323 			 */
8324 			idn_send_cmd_nackresp(domid, mtp, cmdtype,
8325 			    IDNNACK_BADCMD);
8326 		}
8327 	}
8328 }
8329 
8330 /*
8331  * This is a supporting routine for idn_broadcast_cmd() to
8332  * handle processing of the requested command for the local
8333  * domain.  Currently the only support broadcast command
8334  * supported is reaping.
8335  */
8336 /*ARGSUSED2*/
8337 static void
8338 idn_local_cmd(idn_cmd_t cmdtype, uint_t arg1, uint_t arg2, uint_t arg3)
8339 {
8340 	idn_protojob_t	*jp;
8341 	idn_domain_t	*ldp = &idn_domain[idn.localid];
8342 	procname_t	proc = "idn_local_cmd";
8343 
8344 	PR_PROTO("%s: submitting local command %s on domain %d\n",
8345 	    proc, VALID_IDNCMD(cmdtype) ? idncmd_str[cmdtype] : "unknown",
8346 	    idn.localid);
8347 
8348 
8349 	jp = idn_protojob_alloc(KM_SLEEP);
8350 
8351 	jp->j_msg.m_domid    = ldp->domid;
8352 	jp->j_msg.m_msgtype  = IDNP_CMD;
8353 	jp->j_msg.m_cookie   = ldp->dcookie_recv;
8354 	SET_XARGS(jp->j_msg.m_xargs, cmdtype, arg1, arg2, arg3);
8355 
8356 	idn_protojob_submit(ldp->domid, jp);
8357 }
8358 
8359 /*
8360  * Terminate any outstanding commands that may have
8361  * been targeted for the given domain.  A command is
8362  * designated as outstanding if it has an active timer.
8363  *
8364  * serrno = ECANCELED.
8365  */
8366 static void
8367 idn_terminate_cmd(int domid, int serrno)
8368 {
8369 	idn_domain_t	*dp;
8370 	idn_timer_t	*tplist = NULL, *tp;
8371 	procname_t	proc = "idn_terminate_cmd";
8372 
8373 	dp = &idn_domain[domid];
8374 
8375 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
8376 
8377 	IDN_MSGTIMER_GET(dp, IDNP_CMD, tplist, 0);
8378 	/*
8379 	 * At this point the timers are effectively terminated
8380 	 * since when they're t_onq indication is set false.
8381 	 */
8382 	if (tplist == NULL) {
8383 		PR_PROTO("%s:%d: no outstanding cmds found\n",
8384 		    proc, domid);
8385 		/*
8386 		 * There is a window where we may have caught a
8387 		 * request just prior to issuing the actual
8388 		 * command (SLABALLOC).  We're guaranteed if there
8389 		 * was, then he will have at least registered.
8390 		 * So, if we abort the command now, he'll catch
8391 		 * it before going to sleep.
8392 		 * Drop through.
8393 		 */
8394 	}
8395 	ASSERT(tplist ? (tplist->t_back->t_forw == NULL) : 1);
8396 
8397 	for (tp = tplist; tp; tp = tp->t_forw) {
8398 		ASSERT(tp->t_type == IDNP_CMD);
8399 
8400 		PR_PROTO("%s:%d: found outstanding cmd: %s\n",
8401 		    proc, domid, idncmd_str[tp->t_subtype]);
8402 
8403 		switch (tp->t_subtype) {
8404 		case IDNCMD_SLABALLOC:
8405 			/*
8406 			 * Outstanding slaballoc request may have
8407 			 * slab waiters hanging around.  Need to
8408 			 * tell them to bail out.  The given domain
8409 			 * must be the master if we have an outstanding
8410 			 * command to him.  This also presumes that
8411 			 * if there are any waiters they're only in
8412 			 * the local domain's waiting area (i.e. we're
8413 			 * a slave).
8414 			 */
8415 #ifdef DEBUG
8416 			IDN_GLOCK_SHARED();
8417 			ASSERT(domid == IDN_GET_MASTERID());
8418 			ASSERT(idn.localid != IDN_GET_MASTERID());
8419 			IDN_GUNLOCK();
8420 #endif /* DEBUG */
8421 			(void) smr_slabwaiter_abort(idn.localid, serrno);
8422 			break;
8423 
8424 		case IDNCMD_SLABFREE:
8425 		case IDNCMD_SLABREAP:
8426 		case IDNCMD_NODENAME:
8427 			/*
8428 			 * Nothing really waiting for these operations
8429 			 * so no biggy if we just drop.
8430 			 * Note that NODENAME may have an outstanding
8431 			 * buffer, however that will be reclaimed
8432 			 * when we actually unlink from domain.
8433 			 */
8434 			break;
8435 
8436 		default:
8437 			ASSERT(0);
8438 			break;
8439 		}
8440 	}
8441 	/*
8442 	 * As mentioned before the timers are effectively no-op'd
8443 	 * once they're dequeued, however let's cleanup house and
8444 	 * get rid of the useless entries in the timeout queue.
8445 	 */
8446 	if (tplist) {
8447 		IDN_TIMER_STOPALL(tplist);
8448 	}
8449 
8450 	if (idn_domain[idn.localid].dvote.v.master) {
8451 		/*
8452 		 * I'm the master so it's possible I had
8453 		 * outstanding commands (SLABALLOC) waiting
8454 		 * to be satisfied for the given domain.
8455 		 * Since we're forcing an error it's okay
8456 		 * to continue holding onto the drwlock.
8457 		 */
8458 		PR_PROTO("%s:%d: abort slaballoc waiters\n", proc, domid);
8459 		(void) smr_slabwaiter_abort(domid, serrno);
8460 
8461 	} else if (dp->dvote.v.master) {
8462 		PR_PROTO("%s:%d: abort (local domain) slaballoc waiters\n",
8463 		    proc, domid);
8464 		(void) smr_slabwaiter_abort(idn.localid, serrno);
8465 	}
8466 }
8467 
8468 static void
8469 idn_send_acknack(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
8470 {
8471 	idn_domain_t	*dp = &idn_domain[domid];
8472 	procname_t	proc = "idn_send_acknack";
8473 
8474 	ASSERT(mtp ? (mtp->mt_mtype & IDNP_ACKNACK_MASK) : 1);
8475 	ASSERT(domid != IDN_NIL_DOMID);
8476 
8477 #ifdef DEBUG
8478 	{
8479 		STRING(mstr);
8480 		STRING(astr);
8481 
8482 		INUM2STR(mtp->mt_mtype, mstr);
8483 		INUM2STR(mtp->mt_atype, astr);
8484 
8485 		if (mtp->mt_mtype & IDNP_ACK) {
8486 			PR_PROTO("%s:%d: dstate=%s, msg=(%s/%s), "
8487 			    "a1=0x%x, a2=0x%x, a3=0x%x, a4 = 0x%x\n",
8488 			    proc, domid, idnds_str[dp->dstate],
8489 			    astr, mstr, xargs[0], xargs[1],
8490 			    xargs[2], xargs[3]);
8491 		} else {
8492 			idn_nack_t	nack;
8493 
8494 			nack = GET_XARGS_NACK_TYPE(xargs);
8495 			PR_PROTO("%s:%d: dstate=%s, msg=(%s/%s), "
8496 			    "nack=%s(0x%x)\n",
8497 			    proc, domid, idnds_str[dp->dstate],
8498 			    astr, mstr, idnnack_str[nack],
8499 			    (uint_t)nack);
8500 		}
8501 	}
8502 #endif /* DEBUG */
8503 
8504 	(void) IDNXDC(domid, mtp, xargs[0], xargs[1], xargs[2], xargs[3]);
8505 }
8506 
8507 /*ARGSUSED0*/
8508 static void
8509 idn_prealloc_slab(int nslabs)
8510 {
8511 	register int	s, serrno;
8512 	smr_slab_t	*sp;
8513 	idn_domain_t	*ldp = &idn_domain[idn.localid];
8514 	procname_t	proc = "idn_prealloc_slab";
8515 
8516 	IDN_GLOCK_SHARED();
8517 	DSLAB_LOCK_SHARED(idn.localid);
8518 	if ((idn.state != IDNGS_ONLINE) || (ldp->dnslabs > 0)) {
8519 		/*
8520 		 * Not in the proper state or slab already allocated.
8521 		 */
8522 		DSLAB_UNLOCK(idn.localid);
8523 		IDN_GUNLOCK();
8524 		return;
8525 	}
8526 	IDN_GUNLOCK();
8527 	ASSERT(!ldp->dslab);
8528 
8529 	serrno = 0;
8530 	for (s = 0; (s < nslabs) && ((int)ldp->dnslabs < nslabs); s++) {
8531 		/*
8532 		 * Returns with ldp->drwlock dropped.
8533 		 */
8534 		serrno = smr_slab_alloc(idn.localid, &sp);
8535 		if (serrno != 0) {
8536 			PR_PROTO("%s: FAILED to pre-alloc'd "
8537 			    "slab (serrno = %d)\n", proc, serrno);
8538 			break;
8539 		}
8540 		/*
8541 		 * State may have changed since smr_slab_alloc
8542 		 * temporarily drops drwlock.  Make sure we're
8543 		 * still connected.
8544 		 */
8545 		PR_PROTO("%s: SUCCESSFULLY pre-alloc'd slab\n", proc);
8546 
8547 		if (idn.state != IDNGS_ONLINE) {
8548 			PR_PROTO("%s: Lost connection..leaving\n", proc);
8549 			break;
8550 		}
8551 	}
8552 
8553 	DSLAB_UNLOCK(idn.localid);
8554 }
8555 
8556 /*
8557  * Received a request from a remote domain to
8558  * allocate a slab from the master SMR for him.
8559  * Allocate slab and return the response.
8560  */
8561 static void
8562 idn_recv_slaballoc_req(int domid, idn_msgtype_t *mtp, uint_t slab_size)
8563 {
8564 	register idn_domain_t	*dp;
8565 	procname_t		proc = "idn_recv_slaballoc_req";
8566 
8567 	PR_PROTO("%s: slaballoc req from domain %d (size=0x%x)\n",
8568 	    proc, domid, slab_size);
8569 
8570 	dp = &idn_domain[domid];
8571 
8572 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
8573 
8574 	IDN_GLOCK_SHARED();
8575 
8576 	if (idn.localid != IDN_GET_MASTERID()) {
8577 		IDN_GUNLOCK();
8578 		/*
8579 		 * It's a fatal error if the remote domain thinks
8580 		 * we're the master.
8581 		 */
8582 		idn_send_slaballoc_resp(domid, mtp, 0, 0, EACCES);
8583 
8584 	} else if (dp->dstate != IDNDS_CONNECTED) {
8585 
8586 		IDN_GUNLOCK();
8587 		/*
8588 		 * It's a fatal error if we don't yet have a
8589 		 * connection established with the requestor.
8590 		 */
8591 		idn_send_slaballoc_resp(domid, mtp, 0, 0, ENOLINK);
8592 	} else {
8593 		int		serrno;
8594 		smr_slab_t	*sp;
8595 		smr_offset_t	slab_offset;
8596 
8597 		IDN_GUNLOCK();
8598 		DSLAB_LOCK_SHARED(domid);
8599 		IDN_DUNLOCK(domid);
8600 		/*
8601 		 * We're connected and we're the master.
8602 		 * smr_slab_alloc() returns with dp->drwlock dropped.
8603 		 */
8604 		if ((serrno = smr_slab_alloc(domid, &sp)) == 0) {
8605 			/*
8606 			 * Successfully allocated slab for remote slave.
8607 			 */
8608 			slab_offset = IDN_ADDR2OFFSET(sp->sl_start);
8609 			slab_size   = sp->sl_end - sp->sl_start;
8610 			ASSERT((slab_offset != 0) && (slab_size != 0));
8611 		} else {
8612 			slab_offset = slab_size = 0;
8613 		}
8614 		DSLAB_UNLOCK(domid);
8615 		/*
8616 		 * The drwlock is dropped during smr_slab_alloc.
8617 		 * During that time our connection with the given
8618 		 * domain may have changed.  Better check again.
8619 		 */
8620 		IDN_DLOCK_SHARED(domid);
8621 		if ((dp->dstate != IDNDS_CONNECTED) && !serrno) {
8622 			/*
8623 			 * Connection broke.  Keep the slab here.
8624 			 */
8625 			DSLAB_LOCK_EXCL(domid);
8626 			IDN_DUNLOCK(domid);
8627 			smr_slab_free(domid, sp);
8628 			DSLAB_UNLOCK(domid);
8629 			slab_offset = slab_size = 0;
8630 			serrno = ECANCELED;
8631 			IDN_DLOCK_SHARED(domid);
8632 		}
8633 		/*
8634 		 * Send response.
8635 		 * Note that smr_slab_alloc automatically installs
8636 		 * slab into domains respective idn_domain entry
8637 		 * to be associated with that domain.
8638 		 */
8639 		idn_send_slaballoc_resp(domid, mtp, slab_offset, slab_size,
8640 		    serrno);
8641 	}
8642 }
8643 
8644 static void
8645 idn_send_slaballoc_resp(int domid, idn_msgtype_t *mtp, smr_offset_t slab_offset,
8646     uint_t slab_size, int serrno)
8647 {
8648 	procname_t	proc = "idn_send_slaballoc_resp";
8649 
8650 	PR_PROTO("%s: slaballoc resp to domain %d (off=0x%x, size=0x%x) "
8651 	    "[serrno = %d]\n",
8652 	    proc, domid, slab_offset, slab_size, serrno);
8653 
8654 	idn_send_cmdresp(domid, mtp, IDNCMD_SLABALLOC, slab_offset, slab_size,
8655 	    serrno);
8656 }
8657 
8658 /*
8659  * Received the ack or nack to a previous allocation request
8660  * made by the local domain to the master for a slab.  Need
8661  * to "put" the response into the waiting area for any
8662  * waiters.
8663  */
8664 static void
8665 idn_recv_slaballoc_resp(int domid, smr_offset_t slab_offset, uint_t slab_size,
8666     int serrno)
8667 {
8668 	smr_slab_t		*sp = NULL;
8669 	int			rv;
8670 	procname_t		proc = "idn_recv_slaballoc_resp";
8671 
8672 
8673 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
8674 
8675 	PR_PROTO("%s: slaballoc resp from domain %d (off=0x%x, size=0x%x) "
8676 	    "[serrno = %d]\n",
8677 	    proc, domid, slab_offset, slab_size, serrno);
8678 
8679 	if (!serrno) {
8680 		IDN_GLOCK_SHARED();
8681 		if (domid != IDN_GET_MASTERID()) {
8682 			/*
8683 			 * We should only be receiving responses from
8684 			 * our master.  This is either a bogus message
8685 			 * or an old response.  In either case dump it.
8686 			 */
8687 			PR_PROTO("%s: BOGUS slaballoc resp from domid %d "
8688 			    "(master = %d)\n",
8689 			    proc, domid, IDN_GET_MASTERID());
8690 			serrno = EPROTO;
8691 		}
8692 		IDN_GUNLOCK();
8693 
8694 		if (!serrno &&
8695 		    !VALID_NWROFFSET(slab_offset, IDN_SMR_BUFSIZE)) {
8696 			PR_PROTO("%s: slab offset (0x%x) out of range "
8697 			    "(0-0x%lx)\n",
8698 			    proc, slab_offset, MB2B(IDN_NWR_SIZE));
8699 			serrno = EPROTO;
8700 		} else if (!serrno) {
8701 			sp = GETSTRUCT(smr_slab_t, 1);
8702 			sp->sl_start = IDN_OFFSET2ADDR(slab_offset);
8703 			sp->sl_end   = sp->sl_start + slab_size;
8704 			smr_alloc_buflist(sp);
8705 		}
8706 	}
8707 
8708 	/*
8709 	 * Always "put" slabs back to yourself since you're a slave.
8710 	 * Note that we set the forceflag so that even if there are
8711 	 * no waiters we still install the slab for the domain.
8712 	 */
8713 	if (!serrno) {
8714 		DSLAB_LOCK_EXCL(idn.localid);
8715 	}
8716 	rv = smr_slaballoc_put(idn.localid, sp, 1, serrno);
8717 	if (!serrno) {
8718 		DSLAB_UNLOCK(idn.localid);
8719 	}
8720 
8721 	if (rv < 0) {
8722 		/*
8723 		 * Some kind of error trying to install response.
8724 		 * If there was a valid slab sent to us, we'll
8725 		 * just have to send it back.
8726 		 */
8727 		PR_PROTO("%s: failed to install response in waiting area\n",
8728 		    proc);
8729 		if (slab_size != 0) {
8730 			PR_PROTO("%s: sending slab back to domain %d "
8731 			    "(master = %d)\n",
8732 			    proc, domid, IDN_GET_MASTERID());
8733 			idn_send_cmd(domid, IDNCMD_SLABFREE, slab_offset,
8734 			    slab_size, 0);
8735 		}
8736 		if (sp) {
8737 			smr_free_buflist(sp);
8738 			FREESTRUCT(sp, smr_slab_t, 1);
8739 		}
8740 	}
8741 }
8742 
8743 /*
8744  * Note that slab reaping is effectively performed asynchronously
8745  * since the request will be received a protocol server.
8746  */
8747 static void
8748 idn_recv_slabreap_req(int domid, idn_msgtype_t *mtp, int nslabs)
8749 {
8750 	procname_t	proc = "idn_recv_slabreap_req";
8751 
8752 	PR_PROTO("%s: slab reap request (nslabs = %d)\n", proc, nslabs);
8753 
8754 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
8755 
8756 	IDN_GLOCK_SHARED();
8757 	if (domid != IDN_GET_MASTERID()) {
8758 		/*
8759 		 * Only the master can request that slabs be reaped.
8760 		 */
8761 		IDN_GUNLOCK();
8762 		PR_PROTO("%s: only master can request slab reaping\n", proc);
8763 
8764 		idn_send_cmdresp(domid, mtp, IDNCMD_SLABREAP, 0, 0, EACCES);
8765 
8766 		return;
8767 	}
8768 	IDN_GUNLOCK();
8769 
8770 	if (nslabs != 0) {
8771 		IDN_DUNLOCK(domid);
8772 		smr_slab_reap(idn.localid, &nslabs);
8773 		IDN_DLOCK_SHARED(domid);
8774 	}
8775 
8776 	PR_PROTO("%s: slab reap result (nslabs = %d)\n", proc, nslabs);
8777 
8778 	/*
8779 	 * Go ahead and send the reap response back before we start
8780 	 * free'ing off the individual slabs.
8781 	 */
8782 	idn_send_slabreap_resp(domid, mtp, nslabs, 0);
8783 }
8784 
8785 static void
8786 idn_recv_slabreap_resp(int domid, int nslabs, int serrno)
8787 {
8788 	procname_t	proc = "idn_recv_slabreap_resp";
8789 
8790 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
8791 
8792 	if ((idn.localid != IDN_GET_MASTERID()) || (idn.localid == domid)) {
8793 		PR_PROTO("%s: unexpected slabreap resp received "
8794 		    "(domid = %d)\n", proc, domid);
8795 		ASSERT(0);
8796 		return;
8797 	}
8798 	PR_PROTO("%s: recvd reap response from domain %d for %d slabs "
8799 	    "[serrno = %d]\n", proc, domid, nslabs, serrno);
8800 }
8801 
8802 /*
8803  * Not really necessary to send slabreap response.
8804  * XXX - perhaps useful to master for accounting or
8805  *	 throttling of further reaping?
8806  */
8807 static void
8808 idn_send_slabreap_resp(int domid, idn_msgtype_t *mtp, int nslabs, int serrno)
8809 {
8810 	idn_send_cmdresp(domid, mtp, IDNCMD_SLABREAP, nslabs, 0, serrno);
8811 }
8812 
8813 /*
8814  * Slave -> Master ONLY
8815  * Master never sends slabfree request to itself.
8816  */
8817 static void
8818 idn_recv_slabfree_req(int domid, idn_msgtype_t *mtp, smr_offset_t slab_offset,
8819     uint_t slab_size)
8820 {
8821 	smr_slab_t	*sp;
8822 	int		serrno;
8823 	caddr_t		s_start, s_end;
8824 	procname_t	proc = "idn_recv_slabfree_req";
8825 
8826 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
8827 
8828 	if (domid == IDN_GET_MASTERID()) {
8829 		PR_PROTO("%s: unexpected slabfree req received (domid = %d)\n",
8830 		    proc, domid);
8831 		idn_send_slabfree_resp(domid, mtp, slab_offset, slab_size,
8832 		    EACCES);
8833 		return;
8834 	}
8835 	if (slab_size > IDN_SLAB_SIZE) {
8836 		PR_PROTO("%s: unexpected slab size. exp %d, recvd %d\n",
8837 		    proc, IDN_SLAB_SIZE, slab_size);
8838 		idn_send_slabfree_resp(domid, mtp, slab_offset, slab_size,
8839 		    EINVAL);
8840 		return;
8841 	}
8842 	s_start = IDN_OFFSET2ADDR(slab_offset);
8843 	s_end   = s_start + slab_size;
8844 	/*
8845 	 * Master has received a SLABFREE request (effectively a response
8846 	 * to some earlier SLABREAP request.
8847 	 * Find the slab associated with this slab and free it up.
8848 	 */
8849 	DSLAB_LOCK_EXCL(domid);
8850 	if ((sp = smr_slaballoc_get(domid, s_start, s_end)) != NULL) {
8851 		smr_slab_free(domid, sp);
8852 		serrno = 0;
8853 	} else {
8854 		serrno = EINVAL;
8855 	}
8856 	DSLAB_UNLOCK(domid);
8857 
8858 	idn_send_slabfree_resp(domid, mtp, slab_offset, slab_size, serrno);
8859 }
8860 
8861 /*
8862  * Master -> Slave ONLY
8863  */
8864 static void
8865 idn_recv_slabfree_resp(int domid, uint_t slab_offset, uint_t slab_size, int
8866     serrno)
8867 {
8868 	procname_t	proc = "idn_recv_slabfree_resp";
8869 
8870 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
8871 
8872 	if (domid != IDN_GET_MASTERID()) {
8873 		PR_PROTO("%s: unexpected slabfree resp received (domid = %d)\n",
8874 		    proc, domid);
8875 		ASSERT(0);
8876 		return;
8877 	}
8878 	if (slab_size > IDN_SLAB_SIZE) {
8879 		PR_PROTO("%s: unexpected slab size. exp %d, recvd %d\n",
8880 		    proc, IDN_SLAB_SIZE, slab_size);
8881 		ASSERT(0);
8882 		return;
8883 	}
8884 	PR_PROTO("%s: recvd free resp from dom %d "
8885 	    "- slab (off/size) 0x%x/0x%x [serrno = %d]\n",
8886 	    proc, domid, slab_offset, slab_size, serrno);
8887 }
8888 
8889 static void
8890 idn_send_slabfree_resp(int domid, idn_msgtype_t *mtp, uint_t slab_offset,
8891     uint_t slab_size, int serrno)
8892 {
8893 	idn_send_cmdresp(domid, mtp, IDNCMD_SLABFREE, slab_offset, slab_size,
8894 	    serrno);
8895 }
8896 
8897 static void
8898 idn_retry_nodename_req(void *arg)
8899 {
8900 	int	domid = (int)(uintptr_t)arg;
8901 
8902 	idn_send_nodename_req(domid);
8903 }
8904 
8905 static void
8906 idn_send_nodename_req(int domid)
8907 {
8908 	caddr_t		b_bufp;
8909 	smr_offset_t	bufoffset;
8910 	int		serrno;
8911 	idn_domain_t	*dp = &idn_domain[domid];
8912 	procname_t	proc = "idn_send_nodename_req";
8913 
8914 	/*
8915 	 * Need to drop domain lock across
8916 	 * SMR allocation.
8917 	 */
8918 	serrno = smr_buf_alloc(domid, MAXDNAME+1, &b_bufp);
8919 
8920 	IDN_DLOCK_SHARED(domid);
8921 	if (dp->dstate != IDNDS_CONNECTED) {
8922 		/*
8923 		 * Lost connection.
8924 		 */
8925 		PR_PROTO("%s:%d: connection lost [dstate = %s]\n",
8926 		    proc, domid, idnds_str[dp->dstate]);
8927 		IDN_DUNLOCK(domid);
8928 		if (!serrno)
8929 			(void) smr_buf_free(domid, b_bufp, MAXDNAME+1);
8930 		return;
8931 	}
8932 	if (serrno) {
8933 		/*
8934 		 * Failed to allocate buffer, but still have
8935 		 * connection so keep trying.  We may have queried
8936 		 * the master a little too earlier.
8937 		 */
8938 		PR_PROTO("%s:%d: buffer alloc failed [dstate = %s]\n",
8939 		    proc, domid, idnds_str[dp->dstate]);
8940 		(void) timeout(idn_retry_nodename_req, (void *)(uintptr_t)domid,
8941 		    hz);
8942 		IDN_DUNLOCK(domid);
8943 		return;
8944 	}
8945 
8946 	*b_bufp = (char)MAXDNAME;
8947 	bufoffset = IDN_ADDR2OFFSET(b_bufp);
8948 
8949 	idn_send_cmd(domid, IDNCMD_NODENAME, bufoffset, 0, 0);
8950 	IDN_DUNLOCK(domid);
8951 }
8952 
8953 static void
8954 idn_send_nodename_resp(int domid, idn_msgtype_t *mtp, smr_offset_t bufoffset,
8955     int serrno)
8956 {
8957 	idn_send_cmdresp(domid, mtp, IDNCMD_NODENAME, (uint_t)bufoffset, 0,
8958 	    serrno);
8959 }
8960 
8961 static void
8962 idn_recv_nodename_req(int domid, idn_msgtype_t *mtp, smr_offset_t bufoffset)
8963 {
8964 	caddr_t		b_bufp;
8965 	int		length;
8966 	idn_domain_t	*ldp = &idn_domain[idn.localid];
8967 	procname_t	proc = "idn_recv_nodename_req";
8968 
8969 	IDN_DLOCK_EXCL(idn.localid);
8970 	if (!strlen(ldp->dname)) {
8971 		if (!strlen(utsname.nodename)) {
8972 			/*
8973 			 * Local domain's nodename hasn't been
8974 			 * set yet.
8975 			 */
8976 			IDN_DUNLOCK(idn.localid);
8977 			idn_send_cmd_nackresp(domid, mtp, IDNCMD_NODENAME,
8978 			    IDNNACK_RETRY);
8979 			return;
8980 		}
8981 		(void) strncpy(ldp->dname, utsname.nodename, MAXDNAME - 1);
8982 	}
8983 	IDN_DLOCK_DOWNGRADE(idn.localid);
8984 
8985 	if (!VALID_NWROFFSET(bufoffset, IDN_SMR_BUFSIZE)) {
8986 		PR_PROTO("%s:%d: invalid SMR offset received (0x%x)\n",
8987 		    proc, domid, bufoffset);
8988 		IDN_DUNLOCK(idn.localid);
8989 		idn_send_nodename_resp(domid, mtp, bufoffset, EINVAL);
8990 		return;
8991 	}
8992 
8993 	b_bufp = IDN_OFFSET2ADDR(bufoffset);
8994 	length = (int)(*b_bufp++ & 0xff);
8995 
8996 	if (length < strlen(ldp->dname)) {
8997 		PR_PROTO("%s:%d: buffer not big enough (req %lu, got %d)\n",
8998 		    proc, domid, strlen(ldp->dname), length);
8999 		IDN_DUNLOCK(idn.localid);
9000 		idn_send_nodename_resp(domid, mtp, bufoffset, EINVAL);
9001 		return;
9002 	}
9003 
9004 	(void) strncpy(b_bufp, ldp->dname, MAXDNAME);
9005 	b_bufp[MAXDNAME-1] = 0;
9006 	IDN_DUNLOCK(idn.localid);
9007 
9008 	idn_send_nodename_resp(domid, mtp, bufoffset, 0);
9009 }
9010 
9011 static void
9012 idn_recv_nodename_resp(int domid, smr_offset_t bufoffset, int serrno)
9013 {
9014 	caddr_t		b_bufp;
9015 	idn_domain_t	*dp = &idn_domain[domid];
9016 	procname_t	proc = "idn_recv_nodename_resp";
9017 
9018 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
9019 
9020 	if (!VALID_NWROFFSET(bufoffset, IDN_SMR_BUFSIZE)) {
9021 		PR_PROTO("%s:%d: invalid SMR offset received (0x%x)\n",
9022 		    proc, domid, bufoffset);
9023 		return;
9024 	}
9025 
9026 	if (serrno == 0) {
9027 		b_bufp = IDN_OFFSET2ADDR(bufoffset) + 1;
9028 		b_bufp[MAXDNAME-1] = 0;
9029 
9030 		if (strlen(b_bufp) > 0) {
9031 			(void) strncpy(dp->dname, b_bufp, MAXDNAME);
9032 			PR_PROTO("%s:%d: received nodename(%s)\n",
9033 			    proc, domid, dp->dname);
9034 		}
9035 	}
9036 
9037 	(void) smr_buf_free(domid, b_bufp - 1, MAXDNAME + 1);
9038 }
9039 
9040 /*
9041  * The master allocations the SMR management structures.
9042  */
9043 static int
9044 idn_master_init()
9045 {
9046 	idn_domain_t	*ldp = &idn_domain[idn.localid];
9047 	size_t		reserved_size = 0;
9048 	caddr_t		reserved_area = NULL;
9049 	procname_t	proc = "idn_master_init";
9050 
9051 	ASSERT(IDN_GLOCK_IS_EXCL());
9052 	ASSERT(IDN_DLOCK_IS_EXCL(idn.localid));
9053 
9054 	if (idn.mboxarea != NULL) {
9055 		PR_PROTO("%s: master data already initialized\n", proc);
9056 		return (0);
9057 	}
9058 
9059 	PR_PROTO("%s: initializing master data (domid = %d)\n",
9060 	    proc, idn.localid);
9061 
9062 	/*
9063 	 * Reserve an area of the SMR for mailbox usage.
9064 	 * This area is allocated to other domains via
9065 	 * the master.  Round it up to IDN_SMR_BUFSIZE multiple.
9066 	 */
9067 	reserved_size = IDNROUNDUP(IDN_MBOXAREA_SIZE, IDN_SMR_BUFSIZE);
9068 
9069 	PR_PROTO("%s: reserving %lu bytes for mailbox area\n",
9070 	    proc, reserved_size);
9071 
9072 #ifdef DEBUG
9073 	if (reserved_size > (size_t)IDN_SLAB_SIZE) {
9074 		PR_PROTO("%s: WARNING mbox area (%ld) > slab size (%d)\n",
9075 		    proc, reserved_size, IDN_SLAB_SIZE);
9076 	}
9077 #endif /* DEBUG */
9078 	/*
9079 	 * Initialize the pool of slabs and SMR I/O buffers.
9080 	 */
9081 	if (smr_slabpool_init(reserved_size, &reserved_area) != 0) {
9082 		idn_master_deinit();
9083 		return (-1);
9084 	}
9085 
9086 	ASSERT(idn.mboxarea == NULL);
9087 	ASSERT(reserved_area);
9088 
9089 	bzero(reserved_area, reserved_size);
9090 
9091 	idn.mboxarea = (idn_mboxtbl_t *)reserved_area;
9092 	ldp->dmbox.m_tbl = IDN_MBOXAREA_BASE(idn.mboxarea, idn.localid);
9093 	/*
9094 	 * Initialize the SMR pointers in the entire
9095 	 * mailbox table.
9096 	 */
9097 	idn_mboxarea_init(idn.mboxarea, IDN_MBOXAREA_SIZE / IDN_MBOXTBL_SIZE);
9098 
9099 	return (0);
9100 }
9101 
9102 static void
9103 idn_master_deinit()
9104 {
9105 	idn_domain_t	*ldp;
9106 	smr_slab_t	*sp;
9107 	procname_t	proc = "idn_master_deinit";
9108 
9109 	ASSERT(IDN_GLOCK_IS_EXCL());
9110 	ASSERT(IDN_DLOCK_IS_EXCL(idn.localid));
9111 
9112 	if (idn.mboxarea == NULL) {
9113 		PR_PROTO("%s: master data already deinitialized\n", proc);
9114 		return;
9115 	}
9116 
9117 	ldp = &idn_domain[idn.localid];
9118 
9119 	PR_PROTO("%s: deinitializing master data (domid = %d)\n",
9120 	    proc, idn.localid);
9121 
9122 	ldp->dmbox.m_tbl = NULL;
9123 	idn.mboxarea = NULL;
9124 	/*
9125 	 * Master may still be holding onto slabs of his own.
9126 	 */
9127 	DSLAB_LOCK_EXCL(idn.localid);
9128 	sp = ldp->dslab;
9129 	ldp->dslab = NULL;
9130 	ldp->dnslabs = 0;
9131 	if (sp)
9132 		smr_slab_free(idn.localid, sp);
9133 	ldp->dslab_state = DSLAB_STATE_UNKNOWN;
9134 	DSLAB_UNLOCK(idn.localid);
9135 
9136 	smr_slabpool_deinit();
9137 }
9138 
9139 static int
9140 idn_mark_awol(int domid, clock_t *atime)
9141 {
9142 	clock_t		awol;
9143 	idn_domain_t	*dp = &idn_domain[domid];
9144 
9145 	ASSERT(IDN_SYNC_IS_LOCKED());
9146 	ASSERT(IDN_GLOCK_IS_EXCL());
9147 
9148 	if (!DOMAIN_IN_SET(idn.domset.ds_awol, domid)) {
9149 		DOMAINSET_ADD(idn.domset.ds_awol, domid);
9150 		idn.nawols++;
9151 	}
9152 	awol = ddi_get_lbolt();
9153 	if (dp->dawol.a_count++ == 0)
9154 		dp->dawol.a_time = awol;
9155 	dp->dawol.a_last = awol;
9156 	if ((awol - dp->dawol.a_msg) >= (clock_t)(idn_awolmsg_interval * hz))
9157 		dp->dawol.a_msg = awol;
9158 	else
9159 		awol = 0;
9160 
9161 	*atime = awol;
9162 
9163 	idn_awol_event_set(dp->dhw.dh_boardset);
9164 
9165 	return (dp->dawol.a_count);
9166 }
9167 
9168 void
9169 idn_clear_awol(int domid)
9170 {
9171 	idn_domain_t	*dp = &idn_domain[domid];
9172 
9173 	ASSERT(IDN_SYNC_IS_LOCKED());
9174 	ASSERT(IDN_GLOCK_IS_EXCL());
9175 	if (DOMAIN_IN_SET(idn.domset.ds_awol, domid)) {
9176 		DOMAINSET_DEL(idn.domset.ds_awol, domid);
9177 		idn.nawols--;
9178 	}
9179 	if (dp->dawol.a_count > 0) {
9180 		dp->dawol.a_count = 0;
9181 		dp->dawol.a_last = dp->dawol.a_time;
9182 		dp->dawol.a_time = 0;
9183 		dp->dawol.a_msg = 0;
9184 
9185 		idn_awol_event_clear(dp->dhw.dh_boardset);
9186 	}
9187 }
9188 
9189 /*
9190  * A timer expired.
9191  */
9192 void
9193 idn_timer_expired(void *arg)
9194 {
9195 	idn_domain_t	*dp;
9196 	char		*op = "UNKNOWN";
9197 	clock_t		awol = 0;
9198 	int		awolcount, dcpu, domid;
9199 	idn_timer_t	*tp = (idn_timer_t *)arg;
9200 	idn_timerq_t	*tq = NULL;
9201 	uint_t		token;
9202 	char		dname[MAXDNAME];
9203 	procname_t	proc = "idn_timer_expired";
9204 	STRING(str);
9205 
9206 	tq = tp->t_q;
9207 
9208 	ASSERT(tp->t_domid != IDN_NIL_DOMID);
9209 
9210 	IDN_TIMERQ_LOCK(tq);
9211 
9212 	INUM2STR(tp->t_type, str);
9213 
9214 	if (tp->t_onq == 0) {
9215 		PR_TIMER("%s: timer CAUGHT TERMINATION (type = %s)\n",
9216 		    proc, str);
9217 		/*
9218 		 * Timer was dequeued.  Somebody is trying
9219 		 * to shut it down.
9220 		 */
9221 		IDN_TIMERQ_UNLOCK(tq);
9222 		return;
9223 	}
9224 
9225 	IDN_TIMER_DEQUEUE(tq, tp);
9226 
9227 	IDN_TIMERQ_UNLOCK(tq);
9228 
9229 	IDN_SYNC_LOCK();
9230 	IDN_DLOCK_EXCL(tp->t_domid);
9231 
9232 	domid = tp->t_domid;
9233 
9234 	dp = &idn_domain[domid];
9235 	(void) strcpy(dname, dp->dname);
9236 	dcpu = dp->dcpu;
9237 
9238 	IDN_TIMER_EXEC(tp);
9239 
9240 #ifdef DEBUG
9241 	PR_TIMER("%s:%d: [%s] timer EXPIRED (C=0x%x, P=0x%llx, X=0x%llx)\n",
9242 	    proc, tp->t_domid, str, tp->t_cookie,
9243 	    tp->t_posttime, tp->t_exectime);
9244 #endif /* DEBUG */
9245 
9246 	/*
9247 	 * IMPORTANT:
9248 	 * Each case is responsible for dropping SYNC_LOCK & DLOCK.
9249 	 */
9250 	switch (tp->t_type) {
9251 	case IDNP_DATA:
9252 		IDN_SYNC_UNLOCK();
9253 		/*
9254 		 * Timed out waiting for a data packet response.
9255 		 * We can't close domain since he may just be
9256 		 * temporarily AWOL.
9257 		 * Note that dio and diocheck do not get cleared.
9258 		 * This is taken care of when the domain restarts
9259 		 * or is fatally closed.
9260 		 * We only need a reader lock for this.
9261 		 */
9262 		IDN_DLOCK_DOWNGRADE(domid);
9263 		if (dp->diocheck && dp->dmbox.m_send) {
9264 			(void) idn_reclaim_mboxdata(domid, 0, -1);
9265 			if (dp->dio >= IDN_WINDOW_EMAX) {
9266 				idn_msgtype_t	mt;
9267 				/*
9268 				 * Restart timer for another
9269 				 * go around.
9270 				 */
9271 				IDN_MSGTIMER_START(domid, IDNP_DATA, 0,
9272 				    idn_msg_waittime[IDNP_DATA],
9273 				    &mt.mt_cookie);
9274 			} else {
9275 				lock_clear(&dp->diocheck);
9276 			}
9277 		}
9278 		IDN_DUNLOCK(domid);
9279 		break;
9280 
9281 	case IDNP_NEGO:
9282 		/*
9283 		 * If we're not in a NEGO transition, then
9284 		 * just ignore this timeout.
9285 		 */
9286 		if (dp->dxp == &xphase_nego) {
9287 			uint_t		token;
9288 
9289 			IDN_GLOCK_EXCL();
9290 			op = "CONNECT";
9291 			awolcount = idn_mark_awol(domid, &awol);
9292 			IDN_GUNLOCK();
9293 
9294 			idn_nego_cleanup_check(domid, IDN_NIL_DOMID,
9295 			    IDN_NIL_DCPU);
9296 
9297 			IDN_XSTATE_TRANSITION(dp, IDNXS_PEND);
9298 			token = IDN_RETRY_TOKEN(domid, IDNRETRY_NEGO);
9299 			idn_retry_submit(idn_retry_nego, NULL, token,
9300 			    idn_msg_retrytime[(int)IDNRETRY_NEGO]);
9301 		}
9302 		IDN_DUNLOCK(domid);
9303 		IDN_SYNC_UNLOCK();
9304 		break;
9305 
9306 	case IDNP_CMD:
9307 		/*
9308 		 * Timeouts on commands typically mean that the
9309 		 * the master is not responding.  Furthermore, we
9310 		 * can't FORCE a FIN disconnect since at this stage
9311 		 * we are CONNECTED and thus other domains may
9312 		 * have cache entries that we're sharing with them.
9313 		 * Only choice is to completely disconnect from
9314 		 * IDN and try to reestablish connection.
9315 		 *
9316 		 * However, timeouts attempting to get nodename
9317 		 * are not fatal.  Although we don't want to retry
9318 		 * either since each timeout is a lost buffer to
9319 		 * the remote domain.
9320 		 */
9321 		if (tp->t_subtype == (ushort_t)IDNCMD_NODENAME) {
9322 			PR_PROTO("%s:%d: timedout waiting for nodename\n",
9323 			    proc, domid);
9324 			IDN_DUNLOCK(domid);
9325 			IDN_SYNC_UNLOCK();
9326 			break;
9327 		}
9328 
9329 		IDN_GLOCK_EXCL();
9330 		if (idn.state == IDNGS_ONLINE) {
9331 			domainset_t	domset;
9332 			int		masterid = IDN_GET_MASTERID();
9333 
9334 			IDN_GKSTAT_GLOBAL_EVENT(gk_reconfigs,
9335 			    gk_reconfig_last);
9336 
9337 			PR_PROTO("%s:%d: RECONFIG trying old masterid = %d\n",
9338 			    proc, domid, masterid);
9339 
9340 			IDN_GSTATE_TRANSITION(IDNGS_RECONFIG);
9341 			IDN_SET_NEW_MASTERID(masterid);
9342 			IDN_GUNLOCK();
9343 			IDN_DUNLOCK(domid);
9344 
9345 			domset = idn.domset.ds_trans_on |
9346 			    idn.domset.ds_connected;
9347 
9348 			idn_unlink_domainset(domset, IDNFIN_NORMAL,
9349 			    IDNFIN_ARG_NONE, IDNFIN_OPT_RELINK,	BOARDSET_ALL);
9350 		} else {
9351 			IDN_GUNLOCK();
9352 			IDN_DUNLOCK(domid);
9353 		}
9354 		IDN_SYNC_UNLOCK();
9355 		break;
9356 
9357 	case IDNP_CON:
9358 		if (tp->t_subtype == (ushort_t)IDNCON_QUERY) {
9359 			/*
9360 			 * Timed out sending a CON-query.  This is
9361 			 * non-fatal.  We simply need to retry.
9362 			 */
9363 			IDN_GLOCK_EXCL();
9364 			op = "CONNECT";
9365 			awolcount = idn_mark_awol(domid, &awol);
9366 			IDN_GUNLOCK();
9367 			token = IDN_RETRY_TOKEN(domid, IDNRETRY_CONQ);
9368 			idn_retry_submit(idn_retry_query, NULL, token,
9369 			    idn_msg_retrytime[(int)IDNRETRY_CONQ]);
9370 			IDN_DUNLOCK(domid);
9371 			IDN_SYNC_UNLOCK();
9372 			break;
9373 		}
9374 		/*FALLTHROUGH*/
9375 	case IDNP_CFG:
9376 		/*
9377 		 * Any timeouts here we simply try to disconnect
9378 		 * and reestablish the link.  Since we haven't
9379 		 * reached the connected state w.r.t. this domain
9380 		 * we put his fin state to FORCE-HARD in order
9381 		 * to shoot right through without involving other
9382 		 * domains.  Recall that other domains may have
9383 		 * established connections with the given domain
9384 		 * which means any FIN queries to them will always
9385 		 * return connected to the given domain.  Since
9386 		 * neither the given domain nor the local domain
9387 		 * plan on disconnecting from the IDN the connection
9388 		 * to the other domains will remain thereby preventing
9389 		 * the local FIN from ever completing.  Recall that
9390 		 * a FIN depends on all member domains FIN'ing also.
9391 		 */
9392 		IDN_GLOCK_EXCL();
9393 		op = "CONNECT";
9394 		awolcount = idn_mark_awol(domid, &awol);
9395 		IDN_GUNLOCK();
9396 		DOMAINSET_ADD(idn.domset.ds_relink, domid);
9397 		IDN_HISTORY_LOG(IDNH_RELINK, domid, dp->dstate,
9398 		    idn.domset.ds_relink);
9399 		(void) idn_disconnect(domid, IDNFIN_FORCE_SOFT,
9400 		    IDNFIN_ARG_NONE, IDNFIN_SYNC_NO);
9401 		IDN_DUNLOCK(domid);
9402 		IDN_SYNC_UNLOCK();
9403 		break;
9404 
9405 	case IDNP_FIN:
9406 		/*
9407 		 * Timeouts here simply try to retry.
9408 		 */
9409 		IDN_GLOCK_EXCL();
9410 		op = "DISCONNECT";
9411 		awolcount = idn_mark_awol(domid, &awol);
9412 		IDN_GUNLOCK();
9413 		if (tp->t_subtype == (ushort_t)IDNFIN_QUERY) {
9414 			int		d;
9415 			domainset_t	rdyset;
9416 			/*
9417 			 * Timed out sending a FIN-query.  This is
9418 			 * non-fatal.  We simply need to retry.
9419 			 * If we were doing a forced unlink of any
9420 			 * domains, we don't want this awol guy
9421 			 * to hold us up.  Looks for any forced
9422 			 * unlinks and make them "ready" with
9423 			 * respect to this awol domain.
9424 			 */
9425 			rdyset = 0;
9426 			for (d = 0; d < MAX_DOMAINS; d++) {
9427 				if (FIN_IS_FORCE(idn_domain[d].dfin)) {
9428 					DOMAINSET_ADD(rdyset, d);
9429 				}
9430 			}
9431 			if (rdyset)
9432 				(void) idn_sync_register(domid,
9433 				    IDNSYNC_DISCONNECT,
9434 				    rdyset, IDNSYNC_REG_REG);
9435 
9436 			token = IDN_RETRY_TOKEN(domid, IDNRETRY_FINQ);
9437 			idn_retry_submit(idn_retry_query, NULL, token,
9438 			    idn_msg_retrytime[(int)IDNRETRY_FINQ]);
9439 			IDN_DUNLOCK(domid);
9440 			IDN_SYNC_UNLOCK();
9441 			break;
9442 		}
9443 
9444 		if (dp->dfin == IDNFIN_FORCE_SOFT) {
9445 			IDN_FSTATE_TRANSITION(dp, IDNFIN_FORCE_HARD);
9446 		}
9447 		/*
9448 		 * Anybody that was waiting on this domain and
9449 		 * had a hard-force in action gets this guy for
9450 		 * free in their base ready-set.
9451 		 */
9452 		idn_sync_register_awol(domid);
9453 
9454 		dp->dxp = &xphase_fin;
9455 		IDN_XSTATE_TRANSITION(dp, IDNXS_PEND);
9456 		token = IDN_RETRY_TOKEN(domid, IDNRETRY_FIN);
9457 		idn_retry_submit(idn_retry_fin, NULL, token,
9458 		    idn_msg_retrytime[(int)IDNRETRY_FIN]);
9459 		IDN_DUNLOCK(domid);
9460 		IDN_SYNC_UNLOCK();
9461 		break;
9462 
9463 	default:
9464 
9465 		ASSERT(0);
9466 		IDN_DUNLOCK(domid);
9467 		IDN_SYNC_UNLOCK();
9468 		break;
9469 	}
9470 
9471 	IDN_TIMER_FREE(tp);
9472 
9473 	if (awol) {
9474 		if (strlen(dname) > 0) {
9475 			cmn_err(CE_WARN,
9476 			    "IDN: 236: domain (%s) [ID %d] not "
9477 			    "responding to %s [#%d]",
9478 			    dname, domid, op, awolcount);
9479 		} else {
9480 			cmn_err(CE_WARN,
9481 			    "IDN: 236: domain [ID %d, CPU %d] not "
9482 			    "responding to %s [#%d]",
9483 			    domid, dcpu, op, awolcount);
9484 		}
9485 	}
9486 }
9487 
9488 #if 0
9489 static int
9490 idn_retry_check(uint_t token)
9491 {
9492 	int			i, count = 0;
9493 	int			domid = IDN_RETRY_TOKEN2DOMID(token);
9494 	int			key = IDN_RETRY_TOKEN2TYPE(token);
9495 	idn_retry_job_t		*rp;
9496 	idn_retry_queue_t	*qp;
9497 
9498 	qp = &idn.retryqueue;
9499 
9500 	mutex_enter(&qp->rq_mutex);
9501 
9502 	for (i = 0, rp = qp->rq_jobs; i < qp->rq_count; i++, rp = rp->rj_next)
9503 		if ((domid == IDN_RETRY_TOKEN2DOMID(rp->rj_token)) &&
9504 		    ((key == IDN_RETRY_TYPEALL) || (rp->rj_token == token)))
9505 			count++;
9506 
9507 	mutex_exit(&qp->rq_mutex);
9508 
9509 	return (count);
9510 }
9511 #endif /* 0 */
9512 
9513 static void
9514 idn_retry_execute(void *arg)
9515 {
9516 	idn_retry_job_t		*rp = (idn_retry_job_t *)arg;
9517 	idn_retry_queue_t	*qp;
9518 
9519 	qp = &idn.retryqueue;
9520 
9521 	mutex_enter(&qp->rq_mutex);
9522 	if (rp->rj_onq == 0) {
9523 		/*
9524 		 * Job has already been claimed by
9525 		 * retry termination routine.
9526 		 * Bail out.
9527 		 */
9528 		mutex_exit(&qp->rq_mutex);
9529 		return;
9530 	}
9531 	rp->rj_next->rj_prev = rp->rj_prev;
9532 	rp->rj_prev->rj_next = rp->rj_next;
9533 	if (--(qp->rq_count) == 0)
9534 		qp->rq_jobs = NULL;
9535 	else if (qp->rq_jobs == rp)
9536 		qp->rq_jobs = rp->rj_next;
9537 	mutex_exit(&qp->rq_mutex);
9538 
9539 	(*rp->rj_func)(rp->rj_token, rp->rj_arg);
9540 
9541 	IDNRETRY_FREEJOB(rp);
9542 }
9543 
9544 /*
9545  *
9546  */
9547 static void
9548 idn_retry_submit(void (*func)(uint_t token, void *arg), void *arg, uint_t token,
9549     clock_t ticks)
9550 {
9551 	idn_retry_job_t		*rp, *cp;
9552 	idn_retry_queue_t	*qp;
9553 	int			c;
9554 	procname_t		proc = "idn_retry_submit";
9555 
9556 	if (ticks < 0) {
9557 		PR_PROTO("%s: (token = 0x%x) WARNING ticks = %ld\n",
9558 		    proc, token, ticks);
9559 		return;
9560 	}
9561 	if (ticks == 0)		/* At least one tick to get into background */
9562 		ticks++;
9563 
9564 	PR_PROTO("%s: token = 0x%x\n", proc, token);
9565 
9566 	qp = &idn.retryqueue;
9567 
9568 	mutex_enter(&qp->rq_mutex);
9569 	for (c = 0, cp = qp->rq_jobs; c < qp->rq_count; cp = cp->rj_next, c++) {
9570 		if (cp->rj_token == token) {
9571 			PR_PROTO("%s: token = (%d,0x%x) already present\n",
9572 			    proc, IDN_RETRY_TOKEN2DOMID(token),
9573 			    IDN_RETRY_TOKEN2TYPE(token));
9574 			break;
9575 		}
9576 	}
9577 
9578 	if (c < qp->rq_count) {
9579 		mutex_exit(&qp->rq_mutex);
9580 		return;
9581 	}
9582 
9583 	rp = IDNRETRY_ALLOCJOB();
9584 	rp->rj_func = func;
9585 	rp->rj_arg = arg;
9586 	rp->rj_token = token;
9587 	rp->rj_prev = rp->rj_next = rp;
9588 
9589 	if (qp->rq_jobs == NULL) {
9590 		qp->rq_jobs = rp;
9591 	} else {
9592 		rp->rj_next = qp->rq_jobs;
9593 		rp->rj_prev = qp->rq_jobs->rj_prev;
9594 		rp->rj_next->rj_prev = rp;
9595 		rp->rj_prev->rj_next = rp;
9596 	}
9597 	rp->rj_onq = 1;
9598 	qp->rq_count++;
9599 	rp->rj_id = timeout(idn_retry_execute, (caddr_t)rp, ticks);
9600 	mutex_exit(&qp->rq_mutex);
9601 }
9602 
9603 int
9604 idn_retry_terminate(uint_t token)
9605 {
9606 	int			i, domid;
9607 	uint_t			key, count;
9608 	idn_retry_job_t		*rp, *nrp, *fp;
9609 	idn_retry_queue_t	*qp;
9610 	procname_t		proc = "idn_retry_terminate";
9611 
9612 	key = IDN_RETRY_TOKEN2TYPE(token);
9613 	domid = IDN_RETRY_TOKEN2DOMID(token);
9614 	fp = NULL;
9615 	qp = &idn.retryqueue;
9616 
9617 	mutex_enter(&qp->rq_mutex);
9618 	for (i = count = 0, rp = qp->rq_jobs; i < qp->rq_count; i++) {
9619 		nrp = rp->rj_next;
9620 		if ((domid == IDN_RETRY_TOKEN2DOMID(rp->rj_token)) &&
9621 		    ((key == IDN_RETRY_TYPEALL) ||
9622 		    (rp->rj_token == token))) {
9623 			/*
9624 			 * Turn off onq field as a signal to
9625 			 * the execution routine that this
9626 			 * retry has been terminated.  This
9627 			 * is necessary since we can't untimeout
9628 			 * while holding the rq_mutex otherwise
9629 			 * we'll deadlock with the execution
9630 			 * routine.  We'll untimeout these guys
9631 			 * _after_ we drop rq_mutex.
9632 			 */
9633 			rp->rj_onq = 0;
9634 			rp->rj_next->rj_prev = rp->rj_prev;
9635 			rp->rj_prev->rj_next = rp->rj_next;
9636 			if (qp->rq_jobs == rp)
9637 				qp->rq_jobs = rp->rj_next;
9638 			rp->rj_next = fp;
9639 			fp = rp;
9640 			count++;
9641 		}
9642 		rp = nrp;
9643 	}
9644 
9645 	if ((qp->rq_count -= count) == 0)
9646 		qp->rq_jobs = NULL;
9647 
9648 	mutex_exit(&qp->rq_mutex);
9649 
9650 	PR_PROTO("%s: token = (%d,0x%x), dequeued = %d\n",
9651 	    proc, domid, key, count);
9652 
9653 	for (; fp; fp = nrp) {
9654 		(void) untimeout(fp->rj_id);
9655 
9656 		nrp = fp->rj_next;
9657 		IDNRETRY_FREEJOB(fp);
9658 	}
9659 
9660 	return (count);
9661 }
9662 
9663 /*
9664  * -----------------------------------------------------------------------
9665  * The sole purpose of the idn_protocol_server is to manage the IDN
9666  * protocols between the various domains.  These messages do _not_ go
9667  * through the regular streams queues since they are not dependent on
9668  * any user process or module necessarily having the IDN driver open.
9669  * There may be multiple instances of these servers to enhance performance
9670  * of domain management.  Each server is assigned a idn_protoqueue_t
9671  * from which to obtain the work they need to do.
9672  * -----------------------------------------------------------------------
9673  */
9674 int
9675 idn_protocol_init(int nservers)
9676 {
9677 	int		i;
9678 	idn_protojob_t	*jp;
9679 	register idn_protoqueue_t	*protoq;
9680 
9681 	if (nservers <= 0) {
9682 		cmn_err(CE_WARN,
9683 		    "IDN: 237: invalid number (%d) of protocol servers",
9684 		    nservers);
9685 		return (-1);
9686 	}
9687 
9688 	idn.protocol.p_jobpool = kmem_cache_create("idn_protocol_jobcache",
9689 	    sizeof (idn_protojob_t), 0, NULL, NULL, NULL, NULL, NULL, 0);
9690 	if (idn.protocol.p_jobpool == NULL) {
9691 		cmn_err(CE_WARN,
9692 		    "IDN: 238: kmem_cache_create(jobcache) failed");
9693 		return (-1);
9694 	}
9695 
9696 	/*
9697 	 * Initialize static cache for protojob.
9698 	 */
9699 	mutex_init(&idn_protojob_cache_lock, NULL, MUTEX_DRIVER, NULL);
9700 	jp = &idn_protojob_cache[0];
9701 	for (i = 1; i < IDN_DMV_PENDING_MAX; jp = jp->j_next, i++) {
9702 		jp->j_cache = 1;
9703 		jp->j_next = &idn_protojob_cache[i];
9704 	}
9705 	jp->j_cache = 1;
9706 	jp->j_next = NULL;
9707 	idn_protojob_cache_list = &idn_protojob_cache[0];
9708 
9709 	/*
9710 	 * Init morgue semaphore.
9711 	 */
9712 	sema_init(&idn.protocol.p_morgue, 0, NULL, SEMA_DEFAULT, NULL);
9713 	/*
9714 	 * Alloc server queues.
9715 	 */
9716 	idn.protocol.p_serverq = GETSTRUCT(idn_protoqueue_t, nservers);
9717 
9718 	/*
9719 	 * Init server queues.
9720 	 */
9721 	protoq = idn.protocol.p_serverq;
9722 	for (i = 0; i < nservers; protoq++, i++) {
9723 		mutex_init(&protoq->q_mutex, NULL, MUTEX_DRIVER, NULL);
9724 		cv_init(&protoq->q_cv, NULL, CV_DEFAULT, NULL);
9725 		protoq->q_id	  = i;
9726 		protoq->q_joblist = NULL;
9727 		protoq->q_joblist_tail = NULL;
9728 		protoq->q_die	  = 0;
9729 		protoq->q_morgue  = &idn.protocol.p_morgue;
9730 		/*
9731 		 * Create protocol server thread.
9732 		 */
9733 		protoq->q_threadp = thread_create(NULL, 0,
9734 		    idn_protocol_server, (caddr_t)&i, sizeof (i), &p0,
9735 		    TS_RUN, maxclsyspri);
9736 	}
9737 	/*
9738 	 * The servers are kept in the p_server[] array, however
9739 	 * we'll build a linked list of them to facilitate debugging.
9740 	 */
9741 	protoq = idn.protocol.p_serverq;
9742 	for (i = 0; i < (nservers - 1); protoq++, i++)
9743 		protoq->q_next = (protoq + 1);
9744 	protoq->q_next = NULL;
9745 
9746 	idn.nservers = nservers;
9747 
9748 	return (idn.nservers);
9749 }
9750 
9751 void
9752 idn_protocol_deinit()
9753 {
9754 	register int	i;
9755 	int		nservers;
9756 	register idn_protoqueue_t	*protoq;
9757 
9758 	nservers = idn.nservers;
9759 
9760 	if (nservers <= 0)
9761 		return;
9762 
9763 	/*
9764 	 * Make sure the servers are dead.
9765 	 */
9766 	idn_protocol_server_killall();
9767 	ASSERT(idn.nservers == 0);
9768 	/*
9769 	 * Destroy the mutexes.
9770 	 */
9771 	protoq = idn.protocol.p_serverq;
9772 	for (i = 0; i < nservers; protoq++, i++) {
9773 		mutex_destroy(&protoq->q_mutex);
9774 		cv_destroy(&protoq->q_cv);
9775 	}
9776 	/*
9777 	 * Free up the protoqueue memory.
9778 	 */
9779 	FREESTRUCT(idn.protocol.p_serverq, idn_protoqueue_t, nservers);
9780 	idn.protocol.p_serverq = NULL;
9781 	/*
9782 	 * Destroy the morgue semaphore.
9783 	 */
9784 	sema_destroy(&idn.protocol.p_morgue);
9785 
9786 	if (idn.protocol.p_jobpool) {
9787 		kmem_cache_destroy(idn.protocol.p_jobpool);
9788 		idn.protocol.p_jobpool = NULL;
9789 	}
9790 }
9791 
9792 static void
9793 idn_protocol_server(int *id)
9794 {
9795 	idn_protoqueue_t	*pq;
9796 	idn_protojob_t		*jl;
9797 	register idn_protojob_t	*jp;
9798 	procname_t		proc = "idn_protocol_server";
9799 
9800 	if (id == NULL) {
9801 		PR_PROTO("%s: id == NULL, thread exiting\n", proc);
9802 		return;
9803 	}
9804 	ASSERT((*id >= 0) && (*id < idn_protocol_nservers));
9805 
9806 	pq = &idn.protocol.p_serverq[*id];
9807 
9808 	ASSERT(pq->q_id == *id);
9809 
9810 	PR_PROTO("%s: id %d starting up (pq = 0x%p)\n",
9811 	    proc, pq->q_id, (void *)pq);
9812 
9813 	/*CONSTCOND*/
9814 	while (1) {
9815 		mutex_enter(&pq->q_mutex);
9816 
9817 		while (((jl = pq->q_joblist) == NULL) && !pq->q_die)
9818 			cv_wait(&pq->q_cv, &pq->q_mutex);
9819 
9820 		pq->q_joblist = pq->q_joblist_tail = NULL;
9821 
9822 		if (pq->q_die) {
9823 			/*
9824 			 * We've been killed.  Need to check-in
9825 			 * at the morgue.
9826 			 */
9827 			pq->q_threadp = NULL;
9828 			mutex_exit(&pq->q_mutex);
9829 			PR_PROTO("%s: thread (%d) killed...bye bye\n",
9830 			    proc, pq->q_id);
9831 			for (jp = jl; jp; jp = jl) {
9832 				jl = jp->j_next;
9833 				idn_protojob_free(jp);
9834 			}
9835 			sema_v(pq->q_morgue);
9836 			thread_exit();
9837 			/*NOTREACHED*/
9838 		}
9839 		mutex_exit(&pq->q_mutex);
9840 
9841 		/*
9842 		 * We can process the jobs asynchronously while more are
9843 		 * put on.
9844 		 */
9845 		for (jp = jl; jp; jp = jl) {
9846 			jl = jp->j_next;
9847 			idn_recv_proto(&(jp->j_msg));
9848 			idn_protojob_free(jp);
9849 		}
9850 	}
9851 }
9852 
9853 /*
9854  * Kill off all the protocol servers.
9855  */
9856 static void
9857 idn_protocol_server_killall()
9858 {
9859 	register idn_protoqueue_t	*pq;
9860 	int		i;
9861 	procname_t	proc = "idn_protocol_server_killall";
9862 
9863 	PR_PROTO("%s: killing off %d protocol servers\n",
9864 	    proc, idn.nservers);
9865 
9866 	pq = idn.protocol.p_serverq;
9867 	for (i = 0; i < idn.nservers; pq++, i++) {
9868 		mutex_enter(&pq->q_mutex);
9869 		pq->q_die = 1;
9870 		cv_signal(&pq->q_cv);
9871 		mutex_exit(&pq->q_mutex);
9872 	}
9873 
9874 	while (idn.nservers > 0) {
9875 		sema_p(&idn.protocol.p_morgue);
9876 		idn.nservers--;
9877 	}
9878 }
9879 
9880 idn_protojob_t *
9881 idn_protojob_alloc(int kmflag)
9882 {
9883 	idn_protojob_t	*jp;
9884 
9885 	jp = kmem_cache_alloc(idn.protocol.p_jobpool, kmflag);
9886 	if (jp == NULL) {
9887 		mutex_enter(&idn_protojob_cache_lock);
9888 		if ((jp = idn_protojob_cache_list) != NULL)
9889 			idn_protojob_cache_list = jp->j_next;
9890 		mutex_exit(&idn_protojob_cache_lock);
9891 	} else {
9892 		jp->j_cache = 0;
9893 	}
9894 
9895 	return (jp);
9896 }
9897 
9898 static void
9899 idn_protojob_free(idn_protojob_t *jp)
9900 {
9901 	ASSERT(jp);
9902 
9903 	if (jp->j_cache) {
9904 		mutex_enter(&idn_protojob_cache_lock);
9905 		jp->j_next = idn_protojob_cache_list;
9906 		idn_protojob_cache_list = jp;
9907 		mutex_exit(&idn_protojob_cache_lock);
9908 	} else {
9909 		kmem_cache_free(idn.protocol.p_jobpool, (void *)jp);
9910 	}
9911 }
9912 
9913 void
9914 idn_protojob_submit(int cookie, idn_protojob_t *jp)
9915 {
9916 	idn_protoqueue_t	*pq;
9917 	int			serverid;
9918 	procname_t		proc = "idn_protojob_submit";
9919 	STRING(str);
9920 
9921 	if (jp == NULL)
9922 		return;
9923 
9924 	serverid = IDN_PROTOCOL_SERVER_HASH(cookie);
9925 
9926 	pq = &idn.protocol.p_serverq[serverid];
9927 
9928 	INUM2STR(jp->j_msg.m_msgtype, str);
9929 	PR_PROTO("%s: job (d=%d, m=0x%x, %s) submitted to "
9930 	    "protocol server %d\n", proc, jp->j_msg.m_domid,
9931 	    jp->j_msg.m_msgtype, str, serverid);
9932 
9933 	mutex_enter(&pq->q_mutex);
9934 	/*
9935 	 * Can't submit jobs to dying servers.
9936 	 */
9937 	if (!pq->q_die) {
9938 		if (pq->q_joblist_tail) {
9939 			pq->q_joblist_tail->j_next = jp;
9940 			pq->q_joblist_tail = jp;
9941 		} else {
9942 			pq->q_joblist = pq->q_joblist_tail = jp;
9943 		}
9944 		jp->j_next = NULL;
9945 		cv_signal(&pq->q_cv);
9946 	} else {
9947 		PR_PROTO("%s: protocol server dead.  freeing protojob\n",
9948 		    proc);
9949 		idn_protojob_free(jp);
9950 	}
9951 	mutex_exit(&pq->q_mutex);
9952 }
9953 
9954 static void
9955 idn_mboxarea_init(idn_mboxtbl_t *mtp, register int ntbls)
9956 {
9957 	register int	d;
9958 	caddr_t		state_ptr = NULL, mtbasep = (caddr_t)mtp;
9959 	idn_mboxtbl_t	*amtp;
9960 	procname_t	proc = "idn_mboxarea_init";
9961 
9962 	ASSERT(mtp && (ntbls > 0));
9963 
9964 	PR_PROTO("%s: init mboxtbl (0x%p) ntbls = %d\n",
9965 	    proc, (void *)mtp, ntbls);
9966 
9967 	for (d = 0; d < ntbls; d++) {
9968 		register int	pd, sd;
9969 		register int	ch;
9970 
9971 		mtp->mt_header.mh_svr_active = 0;
9972 		mtp->mt_header.mh_svr_ready = 0;
9973 		/*
9974 		 * Initialize the header of each mbox table
9975 		 * with a cookie for identity.
9976 		 */
9977 		/*
9978 		 * Format: 0xc0c0DSCC
9979 		 *	 D = primary domain
9980 		 *	 S = sub-domain of primary
9981 		 *	CC = channel of sub-domain.
9982 		 */
9983 		pd = (d / MAX_DOMAINS) / IDN_MAX_NETS;
9984 		sd = (d / IDN_MAX_NETS) % MAX_DOMAINS;
9985 		ch = d % IDN_MAX_NETS;
9986 
9987 		/*
9988 		 * We point all sub-domains in the same channel
9989 		 * to the same active sync flag since a single server
9990 		 * services all domains in the same channel.
9991 		 */
9992 		amtp = IDN_MBOXTBL_ABS_PTR(mtbasep, pd, 0, ch);
9993 
9994 		state_ptr = (caddr_t)&amtp->mt_header.mh_svr_active;
9995 		mtp->mt_header.mh_svr_active_ptr = IDN_ADDR2OFFSET(state_ptr);
9996 
9997 		state_ptr = (caddr_t)&amtp->mt_header.mh_svr_ready;
9998 		mtp->mt_header.mh_svr_ready_ptr = IDN_ADDR2OFFSET(state_ptr);
9999 
10000 		mtp->mt_header.mh_cookie = IDN_MAKE_MBOXHDR_COOKIE(pd, sd, ch);
10001 
10002 		mtp->mt_header.mh_cksum = IDN_CKSUM_MBOX(&mtp->mt_header);
10003 
10004 		IDN_MBOXTBL_PTR_INC(mtp);
10005 	}
10006 	/*
10007 	 * Now that the master has initialized the entire mailbox
10008 	 * region the referenced memory may not necessarily be up-to-date
10009 	 * with respect to the actual SMR memory due to caching.
10010 	 * In order to make sure future connecting domains get a
10011 	 * consistent picture of the mailbox region, it's necessary
10012 	 * for the master to flush its caches.
10013 	 */
10014 	PR_PROTO("%s: flushing ecache's of local (master) domain\n", proc);
10015 
10016 	idnxf_flushall_ecache();
10017 }
10018 
10019 idn_mainmbox_t *
10020 idn_mainmbox_init(int domid, int mbx)
10021 {
10022 	idn_mainmbox_t	*mmp;
10023 	int		c;
10024 	idn_mainmbox_t	*cmp;
10025 	procname_t	proc = "idn_mainmbox_init";
10026 
10027 	ASSERT(idn_domain[domid].dcpu != IDN_NIL_DCPU);
10028 	ASSERT(IDN_DLOCK_IS_HELD(domid));
10029 
10030 	PR_PROTO("%s: initializing main %s mailbox for domain %d\n",
10031 	    proc, IDNMBOX_IS_RECV(mbx) ? "RECV" : "SEND", domid);
10032 
10033 	cmp = GETSTRUCT(idn_mainmbox_t, IDN_MAX_NETS);
10034 	for (c = 0; c < IDN_MAX_NETS; c++) {
10035 		mmp = &cmp[c];
10036 		mmp->mm_channel = (short)c;
10037 		mutex_init(&mmp->mm_mutex, NULL, MUTEX_DRIVER, NULL);
10038 		mmp->mm_domid = (short)domid;
10039 		mmp->mm_type = mbx;
10040 	}
10041 	mmp = cmp;
10042 	/*
10043 	 * The actual SMR mailbox (mmp->mm_smr_mboxp) gets setup
10044 	 * when the SMR is setup.
10045 	 */
10046 
10047 	return (mmp);
10048 }
10049 
10050 static void
10051 idn_mainmbox_reset(int domid, idn_mainmbox_t *cmp)
10052 {
10053 	idn_mainmbox_t	*mmp;
10054 	int		c;
10055 	procname_t	proc = "idn_mainmbox_reset";
10056 
10057 	ASSERT(IDN_DLOCK_IS_EXCL(domid));
10058 
10059 	PR_PROTO("%s: reseting main %s mailbox for domain %d\n",
10060 	    proc, IDNMBOX_IS_RECV(cmp->mm_type) ? "RECV" : "SEND", domid);
10061 
10062 	for (c = 0; c < IDN_MAX_NETS; c++) {
10063 		mmp = &cmp[c];
10064 
10065 		mmp->mm_channel = (short)c;
10066 		mmp->mm_domid = (short)domid;
10067 		mmp->mm_count = 0;
10068 		mmp->mm_flags = 0;
10069 		mmp->mm_qiget = mmp->mm_qiput = 0;
10070 		mmp->mm_csp = NULL;
10071 		ASSERT(mmp->mm_type == cmp->mm_type);
10072 	}
10073 }
10074 
10075 void
10076 idn_mainmbox_deinit(int domid, idn_mainmbox_t *mmp)
10077 {
10078 	procname_t	proc = "idn_mainmbox_deinit";
10079 
10080 	ASSERT(IDN_DLOCK_IS_HELD(domid));
10081 
10082 	PR_PROTO("%s: deinitializing main %s mailbox for domain %d\n",
10083 	    proc, IDNMBOX_IS_RECV(mmp->mm_type) ? "RECV" : "SEND", domid);
10084 
10085 	ASSERT(idn_domain_is_registered(domid, -1, NULL) == 0);
10086 
10087 	FREESTRUCT(mmp, idn_mainmbox_t, IDN_MAX_NETS);
10088 }
10089 
10090 static void
10091 idn_mainmbox_activate(int domid)
10092 {
10093 	register int	c;
10094 	idn_domain_t	*dp = &idn_domain[domid];
10095 	procname_t	proc = "idn_mainmbox_activate";
10096 
10097 	ASSERT(IDN_DLOCK_IS_HELD(domid));
10098 
10099 	PR_PROTO("%s:%d: activating main mailbox\n", proc, domid);
10100 
10101 	for (c = 0; c < IDN_MAX_NETS; c++)
10102 		idn_mainmbox_chan_register(domid, &dp->dmbox.m_send[c],
10103 		    &dp->dmbox.m_recv[c], c);
10104 }
10105 
10106 /*
10107  * Called upon disabling the SMR to deactivate all the mailboxes
10108  * so that they no longer reference the SMR that's going away.
10109  *
10110  * stopall - Indicates to stop all channel services, across the board.
10111  */
10112 static void
10113 idn_mainmbox_deactivate(ushort_t domset)
10114 {
10115 	int		svr_count;
10116 	procname_t	proc = "idn_mainmbox_deactivate";
10117 
10118 
10119 	if (domset == 0)
10120 		return;
10121 
10122 	PR_PROTO("%s: %s deactivating main mailboxes for domset 0x%x\n",
10123 	    proc, (domset == (ushort_t)-1) ? "STOP-ALL" : "NORMAL", domset);
10124 
10125 	svr_count = idn_mainmbox_chan_unregister(domset, -1);
10126 
10127 	PR_PROTO("%s: deactivated %d chansvrs (domset 0x%x)\n",
10128 	    proc, svr_count, domset);
10129 }
10130 
10131 static void
10132 idn_mainmbox_chan_register(int domid, idn_mainmbox_t *send_mmp,
10133     idn_mainmbox_t *recv_mmp, int channel)
10134 {
10135 	ASSERT(IDN_DLOCK_IS_HELD(domid));
10136 
10137 	/*
10138 	 * Obtain receive mailbox lock first.
10139 	 */
10140 	mutex_enter(&recv_mmp->mm_mutex);
10141 	mutex_enter(&send_mmp->mm_mutex);
10142 
10143 	ASSERT(recv_mmp->mm_channel == (short)channel);
10144 	ASSERT(send_mmp->mm_channel == (short)channel);
10145 
10146 	recv_mmp->mm_csp = &idn.chan_servers[channel];
10147 	recv_mmp->mm_count = 0;
10148 	recv_mmp->mm_dropped = 0;
10149 	recv_mmp->mm_flags = 0;
10150 
10151 	send_mmp->mm_csp = &idn.chan_servers[channel];
10152 	send_mmp->mm_count = 0;
10153 	send_mmp->mm_dropped = 0;
10154 	send_mmp->mm_flags = 0;
10155 
10156 	mutex_exit(&send_mmp->mm_mutex);
10157 	mutex_exit(&recv_mmp->mm_mutex);
10158 
10159 	/*
10160 	 * We have to add ourselves to the respective
10161 	 * channel server's service table.
10162 	 * Note that the channel may not necessarily be
10163 	 * active at this time.
10164 	 */
10165 	ASSERT(idn.chan_servers);
10166 	/*
10167 	 * Have to get the channel server under
10168 	 * control so we can add ourselves.
10169 	 * Returns w/c_mutex.
10170 	 */
10171 	IDN_CHAN_LOCK_GLOBAL(&idn.chan_servers[channel]);
10172 	/*
10173 	 * Add the following domain (mailbox) for monitoring
10174 	 * by the respective channel server.
10175 	 */
10176 	idn_chan_addmbox(channel, DOMAINSET(domid));
10177 
10178 	IDN_CHAN_UNLOCK_GLOBAL(&idn.chan_servers[channel]);
10179 }
10180 
10181 /*
10182  * Unregister the given domain from the specified channel(s) for monitoring.
10183  */
10184 static int
10185 idn_mainmbox_chan_unregister(ushort_t domset, int channel)
10186 {
10187 	int		c, dd_count;
10188 	int		min_chan, max_chan;
10189 	procname_t	proc = "idn_mainmbox_chan_unregister";
10190 
10191 	PR_CHAN("%s: deactivating main mailboxes (channel %d) "
10192 	    "for domset 0x%x\n", proc, channel, domset);
10193 
10194 	if (channel == -1) {
10195 		min_chan = 0;
10196 		max_chan = IDN_MAX_NETS - 1;
10197 	} else {
10198 		min_chan = max_chan = channel;
10199 	}
10200 	/*
10201 	 * Point all the data dispatchers to the same morgue
10202 	 * so we can kill them all at once.
10203 	 */
10204 	dd_count = 0;
10205 	for (c = min_chan; c <= max_chan; c++) {
10206 
10207 		/*
10208 		 * Have to get the channel server under
10209 		 * control so we can remove ourselves.
10210 		 * Returns w/c_mutex held.
10211 		 */
10212 		IDN_CHAN_LOCK_GLOBAL(&idn.chan_servers[c]);
10213 		/*
10214 		 * Delete the following domain (mailbox) from
10215 		 * monitoring by the respective channel server.
10216 		 */
10217 		idn_chan_delmbox(c, (ushort_t)domset);
10218 
10219 		IDN_CHAN_UNLOCK_GLOBAL(&idn.chan_servers[c]);
10220 		dd_count++;
10221 	}
10222 	PR_CHAN("%s: deactivated %d channel mboxes for domset 0x%x, chan %d\n",
10223 	    proc, dd_count, domset, channel);
10224 	return (dd_count);
10225 }
10226 
10227 /*
10228  * Check if the given domain is registered with the given channel(s).
10229  */
10230 int
10231 idn_domain_is_registered(int domid, int channel, idn_chanset_t *chansetp)
10232 {
10233 	int		regcount;
10234 	int		c, min_chan, max_chan;
10235 	idn_chanset_t	chanset;
10236 	procname_t	proc = "idn_domain_is_registered";
10237 
10238 	CHANSET_ZERO(chanset);
10239 
10240 	if (idn.chan_servers == NULL) {
10241 		PR_CHAN("%s: idn.chan_servers == NULL!!\n", proc);
10242 		return (0);
10243 	}
10244 
10245 	if (channel == -1) {
10246 		min_chan = 0;
10247 		max_chan = IDN_MAX_NETS - 1;
10248 	} else {
10249 		min_chan = max_chan = channel;
10250 	}
10251 
10252 	regcount = 0;
10253 
10254 	for (c = min_chan; c <= max_chan; c++) {
10255 		idn_chansvr_t	*csp;
10256 
10257 		csp = &idn.chan_servers[c];
10258 		IDN_CHAN_LOCK_SEND(csp);
10259 		/*
10260 		 * Don't really need recv side lock since registeration
10261 		 * can't change while we're holding send side.
10262 		 * No need to wait for send side to actually suspend
10263 		 * since all we want to do is prevent the registered
10264 		 * information from changing.
10265 		 */
10266 		if (IDN_CHAN_DOMAIN_IS_REGISTERED(csp, domid)) {
10267 			regcount++;
10268 			CHANSET_ADD(chanset, c);
10269 		}
10270 
10271 		IDN_CHAN_UNLOCK_SEND(csp);
10272 	}
10273 
10274 	PR_CHAN("%s: domid %d mbox reg'd with %d channels [0x%x] (req=%d)\n",
10275 	    proc, domid, regcount, chanset, channel);
10276 
10277 	if (chansetp)
10278 		*chansetp = chanset;
10279 
10280 	return (regcount);
10281 }
10282 
10283 static int
10284 idn_mainmbox_flush(int domid, idn_mainmbox_t *mmp)
10285 {
10286 	register int		qi;
10287 	register idn_mboxmsg_t	*mqp;
10288 	int		total_count = 0;
10289 	int		c, count;
10290 	int		mbox_type;
10291 	char		*mbox_str;
10292 	int		lost_io, total_lost_io = 0;
10293 	idn_chanset_t	chanset;
10294 	procname_t	proc = "idn_mainmbox_flush";
10295 
10296 
10297 	if (mmp == NULL)
10298 		return (0);
10299 
10300 	CHANSET_ZERO(chanset);
10301 
10302 	mbox_type = mmp->mm_type;
10303 	ASSERT((mbox_type == IDNMMBOX_TYPE_SEND) ||
10304 	    (mbox_type == IDNMMBOX_TYPE_RECV));
10305 
10306 	mbox_str = (mbox_type == IDNMMBOX_TYPE_SEND) ? "SEND" : "RECV";
10307 
10308 	/*
10309 	 * Determine which channels this domain is registered
10310 	 * with.  If he's not registered with any, then we
10311 	 * can't touch the SMR.
10312 	 */
10313 	(void) idn_domain_is_registered(domid, -1, &chanset);
10314 
10315 	for (c = 0; c < IDN_MAX_NETS; c++) {
10316 		ushort_t	mbox_csum;
10317 
10318 		if (mmp[c].mm_smr_mboxp == NULL)
10319 			continue;
10320 		mutex_enter(&mmp[c].mm_mutex);
10321 		ASSERT(mmp[c].mm_type == mbox_type);
10322 		if (CHAN_IN_SET(chanset, c) == 0) {
10323 			/*
10324 			 * Domain is no longer registered.
10325 			 * DON'T TOUCH THE SMR - IT'S POISON!
10326 			 */
10327 			if (mmp[c].mm_smr_mboxp) {
10328 				PR_CHAN("%s:%d:%s: domain unregistered "
10329 				    "w/chan %d - DUMPING SMR reference\n",
10330 				    proc, domid, mbox_str, c);
10331 				lost_io = IDN_MMBOXINDEX_DIFF(mmp[c].mm_qiput,
10332 				    mmp[c].mm_qiget);
10333 #ifdef DEBUG
10334 				if (mbox_type == IDNMMBOX_TYPE_RECV) {
10335 					PR_CHAN("%s:%d:%s: blowing away %d "
10336 					    "incoming pkts\n",
10337 					    proc, domid, mbox_str, lost_io);
10338 				} else {
10339 					PR_CHAN("%s:%d:%s: blowing away %d/%d "
10340 					    "outstanding pkts\n",
10341 					    proc, domid, mbox_str, lost_io,
10342 					    idn_domain[domid].dio);
10343 				}
10344 #endif /* DEBUG */
10345 			}
10346 			mmp[c].mm_qiput = mmp[c].mm_qiget = 0;
10347 			mmp[c].mm_smr_mboxp = NULL;
10348 			total_lost_io += lost_io;
10349 		}
10350 		if (mmp[c].mm_smr_mboxp) {
10351 			mbox_csum =
10352 			    IDN_CKSUM_MBOX(&mmp[c].mm_smr_mboxp->mt_header);
10353 			if (!VALID_NWRADDR(mmp[c].mm_smr_mboxp, 4) ||
10354 			    !VALID_MBOXHDR(&mmp[c].mm_smr_mboxp->mt_header,
10355 			    c, mbox_csum)) {
10356 				lost_io = IDN_MMBOXINDEX_DIFF(mmp[c].mm_qiput,
10357 				    mmp[c].mm_qiget);
10358 #ifdef DEBUG
10359 				if (mbox_type == IDNMMBOX_TYPE_RECV) {
10360 					PR_CHAN("%s:%d:%s: bad mbox.  blowing "
10361 					    "away %d incoming pkts\n",
10362 					    proc, domid, mbox_str, lost_io);
10363 				} else {
10364 					PR_CHAN("%s:%d:%s: bad mbox.  blowing "
10365 					    "away %d/%d outstanding pkts\n",
10366 					    proc, domid, mbox_str, lost_io,
10367 					    idn_domain[domid].dio);
10368 				}
10369 #endif /* DEBUG */
10370 				mmp[c].mm_smr_mboxp = NULL;
10371 				mmp[c].mm_qiput = mmp[c].mm_qiget = 0;
10372 				total_lost_io += lost_io;
10373 			}
10374 		}
10375 		if (mmp[c].mm_smr_mboxp == NULL) {
10376 			mutex_exit(&mmp[c].mm_mutex);
10377 			continue;
10378 		}
10379 		mqp = &mmp[c].mm_smr_mboxp->mt_queue[0];
10380 		qi = 0;
10381 		count = 0;
10382 		/*
10383 		 * It's quite possible the remote domain may be accessing
10384 		 * these mailbox entries at the exact same time we're
10385 		 * clearing the owner bit.  That's okay.  All we're trying
10386 		 * to do at this point is to minimize the number of packets
10387 		 * the remote domain might try to process unnecessarily.
10388 		 */
10389 		do {
10390 			if (mqp[qi].ms_owner)
10391 				count++;
10392 			mqp[qi].ms_owner = 0;
10393 			IDN_MMBOXINDEX_INC(qi);
10394 		} while (qi);
10395 
10396 		lost_io = IDN_MMBOXINDEX_DIFF(mmp[c].mm_qiput, mmp[c].mm_qiget);
10397 		total_lost_io += lost_io;
10398 
10399 		mmp[c].mm_qiput = mmp[c].mm_qiget = 0;
10400 		mmp[c].mm_smr_mboxp = NULL;
10401 		mutex_exit(&mmp[c].mm_mutex);
10402 
10403 		total_count += count;
10404 
10405 		PR_CHAN("%s:%d:%s: flushed out %d mbox entries for chan %d\n",
10406 		    proc, domid, mbox_str, count, c);
10407 	}
10408 
10409 	if (total_lost_io && (mbox_type == IDNMMBOX_TYPE_SEND)) {
10410 		int	lost_bufs;
10411 		/*
10412 		 * If we lost all our outstanding I/O.  We could
10413 		 * possible could have slabs now with mistakenly
10414 		 * outstanding I/O buffers.  Need to clean them up.
10415 		 * Clean up of leftovers our self.
10416 		 */
10417 		lost_bufs = smr_buf_free_all(domid);
10418 
10419 		PR_CHAN("%s:%d:%s: flushed %d/%d buffers from slabs\n",
10420 		    proc, domid, mbox_str, lost_bufs, total_lost_io);
10421 	}
10422 
10423 	PR_CHAN("%s:%d:%s: flushed total of %d mailbox entries (lost %d)\n",
10424 	    proc, domid, mbox_str, total_count, total_lost_io);
10425 
10426 	return (total_count);
10427 }
10428 
10429 void
10430 idn_chanserver_bind(int net, int cpuid)
10431 {
10432 	int		ocpuid;
10433 	cpu_t		*cp;
10434 	idn_chansvr_t	*csp;
10435 	kthread_id_t	tp;
10436 	procname_t	proc = "idn_chanserver_bind";
10437 
10438 	csp = &idn.chan_servers[net];
10439 	IDN_CHAN_LOCK_GLOBAL(csp);
10440 
10441 	mutex_enter(&cpu_lock);		/* protect checking cpu_ready_set */
10442 	ocpuid = csp->ch_bound_cpuid;
10443 	cp = cpu_get(cpuid);
10444 	if ((cpuid != -1) && ((cp == NULL) || !cpu_is_online(cp))) {
10445 		mutex_exit(&cpu_lock);
10446 		cmn_err(CE_WARN,
10447 		    "IDN: 239: invalid CPU ID (%d) specified for "
10448 		    "IDN net %d",
10449 		    cpuid, net);
10450 		IDN_CHAN_UNLOCK_GLOBAL(csp);
10451 		return;
10452 	}
10453 	if ((tp = csp->ch_recv_threadp) == NULL) {
10454 		/*
10455 		 * Thread is not yet active.  Set ch_bound_cpuid
10456 		 * so when thread activates it will automatically
10457 		 * bind itself.
10458 		 */
10459 		csp->ch_bound_cpuid = -1;
10460 		csp->ch_bound_cpuid_pending = cpuid;
10461 	} else {
10462 		if (ocpuid != -1) {
10463 			thread_affinity_clear(tp);
10464 			csp->ch_bound_cpuid = -1;
10465 		}
10466 		if (cpuid >= 0) {
10467 			thread_affinity_set(tp, cpuid);
10468 			csp->ch_bound_cpuid = cpuid;
10469 		}
10470 		csp->ch_bound_cpuid_pending = -1;
10471 	}
10472 	mutex_exit(&cpu_lock);
10473 
10474 	PR_CHAN("%s: bound net/channel (%d) from cpuid %d to%scpuid %d\n",
10475 	    proc, net, ocpuid, tp ? " " : " (pending) ", cpuid);
10476 
10477 	IDN_CHAN_UNLOCK_GLOBAL(csp);
10478 }
10479 
10480 #ifdef DEBUG
10481 static idn_mboxhdr_t	*prev_mhp[IDN_MAXMAX_NETS];
10482 #endif /* DEBUG */
10483 /*
10484  * Get access to the respective channel server's synchronization
10485  * header which resides in SMR space.
10486  */
10487 static idn_mboxhdr_t *
10488 idn_chan_server_syncheader(int channel)
10489 {
10490 	idn_domain_t	*ldp = &idn_domain[idn.localid];
10491 	idn_mboxtbl_t	*mtp;
10492 	idn_mboxhdr_t	*mhp;
10493 	ushort_t	mbox_csum;
10494 	procname_t	proc = "idn_chan_server_syncheader";
10495 
10496 	ASSERT(IDN_CHAN_RECV_IS_LOCKED(&idn.chan_servers[channel]));
10497 
10498 	IDN_DLOCK_SHARED(idn.localid);
10499 
10500 	if (ldp->dmbox.m_tbl == NULL) {
10501 		PR_CHAN("%s: local dmbox.m_tbl == NULL\n", proc);
10502 		IDN_DUNLOCK(idn.localid);
10503 		return (NULL);
10504 	}
10505 
10506 	mtp = IDN_MBOXTBL_PTR_CHAN(ldp->dmbox.m_tbl, channel);
10507 	mhp = &mtp->mt_header;
10508 	mbox_csum = IDN_CKSUM_MBOX(&mtp->mt_header);
10509 
10510 #ifdef DEBUG
10511 	if (mhp != prev_mhp[channel]) {
10512 		prev_mhp[channel] = mhp;
10513 		PR_CHAN("%s: chan_server (%d) cookie = 0x%x (exp 0x%x)\n",
10514 		    proc, channel, IDN_GET_MBOXHDR_COOKIE(mhp),
10515 		    IDN_MAKE_MBOXHDR_COOKIE(0, 0, channel));
10516 		PR_CHAN("%s: chan_server (%d) actv_ptr = 0x%x (exp 0x%x)\n",
10517 		    proc, channel, mhp->mh_svr_active_ptr,
10518 		    IDN_ADDR2OFFSET(&mhp->mh_svr_active));
10519 		PR_CHAN("%s: chan_server (%d) ready_ptr = 0x%x (exp 0x%x)\n",
10520 		    proc, channel, mhp->mh_svr_ready_ptr,
10521 		    IDN_ADDR2OFFSET(&mhp->mh_svr_ready));
10522 		PR_CHAN("%s: chan_server (%d) mbox_cksum = 0x%x (exp 0x%x)\n",
10523 		    proc, channel, (int)mhp->mh_cksum, (int)mbox_csum);
10524 	}
10525 #endif /* DEBUG */
10526 
10527 	if ((IDN_ADDR2OFFSET(&mhp->mh_svr_active) !=
10528 	    mhp->mh_svr_active_ptr) ||
10529 	    (IDN_ADDR2OFFSET(&mhp->mh_svr_ready) != mhp->mh_svr_ready_ptr) ||
10530 	    !VALID_MBOXHDR(mhp, channel, mbox_csum)) {
10531 		idn_chansvr_t	*csp;
10532 
10533 		csp = &idn.chan_servers[channel];
10534 		if (IDN_CHANNEL_IS_RECV_CORRUPTED(csp) == 0) {
10535 			IDN_CHANSVC_MARK_RECV_CORRUPTED(csp);
10536 
10537 			cmn_err(CE_WARN,
10538 			    "IDN: 240: (channel %d) SMR CORRUPTED "
10539 			    "- RELINK", channel);
10540 			cmn_err(CE_CONT,
10541 			    "IDN: 240: (channel %d) cookie "
10542 			    "(expected 0x%x, actual 0x%x)\n",
10543 			    channel,
10544 			    IDN_MAKE_MBOXHDR_COOKIE(0, 0, channel),
10545 			    mhp->mh_cookie);
10546 			cmn_err(CE_CONT,
10547 			    "IDN: 240: (channel %d) actv_flg "
10548 			    "(expected 0x%x, actual 0x%x)\n",
10549 			    channel, mhp->mh_svr_active_ptr,
10550 			    IDN_ADDR2OFFSET(&mhp->mh_svr_active));
10551 			cmn_err(CE_CONT,
10552 			    "IDN: 240: (channel %d) ready_flg "
10553 			    "(expected 0x%x, actual 0x%x)\n",
10554 			    channel, mhp->mh_svr_ready_ptr,
10555 			    IDN_ADDR2OFFSET(&mhp->mh_svr_ready));
10556 		}
10557 
10558 		mhp = NULL;
10559 	}
10560 	IDN_DUNLOCK(idn.localid);
10561 
10562 	PR_CHAN("%s: channel(%d) mainhp = 0x%p\n", proc, channel, (void *)mhp);
10563 
10564 	return (mhp);
10565 }
10566 
10567 #define	CHANSVR_SYNC_CACHE(csp, mmp, chan) \
10568 { \
10569 	ASSERT(IDN_CHAN_RECV_IS_LOCKED(csp)); \
10570 	if ((csp)->ch_recv_changed) { \
10571 		register int _d; \
10572 		(csp)->ch_recv_scanset = (csp)->ch_recv_scanset_pending; \
10573 		(csp)->ch_recv_domset = (csp)->ch_recv_domset_pending; \
10574 		for (_d = 0; _d < MAX_DOMAINS; _d++) { \
10575 			if (DOMAIN_IN_SET((csp)->ch_recv_domset, _d)) { \
10576 				(mmp)[_d] = \
10577 				    &idn_domain[_d].dmbox.m_recv[chan]; \
10578 			} else { \
10579 				(mmp)[_d] = NULL; \
10580 			} \
10581 		} \
10582 		(csp)->ch_recv_changed = 0; \
10583 	} \
10584 }
10585 #define	CHANSVR_NEXT_DOMID(csp, i, d) \
10586 { \
10587 	(i) = ((i) + 1) & (MAX_DOMAINS - 1); \
10588 	(d) = (int)(((csp)->ch_recv_scanset >> ((i) << 2)) & 0xf); \
10589 }
10590 #define	CHANSVR_RESET_INDEX(i)	((i) = -1)
10591 
10592 #ifdef DEBUG
10593 static idn_mainmbox_t	*Mmp[IDN_MAXMAX_NETS][MAX_DOMAINS];
10594 #endif /* DEBUG */
10595 
10596 static void
10597 idn_chan_server(idn_chansvr_t **cspp)
10598 {
10599 	idn_mboxhdr_t	*mainhp;
10600 	register idn_chansvr_t		*csp;
10601 	register idn_mboxmsg_t		*mqp;
10602 #ifdef DEBUG
10603 	idn_mainmbox_t			**mmp;
10604 #else
10605 	idn_mainmbox_t			*mmp[MAX_DOMAINS];
10606 #endif /* DEBUG */
10607 	register int	qi;
10608 	struct idn	*sip;
10609 	int		channel;
10610 	int		cpuid;
10611 	int		empty;
10612 	int		tot_pktcount, tot_dropcount;
10613 	register int	index;
10614 	register int	domid;
10615 	register int	idleloops;
10616 	procname_t	proc = "idn_chan_server";
10617 
10618 
10619 #ifdef DEBUG
10620 	mmp = &Mmp[(*cspp)->ch_id][0];
10621 	bzero(mmp, MAX_DOMAINS * sizeof (idn_mainmbox_t *));
10622 #else /* DEBUG */
10623 	bzero(mmp, sizeof (mmp));
10624 #endif /* DEBUG */
10625 
10626 	tot_pktcount = tot_dropcount = 0;
10627 
10628 	ASSERT(cspp && *cspp);
10629 
10630 	csp = *cspp;
10631 	channel = csp->ch_id;
10632 	sip = IDN_INST2SIP(channel);
10633 	ASSERT(sip);
10634 
10635 	PR_CHAN("%s: CHANNEL SERVER (channel %d) GOING ACTIVE...\n",
10636 	    proc, channel);
10637 
10638 	IDN_CHAN_LOCK_RECV(csp);
10639 	IDN_CHAN_RECV_INPROGRESS(csp);
10640 	ASSERT(csp->ch_recv_threadp == curthread);
10641 	mutex_enter(&cpu_lock);
10642 	if ((cpuid = csp->ch_bound_cpuid_pending) != -1) {
10643 		cpu_t	*cp = cpu_get(cpuid);
10644 		/*
10645 		 * We've been requested to bind to
10646 		 * a particular cpu.
10647 		 */
10648 		if ((cp == NULL) || !cpu_is_online(cp)) {
10649 			/*
10650 			 * Cpu seems to have gone away or gone offline
10651 			 * since originally requested.
10652 			 */
10653 			mutex_exit(&cpu_lock);
10654 			cmn_err(CE_WARN,
10655 			    "IDN: 239: invalid CPU ID (%d) specified for "
10656 			    "IDN net %d",
10657 			    cpuid, channel);
10658 		} else {
10659 			csp->ch_bound_cpuid = cpuid;
10660 			affinity_set(csp->ch_bound_cpuid);
10661 			mutex_exit(&cpu_lock);
10662 		}
10663 		csp->ch_bound_cpuid_pending = -1;
10664 	} else {
10665 		mutex_exit(&cpu_lock);
10666 	}
10667 	if (csp->ch_bound_cpuid != -1) {
10668 		PR_CHAN("%s: thread bound to cpuid %d\n",
10669 		    proc, csp->ch_bound_cpuid);
10670 	}
10671 	/*
10672 	 * Only the first (main) mbox header is used for
10673 	 * synchronization with data delivery since there is
10674 	 * only data server for all mailboxes for this
10675 	 * given channel.
10676 	 */
10677 	CHANSVR_SYNC_CACHE(csp, mmp, channel);
10678 
10679 	mainhp = ((csp->ch_recv_domcount > 0) &&
10680 	    IDN_CHANNEL_IS_RECV_ACTIVE(csp))
10681 	    ? idn_chan_server_syncheader(channel) : NULL;
10682 
10683 	if (mainhp && IDN_CHANNEL_IS_RECV_ACTIVE(csp))
10684 		mainhp->mh_svr_active = 1;
10685 
10686 	ASSERT(csp->ch_recv_domcount ?
10687 	    (csp->ch_recv_scanset && csp->ch_recv_domset) : 1);
10688 
10689 	IDN_CHAN_UNLOCK_RECV(csp);
10690 
10691 	empty = 0;
10692 	idleloops = 0;
10693 	CHANSVR_RESET_INDEX(index);
10694 
10695 	/*
10696 	 * ---------------------------------------------
10697 	 */
10698 	/*CONSTCOND*/
10699 	while (1) {
10700 		register int	pktcount;
10701 		register int	dropcount;
10702 		ushort_t		mbox_csum;
10703 		idn_mboxtbl_t	*smr_mboxp;	/* points to SMR space */
10704 		register smr_offset_t	bufoffset;
10705 #ifdef DEBUG
10706 		register smr_pkthdr_t	*hdrp;
10707 		idn_netaddr_t		netaddr;
10708 #endif /* DEBUG */
10709 
10710 		/*
10711 		 * Speed through and find the next available domid.
10712 		 */
10713 		CHANSVR_NEXT_DOMID(csp, index, domid);
10714 
10715 		if (!index) {
10716 			/*
10717 			 * We only check state changes when
10718 			 * we wrap around.  Done for performance.
10719 			 */
10720 			if (!IDN_CHANNEL_IS_RECV_ACTIVE(csp) ||
10721 			    csp->ch_recv.c_checkin ||
10722 			    (idn.state != IDNGS_ONLINE)) {
10723 
10724 				PR_DATA("%s: (channel %d) %s\n",
10725 				    proc, channel,
10726 				    IDN_CHANNEL_IS_DETACHED(csp)
10727 				    ? "DEAD" :
10728 				    IDN_CHANNEL_IS_PENDING(csp)
10729 				    ? "IDLED" :
10730 				    IDN_CHANNEL_IS_ACTIVE(csp)
10731 				    ? "ACTIVE" : "DISABLED");
10732 				goto cc_sleep;
10733 			}
10734 		}
10735 		if (csp->ch_recv.c_checkin)
10736 			goto cc_sleep;
10737 
10738 		if (empty == csp->ch_recv_domcount) {
10739 			empty = 0;
10740 			goto cc_slowdown;
10741 		}
10742 
10743 		ASSERT(mmp[domid] != NULL);
10744 
10745 		mutex_enter(&mmp[domid]->mm_mutex);
10746 		if ((smr_mboxp = mmp[domid]->mm_smr_mboxp) == NULL) {
10747 			/*
10748 			 * Somebody is trying to shut things down.
10749 			 */
10750 			empty++;
10751 			mutex_exit(&mmp[domid]->mm_mutex);
10752 			continue;
10753 		}
10754 		ASSERT(mmp[domid]->mm_channel == (short)channel);
10755 		/*
10756 		 * We don't care if the mm_smr_mboxp is nullified
10757 		 * after this point.  The thread attempting to shut
10758 		 * us down has to formally pause this channel before
10759 		 * anything is official anyway.  So, we can continue
10760 		 * with our local SMR reference until the thread
10761 		 * shutting us down really stops us.
10762 		 *
10763 		 * Need to get the qiget index _before_ we drop the
10764 		 * lock since it might get flushed (idn_mainmbox_flush)
10765 		 * once we drop the mm_mutex.
10766 		 *
10767 		 * We prefer not to hold the mm_mutex across the
10768 		 * idn_recv_mboxdata() call since that may be time-
10769 		 * consuming.
10770 		 */
10771 		qi  = mmp[domid]->mm_qiget;
10772 
10773 		/*
10774 		 * Check the mailbox header if checksum is turned on.
10775 		 */
10776 		mbox_csum = IDN_CKSUM_MBOX(&smr_mboxp->mt_header);
10777 		if (!VALID_MBOXHDR(&smr_mboxp->mt_header, channel, mbox_csum)) {
10778 			IDN_KSTAT_INC(sip, si_mboxcrc);
10779 			IDN_KSTAT_INC(sip, si_ierrors);
10780 			if (!(mmp[domid]->mm_flags & IDNMMBOX_FLAG_CORRUPTED)) {
10781 				cmn_err(CE_WARN,
10782 				    "IDN: 241: [recv] (domain %d, "
10783 				    "channel %d) SMR CORRUPTED - RELINK",
10784 				    domid, channel);
10785 				mmp[domid]->mm_flags |= IDNMMBOX_FLAG_CORRUPTED;
10786 			}
10787 			empty = 0;
10788 			mutex_exit(&mmp[domid]->mm_mutex);
10789 			goto cc_sleep;
10790 		}
10791 		mutex_exit(&mmp[domid]->mm_mutex);
10792 		mqp = &smr_mboxp->mt_queue[0];
10793 
10794 		pktcount = dropcount = 0;
10795 
10796 		if (mqp[qi].ms_owner == 0)
10797 			goto cc_next;
10798 
10799 		bufoffset = IDN_BFRAME2OFFSET(mqp[qi].ms_bframe);
10800 
10801 		if (!VALID_NWROFFSET(bufoffset, IDN_SMR_BUFSIZE)) {
10802 			/* ASSERT(0); */
10803 			mqp[qi].ms_flag |= IDN_MBOXMSG_FLAG_ERR_BADOFFSET;
10804 			mqp[qi].ms_owner = 0;
10805 			IDN_MMBOXINDEX_INC(qi);
10806 			dropcount++;
10807 
10808 			IDN_KSTAT_INC(sip, si_smraddr);
10809 			IDN_KSTAT_INC(sip, si_ierrors);
10810 
10811 		} else {
10812 			PR_DATA("%s: (channel %d) pkt (off 0x%x, "
10813 			    "qiget %d) from domain %d\n",
10814 			    proc, channel, bufoffset, qi, domid);
10815 #ifdef DEBUG
10816 
10817 			hdrp = IDN_BUF2HDR(IDN_OFFSET2ADDR(bufoffset));
10818 			netaddr.netaddr = hdrp->b_netaddr;
10819 			ASSERT(netaddr.net.chan == (ushort_t)channel);
10820 #endif /* DEBUG */
10821 
10822 			if (idn_recv_mboxdata(channel,
10823 			    IDN_OFFSET2ADDR(bufoffset)) < 0) {
10824 				mutex_enter(&mmp[domid]->mm_mutex);
10825 				if (!(mmp[domid]->mm_flags &
10826 				    IDNMMBOX_FLAG_CORRUPTED)) {
10827 					cmn_err(CE_WARN,
10828 					    "IDN: 241: [recv] (domain "
10829 					    "%d, channel %d) SMR "
10830 					    "CORRUPTED - RELINK",
10831 					    domid, channel);
10832 					mmp[domid]->mm_flags |=
10833 					    IDNMMBOX_FLAG_CORRUPTED;
10834 				}
10835 				mutex_exit(&mmp[domid]->mm_mutex);
10836 			}
10837 
10838 			mqp[qi].ms_owner = 0;
10839 			IDN_MMBOXINDEX_INC(qi);
10840 			pktcount++;
10841 		}
10842 
10843 cc_next:
10844 
10845 		mutex_enter(&mmp[domid]->mm_mutex);
10846 		if (mmp[domid]->mm_smr_mboxp) {
10847 			if (dropcount)
10848 				mmp[domid]->mm_dropped += dropcount;
10849 			mmp[domid]->mm_qiget = qi;
10850 			mmp[domid]->mm_count += pktcount;
10851 		}
10852 		mutex_exit(&mmp[domid]->mm_mutex);
10853 
10854 		if (pktcount == 0) {
10855 			empty++;
10856 		} else {
10857 			csp->ch_recv_waittime = IDN_NETSVR_WAIT_MIN;
10858 			empty = 0;
10859 			idleloops = 0;
10860 
10861 			PR_DATA("%s: (channel %d) dom=%d, pktcnt=%d\n",
10862 			    proc, channel, domid, pktcount);
10863 		}
10864 
10865 		continue;
10866 
10867 cc_slowdown:
10868 
10869 #ifdef DEBUG
10870 		if (idleloops == 0) {
10871 			PR_DATA("%s: (channel %d) going SOFT IDLE...\n",
10872 			    proc, channel);
10873 		}
10874 #endif /* DEBUG */
10875 		if (idleloops++ < IDN_NETSVR_SPIN_COUNT) {
10876 			/*
10877 			 * At this level we only busy-wait.
10878 			 * Get back into action.
10879 			 */
10880 			continue;
10881 		}
10882 		idleloops = 0;
10883 
10884 cc_sleep:
10885 
10886 		if (mainhp)
10887 			mainhp->mh_svr_active = 0;
10888 
10889 		IDN_CHAN_LOCK_RECV(csp);
10890 
10891 cc_die:
10892 
10893 		ASSERT(IDN_CHAN_RECV_IS_LOCKED(csp));
10894 
10895 		if (!IDN_CHANNEL_IS_RECV_ACTIVE(csp) &&
10896 		    IDN_CHANNEL_IS_DETACHED(csp)) {
10897 			/*
10898 			 * Time to die...
10899 			 */
10900 			PR_CHAN("%s: (channel %d) serviced %d "
10901 			    "packets, drop = %d\n", proc, channel,
10902 			    tot_pktcount, tot_dropcount);
10903 			PR_CHAN("%s: (channel %d) TERMINATING\n",
10904 			    proc, channel);
10905 			PR_CHAN("%s: (channel %d) ch_morguep = %p\n",
10906 			    proc, channel, (void *)csp->ch_recv_morguep);
10907 
10908 			csp->ch_recv_threadp = NULL;
10909 #ifdef DEBUG
10910 			for (index = 0; index < csp->ch_recv_domcount;
10911 			    index++) {
10912 				if ((int)((csp->ch_recv_scanset >>
10913 				    (index*4)) & 0xf) == domid) {
10914 					PR_DATA("%s: WARNING (channel %d) "
10915 					    "DROPPING domid %d...\n",
10916 					    proc, channel, domid);
10917 				}
10918 			}
10919 #endif /* DEBUG */
10920 			IDN_CHAN_RECV_DONE(csp);
10921 
10922 			sema_v(csp->ch_recv_morguep);
10923 
10924 			IDN_CHAN_UNLOCK_RECV(csp);
10925 
10926 			thread_exit();
10927 			/* not reached */
10928 		}
10929 
10930 		do {
10931 			if (IDN_CHANNEL_IS_DETACHED(csp)) {
10932 				PR_CHAN("%s: (channel %d) going to DIE...\n",
10933 				    proc, channel);
10934 				goto cc_die;
10935 			}
10936 #ifdef DEBUG
10937 			if (IDN_CHANNEL_IS_RECV_ACTIVE(csp) &&
10938 			    (csp->ch_recv_waittime <= IDN_NETSVR_WAIT_MAX)) {
10939 				PR_CHAN("%s: (channel %d) going SOFT IDLE "
10940 				    "(waittime = %d ticks)...\n",
10941 				    proc, channel,
10942 				    csp->ch_recv_waittime);
10943 			} else {
10944 				PR_CHAN("%s: (channel %d) going "
10945 				    "HARD IDLE...\n", proc, channel);
10946 			}
10947 #endif /* DEBUG */
10948 			IDN_CHAN_RECV_DONE(csp);
10949 
10950 			/*
10951 			 * If we're being asked to check-in then
10952 			 * go into a hard sleep.  Want to give the
10953 			 * thread requesting us to checkin a chance.
10954 			 */
10955 			while (csp->ch_recv.c_checkin)
10956 				cv_wait(&csp->ch_recv_cv,
10957 				    &csp->ch_recv.c_mutex);
10958 
10959 			if (csp->ch_recv_waittime > IDN_NETSVR_WAIT_MAX)
10960 				cv_wait(&csp->ch_recv_cv,
10961 				    &csp->ch_recv.c_mutex);
10962 			else
10963 				(void) cv_reltimedwait(&csp->ch_recv_cv,
10964 				    &csp->ch_recv.c_mutex,
10965 				    csp->ch_recv_waittime, TR_CLOCK_TICK);
10966 
10967 			IDN_CHAN_RECV_INPROGRESS(csp);
10968 
10969 			IDN_KSTAT_INC(sip, si_sigsvr);
10970 
10971 			if (csp->ch_recv_waittime <= IDN_NETSVR_WAIT_MAX)
10972 				csp->ch_recv_waittime <<=
10973 				    IDN_NETSVR_WAIT_SHIFT;
10974 
10975 		} while (!IDN_CHANNEL_IS_RECV_ACTIVE(csp));
10976 
10977 		/*
10978 		 * Before we see the world (and touch SMR space),
10979 		 * see if we've been told to die.
10980 		 */
10981 		mainhp = NULL;
10982 		/*
10983 		 * The world may have changed since we were
10984 		 * asleep.  Need to resync cache and check for a
10985 		 * new syncheader.
10986 		 *
10987 		 * Reset chansvr cache against any changes in
10988 		 * mbox fields we need (mm_qiget).
10989 		 */
10990 		CHANSVR_SYNC_CACHE(csp, mmp, channel);
10991 		if (csp->ch_recv_domcount <= 0) {
10992 			/*
10993 			 * Everybody disappeared on us.
10994 			 * Go back to sleep.
10995 			 */
10996 			goto cc_die;
10997 		}
10998 		ASSERT(csp->ch_recv_scanset && csp->ch_recv_domset);
10999 
11000 		mainhp = idn_chan_server_syncheader(channel);
11001 		if (mainhp == NULL) {
11002 			/*
11003 			 * Bummer...we're idling...
11004 			 */
11005 			goto cc_die;
11006 		}
11007 
11008 		mainhp->mh_svr_active = 1;
11009 
11010 		IDN_CHAN_UNLOCK_RECV(csp);
11011 		/*
11012 		 * Reset the domid index after sleeping.
11013 		 */
11014 		CHANSVR_RESET_INDEX(index);
11015 
11016 		empty = 0;
11017 		idleloops = 0;
11018 	}
11019 }
11020 
11021 #if 0
11022 /*
11023  * We maintain a separate function for flushing the STREAMs
11024  * queue of a channel because it must be done outside the
11025  * context of the idn_chan_action routine.  The streams flush
11026  * cannot occur inline with the idn_chan_action because
11027  * the act of flushing may cause IDN send functions to be called
11028  * directly and thus locks to be obtained which could result
11029  * in deadlocks.
11030  */
11031 static void
11032 idn_chan_flush(idn_chansvr_t *csp)
11033 {
11034 	queue_t		*rq;
11035 	struct idn	*sip;
11036 	int		flush_type = 0;
11037 	idn_chaninfo_t	*csend, *crecv;
11038 	procname_t	proc = "idn_chan_flush";
11039 
11040 	csend = &csp->ch_send;
11041 	crecv = &csp->ch_recv;
11042 
11043 	mutex_enter(&crecv->c_mutex);
11044 	mutex_enter(&csend->c_mutex);
11045 
11046 	if (crecv->c_state & IDN_CHANSVC_STATE_FLUSH)
11047 		flush_type |= FLUSHR;
11048 
11049 	if (csend->c_state & IDN_CHANSVC_STATE_FLUSH)
11050 		flush_type |= FLUSHW;
11051 
11052 	if (flush_type) {
11053 		rq = NULL;
11054 		rw_enter(&idn.struprwlock, RW_READER);
11055 		if ((sip = IDN_INST2SIP(csp->ch_id)) != NULL)
11056 			rq = sip->si_ipq;
11057 		rw_exit(&idn.struprwlock);
11058 		if (rq) {
11059 			/*
11060 			 * Flush the STREAM if possible
11061 			 * to get the channel server coherent
11062 			 * enough to respond to us.
11063 			 */
11064 			PR_CHAN("%s: sending FLUSH (%x) to channel %d\n",
11065 			    proc, flush_type, csp->ch_id);
11066 
11067 			(void) putnextctl1(rq, M_FLUSH, flush_type);
11068 		}
11069 		crecv->c_state &= ~IDN_CHANSVC_STATE_FLUSH;
11070 		csend->c_state &= ~IDN_CHANSVC_STATE_FLUSH;
11071 
11072 		if (crecv->c_waiters)
11073 			cv_broadcast(&crecv->c_cv);
11074 	}
11075 
11076 	mutex_exit(&csend->c_mutex);
11077 	mutex_exit(&crecv->c_mutex);
11078 }
11079 #endif /* 0 */
11080 
11081 /*
11082  * Locks are with respect to SEND/RECV locks (c_mutex).
11083  *
11084  * STOP/SUSPEND/DETACH
11085  *	- Entered with locks dropped, leave with locks held.
11086  *	  DETACH - lock dropped manually.
11087  * RESTART/RESUME
11088  *	- Entered with locks held, leave with locks dropped.
11089  * ATTACH
11090  *	- both enter and leave with locks dropped.
11091  */
11092 static void
11093 idn_chan_action(int channel, idn_chanaction_t chanaction, int wait)
11094 {
11095 	uchar_t		clr_state, set_state;
11096 	uint_t		is_running;
11097 	domainset_t	closed_slabwaiters = 0;
11098 	struct idn	*sip;
11099 	idn_chansvr_t	*csp;
11100 	idn_chaninfo_t	*csend, *crecv;
11101 	procname_t	proc = "idn_chan_action";
11102 
11103 	ASSERT((channel >= 0) && (channel < IDN_MAX_NETS));
11104 	ASSERT(idn.chan_servers);
11105 
11106 	csp = &idn.chan_servers[channel];
11107 
11108 	PR_CHAN("%s: requesting %s for channel %d\n",
11109 	    proc, chanaction_str[(int)chanaction], channel);
11110 
11111 	csend = &csp->ch_send;
11112 	crecv = &csp->ch_recv;
11113 
11114 	ASSERT(IDN_CHAN_GLOBAL_IS_LOCKED(csp));
11115 
11116 	clr_state = set_state = 0;
11117 
11118 	switch (chanaction) {
11119 	case IDNCHAN_ACTION_DETACH:
11120 		clr_state = IDN_CHANSVC_STATE_MASK;
11121 		/*FALLTHROUGH*/
11122 
11123 	case IDNCHAN_ACTION_STOP:
11124 		clr_state |= IDN_CHANSVC_STATE_ENABLED;
11125 		/*FALLTHROUGH*/
11126 
11127 	case IDNCHAN_ACTION_SUSPEND:
11128 		clr_state |= IDN_CHANSVC_STATE_ACTIVE;
11129 
11130 		/*
11131 		 * Must maintain this locking order.
11132 		 * Set asynchronous check-in flags.
11133 		 */
11134 		crecv->c_checkin = 1;
11135 		csend->c_checkin = 1;
11136 
11137 		is_running = 0;
11138 		if ((csend->c_inprogress || crecv->c_inprogress) &&
11139 		    wait && (csp->ch_recv_threadp != curthread)) {
11140 
11141 			rw_enter(&idn.struprwlock, RW_READER);
11142 			if ((sip = IDN_INST2SIP(channel)) != NULL) {
11143 				/*
11144 				 * Temporarily turn off the STREAM
11145 				 * to give a chance to breath.
11146 				 */
11147 				is_running = sip->si_flags & IDNRUNNING;
11148 				if (is_running)
11149 					sip->si_flags &= ~IDNRUNNING;
11150 			}
11151 			rw_exit(&idn.struprwlock);
11152 		}
11153 
11154 		mutex_enter(&crecv->c_mutex);
11155 		crecv->c_state &= ~clr_state;
11156 
11157 		mutex_enter(&csend->c_mutex);
11158 		csend->c_state &= ~clr_state;
11159 
11160 		/*
11161 		 * It's possible the channel server could come
11162 		 * through this flow itself due to putting data upstream
11163 		 * that ultimately turned around and came back down for
11164 		 * sending.  If this is the case we certainly don't
11165 		 * want to cv_wait, otherwise we'll obviously deadlock
11166 		 * waiting for ourself.  So, only block if somebody
11167 		 * other than the channel server we're attempting to
11168 		 * suspend/stop.
11169 		 */
11170 		if (wait && (csp->ch_recv_threadp != curthread)) {
11171 			int	do_flush = 0;
11172 
11173 			if (csend->c_inprogress || crecv->c_inprogress)
11174 				do_flush++;
11175 
11176 			if (do_flush) {
11177 				rw_enter(&idn.struprwlock, RW_READER);
11178 				if ((sip = IDN_INST2SIP(channel)) != NULL) {
11179 					/*
11180 					 * Temporarily turn off the STREAM
11181 					 * to give a chance to breath.
11182 					 */
11183 					if (sip->si_flags & IDNRUNNING) {
11184 						is_running = 1;
11185 						sip->si_flags &= ~IDNRUNNING;
11186 					}
11187 				}
11188 				rw_exit(&idn.struprwlock);
11189 			}
11190 
11191 			/*
11192 			 * If we have any senders in-progress
11193 			 * it's possible they're stuck waiting
11194 			 * down in smr_buf_alloc which may never
11195 			 * arrive if we're in an unlink process.
11196 			 * Rather than wait for it to timeout
11197 			 * let's be proactive so we can disconnect
11198 			 * asap.
11199 			 */
11200 			closed_slabwaiters = csp->ch_reg_domset;
11201 			DOMAINSET_ADD(closed_slabwaiters, idn.localid);
11202 			if (closed_slabwaiters)
11203 				smr_slabwaiter_close(closed_slabwaiters);
11204 
11205 			do {
11206 				/*
11207 				 * It's possible due to a STREAMs
11208 				 * loopback from read queue to write queue
11209 				 * that receiver and sender may be same
11210 				 * thread, i.e. receiver's inprogress
11211 				 * flag will never clear until sender's
11212 				 * inprogress flag clears.  So, we wait
11213 				 * for sender's inprogress first.
11214 				 */
11215 				while (csend->c_inprogress) {
11216 					mutex_exit(&crecv->c_mutex);
11217 					while (csend->c_inprogress) {
11218 						csend->c_waiters++;
11219 						cv_wait(&csend->c_cv,
11220 						    &csend->c_mutex);
11221 						csend->c_waiters--;
11222 					}
11223 					/*
11224 					 * Maintain lock ordering.
11225 					 * Eventually we will catch
11226 					 * him due to the flag settings.
11227 					 */
11228 					mutex_exit(&csend->c_mutex);
11229 					mutex_enter(&crecv->c_mutex);
11230 					mutex_enter(&csend->c_mutex);
11231 				}
11232 				if (crecv->c_inprogress) {
11233 					mutex_exit(&csend->c_mutex);
11234 					while (crecv->c_inprogress) {
11235 						crecv->c_waiters++;
11236 						cv_wait(&crecv->c_cv,
11237 						    &crecv->c_mutex);
11238 						crecv->c_waiters--;
11239 					}
11240 					mutex_enter(&csend->c_mutex);
11241 				}
11242 			} while (csend->c_inprogress);
11243 		}
11244 
11245 		if (is_running) {
11246 			/*
11247 			 * Restore the IDNRUNNING bit in
11248 			 * the flags to let them know the
11249 			 * channel is still alive.
11250 			 */
11251 			rw_enter(&idn.struprwlock, RW_READER);
11252 			if ((sip = IDN_INST2SIP(channel)) != NULL)
11253 				sip->si_flags |= IDNRUNNING;
11254 			rw_exit(&idn.struprwlock);
11255 		}
11256 
11257 		if (closed_slabwaiters) {
11258 			/*
11259 			 * We can reopen now since at this point no new
11260 			 * slabwaiters will attempt to come in and wait.
11261 			 */
11262 			smr_slabwaiter_open(csp->ch_reg_domset);
11263 		}
11264 
11265 		crecv->c_checkin = 0;
11266 		csend->c_checkin = 0;
11267 
11268 		/*
11269 		 * ALL leave with locks held.
11270 		 */
11271 		PR_CHAN("%s: action (%s) for channel %d - COMPLETED\n",
11272 		    proc, chanaction_str[(int)chanaction], channel);
11273 		break;
11274 
11275 	case IDNCHAN_ACTION_ATTACH:
11276 		mutex_enter(&crecv->c_mutex);
11277 		mutex_enter(&csend->c_mutex);
11278 		set_state |= csp->ch_state & IDN_CHANSVC_STATE_ATTACHED;
11279 		/*FALLTHROUGH*/
11280 
11281 	case IDNCHAN_ACTION_RESTART:
11282 		set_state |= csp->ch_state & IDN_CHANSVC_STATE_ENABLED;
11283 		/*FALLTHROUGH*/
11284 
11285 	case IDNCHAN_ACTION_RESUME:
11286 		ASSERT(IDN_CHAN_LOCAL_IS_LOCKED(csp));
11287 		set_state |= csp->ch_state & IDN_CHANSVC_STATE_ACTIVE;
11288 
11289 		crecv->c_state |= set_state;
11290 		csend->c_state |= set_state;
11291 
11292 		/*
11293 		 * The channel server itself could come through this
11294 		 * flow, so obviously no point in attempting to wake
11295 		 * ourself up!.
11296 		 */
11297 		if (csp->ch_recv_threadp && (csp->ch_recv_threadp != curthread))
11298 			cv_signal(&csp->ch_recv_cv);
11299 
11300 		PR_CHAN("%s: action (%s) for channel %d - COMPLETED\n",
11301 		    proc, chanaction_str[(int)chanaction], channel);
11302 
11303 		/*
11304 		 * Leaves with lock released.
11305 		 */
11306 		mutex_exit(&csend->c_mutex);
11307 		mutex_exit(&crecv->c_mutex);
11308 		break;
11309 
11310 	default:
11311 		ASSERT(0);
11312 		break;
11313 	}
11314 }
11315 
11316 static void
11317 idn_chan_addmbox(int channel, ushort_t domset)
11318 {
11319 	idn_chansvr_t	*csp;
11320 	register int	d;
11321 	procname_t	proc = "idn_chan_addmbox";
11322 
11323 	PR_CHAN("%s: adding domset 0x%x main mailboxes to channel %d\n",
11324 	    proc, domset, channel);
11325 
11326 	ASSERT(idn.chan_servers);
11327 
11328 	csp = &idn.chan_servers[channel];
11329 
11330 	/*
11331 	 * Adding domains to a channel can be
11332 	 * asynchonous, so we don't bother waiting.
11333 	 */
11334 	IDN_CHANNEL_SUSPEND(channel, 0);
11335 
11336 	/*
11337 	 * Now we have the sending and receiving sides blocked
11338 	 * for this channel.
11339 	 */
11340 	for (d = 0; d < MAX_DOMAINS; d++) {
11341 		if (!DOMAIN_IN_SET(domset, d))
11342 			continue;
11343 		if (IDN_CHAN_DOMAIN_IS_REGISTERED(csp, d)) {
11344 			DOMAINSET_DEL(domset, d);
11345 			continue;
11346 		}
11347 		IDN_CHANSVR_SCANSET_ADD_PENDING(csp, d);
11348 		DOMAINSET_ADD(csp->ch_recv_domset_pending, d);
11349 		IDN_CHAN_DOMAIN_REGISTER(csp, d);
11350 
11351 		PR_CHAN("%s: domain %d (channel %d) RECV (pending) "
11352 		    "scanset = 0x%lx\n", proc, d, channel,
11353 		    csp->ch_recv_scanset_pending);
11354 		PR_CHAN("%s: domain %d (channel %d) domset = 0x%x\n",
11355 		    proc, d, channel, (uint_t)csp->ch_reg_domset);
11356 
11357 		CHECKPOINT_OPENED(IDNSB_CHKPT_CHAN,
11358 		    idn_domain[d].dhw.dh_boardset, 1);
11359 	}
11360 	if (domset)
11361 		csp->ch_recv_changed = 1;
11362 
11363 	IDN_CHANNEL_RESUME(channel);
11364 }
11365 
11366 static void
11367 idn_chan_delmbox(int channel, ushort_t domset)
11368 {
11369 	idn_chansvr_t	*csp;
11370 	register int	d;
11371 	procname_t	proc = "idn_chan_delmbox";
11372 
11373 	PR_CHAN("%s: deleting domset 0x%x main mailboxes from channel %d\n",
11374 	    proc, domset, channel);
11375 
11376 	ASSERT(idn.chan_servers);
11377 
11378 	csp = &idn.chan_servers[channel];
11379 
11380 	/*
11381 	 * Here we have to wait for the channel server
11382 	 * as it's vital that we don't return without guaranteeing
11383 	 * that the given domset is no longer registered.
11384 	 */
11385 	IDN_CHANNEL_SUSPEND(channel, 1);
11386 
11387 	/*
11388 	 * Now we have the sending and receiving sides blocked
11389 	 * for this channel.
11390 	 */
11391 	for (d = 0; d < MAX_DOMAINS; d++) {
11392 		if (!DOMAIN_IN_SET(domset, d))
11393 			continue;
11394 		if (!IDN_CHAN_DOMAIN_IS_REGISTERED(csp, d)) {
11395 			DOMAINSET_DEL(domset, d);
11396 			continue;
11397 		}
11398 		/*
11399 		 * This domain has a mailbox hanging on this channel.
11400 		 * Get him out.
11401 		 *
11402 		 * First remove him from the receive side.
11403 		 */
11404 		ASSERT(csp->ch_recv_domcount > 0);
11405 		IDN_CHANSVR_SCANSET_DEL_PENDING(csp, d);
11406 		DOMAINSET_DEL(csp->ch_recv_domset_pending, d);
11407 		IDN_CHAN_DOMAIN_UNREGISTER(csp, d);
11408 
11409 		PR_CHAN("%s: domain %d (channel %d) RECV (pending) "
11410 		    "scanset = 0x%lx\n", proc, d, channel,
11411 		    csp->ch_recv_scanset_pending);
11412 		PR_CHAN("%s: domain %d (channel %d) domset = 0x%x\n",
11413 		    proc, d, channel, (uint_t)csp->ch_reg_domset);
11414 
11415 		CHECKPOINT_CLOSED(IDNSB_CHKPT_CHAN,
11416 		    idn_domain[d].dhw.dh_boardset, 2);
11417 
11418 	}
11419 	if (domset)
11420 		csp->ch_recv_changed = 1;
11421 
11422 	IDN_CHANNEL_RESUME(channel);
11423 }
11424 
11425 static int
11426 idn_valid_etherheader(struct ether_header *ehp)
11427 {
11428 	uchar_t	*eap;
11429 
11430 	eap = &ehp->ether_dhost.ether_addr_octet[0];
11431 
11432 	if ((eap[IDNETHER_ZERO] != 0) && (eap[IDNETHER_ZERO] != 0xff))
11433 		return (0);
11434 
11435 	if ((eap[IDNETHER_COOKIE1] != IDNETHER_COOKIE1_VAL) &&
11436 	    (eap[IDNETHER_COOKIE1] != 0xff))
11437 		return (0);
11438 
11439 	if ((eap[IDNETHER_COOKIE2] != IDNETHER_COOKIE2_VAL) &&
11440 	    (eap[IDNETHER_COOKIE2] != 0xff))
11441 		return (0);
11442 
11443 	if ((eap[IDNETHER_RESERVED] != IDNETHER_RESERVED_VAL) &&
11444 	    (eap[IDNETHER_RESERVED] != 0xff))
11445 		return (0);
11446 
11447 	if (!VALID_UCHANNEL(eap[IDNETHER_CHANNEL]) &&
11448 	    (eap[IDNETHER_CHANNEL] != 0xff))
11449 		return (0);
11450 
11451 	if (!VALID_UDOMAINID(IDN_NETID2DOMID(eap[IDNETHER_NETID])) &&
11452 	    (eap[IDNETHER_NETID] != 0xff))
11453 		return (0);
11454 
11455 	return (1);
11456 }
11457 
11458 /*
11459  * Packet header has already been filled in.
11460  * RETURNS:	0
11461  *		ENOLINK
11462  *		EPROTO
11463  *		ENOSPC
11464  */
11465 /*ARGSUSED*/
11466 static int
11467 idn_send_mboxdata(int domid, struct idn *sip, int channel, caddr_t bufp)
11468 {
11469 	idn_mainmbox_t	*mmp;
11470 	idn_mboxmsg_t	*mqp;
11471 	smr_pkthdr_t	*hdrp;
11472 	smr_offset_t	bufoffset;
11473 	idn_netaddr_t	dst;
11474 	ushort_t		mbox_csum;
11475 	int		rv = 0;
11476 	int		pktlen, qi;
11477 	procname_t	proc = "idn_send_mboxdata";
11478 
11479 	mmp = idn_domain[domid].dmbox.m_send;
11480 	if (mmp == NULL) {
11481 		PR_DATA("%s: dmbox.m_send == NULL\n", proc);
11482 		IDN_KSTAT_INC(sip, si_linkdown);
11483 		return (ENOLINK);
11484 	}
11485 
11486 	mmp += channel;
11487 	mutex_enter(&mmp->mm_mutex);
11488 
11489 	if (mmp->mm_smr_mboxp == NULL) {
11490 		PR_DATA("%s: (d %d, chn %d) mm_smr_mboxp == NULL\n",
11491 		    proc, domid, channel);
11492 		IDN_KSTAT_INC(sip, si_linkdown);
11493 		rv = ENOLINK;
11494 		goto send_err;
11495 	}
11496 	mbox_csum = IDN_CKSUM_MBOX(&mmp->mm_smr_mboxp->mt_header);
11497 	if (mbox_csum != mmp->mm_smr_mboxp->mt_header.mh_cksum) {
11498 		PR_DATA("%s: (d %d, chn %d) mbox hdr cksum (%d) "
11499 		    "!= actual (%d)\n",
11500 		    proc, domid, channel, mbox_csum,
11501 		    mmp->mm_smr_mboxp->mt_header.mh_cksum);
11502 		if ((mmp->mm_flags & IDNMMBOX_FLAG_CORRUPTED) == 0) {
11503 			cmn_err(CE_WARN,
11504 			    "IDN: 241: [send] (domain %d, "
11505 			    "channel %d) SMR CORRUPTED - RELINK",
11506 			    domid, channel);
11507 			mmp->mm_flags |= IDNMMBOX_FLAG_CORRUPTED;
11508 		}
11509 		IDN_KSTAT_INC(sip, si_mboxcrc);
11510 		IDN_KSTAT_INC(sip, si_oerrors);
11511 		rv = EPROTO;
11512 		goto send_err;
11513 	}
11514 
11515 	bufoffset = IDN_ADDR2OFFSET(bufp);
11516 	hdrp	  = IDN_BUF2HDR(bufp);
11517 	pktlen    = hdrp->b_length;
11518 	dst.netaddr = hdrp->b_netaddr;
11519 	ASSERT(dst.net.chan == (ushort_t)channel);
11520 
11521 	mqp = &mmp->mm_smr_mboxp->mt_queue[0];
11522 	qi  = mmp->mm_qiput;
11523 
11524 	if (mqp[qi].ms_owner) {
11525 		PR_DATA("%s: mailbox FULL (qiput=%d, qiget=%d)\n",
11526 		    proc, mmp->mm_qiput, mmp->mm_qiget);
11527 		IDN_KSTAT_INC(sip, si_txfull);
11528 		rv = ENOSPC;
11529 		goto send_err;
11530 	}
11531 	if (mqp[qi].ms_flag & IDN_MBOXMSG_FLAG_RECLAIM) {
11532 		smr_offset_t	recl_bufoffset;
11533 		/*
11534 		 * Remote domain finished with mailbox entry,
11535 		 * however it has not been reclaimed yet.  A reclaim
11536 		 * was done before coming into this routine, however
11537 		 * timing may have been such that the entry became
11538 		 * free just after the reclamation, but before
11539 		 * entry into here.  Go ahead and reclaim this entry.
11540 		 */
11541 		recl_bufoffset = IDN_BFRAME2OFFSET(mqp[qi].ms_bframe);
11542 
11543 		PR_DATA("%s: attempting reclaim (domain %d) "
11544 		    "(qiput=%d, b_off=0x%x)\n",
11545 		    proc, domid, qi, recl_bufoffset);
11546 
11547 		if (VALID_NWROFFSET(recl_bufoffset, IDN_SMR_BUFSIZE)) {
11548 			int		recl;
11549 			caddr_t		b_bufp;
11550 			smr_pkthdr_t	*b_hdrp;
11551 
11552 			b_bufp = IDN_OFFSET2ADDR(recl_bufoffset);
11553 			b_hdrp = IDN_BUF2HDR(b_bufp);
11554 
11555 			if (IDN_CKSUM_PKT(b_hdrp) != b_hdrp->b_cksum) {
11556 				IDN_KSTAT_INC(sip, si_crc);
11557 				IDN_KSTAT_INC(sip, si_fcs_errors);
11558 				IDN_KSTAT_INC(sip, si_reclaim);
11559 				IDN_KSTAT_INC(sip, si_oerrors);
11560 			}
11561 
11562 			recl = smr_buf_free(domid, b_bufp, b_hdrp->b_length);
11563 #ifdef DEBUG
11564 			if (recl == 0) {
11565 				PR_DATA("%s: SUCCESSFULLY reclaimed buf "
11566 				    "(domain %d)\n", proc, domid);
11567 			} else {
11568 				PR_DATA("%s: WARNING: reclaim failed (FREE) "
11569 				    "(domain %d)\n", proc, domid);
11570 			}
11571 #endif /* DEBUG */
11572 		} else {
11573 			IDN_KSTAT_INC(sip, si_smraddr);
11574 			IDN_KSTAT_INC(sip, si_reclaim);
11575 			PR_DATA("%s: WARNING: reclaim failed (BAD OFFSET) "
11576 			    "(domain %d)\n", proc, domid);
11577 		}
11578 	}
11579 
11580 	if (*mmp->mm_smr_readyp == 0) {
11581 		mmp->mm_qiput = qi;
11582 		IDN_KSTAT_INC(sip, si_linkdown);
11583 		rv = ENOLINK;
11584 		goto send_err;
11585 	}
11586 
11587 	mqp[qi].ms_flag = IDN_MBOXMSG_FLAG_RECLAIM;
11588 	mqp[qi].ms_bframe = IDN_OFFSET2BFRAME(bufoffset);
11589 	/* membar_stst(); */
11590 	mqp[qi].ms_owner = 1;
11591 
11592 	IDN_MMBOXINDEX_INC(qi);
11593 
11594 	mmp->mm_qiput = qi;
11595 
11596 	mmp->mm_count++;
11597 
11598 	if ((*mmp->mm_smr_readyp) && !(*mmp->mm_smr_activep)) {
11599 		idn_msgtype_t	mt;
11600 
11601 		mt.mt_mtype = IDNP_DATA;
11602 		mt.mt_atype = 0;
11603 		IDN_KSTAT_INC(sip, si_xdcall);
11604 		(void) IDNXDC(domid, &mt, (uint_t)dst.net.chan, 0, 0, 0);
11605 	}
11606 	mutex_exit(&mmp->mm_mutex);
11607 	IDN_KSTAT_INC(sip, si_opackets);
11608 	IDN_KSTAT_INC(sip, si_opackets64);
11609 	IDN_KSTAT_ADD(sip, si_xmtbytes, pktlen);
11610 	IDN_KSTAT_ADD(sip, si_obytes64, (uint64_t)pktlen);
11611 
11612 	return (0);
11613 
11614 send_err:
11615 	mmp->mm_dropped++;
11616 
11617 	mutex_exit(&mmp->mm_mutex);
11618 
11619 	return (rv);
11620 }
11621 
11622 static int
11623 idn_recv_mboxdata(int channel, caddr_t bufp)
11624 {
11625 	smr_pkthdr_t	*hdrp;
11626 	struct idn	*sip;
11627 	mblk_t		*mp = nilp(mblk_t);
11628 	int		pktlen;
11629 	int		apktlen;
11630 	int		rv = 0;
11631 	smr_offset_t	bufoffset;
11632 	ushort_t	csum;
11633 	idn_netaddr_t	dst, daddr;
11634 	procname_t	proc = "idn_recv_mboxdata";
11635 
11636 	hdrp = IDN_BUF2HDR(bufp);
11637 
11638 	csum = IDN_CKSUM_PKT(hdrp);
11639 
11640 	sip = IDN_INST2SIP(channel);
11641 	if (sip == NULL) {
11642 		/*LINTED*/
11643 		sip = IDN_INST2SIP(0);
11644 	}
11645 	ASSERT(sip);
11646 
11647 	if (csum != hdrp->b_cksum) {
11648 		PR_DATA("%s: bad checksum(%x) != expected(%x)\n",
11649 		    proc, (uint_t)csum, (uint_t)hdrp->b_cksum);
11650 		IDN_KSTAT_INC(sip, si_crc);
11651 		IDN_KSTAT_INC(sip, si_fcs_errors);
11652 		rv = -1;
11653 		goto recv_err;
11654 	}
11655 
11656 	daddr.net.chan = (ushort_t)channel;
11657 	daddr.net.netid = (ushort_t)idn.localid;
11658 
11659 	dst.netaddr = hdrp->b_netaddr;
11660 	bufoffset = hdrp->b_offset;
11661 
11662 	if (dst.netaddr != daddr.netaddr) {
11663 		PR_DATA("%s: wrong dest netaddr (0x%x), expected (0x%x)\n",
11664 		    proc, dst.netaddr, daddr.netaddr);
11665 		IDN_KSTAT_INC(sip, si_nolink);
11666 		IDN_KSTAT_INC(sip, si_macrcv_errors);
11667 		goto recv_err;
11668 	}
11669 	pktlen  = hdrp->b_length;
11670 	apktlen = pktlen;
11671 
11672 	if ((pktlen <= 0) || (pktlen > IDN_DATA_SIZE)) {
11673 		PR_DATA("%s: invalid packet length (%d) <= 0 || > %lu\n",
11674 		    proc, pktlen, IDN_DATA_SIZE);
11675 		IDN_KSTAT_INC(sip, si_buff);
11676 		IDN_KSTAT_INC(sip, si_toolong_errors);
11677 		goto recv_err;
11678 	}
11679 
11680 	mp = allocb(apktlen + IDN_ALIGNSIZE, BPRI_LO);
11681 	if (mp == nilp(mblk_t)) {
11682 		PR_DATA("%s: allocb(pkt) failed\n", proc);
11683 		IDN_KSTAT_INC(sip, si_allocbfail);
11684 		IDN_KSTAT_INC(sip, si_norcvbuf);	/* MIB II */
11685 		goto recv_err;
11686 	}
11687 	ASSERT(DB_TYPE(mp) == M_DATA);
11688 	/*
11689 	 * Copy data packet into its streams buffer.
11690 	 * Align pointers for maximum bcopy performance.
11691 	 */
11692 	mp->b_rptr = (uchar_t *)IDN_ALIGNPTR(mp->b_rptr, bufoffset);
11693 	bcopy(IDN_BUF2DATA(bufp, bufoffset), mp->b_rptr, apktlen);
11694 	mp->b_wptr = mp->b_rptr + pktlen;
11695 
11696 	if (IDN_CHECKSUM &&
11697 		!idn_valid_etherheader((struct ether_header *)mp->b_rptr)) {
11698 		freeb(mp);
11699 		mp = nilp(mblk_t);
11700 		PR_DATA("%s: etherheader CORRUPTED\n", proc);
11701 		IDN_KSTAT_INC(sip, si_crc);
11702 		IDN_KSTAT_INC(sip, si_fcs_errors);
11703 		rv = -1;
11704 		goto recv_err;
11705 	}
11706 
11707 	idndl_read(NULL, mp);
11708 
11709 recv_err:
11710 
11711 	if (mp == nilp(mblk_t)) {
11712 		IDN_KSTAT_INC(sip, si_ierrors);
11713 	}
11714 
11715 	return (rv);
11716 }
11717 
11718 /*
11719  * When on shutdown path (idn_active_resources) must call
11720  * idn_mainmbox_flush() _BEFORE_ calling idn_reclaim_mboxdata()
11721  * for any final data.  This is necessary incase the mailboxes
11722  * have been unregistered.  If they have then idn_mainmbox_flush()
11723  * will set mm_smr_mboxp to NULL which prevents us from touching
11724  * poison SMR space.
11725  */
11726 int
11727 idn_reclaim_mboxdata(int domid, int channel, int nbufs)
11728 {
11729 	idn_mainmbox_t	*mmp;
11730 	idn_mboxmsg_t	*mqp;
11731 	smr_pkthdr_t	*hdrp;
11732 	idn_domain_t	*dp;
11733 	int		qi;
11734 	int		mi;
11735 	int		reclaim_cnt = 0;
11736 	int		free_cnt;
11737 	ushort_t	csum;
11738 	struct idn	*sip;
11739 	smr_offset_t	reclaim_list, curr, prev;
11740 	procname_t	proc = "idn_reclaim_mboxdata";
11741 
11742 
11743 	sip = IDN_INST2SIP(channel);
11744 	if (sip == NULL) {
11745 		/*LINTED*/
11746 		sip = IDN_INST2SIP(0);
11747 	}
11748 	ASSERT(sip);
11749 
11750 	dp = &idn_domain[domid];
11751 
11752 	PR_DATA("%s: requested %d buffers from domain %d\n",
11753 	    proc, nbufs, domid);
11754 
11755 	if (lock_try(&dp->dreclaim_inprogress) == 0) {
11756 		/*
11757 		 * Reclaim is already in progress, don't
11758 		 * bother.
11759 		 */
11760 		PR_DATA("%s: reclaim already in progress\n", proc);
11761 		return (0);
11762 	}
11763 
11764 	if (dp->dmbox.m_send == NULL)
11765 		return (0);
11766 
11767 	reclaim_list = curr = prev = IDN_NIL_SMROFFSET;
11768 
11769 	mi = (int)dp->dreclaim_index;
11770 	do {
11771 		ushort_t	mbox_csum;
11772 
11773 		mmp = &dp->dmbox.m_send[mi];
11774 		/* do-while continues down */
11775 		ASSERT(mmp);
11776 		if (mutex_tryenter(&mmp->mm_mutex) == 0) {
11777 			/*
11778 			 * This channel is busy, move on.
11779 			 */
11780 			IDN_MBOXCHAN_INC(mi);
11781 			continue;
11782 		}
11783 
11784 		if (mmp->mm_smr_mboxp == NULL) {
11785 			PR_DATA("%s: no smr pointer for domid %d, chan %d\n",
11786 			    proc, domid, (int)mmp->mm_channel);
11787 			ASSERT(mmp->mm_qiget == mmp->mm_qiput);
11788 			mutex_exit(&mmp->mm_mutex);
11789 			IDN_MBOXCHAN_INC(mi);
11790 			continue;
11791 		}
11792 		mbox_csum = IDN_CKSUM_MBOX(&mmp->mm_smr_mboxp->mt_header);
11793 		if (mbox_csum != mmp->mm_smr_mboxp->mt_header.mh_cksum) {
11794 			PR_DATA("%s: (d %d, chn %d) mbox hdr "
11795 			    "cksum (%d) != actual (%d)\n",
11796 			    proc, domid, (int)mmp->mm_channel, mbox_csum,
11797 			    mmp->mm_smr_mboxp->mt_header.mh_cksum);
11798 			IDN_KSTAT_INC(sip, si_mboxcrc);
11799 			IDN_KSTAT_INC(sip, si_oerrors);
11800 			mutex_exit(&mmp->mm_mutex);
11801 			IDN_MBOXCHAN_INC(mi);
11802 			continue;
11803 		}
11804 		mqp = &mmp->mm_smr_mboxp->mt_queue[0];
11805 		qi  = mmp->mm_qiget;
11806 
11807 		while (!mqp[qi].ms_owner &&
11808 		    (mqp[qi].ms_flag & IDN_MBOXMSG_FLAG_RECLAIM) &&
11809 		    nbufs) {
11810 			idn_mboxmsg_t	*msp;
11811 			int		badbuf;
11812 
11813 			badbuf = 0;
11814 			msp = &mqp[qi];
11815 
11816 			if (msp->ms_flag & IDN_MBOXMSG_FLAG_ERRMASK) {
11817 				PR_DATA("%s: msg.flag ERROR(0x%x) (off=0x%x, "
11818 				    "domid=%d, qiget=%d)\n", proc,
11819 				    (uint_t)(msp->ms_flag &
11820 				    IDN_MBOXMSG_FLAG_ERRMASK),
11821 				    IDN_BFRAME2OFFSET(msp->ms_bframe),
11822 				    domid, qi);
11823 			}
11824 			prev = curr;
11825 			curr = IDN_BFRAME2OFFSET(mqp[qi].ms_bframe);
11826 
11827 			if (!VALID_NWROFFSET(curr, IDN_SMR_BUFSIZE)) {
11828 				badbuf = 1;
11829 				IDN_KSTAT_INC(sip, si_reclaim);
11830 			} else {
11831 				/*
11832 				 * Put the buffers onto a list that will be
11833 				 * formally reclaimed down below.  This allows
11834 				 * us to free up mboxq entries as fast as
11835 				 * possible.
11836 				 */
11837 				hdrp = IDN_BUF2HDR(IDN_OFFSET2ADDR(curr));
11838 				csum = IDN_CKSUM_PKT(hdrp);
11839 
11840 				if (csum != hdrp->b_cksum) {
11841 					badbuf = 1;
11842 					IDN_KSTAT_INC(sip, si_crc);
11843 					IDN_KSTAT_INC(sip, si_fcs_errors);
11844 					IDN_KSTAT_INC(sip, si_reclaim);
11845 					if (!(mmp->mm_flags &
11846 					    IDNMMBOX_FLAG_CORRUPTED)) {
11847 						cmn_err(CE_WARN,
11848 						    "IDN: 241: [send] "
11849 						    "(domain %d, channel "
11850 						    "%d) SMR CORRUPTED - "
11851 						    "RELINK",
11852 						    domid, channel);
11853 						mmp->mm_flags |=
11854 						    IDNMMBOX_FLAG_CORRUPTED;
11855 					}
11856 
11857 				} else if (reclaim_list == IDN_NIL_SMROFFSET) {
11858 					reclaim_list = curr;
11859 				} else {
11860 					caddr_t	bufp;
11861 
11862 					bufp = IDN_OFFSET2ADDR(prev);
11863 					hdrp = IDN_BUF2HDR(bufp);
11864 					hdrp->b_next = curr;
11865 				}
11866 			}
11867 
11868 			mqp[qi].ms_flag = 0;
11869 
11870 			IDN_MMBOXINDEX_INC(qi);
11871 
11872 			if (!badbuf) {
11873 				nbufs--;
11874 				reclaim_cnt++;
11875 			}
11876 
11877 			if (qi == mmp->mm_qiget)
11878 				break;
11879 		}
11880 		mmp->mm_qiget = qi;
11881 
11882 		mutex_exit(&mmp->mm_mutex);
11883 
11884 		IDN_MBOXCHAN_INC(mi);
11885 
11886 	} while ((mi != (int)dp->dreclaim_index) && nbufs);
11887 
11888 	dp->dreclaim_index = (uchar_t)mi;
11889 
11890 	if (reclaim_list != IDN_NIL_SMROFFSET) {
11891 		hdrp = IDN_BUF2HDR(IDN_OFFSET2ADDR(curr));
11892 		hdrp->b_next = IDN_NIL_SMROFFSET;
11893 	}
11894 
11895 	PR_DATA("%s: reclaimed %d buffers from domain %d\n",
11896 	    proc, reclaim_cnt, domid);
11897 
11898 	if (reclaim_cnt == 0) {
11899 		lock_clear(&dp->dreclaim_inprogress);
11900 		return (0);
11901 	}
11902 
11903 	/*
11904 	 * Now actually go and reclaim (free) the buffers.
11905 	 */
11906 	free_cnt = 0;
11907 
11908 	for (curr = reclaim_list; curr != IDN_NIL_SMROFFSET; ) {
11909 		caddr_t		bufp;
11910 
11911 		bufp = IDN_OFFSET2ADDR(curr);
11912 		hdrp = IDN_BUF2HDR(bufp);
11913 		csum = IDN_CKSUM_PKT(hdrp);
11914 		if (csum != hdrp->b_cksum) {
11915 			/*
11916 			 * Once corruption is detected we
11917 			 * can't trust our list any further.
11918 			 * These buffers are effectively lost.
11919 			 */
11920 			cmn_err(CE_WARN,
11921 			    "IDN: 241: [send] (domain %d, channel %d) SMR "
11922 			    "CORRUPTED - RELINK", domid, channel);
11923 			break;
11924 		}
11925 
11926 		curr = hdrp->b_next;
11927 
11928 		if (!smr_buf_free(domid, bufp, hdrp->b_length))
11929 			free_cnt++;
11930 	}
11931 
11932 	if ((dp->dio < IDN_WINDOW_EMAX) && dp->diocheck) {
11933 		lock_clear(&dp->diocheck);
11934 		IDN_MSGTIMER_STOP(domid, IDNP_DATA, 0);
11935 	}
11936 
11937 #ifdef DEBUG
11938 	if (free_cnt != reclaim_cnt) {
11939 		PR_DATA("%s: *** WARNING *** freecnt(%d) != reclaim_cnt (%d)\n",
11940 		    proc, free_cnt, reclaim_cnt);
11941 	}
11942 #endif /* DEBUG */
11943 
11944 	lock_clear(&dp->dreclaim_inprogress);
11945 
11946 	return (reclaim_cnt);
11947 }
11948 
11949 void
11950 idn_signal_data_server(int domid, ushort_t channel)
11951 {
11952 	idn_nack_t	nacktype = 0;
11953 	idn_domain_t	*dp;
11954 	idn_chansvr_t	*csp;
11955 	int		c, min_chan, max_chan;
11956 	idn_mainmbox_t	*mmp;
11957 	procname_t	proc = "idn_signal_data_server";
11958 
11959 
11960 	if (domid == IDN_NIL_DOMID)
11961 		return;
11962 
11963 	dp = &idn_domain[domid];
11964 
11965 	if (dp->dawol.a_count > 0) {
11966 		/*
11967 		 * Domain was previously AWOL, but no longer.
11968 		 */
11969 		IDN_SYNC_LOCK();
11970 		IDN_GLOCK_EXCL();
11971 		idn_clear_awol(domid);
11972 		IDN_GUNLOCK();
11973 		IDN_SYNC_UNLOCK();
11974 	}
11975 	/*
11976 	 * Do a precheck before wasting time trying to acquire the lock.
11977 	 */
11978 	if ((dp->dstate != IDNDS_CONNECTED) || !IDN_DLOCK_TRY_SHARED(domid)) {
11979 		/*
11980 		 * Either we're not connected or somebody is busy working
11981 		 * on the domain.  Bail on the signal for now, we'll catch
11982 		 * it on the next go around.
11983 		 */
11984 		return;
11985 	}
11986 	/*
11987 	 * We didn't have the drwlock on the first check of dstate,
11988 	 * but now that we do, make sure the world hasn't changed!
11989 	 */
11990 	if (dp->dstate != IDNDS_CONNECTED) {
11991 		/*
11992 		 * If we reach here, then no connection.
11993 		 * Send no response if this is the case.
11994 		 */
11995 		nacktype = IDNNACK_NOCONN;
11996 		goto send_dresp;
11997 	}
11998 
11999 	/*
12000 	 * No need to worry about locking mainmbox
12001 	 * because we're already holding reader
12002 	 * lock on domain, plus we're just reading
12003 	 * fields in the mainmbox which only change
12004 	 * (or go away) when the writer lock is
12005 	 * held on the domain.
12006 	 */
12007 	if ((mmp = dp->dmbox.m_recv) == NULL) {
12008 		/*
12009 		 * No local mailbox.
12010 		 */
12011 		nacktype = IDNNACK_BADCFG;
12012 		goto send_dresp;
12013 	}
12014 	if ((channel != IDN_BROADCAST_ALLCHAN) && (channel >= IDN_MAX_NETS)) {
12015 		nacktype = IDNNACK_BADCHAN;
12016 		goto send_dresp;
12017 	}
12018 	if (channel == IDN_BROADCAST_ALLCHAN) {
12019 		PR_DATA("%s: requested signal to ALL channels on domain %d\n",
12020 		    proc, domid);
12021 		min_chan = 0;
12022 		max_chan = IDN_MAX_NETS - 1;
12023 	} else {
12024 		PR_DATA("%s: requested signal to channel %d on domain %d\n",
12025 		    proc, channel, domid);
12026 		min_chan = max_chan = (int)channel;
12027 	}
12028 	mmp += min_chan;
12029 	for (c = min_chan; c <= max_chan; mmp++, c++) {
12030 
12031 		/*
12032 		 * We do a quick check for a pending channel.
12033 		 * If pending it will need activation and we rather
12034 		 * do that through a separate (proto) thread.
12035 		 */
12036 		csp = &idn.chan_servers[c];
12037 
12038 		if (csp->ch_recv.c_checkin) {
12039 			PR_DATA("%s: chansvr (%d) for domid %d CHECK-IN\n",
12040 			    proc, c, domid);
12041 			continue;
12042 		}
12043 
12044 		if (IDN_CHAN_TRYLOCK_RECV(csp) == 0) {
12045 			/*
12046 			 * Failed to grab lock, server must be active.
12047 			 */
12048 			PR_DATA("%s: chansvr (%d) for domid %d already actv\n",
12049 			    proc, c, domid);
12050 			continue;
12051 		}
12052 
12053 		if (IDN_CHANNEL_IS_PENDING(csp)) {
12054 			/*
12055 			 * Lock is pending.  Submit asynchronous
12056 			 * job to activate and move-on.
12057 			 */
12058 			IDN_CHAN_UNLOCK_RECV(csp);
12059 			idn_submit_chanactivate_job(c);
12060 			continue;
12061 		}
12062 
12063 		/*
12064 		 * If he ain't active, we ain't talkin'.
12065 		 */
12066 		if (IDN_CHANNEL_IS_RECV_ACTIVE(csp) == 0) {
12067 			IDN_CHAN_UNLOCK_RECV(csp);
12068 			PR_DATA("%s: chansvr (%d) for domid %d inactive\n",
12069 			    proc, c, domid);
12070 			continue;
12071 		}
12072 
12073 		if (mutex_tryenter(&mmp->mm_mutex) == 0) {
12074 			IDN_CHAN_UNLOCK_RECV(csp);
12075 			continue;
12076 		}
12077 
12078 		if (mmp->mm_csp != csp) {
12079 			/*
12080 			 * Not registered.
12081 			 */
12082 			mutex_exit(&mmp->mm_mutex);
12083 			IDN_CHAN_UNLOCK_RECV(csp);
12084 			continue;
12085 
12086 		}
12087 		if (mmp->mm_smr_mboxp == NULL) {
12088 			/*
12089 			 * No SMR mailbox.
12090 			 */
12091 			mutex_exit(&mmp->mm_mutex);
12092 			IDN_CHAN_UNLOCK_RECV(csp);
12093 			continue;
12094 		}
12095 		mutex_exit(&mmp->mm_mutex);
12096 
12097 		if (csp->ch_recv.c_inprogress) {
12098 			/*
12099 			 * Data server is already active.
12100 			 */
12101 			IDN_CHAN_UNLOCK_RECV(csp);
12102 			PR_DATA("%s: chansvr (%d) for domid %d already actv\n",
12103 			    proc, c, domid);
12104 			continue;
12105 		}
12106 		ASSERT(csp == &idn.chan_servers[c]);
12107 
12108 
12109 		PR_DATA("%s: signaling data dispatcher for chan %d dom %d\n",
12110 		    proc, c, domid);
12111 		ASSERT(csp);
12112 		cv_signal(&csp->ch_recv_cv);
12113 		IDN_CHAN_UNLOCK_RECV(csp);
12114 	}
12115 
12116 	if (!nacktype || (channel == IDN_BROADCAST_ALLCHAN)) {
12117 		/*
12118 		 * If there were no real errors or we were
12119 		 * handling multiple channels, then just
12120 		 * return.
12121 		 */
12122 		IDN_DUNLOCK(domid);
12123 		return;
12124 	}
12125 
12126 send_dresp:
12127 
12128 	PR_DATA("%s: sending NACK (%s) back to domain %d (cpu %d)\n",
12129 	    proc, idnnack_str[nacktype], domid, idn_domain[domid].dcpu);
12130 
12131 	idn_send_dataresp(domid, nacktype);
12132 
12133 	IDN_DUNLOCK(domid);
12134 }
12135 
12136 /*ARGSUSED*/
12137 static int
12138 idn_recv_data(int domid, idn_msgtype_t *mtp, idn_xdcargs_t xargs)
12139 {
12140 #ifdef DEBUG
12141 	uint_t		msg = mtp ? mtp->mt_mtype : 0;
12142 	uint_t		msgarg = mtp ? mtp->mt_atype : 0;
12143 	procname_t	proc = "idn_recv_data";
12144 
12145 	PR_PROTO("%s:%d: DATA message received (msg = 0x%x, msgarg = 0x%x)\n",
12146 	    proc, domid, msg, msgarg);
12147 	PR_PROTO("%s:%d: xargs = (0x%x, 0x%x, 0x%x, 0x%x)\n",
12148 	    proc, domid, xargs[0], xargs[1], xargs[2], xargs[3]);
12149 #endif /* DEBUG */
12150 
12151 	return (0);
12152 }
12153 
12154 /*
12155  * Only used when sending a negative response.
12156  */
12157 static void
12158 idn_send_dataresp(int domid, idn_nack_t nacktype)
12159 {
12160 	idn_msgtype_t	mt;
12161 
12162 	ASSERT(IDN_DLOCK_IS_HELD(domid));
12163 
12164 	if (idn_domain[domid].dcpu == IDN_NIL_DCPU)
12165 		return;
12166 
12167 	mt.mt_mtype = IDNP_NACK;
12168 	mt.mt_atype = IDNP_DATA;
12169 
12170 	(void) IDNXDC(domid, &mt, (uint_t)nacktype, 0, 0, 0);
12171 }
12172 
12173 /*
12174  * Checksum routine used in checksum smr_pkthdr_t and idn_mboxhdr_t.
12175  */
12176 static ushort_t
12177 idn_cksum(register ushort_t *hdrp, register int count)
12178 {
12179 	register int		i;
12180 	register ushort_t	sum = 0;
12181 
12182 	for (i = 0; i < count; i++)
12183 		sum += hdrp[i];
12184 
12185 	sum = (sum >> 16) + (sum & 0xffff);
12186 	sum += (sum >> 16);
12187 
12188 	return (~sum);
12189 }
12190 
12191 /*
12192  * ------------------------------------------------
12193  */
12194 
12195 int
12196 idn_open_channel(int channel)
12197 {
12198 	int		masterid;
12199 	idn_chansvr_t	*csp;
12200 	struct idn	*sip;
12201 	procname_t	proc = "idn_open_channel";
12202 
12203 	if (channel >= IDN_MAX_NETS) {
12204 		cmn_err(CE_WARN,
12205 		    "IDN: 242: maximum channels (%d) already open",
12206 		    IDN_MAX_NETS);
12207 		return (-1);
12208 	}
12209 	IDN_GLOCK_EXCL();
12210 
12211 	ASSERT(idn.chan_servers != NULL);
12212 
12213 	csp = &idn.chan_servers[channel];
12214 
12215 	IDN_CHAN_LOCK_GLOBAL(csp);
12216 
12217 	if (IDN_CHANNEL_IS_ATTACHED(csp)) {
12218 		PR_CHAN("%s: channel %d already open\n", proc, channel);
12219 		IDN_CHAN_UNLOCK_GLOBAL(csp);
12220 		IDN_GUNLOCK();
12221 		return (0);
12222 	}
12223 
12224 	/*
12225 	 * Need to zero out the kstats now that we're activating
12226 	 * this channel.
12227 	 */
12228 	for (sip = idn.sip; sip; sip = sip->si_nextp) {
12229 		if (sip->si_dip && (ddi_get_instance(sip->si_dip) == channel)) {
12230 			bzero(&sip->si_kstat, sizeof (sip->si_kstat));
12231 			break;
12232 		}
12233 	}
12234 
12235 	IDN_CHANSVC_MARK_ATTACHED(csp);
12236 	idn.nchannels++;
12237 	CHANSET_ADD(idn.chanset, channel);
12238 	IDN_CHANNEL_ATTACH(channel);
12239 
12240 	IDN_CHAN_UNLOCK_GLOBAL(csp);
12241 
12242 	/*
12243 	 * We increase our window threshold each time a channel
12244 	 * is opened.
12245 	 */
12246 	ASSERT(idn.nchannels > 0);
12247 	IDN_WINDOW_EMAX = IDN_WINDOW_MAX +
12248 	    ((idn.nchannels - 1) * IDN_WINDOW_INCR);
12249 
12250 	PR_CHAN("%s: channel %d is OPEN (nchannels = %d)\n",
12251 	    proc, channel, idn.nchannels);
12252 
12253 	masterid = IDN_GET_MASTERID();
12254 	IDN_GUNLOCK();
12255 
12256 	/*
12257 	 * Check if there is an active master to which
12258 	 * we're connected.  If so, then activate channel.
12259 	 */
12260 	if (masterid != IDN_NIL_DOMID) {
12261 		idn_domain_t	*dp;
12262 
12263 		dp = &idn_domain[masterid];
12264 		IDN_DLOCK_SHARED(masterid);
12265 		if (dp->dvote.v.master && (dp->dstate == IDNDS_CONNECTED))
12266 			(void) idn_activate_channel(CHANSET(channel),
12267 			    IDNCHAN_ONLINE);
12268 		IDN_DUNLOCK(masterid);
12269 	}
12270 
12271 	return (0);
12272 }
12273 
12274 void
12275 idn_close_channel(int channel, idn_chanop_t chanop)
12276 {
12277 	idn_chansvr_t	*csp;
12278 	procname_t	proc = "idn_close_channel";
12279 
12280 
12281 	ASSERT(idn.chan_servers != NULL);
12282 
12283 	csp = &idn.chan_servers[channel];
12284 
12285 	IDN_GLOCK_EXCL();
12286 
12287 	IDN_CHAN_LOCK_GLOBAL(csp);
12288 	if (IDN_CHANNEL_IS_DETACHED(csp)) {
12289 		PR_CHAN("%s: channel %d already closed\n", proc, channel);
12290 		IDN_CHAN_UNLOCK_GLOBAL(csp);
12291 		IDN_GUNLOCK();
12292 		return;
12293 	}
12294 	IDN_CHAN_UNLOCK_GLOBAL(csp);
12295 
12296 	idn_deactivate_channel(CHANSET(channel), chanop);
12297 
12298 	IDN_CHAN_LOCK_GLOBAL(csp);
12299 
12300 	if (chanop == IDNCHAN_HARD_CLOSE) {
12301 		idn.nchannels--;
12302 		CHANSET_DEL(idn.chanset, channel);
12303 		/*
12304 		 * We increase our window threshold each time a channel
12305 		 * is opened.
12306 		 */
12307 		if (idn.nchannels <= 0)
12308 			IDN_WINDOW_EMAX = 0;
12309 		else
12310 			IDN_WINDOW_EMAX = IDN_WINDOW_MAX +
12311 			    ((idn.nchannels - 1) * IDN_WINDOW_INCR);
12312 	}
12313 
12314 	PR_CHAN("%s: channel %d is (%s) CLOSED (nchannels = %d)\n",
12315 	    proc, channel,
12316 	    (chanop == IDNCHAN_SOFT_CLOSE) ? "SOFT"
12317 	    : (chanop == IDNCHAN_HARD_CLOSE) ? "HARD" : "OFFLINE",
12318 	    idn.nchannels);
12319 
12320 	IDN_CHAN_UNLOCK_GLOBAL(csp);
12321 	IDN_GUNLOCK();
12322 }
12323 
12324 static int
12325 idn_activate_channel(idn_chanset_t chanset, idn_chanop_t chanop)
12326 {
12327 	int		c, rv = 0;
12328 	procname_t	proc = "idn_activate_channel";
12329 
12330 	PR_CHAN("%s: chanset = 0x%x, chanop = %s\n",
12331 	    proc, chanset, chanop_str[chanop]);
12332 
12333 	if (idn.state != IDNGS_ONLINE) {
12334 		/*
12335 		 * Can't activate any channels unless local
12336 		 * domain is connected and thus has a master.
12337 		 */
12338 		PR_CHAN("%s: local domain not connected.  no data servers\n",
12339 		    proc);
12340 		return (-1);
12341 	}
12342 
12343 	for (c = 0; c < IDN_MAX_NETS; c++) {
12344 		idn_chansvr_t	*csp;
12345 		idn_mboxhdr_t	*mainhp;
12346 		struct idn	*sip;
12347 
12348 		if (!CHAN_IN_SET(chanset, c))
12349 			continue;
12350 		csp = &idn.chan_servers[c];
12351 
12352 		if (chanop == IDNCHAN_ONLINE) {
12353 			IDN_CHAN_LOCK_GLOBAL(csp);
12354 		} else {
12355 			/*
12356 			 * We don't wait to grab the global lock
12357 			 * if IDNCHAN_OPEN since these occur along
12358 			 * critical data paths and will be retried
12359 			 * anyway if needed.
12360 			 */
12361 			if (IDN_CHAN_TRYLOCK_GLOBAL(csp) == 0) {
12362 				PR_CHAN("%s: failed to acquire global "
12363 				    "lock for channel %d\n",
12364 				    proc, c);
12365 				continue;
12366 			}
12367 		}
12368 
12369 		if (!IDN_CHANNEL_IS_ATTACHED(csp)) {
12370 			PR_CHAN("%s: channel %d NOT open\n", proc, c);
12371 			IDN_CHAN_UNLOCK_GLOBAL(csp);
12372 			continue;
12373 
12374 		}
12375 
12376 		if (IDN_CHANNEL_IS_ACTIVE(csp)) {
12377 
12378 			PR_CHAN("%s: channel %d already active\n", proc, c);
12379 			rv++;
12380 			IDN_CHAN_UNLOCK_GLOBAL(csp);
12381 			continue;
12382 
12383 		}
12384 		/*
12385 		 * Channel activation can happen asynchronously.
12386 		 */
12387 		IDN_CHANNEL_SUSPEND(c, 0);
12388 
12389 		if (IDN_CHANNEL_IS_PENDING(csp) && (chanop == IDNCHAN_OPEN)) {
12390 
12391 			PR_CHAN("%s: ACTIVATING channel %d\n", proc, c);
12392 
12393 			if (idn_activate_channel_services(c) >= 0) {
12394 				PR_CHAN("%s: Setting channel %d ACTIVE\n",
12395 				    proc, c);
12396 				IDN_CHANSVC_MARK_ACTIVE(csp);
12397 				rv++;
12398 			}
12399 		} else if (!IDN_CHANNEL_IS_PENDING(csp) &&
12400 		    (chanop == IDNCHAN_ONLINE)) {
12401 			PR_CHAN("%s: Setting channel %d PENDING\n", proc, c);
12402 
12403 			IDN_CHANSVC_MARK_PENDING(csp);
12404 		}
12405 		/*
12406 		 * Don't syncheader (i.e. touch SMR) unless
12407 		 * channel is at least ENABLED.  For a DISABLED
12408 		 * channel, the SMR may be invalid so do NOT
12409 		 * touch it.
12410 		 */
12411 		if (IDN_CHANNEL_IS_ENABLED(csp) &&
12412 		    ((mainhp = idn_chan_server_syncheader(c)) != NULL)) {
12413 			PR_CHAN("%s: marking chansvr (mhp=0x%p) %d READY\n",
12414 			    proc, (void *)mainhp, c);
12415 			mainhp->mh_svr_ready = 1;
12416 		}
12417 
12418 		IDN_CHANNEL_RESUME(c);
12419 		sip = IDN_INST2SIP(c);
12420 		ASSERT(sip);
12421 		if (sip->si_wantw) {
12422 			mutex_enter(&idn.sipwenlock);
12423 			idndl_wenable(sip);
12424 			mutex_exit(&idn.sipwenlock);
12425 		}
12426 		IDN_CHAN_UNLOCK_GLOBAL(csp);
12427 
12428 	}
12429 	/*
12430 	 * Returns "not active", i.e. value of 0 indicates
12431 	 * no channels are activated.
12432 	 */
12433 	return (rv == 0);
12434 }
12435 
12436 static void
12437 idn_deactivate_channel(idn_chanset_t chanset, idn_chanop_t chanop)
12438 {
12439 	int		c;
12440 	procname_t	proc = "idn_deactivate_channel";
12441 
12442 	PR_CHAN("%s: chanset = 0x%x, chanop = %s\n",
12443 	    proc, chanset, chanop_str[chanop]);
12444 
12445 	for (c = 0; c < IDN_MAX_NETS; c++) {
12446 		idn_chansvr_t	*csp;
12447 		idn_mboxhdr_t	*mainhp;
12448 
12449 		if (!CHAN_IN_SET(chanset, c))
12450 			continue;
12451 
12452 		csp = &idn.chan_servers[c];
12453 
12454 		IDN_CHAN_LOCK_GLOBAL(csp);
12455 
12456 		if (((chanop == IDNCHAN_SOFT_CLOSE) &&
12457 		    !IDN_CHANNEL_IS_ACTIVE(csp)) ||
12458 		    ((chanop == IDNCHAN_HARD_CLOSE) &&
12459 		    IDN_CHANNEL_IS_DETACHED(csp)) ||
12460 		    ((chanop == IDNCHAN_OFFLINE) &&
12461 		    !IDN_CHANNEL_IS_ENABLED(csp))) {
12462 
12463 			ASSERT(!IDN_CHANNEL_IS_RECV_ACTIVE(csp));
12464 			ASSERT(!IDN_CHANNEL_IS_SEND_ACTIVE(csp));
12465 
12466 			PR_CHAN("%s: channel %d already deactivated\n",
12467 			    proc, c);
12468 			IDN_CHAN_UNLOCK_GLOBAL(csp);
12469 			continue;
12470 		}
12471 
12472 		switch (chanop) {
12473 		case IDNCHAN_OFFLINE:
12474 			IDN_CHANSVC_MARK_IDLE(csp);
12475 			IDN_CHANSVC_MARK_DISABLED(csp);
12476 			IDN_CHANNEL_STOP(c, 1);
12477 			mainhp = idn_chan_server_syncheader(c);
12478 			if (mainhp != NULL)
12479 				mainhp->mh_svr_ready = 0;
12480 			break;
12481 
12482 		case IDNCHAN_HARD_CLOSE:
12483 			IDN_CHANSVC_MARK_DETACHED(csp);
12484 			IDN_CHANNEL_DETACH(c, 1);
12485 			mainhp = idn_chan_server_syncheader(c);
12486 			if (mainhp != NULL)
12487 				mainhp->mh_svr_ready = 0;
12488 			break;
12489 
12490 		default:
12491 			IDN_CHANSVC_MARK_IDLE(csp);
12492 			IDN_CHANNEL_SUSPEND(c, 1);
12493 			ASSERT(IDN_CHANNEL_IS_ATTACHED(csp));
12494 			break;
12495 		}
12496 
12497 		lock_clear(&csp->ch_actvlck);
12498 		lock_clear(&csp->ch_initlck);
12499 
12500 		PR_CHAN("%s: DEACTIVATING channel %d (%s)\n", proc, c,
12501 		    chanop_str[chanop]);
12502 		PR_CHAN("%s: removing chanset 0x%x data svrs for "
12503 		    "each domain link\n", proc, chanset);
12504 
12505 		(void) idn_deactivate_channel_services(c, chanop);
12506 	}
12507 	/*
12508 	 * Returns with channels unlocked.
12509 	 */
12510 }
12511 
12512 /*
12513  * The priority of the channel server must be less than that
12514  * of the protocol server since the protocol server tasks
12515  * are (can be) of more importance.
12516  *
12517  * Possible range: 60-99.
12518  */
12519 static pri_t	idn_chansvr_pri = (7 * MAXCLSYSPRI) / 8;
12520 
12521 static int
12522 idn_activate_channel_services(int channel)
12523 {
12524 	idn_chansvr_t	*csp;
12525 	procname_t	proc = "idn_activate_channel_services";
12526 
12527 
12528 	ASSERT((channel >= 0) && (channel < IDN_MAX_NETS));
12529 
12530 	csp = &idn.chan_servers[channel];
12531 
12532 	ASSERT(IDN_CHAN_GLOBAL_IS_LOCKED(csp));
12533 	ASSERT(IDN_CHAN_LOCAL_IS_LOCKED(csp));
12534 
12535 	if (csp->ch_recv_threadp) {
12536 		/*
12537 		 * There's an existing dispatcher!
12538 		 * Must have been idle'd during an earlier
12539 		 * stint.
12540 		 */
12541 		ASSERT(csp->ch_id == (uchar_t)channel);
12542 		PR_CHAN("%s: existing chansvr FOUND for (c=%d)\n",
12543 		    proc, channel);
12544 
12545 		if (IDN_CHANNEL_IS_PENDING(csp) == 0)
12546 			return (-1);
12547 
12548 		PR_CHAN("%s: chansvr (c=%d) Rstate = 0x%x, Sstate = 0x%x\n",
12549 		    proc, channel, csp->ch_recv.c_state,
12550 		    csp->ch_send.c_state);
12551 
12552 		cv_signal(&csp->ch_recv_cv);
12553 
12554 		return (0);
12555 	}
12556 
12557 	if (IDN_CHANNEL_IS_PENDING(csp) == 0)
12558 		return (-1);
12559 
12560 	csp->ch_id = (uchar_t)channel;
12561 
12562 	PR_CHAN("%s: init channel %d server\n", proc, channel);
12563 
12564 	csp->ch_recv_morguep = GETSTRUCT(ksema_t, 1);
12565 	sema_init(csp->ch_recv_morguep, 0, NULL, SEMA_DRIVER, NULL);
12566 
12567 	csp->ch_recv.c_inprogress = 0;
12568 	csp->ch_recv.c_waiters = 0;
12569 	csp->ch_recv.c_checkin = 0;
12570 	csp->ch_recv_changed = 1;
12571 
12572 	csp->ch_recv_domset = csp->ch_reg_domset;
12573 
12574 	csp->ch_recv_waittime = IDN_NETSVR_WAIT_MIN;
12575 
12576 	csp->ch_recv_threadp = thread_create(NULL, 0,
12577 	    idn_chan_server, &csp, sizeof (csp), &p0, TS_RUN, idn_chansvr_pri);
12578 
12579 	csp->ch_send.c_inprogress = 0;
12580 	csp->ch_send.c_waiters = 0;
12581 	csp->ch_send.c_checkin = 0;
12582 
12583 	return (0);
12584 }
12585 
12586 /*
12587  * This routine can handle terminating a set of channel
12588  * servers all at once, however currently only used
12589  * for serial killing, i.e. one-at-a-time.
12590  *
12591  * Entered with RECV locks held on chanset.
12592  * Acquires SEND locks if needed.
12593  * Leaves with all RECV and SEND locks dropped.
12594  */
12595 static int
12596 idn_deactivate_channel_services(int channel, idn_chanop_t chanop)
12597 {
12598 	idn_chansvr_t	*csp;
12599 	int		cs_count;
12600 	int		c;
12601 	idn_chanset_t	chanset;
12602 	ksema_t		*central_morguep = NULL;
12603 	procname_t	proc = "idn_deactivate_channel_services";
12604 
12605 
12606 	ASSERT(idn.chan_servers);
12607 
12608 	PR_CHAN("%s: deactivating channel %d services\n", proc, channel);
12609 
12610 	/*
12611 	 * XXX
12612 	 * Old code allowed us to deactivate multiple channel
12613 	 * servers at once.  Keep for now just in case.
12614 	 */
12615 	chanset = CHANSET(channel);
12616 
12617 	/*
12618 	 * Point all the data dispatchers to the same morgue
12619 	 * so we can kill them all at once.
12620 	 */
12621 	cs_count = 0;
12622 	for (c = 0; c < IDN_MAX_NETS; c++) {
12623 		if (!CHAN_IN_SET(chanset, c))
12624 			continue;
12625 
12626 		csp = &idn.chan_servers[c];
12627 		ASSERT(IDN_CHAN_GLOBAL_IS_LOCKED(csp));
12628 		ASSERT(IDN_CHAN_LOCAL_IS_LOCKED(csp));
12629 
12630 		if (csp->ch_recv_threadp == NULL) {
12631 			/*
12632 			 * No channel server home.
12633 			 * But we're still holding the c_mutex.
12634 			 * At mark him idle incase we start him up.
12635 			 */
12636 			PR_CHAN("%s: no channel server found for chan %d\n",
12637 			    proc, c);
12638 			IDN_CHAN_UNLOCK_LOCAL(csp);
12639 			IDN_CHAN_UNLOCK_GLOBAL(csp);
12640 			continue;
12641 		}
12642 		ASSERT(csp->ch_id == (uchar_t)c);
12643 
12644 		/*
12645 		 * Okay, now we've blocked the send and receive sides.
12646 		 */
12647 
12648 		if ((chanop == IDNCHAN_SOFT_CLOSE) ||
12649 		    (chanop == IDNCHAN_OFFLINE)) {
12650 			/*
12651 			 * We set turned off the ACTIVE flag, but there's
12652 			 * no guarantee he stopped because of it.  He may
12653 			 * have already been sleeping.  We need to be
12654 			 * sure he recognizes the IDLE, so we need to
12655 			 * signal him and give him a chance to see it.
12656 			 */
12657 			cv_signal(&csp->ch_recv_cv);
12658 			IDN_CHAN_UNLOCK_LOCAL(csp);
12659 			IDN_CHAN_UNLOCK_GLOBAL(csp);
12660 			cs_count++;
12661 			continue;
12662 		}
12663 
12664 		PR_CHAN("%s: pointing chansvr %d to morgue (0x%p)\n",
12665 		    proc, c, central_morguep ? central_morguep
12666 		    : (void *)(csp->ch_recv_morguep));
12667 
12668 		if (central_morguep == NULL) {
12669 			central_morguep = csp->ch_recv_morguep;
12670 		} else {
12671 			sema_destroy(csp->ch_recv_morguep);
12672 			FREESTRUCT(csp->ch_recv_morguep, ksema_t, 1);
12673 
12674 			csp->ch_recv_morguep = central_morguep;
12675 		}
12676 		cv_signal(&csp->ch_recv_cv);
12677 		if (csp->ch_recv.c_waiters > 0)
12678 			cv_broadcast(&csp->ch_recv.c_cv);
12679 		/*
12680 		 * Save any existing binding for next reincarnation.
12681 		 * Note that we're holding the local and global
12682 		 * locks so we're protected against others touchers
12683 		 * of the ch_bound_cpuid fields.
12684 		 */
12685 		csp->ch_bound_cpuid_pending = csp->ch_bound_cpuid;
12686 		csp->ch_bound_cpuid = -1;
12687 		IDN_CHAN_UNLOCK_LOCAL(csp);
12688 		IDN_CHAN_UNLOCK_GLOBAL(csp);
12689 		cs_count++;
12690 	}
12691 	PR_CHAN("%s: signaled %d chansvrs for chanset 0x%x\n",
12692 	    proc, cs_count, chanset);
12693 
12694 	if ((chanop == IDNCHAN_SOFT_CLOSE) || (chanop == IDNCHAN_OFFLINE))
12695 		return (cs_count);
12696 
12697 	PR_CHAN("%s: waiting for %d (chnset=0x%x) chan svrs to term\n",
12698 	    proc, cs_count, chanset);
12699 	PR_CHAN("%s: morguep = 0x%p\n", proc, (void *)central_morguep);
12700 
12701 	ASSERT((cs_count > 0) ? (central_morguep != NULL) : 1);
12702 	while (cs_count-- > 0)
12703 		sema_p(central_morguep);
12704 
12705 	if (central_morguep) {
12706 		sema_destroy(central_morguep);
12707 		FREESTRUCT(central_morguep, ksema_t, 1);
12708 	}
12709 
12710 	return (cs_count);
12711 }
12712 
12713 int
12714 idn_chanservers_init()
12715 {
12716 	int		c;
12717 	idn_chansvr_t	*csp;
12718 
12719 
12720 	if (idn.chan_servers)
12721 		return (0);
12722 
12723 	idn.chan_servers = GETSTRUCT(idn_chansvr_t, IDN_MAXMAX_NETS);
12724 
12725 	for (c = 0; c < IDN_MAXMAX_NETS; c++) {
12726 		csp = &idn.chan_servers[c];
12727 		mutex_init(&csp->ch_send.c_mutex, NULL, MUTEX_DEFAULT, NULL);
12728 		mutex_init(&csp->ch_recv.c_mutex, NULL, MUTEX_DEFAULT, NULL);
12729 		cv_init(&csp->ch_send.c_cv, NULL, CV_DRIVER, NULL);
12730 		cv_init(&csp->ch_recv.c_cv, NULL, CV_DRIVER, NULL);
12731 		cv_init(&csp->ch_recv_cv, NULL, CV_DRIVER, NULL);
12732 		csp->ch_bound_cpuid = -1;
12733 		csp->ch_bound_cpuid_pending = -1;
12734 	}
12735 
12736 	return (c);
12737 }
12738 
12739 void
12740 idn_chanservers_deinit()
12741 {
12742 	int		c;
12743 	idn_chansvr_t	*csp;
12744 
12745 
12746 	if (idn.chan_servers == NULL)
12747 		return;
12748 
12749 	for (c = 0; c < IDN_MAXMAX_NETS; c++) {
12750 		csp = &idn.chan_servers[c];
12751 
12752 		mutex_destroy(&csp->ch_send.c_mutex);
12753 		mutex_destroy(&csp->ch_recv.c_mutex);
12754 		cv_destroy(&csp->ch_send.c_cv);
12755 		cv_destroy(&csp->ch_recv.c_cv);
12756 		cv_destroy(&csp->ch_recv_cv);
12757 	}
12758 
12759 	FREESTRUCT(idn.chan_servers, idn_chansvr_t, IDN_MAXMAX_NETS);
12760 	idn.chan_servers = NULL;
12761 }
12762 
12763 static void
12764 idn_exec_chanactivate(void *chn)
12765 {
12766 	int		not_active, channel;
12767 	idn_chansvr_t	*csp;
12768 
12769 	channel = (int)(uintptr_t)chn;
12770 
12771 	IDN_GLOCK_SHARED();
12772 	if (idn.chan_servers == NULL) {
12773 		IDN_GUNLOCK();
12774 		return;
12775 	}
12776 	csp = &idn.chan_servers[channel];
12777 
12778 	if (IDN_CHAN_TRYLOCK_GLOBAL(csp) == 0) {
12779 		/*
12780 		 * If we can't grab the global lock, then
12781 		 * something is up, skip out.
12782 		 */
12783 		IDN_GUNLOCK();
12784 		return;
12785 	}
12786 	IDN_GUNLOCK();
12787 
12788 	if (IDN_CHANNEL_IS_PENDING(csp) && lock_try(&csp->ch_actvlck)) {
12789 		IDN_CHAN_UNLOCK_GLOBAL(csp);
12790 		not_active = idn_activate_channel(CHANSET(channel),
12791 		    IDNCHAN_OPEN);
12792 		if (not_active)
12793 			lock_clear(&csp->ch_actvlck);
12794 	} else {
12795 		IDN_CHAN_UNLOCK_GLOBAL(csp);
12796 	}
12797 }
12798 
12799 /*
12800  * Delayed activation of channel.  We don't want to do this within
12801  * idn_signal_data_server() since that's called within the context
12802  * of an XDC handler so we submit it as a timeout() call to be short
12803  * as soon as possible.
12804  * The ch_initlck & ch_actvlck are used to synchronize activation
12805  * of the channel so that we don't have multiple idn_activate_channel's
12806  * attempting to activate the same channel.
12807  */
12808 static void
12809 idn_submit_chanactivate_job(int channel)
12810 {
12811 	idn_chansvr_t	*csp;
12812 
12813 	if (idn.chan_servers == NULL)
12814 		return;
12815 	csp = &idn.chan_servers[channel];
12816 
12817 	if (lock_try(&csp->ch_initlck) == 0)
12818 		return;
12819 
12820 	(void) timeout(idn_exec_chanactivate, (caddr_t)(uintptr_t)channel, 1);
12821 }
12822 
12823 /*ARGSUSED0*/
12824 static void
12825 idn_xmit_monitor(void *unused)
12826 {
12827 	int		c, d;
12828 	idn_chansvr_t	*csp;
12829 	idn_chanset_t	wake_set;
12830 	domainset_t	conset;
12831 	smr_slab_t	*sp;
12832 	procname_t	proc = "idn_xmit_monitor";
12833 
12834 	CHANSET_ZERO(wake_set);
12835 
12836 	mutex_enter(&idn.xmit_lock);
12837 	if ((idn.xmit_tid == NULL) || !idn.xmit_chanset_wanted) {
12838 		idn.xmit_tid = NULL;
12839 		mutex_exit(&idn.xmit_lock);
12840 		PR_XMON("%s: bailing out\n", proc);
12841 		return;
12842 	}
12843 
12844 	/*
12845 	 * No point in transmitting unless state
12846 	 * is ONLINE.
12847 	 */
12848 	if (idn.state != IDNGS_ONLINE)
12849 		goto retry;
12850 
12851 	conset = idn.domset.ds_connected;
12852 
12853 	/*
12854 	 * Try and reclaim some buffers if possible.
12855 	 */
12856 	for (d = 0; d < MAX_DOMAINS; d++) {
12857 		if (!DOMAIN_IN_SET(conset, d))
12858 			continue;
12859 
12860 		if (!IDN_DLOCK_TRY_SHARED(d))
12861 			continue;
12862 
12863 		if (idn_domain[d].dcpu != IDN_NIL_DCPU)
12864 			(void) idn_reclaim_mboxdata(d, 0, -1);
12865 
12866 		IDN_DUNLOCK(d);
12867 	}
12868 
12869 	/*
12870 	 * Now check if we were successful in getting
12871 	 * any buffers.
12872 	 */
12873 	DSLAB_LOCK_SHARED(idn.localid);
12874 	sp = idn_domain[idn.localid].dslab;
12875 	for (; sp; sp = sp->sl_next)
12876 		if (sp->sl_free)
12877 			break;
12878 	DSLAB_UNLOCK(idn.localid);
12879 
12880 	/*
12881 	 * If there are no buffers available,
12882 	 * no point in reenabling the queues.
12883 	 */
12884 	if (sp == NULL)
12885 		goto retry;
12886 
12887 	CHANSET_ZERO(wake_set);
12888 	for (c = 0; c < IDN_MAX_NETS; c++) {
12889 		int		pending_bits;
12890 		struct idn	*sip;
12891 
12892 		if (!CHAN_IN_SET(idn.xmit_chanset_wanted, c))
12893 			continue;
12894 
12895 		csp = &idn.chan_servers[c];
12896 		if (!IDN_CHAN_TRYLOCK_GLOBAL(csp))
12897 			continue;
12898 
12899 		pending_bits = csp->ch_state & IDN_CHANSVC_PENDING_BITS;
12900 
12901 		sip = IDN_INST2SIP(c);
12902 
12903 		if (!csp->ch_send.c_checkin &&
12904 		    (pending_bits == IDN_CHANSVC_PENDING_BITS) &&
12905 		    sip && (sip->si_flags & IDNRUNNING)) {
12906 
12907 			IDN_CHAN_UNLOCK_GLOBAL(csp);
12908 			CHANSET_ADD(wake_set, c);
12909 
12910 			PR_XMON("%s: QENABLE for channel %d\n", proc, c);
12911 
12912 			rw_enter(&idn.struprwlock, RW_READER);
12913 			mutex_enter(&idn.sipwenlock);
12914 			idndl_wenable(sip);
12915 			mutex_exit(&idn.sipwenlock);
12916 			rw_exit(&idn.struprwlock);
12917 		} else {
12918 			IDN_CHAN_UNLOCK_GLOBAL(csp);
12919 		}
12920 	}
12921 
12922 	/*
12923 	 * Clear the channels we enabled.
12924 	 */
12925 	idn.xmit_chanset_wanted &= ~wake_set;
12926 
12927 retry:
12928 
12929 	if (idn.xmit_chanset_wanted == 0)
12930 		idn.xmit_tid = NULL;
12931 	else
12932 		idn.xmit_tid = timeout(idn_xmit_monitor, NULL,
12933 		    idn_xmit_monitor_freq);
12934 
12935 	mutex_exit(&idn.xmit_lock);
12936 }
12937 
12938 void
12939 idn_xmit_monitor_kickoff(int chan_wanted)
12940 {
12941 	procname_t	proc = "idn_xmit_monitor_kickoff";
12942 
12943 	mutex_enter(&idn.xmit_lock);
12944 
12945 	if (chan_wanted < 0) {
12946 		/*
12947 		 * Wants all channels.
12948 		 */
12949 		idn.xmit_chanset_wanted = CHANSET_ALL;
12950 	} else {
12951 		CHANSET_ADD(idn.xmit_chanset_wanted, chan_wanted);
12952 	}
12953 
12954 	if (idn.xmit_tid != (timeout_id_t)NULL) {
12955 		/*
12956 		 * A monitor is already running, so
12957 		 * he will catch the new "wants" when
12958 		 * he comes around.
12959 		 */
12960 		mutex_exit(&idn.xmit_lock);
12961 		return;
12962 	}
12963 
12964 	PR_XMON("%s: xmit_mon kicked OFF (chanset = 0x%x)\n",
12965 	    proc, idn.xmit_chanset_wanted);
12966 
12967 	idn.xmit_tid = timeout(idn_xmit_monitor, NULL, idn_xmit_monitor_freq);
12968 
12969 	mutex_exit(&idn.xmit_lock);
12970 }
12971