xref: /titanic_41/usr/src/uts/common/rpc/clnt_cots.c (revision ef26a926e54f934431066685d069cbf73bba1534)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2009 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  */
25 
26 /*
27  * Copyright (c) 1983, 1984, 1985, 1986, 1987, 1988, 1989 AT&T
28  *		All Rights Reserved
29  */
30 
31 /*
32  * Portions of this source code were derived from Berkeley 4.3 BSD
33  * under license from the Regents of the University of California.
34  */
35 
36 
37 /*
38  * Implements a kernel based, client side RPC over Connection Oriented
39  * Transports (COTS).
40  */
41 
42 /*
43  * Much of this file has been re-written to let NFS work better over slow
44  * transports. A description follows.
45  *
46  * One of the annoying things about kRPC/COTS is that it will temporarily
47  * create more than one connection between a client and server. This
48  * happens because when a connection is made, the end-points entry in the
49  * linked list of connections (headed by cm_hd), is removed so that other
50  * threads don't mess with it. Went ahead and bit the bullet by keeping
51  * the endpoint on the connection list and introducing state bits,
52  * condition variables etc. to the connection entry data structure (struct
53  * cm_xprt).
54  *
55  * Here is a summary of the changes to cm-xprt:
56  *
57  *	x_ctime is the timestamp of when the endpoint was last
58  *	connected or disconnected. If an end-point is ever disconnected
59  *	or re-connected, then any outstanding RPC request is presumed
60  *	lost, telling clnt_cots_kcallit that it needs to re-send the
61  *	request, not just wait for the original request's reply to
62  *	arrive.
63  *
64  *	x_thread flag which tells us if a thread is doing a connection attempt.
65  *
66  *	x_waitdis flag which tells us we are waiting a disconnect ACK.
67  *
68  *	x_needdis flag which tells us we need to send a T_DISCONN_REQ
69  *	to kill the connection.
70  *
71  *	x_needrel flag which tells us we need to send a T_ORDREL_REQ to
72  *	gracefully close the connection.
73  *
74  *	#defined bitmasks for the all the b_* bits so that more
75  *	efficient (and at times less clumsy) masks can be used to
76  *	manipulated state in cases where multiple bits have to
77  *	set/cleared/checked in the same critical section.
78  *
79  *	x_conn_cv and x_dis-_cv are new condition variables to let
80  *	threads knows when the connection attempt is done, and to let
81  *	the connecting thread know when the disconnect handshake is
82  *	done.
83  *
84  * Added the CONN_HOLD() macro so that all reference holds have the same
85  * look and feel.
86  *
87  * In the private (cku_private) portion of the client handle,
88  *
89  *	cku_flags replaces the cku_sent a boolean. cku_flags keeps
90  *	track of whether a request as been sent, and whether the
91  *	client's handles call record is on the dispatch list (so that
92  *	the reply can be matched by XID to the right client handle).
93  *	The idea of CKU_ONQUEUE is that we can exit clnt_cots_kcallit()
94  *	and still have the response find the right client handle so
95  *	that the retry of CLNT_CALL() gets the result. Testing, found
96  *	situations where if the timeout was increased, performance
97  *	degraded. This was due to us hitting a window where the thread
98  *	was back in rfscall() (probably printing server not responding)
99  *	while the response came back but no place to put it.
100  *
101  *	cku_ctime is just a cache of x_ctime. If they match,
102  *	clnt_cots_kcallit() won't to send a retry (unless the maximum
103  *	receive count limit as been reached). If the don't match, then
104  *	we assume the request has been lost, and a retry of the request
105  *	is needed.
106  *
107  *	cku_recv_attempts counts the number of receive count attempts
108  *	after one try is sent on the wire.
109  *
110  * Added the clnt_delay() routine so that interruptible and
111  * noninterruptible delays are possible.
112  *
113  * CLNT_MIN_TIMEOUT has been bumped to 10 seconds from 3. This is used to
114  * control how long the client delays before returned after getting
115  * ECONNREFUSED. At 3 seconds, 8 client threads per mount really does bash
116  * a server that may be booting and not yet started nfsd.
117  *
118  * CLNT_MAXRECV_WITHOUT_RETRY is a new macro (value of 3) (with a tunable)
119  * Why don't we just wait forever (receive an infinite # of times)?
120  * Because the server may have rebooted. More insidious is that some
121  * servers (ours) will drop NFS/TCP requests in some cases. This is bad,
122  * but it is a reality.
123  *
124  * The case of a server doing orderly release really messes up the
125  * client's recovery, especially if the server's TCP implementation is
126  * buggy.  It was found was that the kRPC/COTS client was breaking some
127  * TPI rules, such as not waiting for the acknowledgement of a
128  * T_DISCON_REQ (hence the added case statements T_ERROR_ACK, T_OK_ACK and
129  * T_DISCON_REQ in clnt_dispatch_notifyall()).
130  *
131  * One of things that we've seen is that a kRPC TCP endpoint goes into
132  * TIMEWAIT and a thus a reconnect takes a long time to satisfy because
133  * that the TIMEWAIT state takes a while to finish.  If a server sends a
134  * T_ORDREL_IND, there is little point in an RPC client doing a
135  * T_ORDREL_REQ, because the RPC request isn't going to make it (the
136  * server is saying that it won't accept any more data). So kRPC was
137  * changed to send a T_DISCON_REQ when we get a T_ORDREL_IND. So now the
138  * connection skips the TIMEWAIT state and goes straight to a bound state
139  * that kRPC can quickly switch to connected.
140  *
141  * Code that issues TPI request must use waitforack() to wait for the
142  * corresponding ack (assuming there is one) in any future modifications.
143  * This works around problems that may be introduced by breaking TPI rules
144  * (by submitting new calls before earlier requests have been acked) in the
145  * case of a signal or other early return.  waitforack() depends on
146  * clnt_dispatch_notifyconn() to issue the wakeup when the ack
147  * arrives, so adding new TPI calls may require corresponding changes
148  * to clnt_dispatch_notifyconn(). Presently, the timeout period is based on
149  * CLNT_MIN_TIMEOUT which is 10 seconds. If you modify this value, be sure
150  * not to set it too low or TPI ACKS will be lost.
151  */
152 
153 #include <sys/param.h>
154 #include <sys/types.h>
155 #include <sys/user.h>
156 #include <sys/systm.h>
157 #include <sys/sysmacros.h>
158 #include <sys/proc.h>
159 #include <sys/socket.h>
160 #include <sys/file.h>
161 #include <sys/stream.h>
162 #include <sys/strsubr.h>
163 #include <sys/stropts.h>
164 #include <sys/strsun.h>
165 #include <sys/timod.h>
166 #include <sys/tiuser.h>
167 #include <sys/tihdr.h>
168 #include <sys/t_kuser.h>
169 #include <sys/fcntl.h>
170 #include <sys/errno.h>
171 #include <sys/kmem.h>
172 #include <sys/debug.h>
173 #include <sys/systm.h>
174 #include <sys/kstat.h>
175 #include <sys/t_lock.h>
176 #include <sys/ddi.h>
177 #include <sys/cmn_err.h>
178 #include <sys/time.h>
179 #include <sys/isa_defs.h>
180 #include <sys/callb.h>
181 #include <sys/sunddi.h>
182 #include <sys/atomic.h>
183 #include <sys/sdt.h>
184 
185 #include <netinet/in.h>
186 #include <netinet/tcp.h>
187 
188 #include <rpc/types.h>
189 #include <rpc/xdr.h>
190 #include <rpc/auth.h>
191 #include <rpc/clnt.h>
192 #include <rpc/rpc_msg.h>
193 
194 #define	COTS_DEFAULT_ALLOCSIZE	2048
195 
196 #define	WIRE_HDR_SIZE	20	/* serialized call header, sans proc number */
197 #define	MSG_OFFSET	128	/* offset of call into the mblk */
198 
199 const char *kinet_ntop6(uchar_t *, char *, size_t);
200 
201 static int	clnt_cots_ksettimers(CLIENT *, struct rpc_timers *,
202     struct rpc_timers *, int, void(*)(int, int, caddr_t), caddr_t, uint32_t);
203 static enum clnt_stat	clnt_cots_kcallit(CLIENT *, rpcproc_t, xdrproc_t,
204     caddr_t, xdrproc_t, caddr_t, struct timeval);
205 static void	clnt_cots_kabort(CLIENT *);
206 static void	clnt_cots_kerror(CLIENT *, struct rpc_err *);
207 static bool_t	clnt_cots_kfreeres(CLIENT *, xdrproc_t, caddr_t);
208 static void	clnt_cots_kdestroy(CLIENT *);
209 static bool_t	clnt_cots_kcontrol(CLIENT *, int, char *);
210 
211 
212 /* List of transports managed by the connection manager. */
213 struct cm_xprt {
214 	TIUSER		*x_tiptr;	/* transport handle */
215 	queue_t		*x_wq;		/* send queue */
216 	clock_t		x_time;		/* last time we handed this xprt out */
217 	clock_t		x_ctime;	/* time we went to CONNECTED */
218 	int		x_tidu_size;    /* TIDU size of this transport */
219 	union {
220 	    struct {
221 		unsigned int
222 #ifdef	_BIT_FIELDS_HTOL
223 		b_closing:	1,	/* we've sent a ord rel on this conn */
224 		b_dead:		1,	/* transport is closed or disconn */
225 		b_doomed:	1,	/* too many conns, let this go idle */
226 		b_connected:	1,	/* this connection is connected */
227 
228 		b_ordrel:	1,	/* do an orderly release? */
229 		b_thread:	1,	/* thread doing connect */
230 		b_waitdis:	1,	/* waiting for disconnect ACK */
231 		b_needdis:	1,	/* need T_DISCON_REQ */
232 
233 		b_needrel:	1,	/* need T_ORDREL_REQ */
234 		b_early_disc:	1,	/* got a T_ORDREL_IND or T_DISCON_IND */
235 					/* disconnect during connect */
236 
237 		b_pad:		22;
238 
239 #endif
240 
241 #ifdef	_BIT_FIELDS_LTOH
242 		b_pad:		22,
243 
244 		b_early_disc:	1,	/* got a T_ORDREL_IND or T_DISCON_IND */
245 					/* disconnect during connect */
246 		b_needrel:	1,	/* need T_ORDREL_REQ */
247 
248 		b_needdis:	1,	/* need T_DISCON_REQ */
249 		b_waitdis:	1,	/* waiting for disconnect ACK */
250 		b_thread:	1,	/* thread doing connect */
251 		b_ordrel:	1,	/* do an orderly release? */
252 
253 		b_connected:	1,	/* this connection is connected */
254 		b_doomed:	1,	/* too many conns, let this go idle */
255 		b_dead:		1,	/* transport is closed or disconn */
256 		b_closing:	1;	/* we've sent a ord rel on this conn */
257 #endif
258 	    } bit;	    unsigned int word;
259 
260 #define	x_closing	x_state.bit.b_closing
261 #define	x_dead		x_state.bit.b_dead
262 #define	x_doomed	x_state.bit.b_doomed
263 #define	x_connected	x_state.bit.b_connected
264 
265 #define	x_ordrel	x_state.bit.b_ordrel
266 #define	x_thread	x_state.bit.b_thread
267 #define	x_waitdis	x_state.bit.b_waitdis
268 #define	x_needdis	x_state.bit.b_needdis
269 
270 #define	x_needrel	x_state.bit.b_needrel
271 #define	x_early_disc    x_state.bit.b_early_disc
272 
273 #define	x_state_flags	x_state.word
274 
275 #define	X_CLOSING	0x80000000
276 #define	X_DEAD		0x40000000
277 #define	X_DOOMED	0x20000000
278 #define	X_CONNECTED	0x10000000
279 
280 #define	X_ORDREL	0x08000000
281 #define	X_THREAD	0x04000000
282 #define	X_WAITDIS	0x02000000
283 #define	X_NEEDDIS	0x01000000
284 
285 #define	X_NEEDREL	0x00800000
286 #define	X_EARLYDISC	0x00400000
287 
288 #define	X_BADSTATES	(X_CLOSING | X_DEAD | X_DOOMED)
289 
290 	}		x_state;
291 	int		x_ref;		/* number of users of this xprt */
292 	int		x_family;	/* address family of transport */
293 	dev_t		x_rdev;		/* device number of transport */
294 	struct cm_xprt	*x_next;
295 
296 	struct netbuf	x_server;	/* destination address */
297 	struct netbuf	x_src;		/* src address (for retries) */
298 	kmutex_t	x_lock;		/* lock on this entry */
299 	kcondvar_t	x_cv;		/* to signal when can be closed */
300 	kcondvar_t	x_conn_cv;	/* to signal when connection attempt */
301 					/* is complete */
302 	kstat_t		*x_ksp;
303 
304 	kcondvar_t	x_dis_cv;	/* to signal when disconnect attempt */
305 					/* is complete */
306 	zoneid_t	x_zoneid;	/* zone this xprt belongs to */
307 };
308 
309 typedef struct cm_kstat_xprt {
310 	kstat_named_t	x_wq;
311 	kstat_named_t	x_server;
312 	kstat_named_t	x_family;
313 	kstat_named_t	x_rdev;
314 	kstat_named_t	x_time;
315 	kstat_named_t	x_state;
316 	kstat_named_t	x_ref;
317 	kstat_named_t	x_port;
318 } cm_kstat_xprt_t;
319 
320 static cm_kstat_xprt_t cm_kstat_template = {
321 	{ "write_queue", KSTAT_DATA_UINT32 },
322 	{ "server",	KSTAT_DATA_STRING },
323 	{ "addr_family", KSTAT_DATA_UINT32 },
324 	{ "device",	KSTAT_DATA_UINT32 },
325 	{ "time_stamp",	KSTAT_DATA_UINT32 },
326 	{ "status",	KSTAT_DATA_UINT32 },
327 	{ "ref_count",	KSTAT_DATA_INT32 },
328 	{ "port",	KSTAT_DATA_UINT32 },
329 };
330 
331 /*
332  * The inverse of this is connmgr_release().
333  */
334 #define	CONN_HOLD(Cm_entry)	{\
335 	mutex_enter(&(Cm_entry)->x_lock);	\
336 	(Cm_entry)->x_ref++;	\
337 	mutex_exit(&(Cm_entry)->x_lock);	\
338 }
339 
340 
341 /*
342  * Private data per rpc handle.  This structure is allocated by
343  * clnt_cots_kcreate, and freed by clnt_cots_kdestroy.
344  */
345 typedef struct cku_private_s {
346 	CLIENT			cku_client;	/* client handle */
347 	calllist_t		cku_call;	/* for dispatching calls */
348 	struct rpc_err		cku_err;	/* error status */
349 
350 	struct netbuf		cku_srcaddr;	/* source address for retries */
351 	int			cku_addrfmly;  /* for binding port */
352 	struct netbuf		cku_addr;	/* remote address */
353 	dev_t			cku_device;	/* device to use */
354 	uint_t			cku_flags;
355 #define	CKU_ONQUEUE		0x1
356 #define	CKU_SENT		0x2
357 
358 	bool_t			cku_progress;	/* for CLSET_PROGRESS */
359 	uint32_t		cku_xid;	/* current XID */
360 	clock_t			cku_ctime;	/* time stamp of when */
361 						/* connection was created */
362 	uint_t			cku_recv_attempts;
363 	XDR			cku_outxdr;	/* xdr routine for output */
364 	XDR			cku_inxdr;	/* xdr routine for input */
365 	char			cku_rpchdr[WIRE_HDR_SIZE + 4];
366 						/* pre-serialized rpc header */
367 
368 	uint_t			cku_outbuflen;	/* default output mblk length */
369 	struct cred		*cku_cred;	/* credentials */
370 	bool_t			cku_nodelayonerr;
371 						/* for CLSET_NODELAYONERR */
372 	int			cku_useresvport; /* Use reserved port */
373 	struct rpc_cots_client	*cku_stats;	/* stats for zone */
374 } cku_private_t;
375 
376 static struct cm_xprt *connmgr_wrapconnect(struct cm_xprt *,
377 	const struct timeval *, struct netbuf *, int, struct netbuf *,
378 	struct rpc_err *, bool_t, bool_t, cred_t *);
379 
380 static bool_t	connmgr_connect(struct cm_xprt *, queue_t *, struct netbuf *,
381 				int, calllist_t *, int *, bool_t reconnect,
382 				const struct timeval *, bool_t, cred_t *);
383 
384 static void	*connmgr_opt_getoff(mblk_t *mp, t_uscalar_t offset,
385 				t_uscalar_t length, uint_t align_size);
386 static bool_t	connmgr_setbufsz(calllist_t *e, queue_t *wq, cred_t *cr);
387 static bool_t	connmgr_getopt_int(queue_t *wq, int level, int name, int *val,
388 				calllist_t *e, cred_t *cr);
389 static bool_t	connmgr_setopt_int(queue_t *wq, int level, int name, int val,
390 				calllist_t *e, cred_t *cr);
391 static bool_t	connmgr_setopt(queue_t *, int, int, calllist_t *, cred_t *cr);
392 static void	connmgr_sndrel(struct cm_xprt *);
393 static void	connmgr_snddis(struct cm_xprt *);
394 static void	connmgr_close(struct cm_xprt *);
395 static void	connmgr_release(struct cm_xprt *);
396 static struct cm_xprt *connmgr_wrapget(struct netbuf *, const struct timeval *,
397 	cku_private_t *);
398 
399 static struct cm_xprt *connmgr_get(struct netbuf *, const struct timeval *,
400 	struct netbuf *, int, struct netbuf *, struct rpc_err *, dev_t,
401 	bool_t, int, cred_t *);
402 
403 static void connmgr_cancelconn(struct cm_xprt *);
404 static enum clnt_stat connmgr_cwait(struct cm_xprt *, const struct timeval *,
405 	bool_t);
406 static void connmgr_dis_and_wait(struct cm_xprt *);
407 
408 static int	clnt_dispatch_send(queue_t *, mblk_t *, calllist_t *, uint_t,
409 					uint_t);
410 
411 static int clnt_delay(clock_t, bool_t);
412 
413 static int waitforack(calllist_t *, t_scalar_t, const struct timeval *, bool_t);
414 
415 /*
416  * Operations vector for TCP/IP based RPC
417  */
418 static struct clnt_ops tcp_ops = {
419 	clnt_cots_kcallit,	/* do rpc call */
420 	clnt_cots_kabort,	/* abort call */
421 	clnt_cots_kerror,	/* return error status */
422 	clnt_cots_kfreeres,	/* free results */
423 	clnt_cots_kdestroy,	/* destroy rpc handle */
424 	clnt_cots_kcontrol,	/* the ioctl() of rpc */
425 	clnt_cots_ksettimers,	/* set retry timers */
426 };
427 
428 static int rpc_kstat_instance = 0;  /* keeps the current instance */
429 				/* number for the next kstat_create */
430 
431 static struct cm_xprt *cm_hd = NULL;
432 static kmutex_t connmgr_lock;	/* for connection mngr's list of transports */
433 
434 extern kmutex_t clnt_max_msg_lock;
435 
436 static calllist_t *clnt_pending = NULL;
437 extern kmutex_t clnt_pending_lock;
438 
439 static int clnt_cots_hash_size = DEFAULT_HASH_SIZE;
440 
441 static call_table_t *cots_call_ht;
442 
443 static const struct rpc_cots_client {
444 	kstat_named_t	rccalls;
445 	kstat_named_t	rcbadcalls;
446 	kstat_named_t	rcbadxids;
447 	kstat_named_t	rctimeouts;
448 	kstat_named_t	rcnewcreds;
449 	kstat_named_t	rcbadverfs;
450 	kstat_named_t	rctimers;
451 	kstat_named_t	rccantconn;
452 	kstat_named_t	rcnomem;
453 	kstat_named_t	rcintrs;
454 } cots_rcstat_tmpl = {
455 	{ "calls",	KSTAT_DATA_UINT64 },
456 	{ "badcalls",	KSTAT_DATA_UINT64 },
457 	{ "badxids",	KSTAT_DATA_UINT64 },
458 	{ "timeouts",	KSTAT_DATA_UINT64 },
459 	{ "newcreds",	KSTAT_DATA_UINT64 },
460 	{ "badverfs",	KSTAT_DATA_UINT64 },
461 	{ "timers",	KSTAT_DATA_UINT64 },
462 	{ "cantconn",	KSTAT_DATA_UINT64 },
463 	{ "nomem",	KSTAT_DATA_UINT64 },
464 	{ "interrupts", KSTAT_DATA_UINT64 }
465 };
466 
467 #define	COTSRCSTAT_INCR(p, x)	\
468 	atomic_inc_64(&(p)->x.value.ui64)
469 
470 #define	CLNT_MAX_CONNS	1	/* concurrent connections between clnt/srvr */
471 int clnt_max_conns = CLNT_MAX_CONNS;
472 
473 #define	CLNT_MIN_TIMEOUT	10	/* seconds to wait after we get a */
474 					/* connection reset */
475 #define	CLNT_MIN_CONNTIMEOUT	5	/* seconds to wait for a connection */
476 
477 
478 int clnt_cots_min_tout = CLNT_MIN_TIMEOUT;
479 int clnt_cots_min_conntout = CLNT_MIN_CONNTIMEOUT;
480 
481 /*
482  * Limit the number of times we will attempt to receive a reply without
483  * re-sending a response.
484  */
485 #define	CLNT_MAXRECV_WITHOUT_RETRY	3
486 uint_t clnt_cots_maxrecv	= CLNT_MAXRECV_WITHOUT_RETRY;
487 
488 uint_t *clnt_max_msg_sizep;
489 void (*clnt_stop_idle)(queue_t *wq);
490 
491 #define	ptoh(p)		(&((p)->cku_client))
492 #define	htop(h)		((cku_private_t *)((h)->cl_private))
493 
494 /*
495  * Times to retry
496  */
497 #define	REFRESHES	2	/* authentication refreshes */
498 
499 /*
500  * The following is used to determine the global default behavior for
501  * COTS when binding to a local port.
502  *
503  * If the value is set to 1 the default will be to select a reserved
504  * (aka privileged) port, if the value is zero the default will be to
505  * use non-reserved ports.  Users of kRPC may override this by using
506  * CLNT_CONTROL() and CLSET_BINDRESVPORT.
507  */
508 int clnt_cots_do_bindresvport = 1;
509 
510 static zone_key_t zone_cots_key;
511 
512 /*
513  * Defaults TCP send and receive buffer size for RPC connections.
514  * These values can be tuned by /etc/system.
515  */
516 int rpc_send_bufsz = 1024*1024;
517 int rpc_recv_bufsz = 1024*1024;
518 /*
519  * To use system-wide default for TCP send and receive buffer size,
520  * use /etc/system to set rpc_default_tcp_bufsz to 1:
521  *
522  * set rpcmod:rpc_default_tcp_bufsz=1
523  */
524 int rpc_default_tcp_bufsz = 0;
525 
526 /*
527  * We need to do this after all kernel threads in the zone have exited.
528  */
529 /* ARGSUSED */
530 static void
531 clnt_zone_destroy(zoneid_t zoneid, void *unused)
532 {
533 	struct cm_xprt **cmp;
534 	struct cm_xprt *cm_entry;
535 	struct cm_xprt *freelist = NULL;
536 
537 	mutex_enter(&connmgr_lock);
538 	cmp = &cm_hd;
539 	while ((cm_entry = *cmp) != NULL) {
540 		if (cm_entry->x_zoneid == zoneid) {
541 			*cmp = cm_entry->x_next;
542 			cm_entry->x_next = freelist;
543 			freelist = cm_entry;
544 		} else {
545 			cmp = &cm_entry->x_next;
546 		}
547 	}
548 	mutex_exit(&connmgr_lock);
549 	while ((cm_entry = freelist) != NULL) {
550 		freelist = cm_entry->x_next;
551 		connmgr_close(cm_entry);
552 	}
553 }
554 
555 int
556 clnt_cots_kcreate(dev_t dev, struct netbuf *addr, int family, rpcprog_t prog,
557 	rpcvers_t vers, uint_t max_msgsize, cred_t *cred, CLIENT **ncl)
558 {
559 	CLIENT *h;
560 	cku_private_t *p;
561 	struct rpc_msg call_msg;
562 	struct rpcstat *rpcstat;
563 
564 	RPCLOG(8, "clnt_cots_kcreate: prog %u\n", prog);
565 
566 	rpcstat = zone_getspecific(rpcstat_zone_key, rpc_zone());
567 	ASSERT(rpcstat != NULL);
568 
569 	/* Allocate and intialize the client handle. */
570 	p = kmem_zalloc(sizeof (*p), KM_SLEEP);
571 
572 	h = ptoh(p);
573 
574 	h->cl_private = (caddr_t)p;
575 	h->cl_auth = authkern_create();
576 	h->cl_ops = &tcp_ops;
577 
578 	cv_init(&p->cku_call.call_cv, NULL, CV_DEFAULT, NULL);
579 	mutex_init(&p->cku_call.call_lock, NULL, MUTEX_DEFAULT, NULL);
580 
581 	/*
582 	 * If the current sanity check size in rpcmod is smaller
583 	 * than the size needed, then increase the sanity check.
584 	 */
585 	if (max_msgsize != 0 && clnt_max_msg_sizep != NULL &&
586 	    max_msgsize > *clnt_max_msg_sizep) {
587 		mutex_enter(&clnt_max_msg_lock);
588 		if (max_msgsize > *clnt_max_msg_sizep)
589 			*clnt_max_msg_sizep = max_msgsize;
590 		mutex_exit(&clnt_max_msg_lock);
591 	}
592 
593 	p->cku_outbuflen = COTS_DEFAULT_ALLOCSIZE;
594 
595 	/* Preserialize the call message header */
596 
597 	call_msg.rm_xid = 0;
598 	call_msg.rm_direction = CALL;
599 	call_msg.rm_call.cb_rpcvers = RPC_MSG_VERSION;
600 	call_msg.rm_call.cb_prog = prog;
601 	call_msg.rm_call.cb_vers = vers;
602 
603 	xdrmem_create(&p->cku_outxdr, p->cku_rpchdr, WIRE_HDR_SIZE, XDR_ENCODE);
604 
605 	if (!xdr_callhdr(&p->cku_outxdr, &call_msg)) {
606 		RPCLOG0(1, "clnt_cots_kcreate - Fatal header serialization "
607 		    "error\n");
608 		auth_destroy(h->cl_auth);
609 		kmem_free(p, sizeof (cku_private_t));
610 		RPCLOG0(1, "clnt_cots_kcreate: create failed error EINVAL\n");
611 		return (EINVAL);		/* XXX */
612 	}
613 
614 	/*
615 	 * The zalloc initialized the fields below.
616 	 * p->cku_xid = 0;
617 	 * p->cku_flags = 0;
618 	 * p->cku_srcaddr.len = 0;
619 	 * p->cku_srcaddr.maxlen = 0;
620 	 */
621 
622 	p->cku_cred = cred;
623 	p->cku_device = dev;
624 	p->cku_addrfmly = family;
625 	p->cku_addr.buf = kmem_zalloc(addr->maxlen, KM_SLEEP);
626 	p->cku_addr.maxlen = addr->maxlen;
627 	p->cku_addr.len = addr->len;
628 	bcopy(addr->buf, p->cku_addr.buf, addr->len);
629 	p->cku_stats = rpcstat->rpc_cots_client;
630 	p->cku_useresvport = -1; /* value is has not been set */
631 
632 	*ncl = h;
633 	return (0);
634 }
635 
636 /*ARGSUSED*/
637 static void
638 clnt_cots_kabort(CLIENT *h)
639 {
640 }
641 
642 /*
643  * Return error info on this handle.
644  */
645 static void
646 clnt_cots_kerror(CLIENT *h, struct rpc_err *err)
647 {
648 	/* LINTED pointer alignment */
649 	cku_private_t *p = htop(h);
650 
651 	*err = p->cku_err;
652 }
653 
654 static bool_t
655 clnt_cots_kfreeres(CLIENT *h, xdrproc_t xdr_res, caddr_t res_ptr)
656 {
657 	/* LINTED pointer alignment */
658 	cku_private_t *p = htop(h);
659 	XDR *xdrs;
660 
661 	xdrs = &(p->cku_outxdr);
662 	xdrs->x_op = XDR_FREE;
663 	return ((*xdr_res)(xdrs, res_ptr));
664 }
665 
666 static bool_t
667 clnt_cots_kcontrol(CLIENT *h, int cmd, char *arg)
668 {
669 	cku_private_t *p = htop(h);
670 
671 	switch (cmd) {
672 	case CLSET_PROGRESS:
673 		p->cku_progress = TRUE;
674 		return (TRUE);
675 
676 	case CLSET_XID:
677 		if (arg == NULL)
678 			return (FALSE);
679 
680 		p->cku_xid = *((uint32_t *)arg);
681 		return (TRUE);
682 
683 	case CLGET_XID:
684 		if (arg == NULL)
685 			return (FALSE);
686 
687 		*((uint32_t *)arg) = p->cku_xid;
688 		return (TRUE);
689 
690 	case CLSET_NODELAYONERR:
691 		if (arg == NULL)
692 			return (FALSE);
693 
694 		if (*((bool_t *)arg) == TRUE) {
695 			p->cku_nodelayonerr = TRUE;
696 			return (TRUE);
697 		}
698 		if (*((bool_t *)arg) == FALSE) {
699 			p->cku_nodelayonerr = FALSE;
700 			return (TRUE);
701 		}
702 		return (FALSE);
703 
704 	case CLGET_NODELAYONERR:
705 		if (arg == NULL)
706 			return (FALSE);
707 
708 		*((bool_t *)arg) = p->cku_nodelayonerr;
709 		return (TRUE);
710 
711 	case CLSET_BINDRESVPORT:
712 		if (arg == NULL)
713 			return (FALSE);
714 
715 		if (*(int *)arg != 1 && *(int *)arg != 0)
716 			return (FALSE);
717 
718 		p->cku_useresvport = *(int *)arg;
719 
720 		return (TRUE);
721 
722 	case CLGET_BINDRESVPORT:
723 		if (arg == NULL)
724 			return (FALSE);
725 
726 		*(int *)arg = p->cku_useresvport;
727 
728 		return (TRUE);
729 
730 	default:
731 		return (FALSE);
732 	}
733 }
734 
735 /*
736  * Destroy rpc handle.  Frees the space used for output buffer,
737  * private data, and handle structure.
738  */
739 static void
740 clnt_cots_kdestroy(CLIENT *h)
741 {
742 	/* LINTED pointer alignment */
743 	cku_private_t *p = htop(h);
744 	calllist_t *call = &p->cku_call;
745 
746 	RPCLOG(8, "clnt_cots_kdestroy h: %p\n", (void *)h);
747 	RPCLOG(8, "clnt_cots_kdestroy h: xid=0x%x\n", p->cku_xid);
748 
749 	if (p->cku_flags & CKU_ONQUEUE) {
750 		RPCLOG(64, "clnt_cots_kdestroy h: removing call for xid 0x%x "
751 		    "from dispatch list\n", p->cku_xid);
752 		call_table_remove(call);
753 	}
754 
755 	if (call->call_reply)
756 		freemsg(call->call_reply);
757 	cv_destroy(&call->call_cv);
758 	mutex_destroy(&call->call_lock);
759 
760 	kmem_free(p->cku_srcaddr.buf, p->cku_srcaddr.maxlen);
761 	kmem_free(p->cku_addr.buf, p->cku_addr.maxlen);
762 	kmem_free(p, sizeof (*p));
763 }
764 
765 static int clnt_cots_pulls;
766 #define	RM_HDR_SIZE	4	/* record mark header size */
767 
768 /*
769  * Call remote procedure.
770  */
771 static enum clnt_stat
772 clnt_cots_kcallit(CLIENT *h, rpcproc_t procnum, xdrproc_t xdr_args,
773     caddr_t argsp, xdrproc_t xdr_results, caddr_t resultsp, struct timeval wait)
774 {
775 	/* LINTED pointer alignment */
776 	cku_private_t *p = htop(h);
777 	calllist_t *call = &p->cku_call;
778 	XDR *xdrs;
779 	struct rpc_msg reply_msg;
780 	mblk_t *mp;
781 #ifdef	RPCDEBUG
782 	clock_t time_sent;
783 #endif
784 	struct netbuf *retryaddr;
785 	struct cm_xprt *cm_entry = NULL;
786 	queue_t *wq;
787 	int len, waitsecs, max_waitsecs;
788 	int mpsize;
789 	int refreshes = REFRESHES;
790 	int interrupted;
791 	int tidu_size;
792 	enum clnt_stat status;
793 	struct timeval cwait;
794 	bool_t delay_first = FALSE;
795 	clock_t ticks, now;
796 
797 	RPCLOG(2, "clnt_cots_kcallit, procnum %u\n", procnum);
798 	COTSRCSTAT_INCR(p->cku_stats, rccalls);
799 
800 	RPCLOG(2, "clnt_cots_kcallit: wait.tv_sec: %ld\n", wait.tv_sec);
801 	RPCLOG(2, "clnt_cots_kcallit: wait.tv_usec: %ld\n", wait.tv_usec);
802 	/*
803 	 * Bug ID 1240234:
804 	 * Look out for zero length timeouts. We don't want to
805 	 * wait zero seconds for a connection to be established.
806 	 */
807 	if (wait.tv_sec < clnt_cots_min_conntout) {
808 		cwait.tv_sec = clnt_cots_min_conntout;
809 		cwait.tv_usec = 0;
810 		RPCLOG(8, "clnt_cots_kcallit: wait.tv_sec (%ld) too low,",
811 		    wait.tv_sec);
812 		RPCLOG(8, " setting to: %d\n", clnt_cots_min_conntout);
813 	} else {
814 		cwait = wait;
815 	}
816 
817 call_again:
818 	if (cm_entry) {
819 		connmgr_release(cm_entry);
820 		cm_entry = NULL;
821 	}
822 
823 	mp = NULL;
824 
825 	/*
826 	 * If the call is not a retry, allocate a new xid and cache it
827 	 * for future retries.
828 	 * Bug ID 1246045:
829 	 * Treat call as a retry for purposes of binding the source
830 	 * port only if we actually attempted to send anything on
831 	 * the previous call.
832 	 */
833 	if (p->cku_xid == 0) {
834 		p->cku_xid = alloc_xid();
835 		call->call_zoneid = rpc_zoneid();
836 
837 		/*
838 		 * We need to ASSERT here that our xid != 0 because this
839 		 * determines whether or not our call record gets placed on
840 		 * the hash table or the linked list.  By design, we mandate
841 		 * that RPC calls over cots must have xid's != 0, so we can
842 		 * ensure proper management of the hash table.
843 		 */
844 		ASSERT(p->cku_xid != 0);
845 
846 		retryaddr = NULL;
847 		p->cku_flags &= ~CKU_SENT;
848 
849 		if (p->cku_flags & CKU_ONQUEUE) {
850 			RPCLOG(8, "clnt_cots_kcallit: new call, dequeuing old"
851 			    " one (%p)\n", (void *)call);
852 			call_table_remove(call);
853 			p->cku_flags &= ~CKU_ONQUEUE;
854 			RPCLOG(64, "clnt_cots_kcallit: removing call from "
855 			    "dispatch list because xid was zero (now 0x%x)\n",
856 			    p->cku_xid);
857 		}
858 
859 		if (call->call_reply != NULL) {
860 			freemsg(call->call_reply);
861 			call->call_reply = NULL;
862 		}
863 	} else if (p->cku_srcaddr.buf == NULL || p->cku_srcaddr.len == 0) {
864 		retryaddr = NULL;
865 
866 	} else if (p->cku_flags & CKU_SENT) {
867 		retryaddr = &p->cku_srcaddr;
868 
869 	} else {
870 		/*
871 		 * Bug ID 1246045: Nothing was sent, so set retryaddr to
872 		 * NULL and let connmgr_get() bind to any source port it
873 		 * can get.
874 		 */
875 		retryaddr = NULL;
876 	}
877 
878 	RPCLOG(64, "clnt_cots_kcallit: xid = 0x%x", p->cku_xid);
879 	RPCLOG(64, " flags = 0x%x\n", p->cku_flags);
880 
881 	p->cku_err.re_status = RPC_TIMEDOUT;
882 	p->cku_err.re_errno = p->cku_err.re_terrno = 0;
883 
884 	cm_entry = connmgr_wrapget(retryaddr, &cwait, p);
885 
886 	if (cm_entry == NULL) {
887 		RPCLOG(1, "clnt_cots_kcallit: can't connect status %s\n",
888 		    clnt_sperrno(p->cku_err.re_status));
889 
890 		/*
891 		 * The reasons why we fail to create a connection are
892 		 * varied. In most cases we don't want the caller to
893 		 * immediately retry. This could have one or more
894 		 * bad effects. This includes flooding the net with
895 		 * connect requests to ports with no listener; a hard
896 		 * kernel loop due to all the "reserved" TCP ports being
897 		 * in use.
898 		 */
899 		delay_first = TRUE;
900 
901 		/*
902 		 * Even if we end up returning EINTR, we still count a
903 		 * a "can't connect", because the connection manager
904 		 * might have been committed to waiting for or timing out on
905 		 * a connection.
906 		 */
907 		COTSRCSTAT_INCR(p->cku_stats, rccantconn);
908 		switch (p->cku_err.re_status) {
909 		case RPC_INTR:
910 			p->cku_err.re_errno = EINTR;
911 
912 			/*
913 			 * No need to delay because a UNIX signal(2)
914 			 * interrupted us. The caller likely won't
915 			 * retry the CLNT_CALL() and even if it does,
916 			 * we assume the caller knows what it is doing.
917 			 */
918 			delay_first = FALSE;
919 			break;
920 
921 		case RPC_TIMEDOUT:
922 			p->cku_err.re_errno = ETIMEDOUT;
923 
924 			/*
925 			 * No need to delay because timed out already
926 			 * on the connection request and assume that the
927 			 * transport time out is longer than our minimum
928 			 * timeout, or least not too much smaller.
929 			 */
930 			delay_first = FALSE;
931 			break;
932 
933 		case RPC_SYSTEMERROR:
934 		case RPC_TLIERROR:
935 			/*
936 			 * We want to delay here because a transient
937 			 * system error has a better chance of going away
938 			 * if we delay a bit. If it's not transient, then
939 			 * we don't want end up in a hard kernel loop
940 			 * due to retries.
941 			 */
942 			ASSERT(p->cku_err.re_errno != 0);
943 			break;
944 
945 
946 		case RPC_CANTCONNECT:
947 			/*
948 			 * RPC_CANTCONNECT is set on T_ERROR_ACK which
949 			 * implies some error down in the TCP layer or
950 			 * below. If cku_nodelayonerror is set then we
951 			 * assume the caller knows not to try too hard.
952 			 */
953 			RPCLOG0(8, "clnt_cots_kcallit: connection failed,");
954 			RPCLOG0(8, " re_status=RPC_CANTCONNECT,");
955 			RPCLOG(8, " re_errno=%d,", p->cku_err.re_errno);
956 			RPCLOG(8, " cku_nodelayonerr=%d", p->cku_nodelayonerr);
957 			if (p->cku_nodelayonerr == TRUE)
958 				delay_first = FALSE;
959 
960 			p->cku_err.re_errno = EIO;
961 
962 			break;
963 
964 		case RPC_XPRTFAILED:
965 			/*
966 			 * We want to delay here because we likely
967 			 * got a refused connection.
968 			 */
969 			if (p->cku_err.re_errno == 0)
970 				p->cku_err.re_errno = EIO;
971 
972 			RPCLOG(1, "clnt_cots_kcallit: transport failed: %d\n",
973 			    p->cku_err.re_errno);
974 
975 			break;
976 
977 		default:
978 			/*
979 			 * We delay here because it is better to err
980 			 * on the side of caution. If we got here then
981 			 * status could have been RPC_SUCCESS, but we
982 			 * know that we did not get a connection, so
983 			 * force the rpc status to RPC_CANTCONNECT.
984 			 */
985 			p->cku_err.re_status = RPC_CANTCONNECT;
986 			p->cku_err.re_errno = EIO;
987 			break;
988 		}
989 		if (delay_first == TRUE)
990 			ticks = clnt_cots_min_tout * drv_usectohz(1000000);
991 		goto cots_done;
992 	}
993 
994 	/*
995 	 * If we've never sent any request on this connection (send count
996 	 * is zero, or the connection has been reset), cache the
997 	 * the connection's create time and send a request (possibly a retry)
998 	 */
999 	if ((p->cku_flags & CKU_SENT) == 0 ||
1000 	    p->cku_ctime != cm_entry->x_ctime) {
1001 		p->cku_ctime = cm_entry->x_ctime;
1002 
1003 	} else if ((p->cku_flags & CKU_SENT) && (p->cku_flags & CKU_ONQUEUE) &&
1004 	    (call->call_reply != NULL ||
1005 	    p->cku_recv_attempts < clnt_cots_maxrecv)) {
1006 
1007 		/*
1008 		 * If we've sent a request and our call is on the dispatch
1009 		 * queue and we haven't made too many receive attempts, then
1010 		 * don't re-send, just receive.
1011 		 */
1012 		p->cku_recv_attempts++;
1013 		goto read_again;
1014 	}
1015 
1016 	/*
1017 	 * Now we create the RPC request in a STREAMS message.  We have to do
1018 	 * this after the call to connmgr_get so that we have the correct
1019 	 * TIDU size for the transport.
1020 	 */
1021 	tidu_size = cm_entry->x_tidu_size;
1022 	len = MSG_OFFSET + MAX(tidu_size, RM_HDR_SIZE + WIRE_HDR_SIZE);
1023 
1024 	while ((mp = allocb(len, BPRI_MED)) == NULL) {
1025 		if (strwaitbuf(len, BPRI_MED)) {
1026 			p->cku_err.re_status = RPC_SYSTEMERROR;
1027 			p->cku_err.re_errno = ENOSR;
1028 			COTSRCSTAT_INCR(p->cku_stats, rcnomem);
1029 			goto cots_done;
1030 		}
1031 	}
1032 	xdrs = &p->cku_outxdr;
1033 	xdrmblk_init(xdrs, mp, XDR_ENCODE, tidu_size);
1034 	mpsize = MBLKSIZE(mp);
1035 	ASSERT(mpsize >= len);
1036 	ASSERT(mp->b_rptr == mp->b_datap->db_base);
1037 
1038 	/*
1039 	 * If the size of mblk is not appreciably larger than what we
1040 	 * asked, then resize the mblk to exactly len bytes. The reason for
1041 	 * this: suppose len is 1600 bytes, the tidu is 1460 bytes
1042 	 * (from TCP over ethernet), and the arguments to the RPC require
1043 	 * 2800 bytes. Ideally we want the protocol to render two
1044 	 * ~1400 byte segments over the wire. However if allocb() gives us a 2k
1045 	 * mblk, and we allocate a second mblk for the remainder, the protocol
1046 	 * module may generate 3 segments over the wire:
1047 	 * 1460 bytes for the first, 448 (2048 - 1600) for the second, and
1048 	 * 892 for the third. If we "waste" 448 bytes in the first mblk,
1049 	 * the XDR encoding will generate two ~1400 byte mblks, and the
1050 	 * protocol module is more likely to produce properly sized segments.
1051 	 */
1052 	if ((mpsize >> 1) <= len)
1053 		mp->b_rptr += (mpsize - len);
1054 
1055 	/*
1056 	 * Adjust b_rptr to reserve space for the non-data protocol headers
1057 	 * any downstream modules might like to add, and for the
1058 	 * record marking header.
1059 	 */
1060 	mp->b_rptr += (MSG_OFFSET + RM_HDR_SIZE);
1061 
1062 	if (h->cl_auth->ah_cred.oa_flavor != RPCSEC_GSS) {
1063 		/* Copy in the preserialized RPC header information. */
1064 		bcopy(p->cku_rpchdr, mp->b_rptr, WIRE_HDR_SIZE);
1065 
1066 		/* Use XDR_SETPOS() to set the b_wptr to past the RPC header. */
1067 		XDR_SETPOS(xdrs, (uint_t)(mp->b_rptr - mp->b_datap->db_base +
1068 		    WIRE_HDR_SIZE));
1069 
1070 		ASSERT((mp->b_wptr - mp->b_rptr) == WIRE_HDR_SIZE);
1071 
1072 		/* Serialize the procedure number and the arguments. */
1073 		if ((!XDR_PUTINT32(xdrs, (int32_t *)&procnum)) ||
1074 		    (!AUTH_MARSHALL(h->cl_auth, xdrs, p->cku_cred)) ||
1075 		    (!(*xdr_args)(xdrs, argsp))) {
1076 			p->cku_err.re_status = RPC_CANTENCODEARGS;
1077 			p->cku_err.re_errno = EIO;
1078 			goto cots_done;
1079 		}
1080 
1081 		(*(uint32_t *)(mp->b_rptr)) = p->cku_xid;
1082 	} else {
1083 		uint32_t *uproc = (uint32_t *)&p->cku_rpchdr[WIRE_HDR_SIZE];
1084 		IXDR_PUT_U_INT32(uproc, procnum);
1085 
1086 		(*(uint32_t *)(&p->cku_rpchdr[0])) = p->cku_xid;
1087 
1088 		/* Use XDR_SETPOS() to set the b_wptr. */
1089 		XDR_SETPOS(xdrs, (uint_t)(mp->b_rptr - mp->b_datap->db_base));
1090 
1091 		/* Serialize the procedure number and the arguments. */
1092 		if (!AUTH_WRAP(h->cl_auth, p->cku_rpchdr, WIRE_HDR_SIZE+4,
1093 		    xdrs, xdr_args, argsp)) {
1094 			p->cku_err.re_status = RPC_CANTENCODEARGS;
1095 			p->cku_err.re_errno = EIO;
1096 			goto cots_done;
1097 		}
1098 	}
1099 
1100 	RPCLOG(2, "clnt_cots_kcallit: connected, sending call, tidu_size %d\n",
1101 	    tidu_size);
1102 
1103 	wq = cm_entry->x_wq;
1104 	waitsecs = 0;
1105 
1106 dispatch_again:
1107 	status = clnt_dispatch_send(wq, mp, call, p->cku_xid,
1108 	    (p->cku_flags & CKU_ONQUEUE));
1109 
1110 	if ((status == RPC_CANTSEND) && (call->call_reason == ENOBUFS)) {
1111 		/*
1112 		 * QFULL condition, allow some time for queue to drain
1113 		 * and try again. Give up after waiting for all timeout
1114 		 * specified for the call, or zone is going away.
1115 		 */
1116 		max_waitsecs = wait.tv_sec ? wait.tv_sec : clnt_cots_min_tout;
1117 		if ((waitsecs++ < max_waitsecs) &&
1118 		    !(zone_status_get(curproc->p_zone) >=
1119 		    ZONE_IS_SHUTTING_DOWN)) {
1120 
1121 			/* wait 1 sec for queue to drain */
1122 			if (clnt_delay(drv_usectohz(1000000),
1123 			    h->cl_nosignal) == EINTR) {
1124 				p->cku_err.re_errno = EINTR;
1125 				p->cku_err.re_status = RPC_INTR;
1126 
1127 				goto cots_done;
1128 			}
1129 
1130 			/* and try again */
1131 			goto dispatch_again;
1132 		}
1133 		p->cku_err.re_status = status;
1134 		p->cku_err.re_errno = call->call_reason;
1135 		DTRACE_PROBE(krpc__e__clntcots__kcallit__cantsend);
1136 
1137 		goto cots_done;
1138 	}
1139 
1140 	if (waitsecs) {
1141 		/* adjust timeout to account for time wait to send */
1142 		wait.tv_sec -= waitsecs;
1143 		if (wait.tv_sec < 0) {
1144 			/* pick up reply on next retry */
1145 			wait.tv_sec = 0;
1146 		}
1147 		DTRACE_PROBE2(clnt_cots__sendwait, CLIENT *, h,
1148 		    int, waitsecs);
1149 	}
1150 
1151 	RPCLOG(64, "clnt_cots_kcallit: sent call for xid 0x%x\n",
1152 	    (uint_t)p->cku_xid);
1153 	p->cku_flags = (CKU_ONQUEUE|CKU_SENT);
1154 	p->cku_recv_attempts = 1;
1155 
1156 #ifdef	RPCDEBUG
1157 	time_sent = ddi_get_lbolt();
1158 #endif
1159 
1160 	/*
1161 	 * Wait for a reply or a timeout.  If there is no error or timeout,
1162 	 * (both indicated by call_status), call->call_reply will contain
1163 	 * the RPC reply message.
1164 	 */
1165 read_again:
1166 	mutex_enter(&call->call_lock);
1167 	interrupted = 0;
1168 	if (call->call_status == RPC_TIMEDOUT) {
1169 		/*
1170 		 * Indicate that the lwp is not to be stopped while waiting
1171 		 * for this network traffic.  This is to avoid deadlock while
1172 		 * debugging a process via /proc and also to avoid recursive
1173 		 * mutex_enter()s due to NFS page faults while stopping
1174 		 * (NFS holds locks when it calls here).
1175 		 */
1176 		clock_t cv_wait_ret;
1177 		clock_t timout;
1178 		clock_t oldlbolt;
1179 
1180 		klwp_t *lwp = ttolwp(curthread);
1181 
1182 		if (lwp != NULL)
1183 			lwp->lwp_nostop++;
1184 
1185 		oldlbolt = ddi_get_lbolt();
1186 		timout = wait.tv_sec * drv_usectohz(1000000) +
1187 		    drv_usectohz(wait.tv_usec) + oldlbolt;
1188 		/*
1189 		 * Iterate until the call_status is changed to something
1190 		 * other that RPC_TIMEDOUT, or if cv_timedwait_sig() returns
1191 		 * something <=0 zero. The latter means that we timed
1192 		 * out.
1193 		 */
1194 		if (h->cl_nosignal)
1195 			while ((cv_wait_ret = cv_timedwait(&call->call_cv,
1196 			    &call->call_lock, timout)) > 0 &&
1197 			    call->call_status == RPC_TIMEDOUT)
1198 				;
1199 		else
1200 			while ((cv_wait_ret = cv_timedwait_sig(
1201 			    &call->call_cv,
1202 			    &call->call_lock, timout)) > 0 &&
1203 			    call->call_status == RPC_TIMEDOUT)
1204 				;
1205 
1206 		switch (cv_wait_ret) {
1207 		case 0:
1208 			/*
1209 			 * If we got out of the above loop with
1210 			 * cv_timedwait_sig() returning 0, then we were
1211 			 * interrupted regardless what call_status is.
1212 			 */
1213 			interrupted = 1;
1214 			break;
1215 		case -1:
1216 			/* cv_timedwait_sig() timed out */
1217 			break;
1218 		default:
1219 
1220 			/*
1221 			 * We were cv_signaled(). If we didn't
1222 			 * get a successful call_status and returned
1223 			 * before time expired, delay up to clnt_cots_min_tout
1224 			 * seconds so that the caller doesn't immediately
1225 			 * try to call us again and thus force the
1226 			 * same condition that got us here (such
1227 			 * as a RPC_XPRTFAILED due to the server not
1228 			 * listening on the end-point.
1229 			 */
1230 			if (call->call_status != RPC_SUCCESS) {
1231 				clock_t curlbolt;
1232 				clock_t diff;
1233 
1234 				curlbolt = ddi_get_lbolt();
1235 				ticks = clnt_cots_min_tout *
1236 				    drv_usectohz(1000000);
1237 				diff = curlbolt - oldlbolt;
1238 				if (diff < ticks) {
1239 					delay_first = TRUE;
1240 					if (diff > 0)
1241 						ticks -= diff;
1242 				}
1243 			}
1244 			break;
1245 		}
1246 
1247 		if (lwp != NULL)
1248 			lwp->lwp_nostop--;
1249 	}
1250 	/*
1251 	 * Get the reply message, if any.  This will be freed at the end
1252 	 * whether or not an error occurred.
1253 	 */
1254 	mp = call->call_reply;
1255 	call->call_reply = NULL;
1256 
1257 	/*
1258 	 * call_err is the error info when the call is on dispatch queue.
1259 	 * cku_err is the error info returned to the caller.
1260 	 * Sync cku_err with call_err for local message processing.
1261 	 */
1262 
1263 	status = call->call_status;
1264 	p->cku_err = call->call_err;
1265 	mutex_exit(&call->call_lock);
1266 
1267 	if (status != RPC_SUCCESS) {
1268 		switch (status) {
1269 		case RPC_TIMEDOUT:
1270 			now = ddi_get_lbolt();
1271 			if (interrupted) {
1272 				COTSRCSTAT_INCR(p->cku_stats, rcintrs);
1273 				p->cku_err.re_status = RPC_INTR;
1274 				p->cku_err.re_errno = EINTR;
1275 				RPCLOG(1, "clnt_cots_kcallit: xid 0x%x",
1276 				    p->cku_xid);
1277 				RPCLOG(1, "signal interrupted at %ld", now);
1278 				RPCLOG(1, ", was sent at %ld\n", time_sent);
1279 			} else {
1280 				COTSRCSTAT_INCR(p->cku_stats, rctimeouts);
1281 				p->cku_err.re_errno = ETIMEDOUT;
1282 				RPCLOG(1, "clnt_cots_kcallit: timed out at %ld",
1283 				    now);
1284 				RPCLOG(1, ", was sent at %ld\n", time_sent);
1285 			}
1286 			break;
1287 
1288 		case RPC_XPRTFAILED:
1289 			if (p->cku_err.re_errno == 0)
1290 				p->cku_err.re_errno = EIO;
1291 
1292 			RPCLOG(1, "clnt_cots_kcallit: transport failed: %d\n",
1293 			    p->cku_err.re_errno);
1294 			break;
1295 
1296 		case RPC_SYSTEMERROR:
1297 			ASSERT(p->cku_err.re_errno);
1298 			RPCLOG(1, "clnt_cots_kcallit: system error: %d\n",
1299 			    p->cku_err.re_errno);
1300 			break;
1301 
1302 		default:
1303 			p->cku_err.re_status = RPC_SYSTEMERROR;
1304 			p->cku_err.re_errno = EIO;
1305 			RPCLOG(1, "clnt_cots_kcallit: error: %s\n",
1306 			    clnt_sperrno(status));
1307 			break;
1308 		}
1309 		if (p->cku_err.re_status != RPC_TIMEDOUT) {
1310 
1311 			if (p->cku_flags & CKU_ONQUEUE) {
1312 				call_table_remove(call);
1313 				p->cku_flags &= ~CKU_ONQUEUE;
1314 			}
1315 
1316 			RPCLOG(64, "clnt_cots_kcallit: non TIMEOUT so xid 0x%x "
1317 			    "taken off dispatch list\n", p->cku_xid);
1318 			if (call->call_reply) {
1319 				freemsg(call->call_reply);
1320 				call->call_reply = NULL;
1321 			}
1322 		} else if (wait.tv_sec != 0) {
1323 			/*
1324 			 * We've sent the request over TCP and so we have
1325 			 * every reason to believe it will get
1326 			 * delivered. In which case returning a timeout is not
1327 			 * appropriate.
1328 			 */
1329 			if (p->cku_progress == TRUE &&
1330 			    p->cku_recv_attempts < clnt_cots_maxrecv) {
1331 				p->cku_err.re_status = RPC_INPROGRESS;
1332 			}
1333 		}
1334 		goto cots_done;
1335 	}
1336 
1337 	xdrs = &p->cku_inxdr;
1338 	xdrmblk_init(xdrs, mp, XDR_DECODE, 0);
1339 
1340 	reply_msg.rm_direction = REPLY;
1341 	reply_msg.rm_reply.rp_stat = MSG_ACCEPTED;
1342 	reply_msg.acpted_rply.ar_stat = SUCCESS;
1343 
1344 	reply_msg.acpted_rply.ar_verf = _null_auth;
1345 	/*
1346 	 *  xdr_results will be done in AUTH_UNWRAP.
1347 	 */
1348 	reply_msg.acpted_rply.ar_results.where = NULL;
1349 	reply_msg.acpted_rply.ar_results.proc = xdr_void;
1350 
1351 	if (xdr_replymsg(xdrs, &reply_msg)) {
1352 		enum clnt_stat re_status;
1353 
1354 		_seterr_reply(&reply_msg, &p->cku_err);
1355 
1356 		re_status = p->cku_err.re_status;
1357 		if (re_status == RPC_SUCCESS) {
1358 			/*
1359 			 * Reply is good, check auth.
1360 			 */
1361 			if (!AUTH_VALIDATE(h->cl_auth,
1362 			    &reply_msg.acpted_rply.ar_verf)) {
1363 				COTSRCSTAT_INCR(p->cku_stats, rcbadverfs);
1364 				RPCLOG0(1, "clnt_cots_kcallit: validation "
1365 				    "failure\n");
1366 				freemsg(mp);
1367 				(void) xdr_rpc_free_verifier(xdrs, &reply_msg);
1368 				mutex_enter(&call->call_lock);
1369 				if (call->call_reply == NULL)
1370 					call->call_status = RPC_TIMEDOUT;
1371 				mutex_exit(&call->call_lock);
1372 				goto read_again;
1373 			} else if (!AUTH_UNWRAP(h->cl_auth, xdrs,
1374 			    xdr_results, resultsp)) {
1375 				RPCLOG0(1, "clnt_cots_kcallit: validation "
1376 				    "failure (unwrap)\n");
1377 				p->cku_err.re_status = RPC_CANTDECODERES;
1378 				p->cku_err.re_errno = EIO;
1379 			}
1380 		} else {
1381 			/* set errno in case we can't recover */
1382 			if (re_status != RPC_VERSMISMATCH &&
1383 			    re_status != RPC_AUTHERROR &&
1384 			    re_status != RPC_PROGVERSMISMATCH)
1385 				p->cku_err.re_errno = EIO;
1386 
1387 			if (re_status == RPC_AUTHERROR) {
1388 				/*
1389 				 * Maybe our credential need to be refreshed
1390 				 */
1391 				if (cm_entry) {
1392 					/*
1393 					 * There is the potential that the
1394 					 * cm_entry has/will be marked dead,
1395 					 * so drop the connection altogether,
1396 					 * force REFRESH to establish new
1397 					 * connection.
1398 					 */
1399 					connmgr_cancelconn(cm_entry);
1400 					cm_entry = NULL;
1401 				}
1402 
1403 				(void) xdr_rpc_free_verifier(xdrs,
1404 				    &reply_msg);
1405 
1406 				if (p->cku_flags & CKU_ONQUEUE) {
1407 					call_table_remove(call);
1408 					p->cku_flags &= ~CKU_ONQUEUE;
1409 				}
1410 				RPCLOG(64,
1411 				    "clnt_cots_kcallit: AUTH_ERROR, xid"
1412 				    " 0x%x removed off dispatch list\n",
1413 				    p->cku_xid);
1414 				if (call->call_reply) {
1415 					freemsg(call->call_reply);
1416 					call->call_reply = NULL;
1417 				}
1418 
1419 				if ((refreshes > 0) &&
1420 				    AUTH_REFRESH(h->cl_auth, &reply_msg,
1421 				    p->cku_cred)) {
1422 					refreshes--;
1423 					freemsg(mp);
1424 					mp = NULL;
1425 
1426 					COTSRCSTAT_INCR(p->cku_stats,
1427 					    rcbadcalls);
1428 					COTSRCSTAT_INCR(p->cku_stats,
1429 					    rcnewcreds);
1430 					goto call_again;
1431 				}
1432 
1433 				/*
1434 				 * We have used the client handle to
1435 				 * do an AUTH_REFRESH and the RPC status may
1436 				 * be set to RPC_SUCCESS; Let's make sure to
1437 				 * set it to RPC_AUTHERROR.
1438 				 */
1439 				p->cku_err.re_status = RPC_AUTHERROR;
1440 
1441 				/*
1442 				 * Map recoverable and unrecoverable
1443 				 * authentication errors to appropriate errno
1444 				 */
1445 				switch (p->cku_err.re_why) {
1446 				case AUTH_TOOWEAK:
1447 					/*
1448 					 * This could be a failure where the
1449 					 * server requires use of a reserved
1450 					 * port,  check and optionally set the
1451 					 * client handle useresvport trying
1452 					 * one more time. Next go round we
1453 					 * fall out with the tooweak error.
1454 					 */
1455 					if (p->cku_useresvport != 1) {
1456 						p->cku_useresvport = 1;
1457 						p->cku_xid = 0;
1458 						freemsg(mp);
1459 						mp = NULL;
1460 						goto call_again;
1461 					}
1462 					/* FALLTHRU */
1463 				case AUTH_BADCRED:
1464 				case AUTH_BADVERF:
1465 				case AUTH_INVALIDRESP:
1466 				case AUTH_FAILED:
1467 				case RPCSEC_GSS_NOCRED:
1468 				case RPCSEC_GSS_FAILED:
1469 						p->cku_err.re_errno = EACCES;
1470 						break;
1471 				case AUTH_REJECTEDCRED:
1472 				case AUTH_REJECTEDVERF:
1473 				default:	p->cku_err.re_errno = EIO;
1474 						break;
1475 				}
1476 				RPCLOG(1, "clnt_cots_kcallit : authentication"
1477 				    " failed with RPC_AUTHERROR of type %d\n",
1478 				    (int)p->cku_err.re_why);
1479 				goto cots_done;
1480 			}
1481 		}
1482 	} else {
1483 		/* reply didn't decode properly. */
1484 		p->cku_err.re_status = RPC_CANTDECODERES;
1485 		p->cku_err.re_errno = EIO;
1486 		RPCLOG0(1, "clnt_cots_kcallit: decode failure\n");
1487 	}
1488 
1489 	(void) xdr_rpc_free_verifier(xdrs, &reply_msg);
1490 
1491 	if (p->cku_flags & CKU_ONQUEUE) {
1492 		call_table_remove(call);
1493 		p->cku_flags &= ~CKU_ONQUEUE;
1494 	}
1495 
1496 	RPCLOG(64, "clnt_cots_kcallit: xid 0x%x taken off dispatch list",
1497 	    p->cku_xid);
1498 	RPCLOG(64, " status is %s\n", clnt_sperrno(p->cku_err.re_status));
1499 cots_done:
1500 	if (cm_entry)
1501 		connmgr_release(cm_entry);
1502 
1503 	if (mp != NULL)
1504 		freemsg(mp);
1505 	if ((p->cku_flags & CKU_ONQUEUE) == 0 && call->call_reply) {
1506 		freemsg(call->call_reply);
1507 		call->call_reply = NULL;
1508 	}
1509 	if (p->cku_err.re_status != RPC_SUCCESS) {
1510 		RPCLOG0(1, "clnt_cots_kcallit: tail-end failure\n");
1511 		COTSRCSTAT_INCR(p->cku_stats, rcbadcalls);
1512 	}
1513 
1514 	/*
1515 	 * No point in delaying if the zone is going away.
1516 	 */
1517 	if (delay_first == TRUE &&
1518 	    !(zone_status_get(curproc->p_zone) >= ZONE_IS_SHUTTING_DOWN)) {
1519 		if (clnt_delay(ticks, h->cl_nosignal) == EINTR) {
1520 			p->cku_err.re_errno = EINTR;
1521 			p->cku_err.re_status = RPC_INTR;
1522 		}
1523 	}
1524 	return (p->cku_err.re_status);
1525 }
1526 
1527 /*
1528  * Kinit routine for cots.  This sets up the correct operations in
1529  * the client handle, as the handle may have previously been a clts
1530  * handle, and clears the xid field so there is no way a new call
1531  * could be mistaken for a retry.  It also sets in the handle the
1532  * information that is passed at create/kinit time but needed at
1533  * call time, as cots creates the transport at call time - device,
1534  * address of the server, protocol family.
1535  */
1536 void
1537 clnt_cots_kinit(CLIENT *h, dev_t dev, int family, struct netbuf *addr,
1538 	int max_msgsize, cred_t *cred)
1539 {
1540 	/* LINTED pointer alignment */
1541 	cku_private_t *p = htop(h);
1542 	calllist_t *call = &p->cku_call;
1543 
1544 	h->cl_ops = &tcp_ops;
1545 	if (p->cku_flags & CKU_ONQUEUE) {
1546 		call_table_remove(call);
1547 		p->cku_flags &= ~CKU_ONQUEUE;
1548 		RPCLOG(64, "clnt_cots_kinit: removing call for xid 0x%x from"
1549 		    " dispatch list\n", p->cku_xid);
1550 	}
1551 
1552 	if (call->call_reply != NULL) {
1553 		freemsg(call->call_reply);
1554 		call->call_reply = NULL;
1555 	}
1556 
1557 	call->call_bucket = NULL;
1558 	call->call_hash = 0;
1559 
1560 	/*
1561 	 * We don't clear cku_flags here, because clnt_cots_kcallit()
1562 	 * takes care of handling the cku_flags reset.
1563 	 */
1564 	p->cku_xid = 0;
1565 	p->cku_device = dev;
1566 	p->cku_addrfmly = family;
1567 	p->cku_cred = cred;
1568 
1569 	if (p->cku_addr.maxlen < addr->len) {
1570 		if (p->cku_addr.maxlen != 0 && p->cku_addr.buf != NULL)
1571 			kmem_free(p->cku_addr.buf, p->cku_addr.maxlen);
1572 		p->cku_addr.buf = kmem_zalloc(addr->maxlen, KM_SLEEP);
1573 		p->cku_addr.maxlen = addr->maxlen;
1574 	}
1575 
1576 	p->cku_addr.len = addr->len;
1577 	bcopy(addr->buf, p->cku_addr.buf, addr->len);
1578 
1579 	/*
1580 	 * If the current sanity check size in rpcmod is smaller
1581 	 * than the size needed, then increase the sanity check.
1582 	 */
1583 	if (max_msgsize != 0 && clnt_max_msg_sizep != NULL &&
1584 	    max_msgsize > *clnt_max_msg_sizep) {
1585 		mutex_enter(&clnt_max_msg_lock);
1586 		if (max_msgsize > *clnt_max_msg_sizep)
1587 			*clnt_max_msg_sizep = max_msgsize;
1588 		mutex_exit(&clnt_max_msg_lock);
1589 	}
1590 }
1591 
1592 /*
1593  * ksettimers is a no-op for cots, with the exception of setting the xid.
1594  */
1595 /* ARGSUSED */
1596 static int
1597 clnt_cots_ksettimers(CLIENT *h, struct rpc_timers *t, struct rpc_timers *all,
1598 	int minimum, void (*feedback)(int, int, caddr_t), caddr_t arg,
1599 	uint32_t xid)
1600 {
1601 	/* LINTED pointer alignment */
1602 	cku_private_t *p = htop(h);
1603 
1604 	if (xid)
1605 		p->cku_xid = xid;
1606 	COTSRCSTAT_INCR(p->cku_stats, rctimers);
1607 	return (0);
1608 }
1609 
1610 extern void rpc_poptimod(struct vnode *);
1611 extern int kstr_push(struct vnode *, char *);
1612 
1613 int
1614 conn_kstat_update(kstat_t *ksp, int rw)
1615 {
1616 	struct cm_xprt *cm_entry;
1617 	struct cm_kstat_xprt *cm_ksp_data;
1618 	uchar_t *b;
1619 	char *fbuf;
1620 
1621 	if (rw == KSTAT_WRITE)
1622 		return (EACCES);
1623 	if (ksp == NULL || ksp->ks_private == NULL)
1624 		return (EIO);
1625 	cm_entry  = (struct cm_xprt *)ksp->ks_private;
1626 	cm_ksp_data = (struct cm_kstat_xprt *)ksp->ks_data;
1627 
1628 	cm_ksp_data->x_wq.value.ui32 = (uint32_t)(uintptr_t)cm_entry->x_wq;
1629 	cm_ksp_data->x_family.value.ui32 = cm_entry->x_family;
1630 	cm_ksp_data->x_rdev.value.ui32 = (uint32_t)cm_entry->x_rdev;
1631 	cm_ksp_data->x_time.value.ui32 = cm_entry->x_time;
1632 	cm_ksp_data->x_ref.value.ui32 = cm_entry->x_ref;
1633 	cm_ksp_data->x_state.value.ui32 = cm_entry->x_state_flags;
1634 
1635 	if (cm_entry->x_server.buf) {
1636 		fbuf = cm_ksp_data->x_server.value.str.addr.ptr;
1637 		if (cm_entry->x_family == AF_INET &&
1638 		    cm_entry->x_server.len ==
1639 		    sizeof (struct sockaddr_in)) {
1640 			struct sockaddr_in  *sa;
1641 			sa = (struct sockaddr_in *)
1642 				cm_entry->x_server.buf;
1643 			b = (uchar_t *)&sa->sin_addr;
1644 			(void) sprintf(fbuf,
1645 			    "%03d.%03d.%03d.%03d", b[0] & 0xFF, b[1] & 0xFF,
1646 			    b[2] & 0xFF, b[3] & 0xFF);
1647 			cm_ksp_data->x_port.value.ui32 =
1648 				(uint32_t)sa->sin_port;
1649 		} else if (cm_entry->x_family == AF_INET6 &&
1650 				cm_entry->x_server.len >=
1651 				sizeof (struct sockaddr_in6)) {
1652 			/* extract server IP address & port */
1653 			struct sockaddr_in6 *sin6;
1654 			sin6 = (struct sockaddr_in6 *)cm_entry->x_server.buf;
1655 			(void) kinet_ntop6((uchar_t *)&sin6->sin6_addr, fbuf,
1656 				INET6_ADDRSTRLEN);
1657 			cm_ksp_data->x_port.value.ui32 = sin6->sin6_port;
1658 		} else {
1659 			struct sockaddr_in  *sa;
1660 
1661 			sa = (struct sockaddr_in *)cm_entry->x_server.buf;
1662 			b = (uchar_t *)&sa->sin_addr;
1663 			(void) sprintf(fbuf,
1664 			    "%03d.%03d.%03d.%03d", b[0] & 0xFF, b[1] & 0xFF,
1665 			    b[2] & 0xFF, b[3] & 0xFF);
1666 		}
1667 		KSTAT_NAMED_STR_BUFLEN(&cm_ksp_data->x_server) =
1668 		    strlen(fbuf) + 1;
1669 	}
1670 
1671 	return (0);
1672 }
1673 
1674 
1675 /*
1676  * We want a version of delay which is interruptible by a UNIX signal
1677  * Return EINTR if an interrupt occured.
1678  */
1679 static int
1680 clnt_delay(clock_t ticks, bool_t nosignal)
1681 {
1682 	if (nosignal == TRUE) {
1683 		delay(ticks);
1684 		return (0);
1685 	}
1686 	return (delay_sig(ticks));
1687 }
1688 
1689 /*
1690  * Wait for a connection until a timeout, or until we are
1691  * signalled that there has been a connection state change.
1692  */
1693 static enum clnt_stat
1694 connmgr_cwait(struct cm_xprt *cm_entry, const struct timeval *waitp,
1695 	bool_t nosignal)
1696 {
1697 	bool_t interrupted;
1698 	clock_t timout, cv_stat;
1699 	enum clnt_stat clstat;
1700 	unsigned int old_state;
1701 
1702 	ASSERT(MUTEX_HELD(&connmgr_lock));
1703 	/*
1704 	 * We wait for the transport connection to be made, or an
1705 	 * indication that it could not be made.
1706 	 */
1707 	clstat = RPC_TIMEDOUT;
1708 	interrupted = FALSE;
1709 
1710 	old_state = cm_entry->x_state_flags;
1711 	/*
1712 	 * Now loop until cv_timedwait{_sig} returns because of
1713 	 * a signal(0) or timeout(-1) or cv_signal(>0). But it may be
1714 	 * cv_signalled for various other reasons too. So loop
1715 	 * until there is a state change on the connection.
1716 	 */
1717 
1718 	timout = waitp->tv_sec * drv_usectohz(1000000) +
1719 	    drv_usectohz(waitp->tv_usec) + ddi_get_lbolt();
1720 
1721 	if (nosignal) {
1722 		while ((cv_stat = cv_timedwait(&cm_entry->x_conn_cv,
1723 		    &connmgr_lock, timout)) > 0 &&
1724 		    cm_entry->x_state_flags == old_state)
1725 			;
1726 	} else {
1727 		while ((cv_stat = cv_timedwait_sig(&cm_entry->x_conn_cv,
1728 		    &connmgr_lock, timout)) > 0 &&
1729 		    cm_entry->x_state_flags == old_state)
1730 			;
1731 
1732 		if (cv_stat == 0) /* got intr signal? */
1733 			interrupted = TRUE;
1734 	}
1735 
1736 	if ((cm_entry->x_state_flags & (X_BADSTATES|X_CONNECTED)) ==
1737 	    X_CONNECTED) {
1738 		clstat = RPC_SUCCESS;
1739 	} else {
1740 		if (interrupted == TRUE)
1741 			clstat = RPC_INTR;
1742 		RPCLOG(1, "connmgr_cwait: can't connect, error: %s\n",
1743 		    clnt_sperrno(clstat));
1744 	}
1745 
1746 	return (clstat);
1747 }
1748 
1749 /*
1750  * Primary interface for how RPC grabs a connection.
1751  */
1752 static struct cm_xprt *
1753 connmgr_wrapget(
1754 	struct netbuf *retryaddr,
1755 	const struct timeval *waitp,
1756 	cku_private_t *p)
1757 {
1758 	struct cm_xprt *cm_entry;
1759 
1760 	cm_entry = connmgr_get(retryaddr, waitp, &p->cku_addr, p->cku_addrfmly,
1761 	    &p->cku_srcaddr, &p->cku_err, p->cku_device,
1762 	    p->cku_client.cl_nosignal, p->cku_useresvport, p->cku_cred);
1763 
1764 	if (cm_entry == NULL) {
1765 		/*
1766 		 * Re-map the call status to RPC_INTR if the err code is
1767 		 * EINTR. This can happen if calls status is RPC_TLIERROR.
1768 		 * However, don't re-map if signalling has been turned off.
1769 		 * XXX Really need to create a separate thread whenever
1770 		 * there isn't an existing connection.
1771 		 */
1772 		if (p->cku_err.re_errno == EINTR) {
1773 			if (p->cku_client.cl_nosignal == TRUE)
1774 				p->cku_err.re_errno = EIO;
1775 			else
1776 				p->cku_err.re_status = RPC_INTR;
1777 		}
1778 	}
1779 
1780 	return (cm_entry);
1781 }
1782 
1783 /*
1784  * Obtains a transport to the server specified in addr.  If a suitable transport
1785  * does not already exist in the list of cached transports, a new connection
1786  * is created, connected, and added to the list. The connection is for sending
1787  * only - the reply message may come back on another transport connection.
1788  *
1789  * To implement round-robin load balancing with multiple client connections,
1790  * the last entry on the list is always selected. Once the entry is selected
1791  * it's re-inserted to the head of the list.
1792  */
1793 static struct cm_xprt *
1794 connmgr_get(
1795 	struct netbuf	*retryaddr,
1796 	const struct timeval	*waitp,	/* changed to a ptr to converse stack */
1797 	struct netbuf	*destaddr,
1798 	int		addrfmly,
1799 	struct netbuf	*srcaddr,
1800 	struct rpc_err	*rpcerr,
1801 	dev_t		device,
1802 	bool_t		nosignal,
1803 	int		useresvport,
1804 	cred_t		*cr)
1805 {
1806 	struct cm_xprt *cm_entry;
1807 	struct cm_xprt *lru_entry;
1808 	struct cm_xprt **cmp, **prev;
1809 	queue_t *wq;
1810 	TIUSER *tiptr;
1811 	int i;
1812 	int retval;
1813 	int tidu_size;
1814 	bool_t	connected;
1815 	zoneid_t zoneid = rpc_zoneid();
1816 
1817 	/*
1818 	 * If the call is not a retry, look for a transport entry that
1819 	 * goes to the server of interest.
1820 	 */
1821 	mutex_enter(&connmgr_lock);
1822 
1823 	if (retryaddr == NULL) {
1824 use_new_conn:
1825 		i = 0;
1826 		cm_entry = lru_entry = NULL;
1827 
1828 		prev = cmp = &cm_hd;
1829 		while ((cm_entry = *cmp) != NULL) {
1830 			ASSERT(cm_entry != cm_entry->x_next);
1831 			/*
1832 			 * Garbage collect conections that are marked
1833 			 * for needs disconnect.
1834 			 */
1835 			if (cm_entry->x_needdis) {
1836 				CONN_HOLD(cm_entry);
1837 				connmgr_dis_and_wait(cm_entry);
1838 				connmgr_release(cm_entry);
1839 				/*
1840 				 * connmgr_lock could have been
1841 				 * dropped for the disconnect
1842 				 * processing so start over.
1843 				 */
1844 				goto use_new_conn;
1845 			}
1846 
1847 			/*
1848 			 * Garbage collect the dead connections that have
1849 			 * no threads working on them.
1850 			 */
1851 			if ((cm_entry->x_state_flags & (X_DEAD|X_THREAD)) ==
1852 			    X_DEAD) {
1853 				mutex_enter(&cm_entry->x_lock);
1854 				if (cm_entry->x_ref != 0) {
1855 					/*
1856 					 * Currently in use.
1857 					 * Cleanup later.
1858 					 */
1859 					cmp = &cm_entry->x_next;
1860 					mutex_exit(&cm_entry->x_lock);
1861 					continue;
1862 				}
1863 				mutex_exit(&cm_entry->x_lock);
1864 				*cmp = cm_entry->x_next;
1865 				mutex_exit(&connmgr_lock);
1866 				connmgr_close(cm_entry);
1867 				mutex_enter(&connmgr_lock);
1868 				goto use_new_conn;
1869 			}
1870 
1871 
1872 			if ((cm_entry->x_state_flags & X_BADSTATES) == 0 &&
1873 			    cm_entry->x_zoneid == zoneid &&
1874 			    cm_entry->x_rdev == device &&
1875 			    destaddr->len == cm_entry->x_server.len &&
1876 			    bcmp(destaddr->buf, cm_entry->x_server.buf,
1877 			    destaddr->len) == 0) {
1878 				/*
1879 				 * If the matching entry isn't connected,
1880 				 * attempt to reconnect it.
1881 				 */
1882 				if (cm_entry->x_connected == FALSE) {
1883 					/*
1884 					 * We don't go through trying
1885 					 * to find the least recently
1886 					 * used connected because
1887 					 * connmgr_reconnect() briefly
1888 					 * dropped the connmgr_lock,
1889 					 * allowing a window for our
1890 					 * accounting to be messed up.
1891 					 * In any case, a re-connected
1892 					 * connection is as good as
1893 					 * a LRU connection.
1894 					 */
1895 					return (connmgr_wrapconnect(cm_entry,
1896 					    waitp, destaddr, addrfmly, srcaddr,
1897 					    rpcerr, TRUE, nosignal, cr));
1898 				}
1899 				i++;
1900 
1901 				/* keep track of the last entry */
1902 				lru_entry = cm_entry;
1903 				prev = cmp;
1904 			}
1905 			cmp = &cm_entry->x_next;
1906 		}
1907 
1908 		if (i > clnt_max_conns) {
1909 			RPCLOG(8, "connmgr_get: too many conns, dooming entry"
1910 			    " %p\n", (void *)lru_entry->x_tiptr);
1911 			lru_entry->x_doomed = TRUE;
1912 			goto use_new_conn;
1913 		}
1914 
1915 		/*
1916 		 * If we are at the maximum number of connections to
1917 		 * the server, hand back the least recently used one.
1918 		 */
1919 		if (i == clnt_max_conns) {
1920 			/*
1921 			 * Copy into the handle the source address of
1922 			 * the connection, which we will use in case of
1923 			 * a later retry.
1924 			 */
1925 			if (srcaddr->len != lru_entry->x_src.len) {
1926 				if (srcaddr->len > 0)
1927 					kmem_free(srcaddr->buf,
1928 					    srcaddr->maxlen);
1929 				srcaddr->buf = kmem_zalloc(
1930 				    lru_entry->x_src.len, KM_SLEEP);
1931 				srcaddr->maxlen = srcaddr->len =
1932 				    lru_entry->x_src.len;
1933 			}
1934 			bcopy(lru_entry->x_src.buf, srcaddr->buf, srcaddr->len);
1935 			RPCLOG(2, "connmgr_get: call going out on %p\n",
1936 			    (void *)lru_entry);
1937 			lru_entry->x_time = ddi_get_lbolt();
1938 			CONN_HOLD(lru_entry);
1939 
1940 			if ((i > 1) && (prev != &cm_hd)) {
1941 				/*
1942 				 * remove and re-insert entry at head of list.
1943 				 */
1944 				*prev = lru_entry->x_next;
1945 				lru_entry->x_next = cm_hd;
1946 				cm_hd = lru_entry;
1947 			}
1948 
1949 			mutex_exit(&connmgr_lock);
1950 			return (lru_entry);
1951 		}
1952 
1953 	} else {
1954 		/*
1955 		 * This is the retry case (retryaddr != NULL).  Retries must
1956 		 * be sent on the same source port as the original call.
1957 		 */
1958 
1959 		/*
1960 		 * Walk the list looking for a connection with a source address
1961 		 * that matches the retry address.
1962 		 */
1963 start_retry_loop:
1964 		cmp = &cm_hd;
1965 		while ((cm_entry = *cmp) != NULL) {
1966 			ASSERT(cm_entry != cm_entry->x_next);
1967 
1968 			/*
1969 			 * determine if this connection matches the passed
1970 			 * in retry address.  If it does not match, advance
1971 			 * to the next element on the list.
1972 			 */
1973 			if (zoneid != cm_entry->x_zoneid ||
1974 			    device != cm_entry->x_rdev ||
1975 			    retryaddr->len != cm_entry->x_src.len ||
1976 			    bcmp(retryaddr->buf, cm_entry->x_src.buf,
1977 			    retryaddr->len) != 0) {
1978 				cmp = &cm_entry->x_next;
1979 				continue;
1980 			}
1981 			/*
1982 			 * Garbage collect conections that are marked
1983 			 * for needs disconnect.
1984 			 */
1985 			if (cm_entry->x_needdis) {
1986 				CONN_HOLD(cm_entry);
1987 				connmgr_dis_and_wait(cm_entry);
1988 				connmgr_release(cm_entry);
1989 				/*
1990 				 * connmgr_lock could have been
1991 				 * dropped for the disconnect
1992 				 * processing so start over.
1993 				 */
1994 				goto start_retry_loop;
1995 			}
1996 			/*
1997 			 * Garbage collect the dead connections that have
1998 			 * no threads working on them.
1999 			 */
2000 			if ((cm_entry->x_state_flags & (X_DEAD|X_THREAD)) ==
2001 			    X_DEAD) {
2002 				mutex_enter(&cm_entry->x_lock);
2003 				if (cm_entry->x_ref != 0) {
2004 					/*
2005 					 * Currently in use.
2006 					 * Cleanup later.
2007 					 */
2008 					cmp = &cm_entry->x_next;
2009 					mutex_exit(&cm_entry->x_lock);
2010 					continue;
2011 				}
2012 				mutex_exit(&cm_entry->x_lock);
2013 				*cmp = cm_entry->x_next;
2014 				mutex_exit(&connmgr_lock);
2015 				connmgr_close(cm_entry);
2016 				mutex_enter(&connmgr_lock);
2017 				goto start_retry_loop;
2018 			}
2019 
2020 			/*
2021 			 * Sanity check: if the connection with our source
2022 			 * port is going to some other server, something went
2023 			 * wrong, as we never delete connections (i.e. release
2024 			 * ports) unless they have been idle.  In this case,
2025 			 * it is probably better to send the call out using
2026 			 * a new source address than to fail it altogether,
2027 			 * since that port may never be released.
2028 			 */
2029 			if (destaddr->len != cm_entry->x_server.len ||
2030 			    bcmp(destaddr->buf, cm_entry->x_server.buf,
2031 			    destaddr->len) != 0) {
2032 				RPCLOG(1, "connmgr_get: tiptr %p"
2033 				    " is going to a different server"
2034 				    " with the port that belongs"
2035 				    " to us!\n", (void *)cm_entry->x_tiptr);
2036 				retryaddr = NULL;
2037 				goto use_new_conn;
2038 			}
2039 
2040 			/*
2041 			 * If the connection of interest is not connected and we
2042 			 * can't reconnect it, then the server is probably
2043 			 * still down.  Return NULL to the caller and let it
2044 			 * retry later if it wants to.  We have a delay so the
2045 			 * machine doesn't go into a tight retry loop.  If the
2046 			 * entry was already connected, or the reconnected was
2047 			 * successful, return this entry.
2048 			 */
2049 			if (cm_entry->x_connected == FALSE) {
2050 				return (connmgr_wrapconnect(cm_entry,
2051 				    waitp, destaddr, addrfmly, NULL,
2052 				    rpcerr, TRUE, nosignal, cr));
2053 			} else {
2054 				CONN_HOLD(cm_entry);
2055 
2056 				cm_entry->x_time = ddi_get_lbolt();
2057 				mutex_exit(&connmgr_lock);
2058 				RPCLOG(2, "connmgr_get: found old "
2059 				    "transport %p for retry\n",
2060 				    (void *)cm_entry);
2061 				return (cm_entry);
2062 			}
2063 		}
2064 
2065 		/*
2066 		 * We cannot find an entry in the list for this retry.
2067 		 * Either the entry has been removed temporarily to be
2068 		 * reconnected by another thread, or the original call
2069 		 * got a port but never got connected,
2070 		 * and hence the transport never got put in the
2071 		 * list.  Fall through to the "create new connection" code -
2072 		 * the former case will fail there trying to rebind the port,
2073 		 * and the later case (and any other pathological cases) will
2074 		 * rebind and reconnect and not hang the client machine.
2075 		 */
2076 		RPCLOG0(8, "connmgr_get: no entry in list for retry\n");
2077 	}
2078 	/*
2079 	 * Set up a transport entry in the connection manager's list.
2080 	 */
2081 	cm_entry = (struct cm_xprt *)
2082 	    kmem_zalloc(sizeof (struct cm_xprt), KM_SLEEP);
2083 
2084 	cm_entry->x_server.buf = kmem_zalloc(destaddr->len, KM_SLEEP);
2085 	bcopy(destaddr->buf, cm_entry->x_server.buf, destaddr->len);
2086 	cm_entry->x_server.len = cm_entry->x_server.maxlen = destaddr->len;
2087 
2088 	cm_entry->x_state_flags = X_THREAD;
2089 	cm_entry->x_ref = 1;
2090 	cm_entry->x_family = addrfmly;
2091 	cm_entry->x_rdev = device;
2092 	cm_entry->x_zoneid = zoneid;
2093 	mutex_init(&cm_entry->x_lock, NULL, MUTEX_DEFAULT, NULL);
2094 	cv_init(&cm_entry->x_cv, NULL, CV_DEFAULT, NULL);
2095 	cv_init(&cm_entry->x_conn_cv, NULL, CV_DEFAULT, NULL);
2096 	cv_init(&cm_entry->x_dis_cv, NULL, CV_DEFAULT, NULL);
2097 
2098 	/*
2099 	 * Note that we add this partially initialized entry to the
2100 	 * connection list. This is so that we don't have connections to
2101 	 * the same server.
2102 	 *
2103 	 * Note that x_src is not initialized at this point. This is because
2104 	 * retryaddr might be NULL in which case x_src is whatever
2105 	 * t_kbind/bindresvport gives us. If another thread wants a
2106 	 * connection to the same server, seemingly we have an issue, but we
2107 	 * don't. If the other thread comes in with retryaddr == NULL, then it
2108 	 * will never look at x_src, and it will end up waiting in
2109 	 * connmgr_cwait() for the first thread to finish the connection
2110 	 * attempt. If the other thread comes in with retryaddr != NULL, then
2111 	 * that means there was a request sent on a connection, in which case
2112 	 * the the connection should already exist. Thus the first thread
2113 	 * never gets here ... it finds the connection it its server in the
2114 	 * connection list.
2115 	 *
2116 	 * But even if theory is wrong, in the retryaddr != NULL case, the 2nd
2117 	 * thread will skip us because x_src.len == 0.
2118 	 */
2119 	cm_entry->x_next = cm_hd;
2120 	cm_hd = cm_entry;
2121 	mutex_exit(&connmgr_lock);
2122 
2123 	/*
2124 	 * Either we didn't find an entry to the server of interest, or we
2125 	 * don't have the maximum number of connections to that server -
2126 	 * create a new connection.
2127 	 */
2128 	RPCLOG0(8, "connmgr_get: creating new connection\n");
2129 	rpcerr->re_status = RPC_TLIERROR;
2130 
2131 	i = t_kopen(NULL, device, FREAD|FWRITE|FNDELAY, &tiptr, zone_kcred());
2132 	if (i) {
2133 		RPCLOG(1, "connmgr_get: can't open cots device, error %d\n", i);
2134 		rpcerr->re_errno = i;
2135 		connmgr_cancelconn(cm_entry);
2136 		return (NULL);
2137 	}
2138 	rpc_poptimod(tiptr->fp->f_vnode);
2139 
2140 	if (i = strioctl(tiptr->fp->f_vnode, I_PUSH, (intptr_t)"rpcmod", 0,
2141 	    K_TO_K, kcred, &retval)) {
2142 		RPCLOG(1, "connmgr_get: can't push cots module, %d\n", i);
2143 		(void) t_kclose(tiptr, 1);
2144 		rpcerr->re_errno = i;
2145 		connmgr_cancelconn(cm_entry);
2146 		return (NULL);
2147 	}
2148 
2149 	if (i = strioctl(tiptr->fp->f_vnode, RPC_CLIENT, 0, 0, K_TO_K,
2150 	    kcred, &retval)) {
2151 		RPCLOG(1, "connmgr_get: can't set client status with cots "
2152 		    "module, %d\n", i);
2153 		(void) t_kclose(tiptr, 1);
2154 		rpcerr->re_errno = i;
2155 		connmgr_cancelconn(cm_entry);
2156 		return (NULL);
2157 	}
2158 
2159 	mutex_enter(&connmgr_lock);
2160 
2161 	wq = tiptr->fp->f_vnode->v_stream->sd_wrq->q_next;
2162 	cm_entry->x_wq = wq;
2163 
2164 	mutex_exit(&connmgr_lock);
2165 
2166 	if (i = strioctl(tiptr->fp->f_vnode, I_PUSH, (intptr_t)"timod", 0,
2167 	    K_TO_K, kcred, &retval)) {
2168 		RPCLOG(1, "connmgr_get: can't push timod, %d\n", i);
2169 		(void) t_kclose(tiptr, 1);
2170 		rpcerr->re_errno = i;
2171 		connmgr_cancelconn(cm_entry);
2172 		return (NULL);
2173 	}
2174 
2175 	/*
2176 	 * If the caller has not specified reserved port usage then
2177 	 * take the system default.
2178 	 */
2179 	if (useresvport == -1)
2180 		useresvport = clnt_cots_do_bindresvport;
2181 
2182 	if ((useresvport || retryaddr != NULL) &&
2183 	    (addrfmly == AF_INET || addrfmly == AF_INET6)) {
2184 		bool_t alloc_src = FALSE;
2185 
2186 		if (srcaddr->len != destaddr->len) {
2187 			kmem_free(srcaddr->buf, srcaddr->maxlen);
2188 			srcaddr->buf = kmem_zalloc(destaddr->len, KM_SLEEP);
2189 			srcaddr->maxlen = destaddr->len;
2190 			srcaddr->len = destaddr->len;
2191 			alloc_src = TRUE;
2192 		}
2193 
2194 		if ((i = bindresvport(tiptr, retryaddr, srcaddr, TRUE)) != 0) {
2195 			(void) t_kclose(tiptr, 1);
2196 			RPCLOG(1, "connmgr_get: couldn't bind, retryaddr: "
2197 			    "%p\n", (void *)retryaddr);
2198 
2199 			/*
2200 			 * 1225408: If we allocated a source address, then it
2201 			 * is either garbage or all zeroes. In that case
2202 			 * we need to clear srcaddr.
2203 			 */
2204 			if (alloc_src == TRUE) {
2205 				kmem_free(srcaddr->buf, srcaddr->maxlen);
2206 				srcaddr->maxlen = srcaddr->len = 0;
2207 				srcaddr->buf = NULL;
2208 			}
2209 			rpcerr->re_errno = i;
2210 			connmgr_cancelconn(cm_entry);
2211 			return (NULL);
2212 		}
2213 	} else {
2214 		if ((i = t_kbind(tiptr, NULL, NULL)) != 0) {
2215 			RPCLOG(1, "clnt_cots_kcreate: t_kbind: %d\n", i);
2216 			(void) t_kclose(tiptr, 1);
2217 			rpcerr->re_errno = i;
2218 			connmgr_cancelconn(cm_entry);
2219 			return (NULL);
2220 		}
2221 	}
2222 
2223 	{
2224 		/*
2225 		 * Keep the kernel stack lean. Don't move this call
2226 		 * declaration to the top of this function because a
2227 		 * call is declared in connmgr_wrapconnect()
2228 		 */
2229 		calllist_t call;
2230 
2231 		bzero(&call, sizeof (call));
2232 		cv_init(&call.call_cv, NULL, CV_DEFAULT, NULL);
2233 
2234 		/*
2235 		 * This is a bound end-point so don't close it's stream.
2236 		 */
2237 		connected = connmgr_connect(cm_entry, wq, destaddr, addrfmly,
2238 		    &call, &tidu_size, FALSE, waitp, nosignal, cr);
2239 		*rpcerr = call.call_err;
2240 		cv_destroy(&call.call_cv);
2241 
2242 	}
2243 
2244 	mutex_enter(&connmgr_lock);
2245 
2246 	/*
2247 	 * Set up a transport entry in the connection manager's list.
2248 	 */
2249 	cm_entry->x_src.buf = kmem_zalloc(srcaddr->len, KM_SLEEP);
2250 	bcopy(srcaddr->buf, cm_entry->x_src.buf, srcaddr->len);
2251 	cm_entry->x_src.len = cm_entry->x_src.maxlen = srcaddr->len;
2252 
2253 	cm_entry->x_tiptr = tiptr;
2254 	cm_entry->x_time = ddi_get_lbolt();
2255 
2256 	if (tiptr->tp_info.servtype == T_COTS_ORD)
2257 		cm_entry->x_ordrel = TRUE;
2258 	else
2259 		cm_entry->x_ordrel = FALSE;
2260 
2261 	cm_entry->x_tidu_size = tidu_size;
2262 
2263 	if (cm_entry->x_early_disc) {
2264 		/*
2265 		 * We need to check if a disconnect request has come
2266 		 * while we are connected, if so, then we need to
2267 		 * set rpcerr->re_status appropriately before returning
2268 		 * NULL to caller.
2269 		 */
2270 		if (rpcerr->re_status == RPC_SUCCESS)
2271 			rpcerr->re_status = RPC_XPRTFAILED;
2272 		cm_entry->x_connected = FALSE;
2273 	} else
2274 		cm_entry->x_connected = connected;
2275 
2276 	/*
2277 	 * There could be a discrepancy here such that
2278 	 * x_early_disc is TRUE yet connected is TRUE as well
2279 	 * and the connection is actually connected. In that case
2280 	 * lets be conservative and declare the connection as not
2281 	 * connected.
2282 	 */
2283 	cm_entry->x_early_disc = FALSE;
2284 	cm_entry->x_needdis = (cm_entry->x_connected == FALSE);
2285 	cm_entry->x_ctime = ddi_get_lbolt();
2286 
2287 	/*
2288 	 * Notify any threads waiting that the connection attempt is done.
2289 	 */
2290 	cm_entry->x_thread = FALSE;
2291 	cv_broadcast(&cm_entry->x_conn_cv);
2292 
2293 	if (cm_entry->x_connected == FALSE) {
2294 		mutex_exit(&connmgr_lock);
2295 		connmgr_release(cm_entry);
2296 		return (NULL);
2297 	}
2298 
2299 	mutex_exit(&connmgr_lock);
2300 
2301 	return (cm_entry);
2302 }
2303 
2304 /*
2305  * Keep the cm_xprt entry on the connecton list when making a connection. This
2306  * is to prevent multiple connections to a slow server from appearing.
2307  * We use the bit field x_thread to tell if a thread is doing a connection
2308  * which keeps other interested threads from messing with connection.
2309  * Those other threads just wait if x_thread is set.
2310  *
2311  * If x_thread is not set, then we do the actual work of connecting via
2312  * connmgr_connect().
2313  *
2314  * mutex convention: called with connmgr_lock held, returns with it released.
2315  */
2316 static struct cm_xprt *
2317 connmgr_wrapconnect(
2318 	struct cm_xprt	*cm_entry,
2319 	const struct timeval	*waitp,
2320 	struct netbuf	*destaddr,
2321 	int		addrfmly,
2322 	struct netbuf	*srcaddr,
2323 	struct rpc_err	*rpcerr,
2324 	bool_t		reconnect,
2325 	bool_t		nosignal,
2326 	cred_t		*cr)
2327 {
2328 	ASSERT(MUTEX_HELD(&connmgr_lock));
2329 	/*
2330 	 * Hold this entry as we are about to drop connmgr_lock.
2331 	 */
2332 	CONN_HOLD(cm_entry);
2333 
2334 	/*
2335 	 * If there is a thread already making a connection for us, then
2336 	 * wait for it to complete the connection.
2337 	 */
2338 	if (cm_entry->x_thread == TRUE) {
2339 		rpcerr->re_status = connmgr_cwait(cm_entry, waitp, nosignal);
2340 
2341 		if (rpcerr->re_status != RPC_SUCCESS) {
2342 			mutex_exit(&connmgr_lock);
2343 			connmgr_release(cm_entry);
2344 			return (NULL);
2345 		}
2346 	} else {
2347 		bool_t connected;
2348 		calllist_t call;
2349 
2350 		cm_entry->x_thread = TRUE;
2351 
2352 		while (cm_entry->x_needrel == TRUE) {
2353 			cm_entry->x_needrel = FALSE;
2354 
2355 			connmgr_sndrel(cm_entry);
2356 			delay(drv_usectohz(1000000));
2357 
2358 			mutex_enter(&connmgr_lock);
2359 		}
2360 
2361 		/*
2362 		 * If we need to send a T_DISCON_REQ, send one.
2363 		 */
2364 		connmgr_dis_and_wait(cm_entry);
2365 
2366 		mutex_exit(&connmgr_lock);
2367 
2368 		bzero(&call, sizeof (call));
2369 		cv_init(&call.call_cv, NULL, CV_DEFAULT, NULL);
2370 
2371 		connected = connmgr_connect(cm_entry, cm_entry->x_wq,
2372 		    destaddr, addrfmly, &call, &cm_entry->x_tidu_size,
2373 		    reconnect, waitp, nosignal, cr);
2374 
2375 		*rpcerr = call.call_err;
2376 		cv_destroy(&call.call_cv);
2377 
2378 		mutex_enter(&connmgr_lock);
2379 
2380 
2381 		if (cm_entry->x_early_disc) {
2382 			/*
2383 			 * We need to check if a disconnect request has come
2384 			 * while we are connected, if so, then we need to
2385 			 * set rpcerr->re_status appropriately before returning
2386 			 * NULL to caller.
2387 			 */
2388 			if (rpcerr->re_status == RPC_SUCCESS)
2389 				rpcerr->re_status = RPC_XPRTFAILED;
2390 			cm_entry->x_connected = FALSE;
2391 		} else
2392 			cm_entry->x_connected = connected;
2393 
2394 		/*
2395 		 * There could be a discrepancy here such that
2396 		 * x_early_disc is TRUE yet connected is TRUE as well
2397 		 * and the connection is actually connected. In that case
2398 		 * lets be conservative and declare the connection as not
2399 		 * connected.
2400 		 */
2401 
2402 		cm_entry->x_early_disc = FALSE;
2403 		cm_entry->x_needdis = (cm_entry->x_connected == FALSE);
2404 
2405 
2406 		/*
2407 		 * connmgr_connect() may have given up before the connection
2408 		 * actually timed out. So ensure that before the next
2409 		 * connection attempt we do a disconnect.
2410 		 */
2411 		cm_entry->x_ctime = ddi_get_lbolt();
2412 		cm_entry->x_thread = FALSE;
2413 
2414 		cv_broadcast(&cm_entry->x_conn_cv);
2415 
2416 		if (cm_entry->x_connected == FALSE) {
2417 			mutex_exit(&connmgr_lock);
2418 			connmgr_release(cm_entry);
2419 			return (NULL);
2420 		}
2421 	}
2422 
2423 	if (srcaddr != NULL) {
2424 		/*
2425 		 * Copy into the handle the
2426 		 * source address of the
2427 		 * connection, which we will use
2428 		 * in case of a later retry.
2429 		 */
2430 		if (srcaddr->len != cm_entry->x_src.len) {
2431 			if (srcaddr->maxlen > 0)
2432 				kmem_free(srcaddr->buf, srcaddr->maxlen);
2433 			srcaddr->buf = kmem_zalloc(cm_entry->x_src.len,
2434 			    KM_SLEEP);
2435 			srcaddr->maxlen = srcaddr->len =
2436 			    cm_entry->x_src.len;
2437 		}
2438 		bcopy(cm_entry->x_src.buf, srcaddr->buf, srcaddr->len);
2439 	}
2440 	cm_entry->x_time = ddi_get_lbolt();
2441 	mutex_exit(&connmgr_lock);
2442 	return (cm_entry);
2443 }
2444 
2445 /*
2446  * If we need to send a T_DISCON_REQ, send one.
2447  */
2448 static void
2449 connmgr_dis_and_wait(struct cm_xprt *cm_entry)
2450 {
2451 	ASSERT(MUTEX_HELD(&connmgr_lock));
2452 	for (;;) {
2453 		while (cm_entry->x_needdis == TRUE) {
2454 			RPCLOG(8, "connmgr_dis_and_wait: need "
2455 			    "T_DISCON_REQ for connection 0x%p\n",
2456 			    (void *)cm_entry);
2457 			cm_entry->x_needdis = FALSE;
2458 			cm_entry->x_waitdis = TRUE;
2459 
2460 			connmgr_snddis(cm_entry);
2461 
2462 			mutex_enter(&connmgr_lock);
2463 		}
2464 
2465 		if (cm_entry->x_waitdis == TRUE) {
2466 			clock_t timout;
2467 
2468 			RPCLOG(8, "connmgr_dis_and_wait waiting for "
2469 			    "T_DISCON_REQ's ACK for connection %p\n",
2470 			    (void *)cm_entry);
2471 
2472 			timout = clnt_cots_min_conntout * drv_usectohz(1000000);
2473 
2474 			/*
2475 			 * The TPI spec says that the T_DISCON_REQ
2476 			 * will get acknowledged, but in practice
2477 			 * the ACK may never get sent. So don't
2478 			 * block forever.
2479 			 */
2480 			(void) cv_reltimedwait(&cm_entry->x_dis_cv,
2481 			    &connmgr_lock, timout, TR_CLOCK_TICK);
2482 		}
2483 		/*
2484 		 * If we got the ACK, break. If we didn't,
2485 		 * then send another T_DISCON_REQ.
2486 		 */
2487 		if (cm_entry->x_waitdis == FALSE) {
2488 			break;
2489 		} else {
2490 			RPCLOG(8, "connmgr_dis_and_wait: did"
2491 			    "not get T_DISCON_REQ's ACK for "
2492 			    "connection  %p\n", (void *)cm_entry);
2493 			cm_entry->x_needdis = TRUE;
2494 		}
2495 	}
2496 }
2497 
2498 static void
2499 connmgr_cancelconn(struct cm_xprt *cm_entry)
2500 {
2501 	/*
2502 	 * Mark the connection table entry as dead; the next thread that
2503 	 * goes through connmgr_release() will notice this and deal with it.
2504 	 */
2505 	mutex_enter(&connmgr_lock);
2506 	cm_entry->x_dead = TRUE;
2507 
2508 	/*
2509 	 * Notify any threads waiting for the connection that it isn't
2510 	 * going to happen.
2511 	 */
2512 	cm_entry->x_thread = FALSE;
2513 	cv_broadcast(&cm_entry->x_conn_cv);
2514 	mutex_exit(&connmgr_lock);
2515 
2516 	connmgr_release(cm_entry);
2517 }
2518 
2519 static void
2520 connmgr_close(struct cm_xprt *cm_entry)
2521 {
2522 	mutex_enter(&cm_entry->x_lock);
2523 	while (cm_entry->x_ref != 0) {
2524 		/*
2525 		 * Must be a noninterruptible wait.
2526 		 */
2527 		cv_wait(&cm_entry->x_cv, &cm_entry->x_lock);
2528 	}
2529 
2530 	if (cm_entry->x_tiptr != NULL)
2531 		(void) t_kclose(cm_entry->x_tiptr, 1);
2532 
2533 	mutex_exit(&cm_entry->x_lock);
2534 	if (cm_entry->x_ksp != NULL) {
2535 		mutex_enter(&connmgr_lock);
2536 		cm_entry->x_ksp->ks_private = NULL;
2537 		mutex_exit(&connmgr_lock);
2538 
2539 		/*
2540 		 * Must free the buffer we allocated for the
2541 		 * server address in the update function
2542 		 */
2543 		if (((struct cm_kstat_xprt *)(cm_entry->x_ksp->ks_data))->
2544 		    x_server.value.str.addr.ptr != NULL)
2545 			kmem_free(((struct cm_kstat_xprt *)(cm_entry->x_ksp->
2546 			    ks_data))->x_server.value.str.addr.ptr,
2547 			    INET6_ADDRSTRLEN);
2548 		kmem_free(cm_entry->x_ksp->ks_data,
2549 		    cm_entry->x_ksp->ks_data_size);
2550 		kstat_delete(cm_entry->x_ksp);
2551 	}
2552 
2553 	mutex_destroy(&cm_entry->x_lock);
2554 	cv_destroy(&cm_entry->x_cv);
2555 	cv_destroy(&cm_entry->x_conn_cv);
2556 	cv_destroy(&cm_entry->x_dis_cv);
2557 
2558 	if (cm_entry->x_server.buf != NULL)
2559 		kmem_free(cm_entry->x_server.buf, cm_entry->x_server.maxlen);
2560 	if (cm_entry->x_src.buf != NULL)
2561 		kmem_free(cm_entry->x_src.buf, cm_entry->x_src.maxlen);
2562 	kmem_free(cm_entry, sizeof (struct cm_xprt));
2563 }
2564 
2565 /*
2566  * Called by KRPC after sending the call message to release the connection
2567  * it was using.
2568  */
2569 static void
2570 connmgr_release(struct cm_xprt *cm_entry)
2571 {
2572 	mutex_enter(&cm_entry->x_lock);
2573 	cm_entry->x_ref--;
2574 	if (cm_entry->x_ref == 0)
2575 		cv_signal(&cm_entry->x_cv);
2576 	mutex_exit(&cm_entry->x_lock);
2577 }
2578 
2579 /*
2580  * Set TCP receive and xmit buffer size for RPC connections.
2581  */
2582 static bool_t
2583 connmgr_setbufsz(calllist_t *e, queue_t *wq, cred_t *cr)
2584 {
2585 	int ok = FALSE;
2586 	int val;
2587 
2588 	if (rpc_default_tcp_bufsz)
2589 		return (FALSE);
2590 
2591 	/*
2592 	 * Only set new buffer size if it's larger than the system
2593 	 * default buffer size. If smaller buffer size is needed
2594 	 * then use /etc/system to set rpc_default_tcp_bufsz to 1.
2595 	 */
2596 	ok = connmgr_getopt_int(wq, SOL_SOCKET, SO_RCVBUF, &val, e, cr);
2597 	if ((ok == TRUE) && (val < rpc_send_bufsz)) {
2598 		ok = connmgr_setopt_int(wq, SOL_SOCKET, SO_RCVBUF,
2599 		    rpc_send_bufsz, e, cr);
2600 		DTRACE_PROBE2(krpc__i__connmgr_rcvbufsz,
2601 		    int, ok, calllist_t *, e);
2602 	}
2603 
2604 	ok = connmgr_getopt_int(wq, SOL_SOCKET, SO_SNDBUF, &val, e, cr);
2605 	if ((ok == TRUE) && (val < rpc_recv_bufsz)) {
2606 		ok = connmgr_setopt_int(wq, SOL_SOCKET, SO_SNDBUF,
2607 		    rpc_recv_bufsz, e, cr);
2608 		DTRACE_PROBE2(krpc__i__connmgr_sndbufsz,
2609 		    int, ok, calllist_t *, e);
2610 	}
2611 	return (TRUE);
2612 }
2613 
2614 /*
2615  * Given an open stream, connect to the remote.  Returns true if connected,
2616  * false otherwise.
2617  */
2618 static bool_t
2619 connmgr_connect(
2620 	struct cm_xprt		*cm_entry,
2621 	queue_t			*wq,
2622 	struct netbuf		*addr,
2623 	int			addrfmly,
2624 	calllist_t 		*e,
2625 	int 			*tidu_ptr,
2626 	bool_t 			reconnect,
2627 	const struct timeval 	*waitp,
2628 	bool_t 			nosignal,
2629 	cred_t			*cr)
2630 {
2631 	mblk_t *mp;
2632 	struct T_conn_req *tcr;
2633 	struct T_info_ack *tinfo;
2634 	int interrupted, error;
2635 	int tidu_size, kstat_instance;
2636 
2637 	/* if it's a reconnect, flush any lingering data messages */
2638 	if (reconnect)
2639 		(void) putctl1(wq, M_FLUSH, FLUSHRW);
2640 
2641 	/*
2642 	 * Note: if the receiver uses SCM_UCRED/getpeerucred the pid will
2643 	 * appear as -1.
2644 	 */
2645 	mp = allocb_cred(sizeof (*tcr) + addr->len, cr, NOPID);
2646 	if (mp == NULL) {
2647 		/*
2648 		 * This is unfortunate, but we need to look up the stats for
2649 		 * this zone to increment the "memory allocation failed"
2650 		 * counter.  curproc->p_zone is safe since we're initiating a
2651 		 * connection and not in some strange streams context.
2652 		 */
2653 		struct rpcstat *rpcstat;
2654 
2655 		rpcstat = zone_getspecific(rpcstat_zone_key, rpc_zone());
2656 		ASSERT(rpcstat != NULL);
2657 
2658 		RPCLOG0(1, "connmgr_connect: cannot alloc mp for "
2659 		    "sending conn request\n");
2660 		COTSRCSTAT_INCR(rpcstat->rpc_cots_client, rcnomem);
2661 		e->call_status = RPC_SYSTEMERROR;
2662 		e->call_reason = ENOSR;
2663 		return (FALSE);
2664 	}
2665 
2666 	/* Set TCP buffer size for RPC connections if needed */
2667 	if (addrfmly == AF_INET || addrfmly == AF_INET6)
2668 		(void) connmgr_setbufsz(e, wq, cr);
2669 
2670 	mp->b_datap->db_type = M_PROTO;
2671 	tcr = (struct T_conn_req *)mp->b_rptr;
2672 	bzero(tcr, sizeof (*tcr));
2673 	tcr->PRIM_type = T_CONN_REQ;
2674 	tcr->DEST_length = addr->len;
2675 	tcr->DEST_offset = sizeof (struct T_conn_req);
2676 	mp->b_wptr = mp->b_rptr + sizeof (*tcr);
2677 
2678 	bcopy(addr->buf, mp->b_wptr, tcr->DEST_length);
2679 	mp->b_wptr += tcr->DEST_length;
2680 
2681 	RPCLOG(8, "connmgr_connect: sending conn request on queue "
2682 	    "%p", (void *)wq);
2683 	RPCLOG(8, " call %p\n", (void *)wq);
2684 	/*
2685 	 * We use the entry in the handle that is normally used for
2686 	 * waiting for RPC replies to wait for the connection accept.
2687 	 */
2688 	if (clnt_dispatch_send(wq, mp, e, 0, 0) != RPC_SUCCESS) {
2689 		DTRACE_PROBE(krpc__e__connmgr__connect__cantsend);
2690 		freemsg(mp);
2691 		return (FALSE);
2692 	}
2693 
2694 	mutex_enter(&clnt_pending_lock);
2695 
2696 	/*
2697 	 * We wait for the transport connection to be made, or an
2698 	 * indication that it could not be made.
2699 	 */
2700 	interrupted = 0;
2701 
2702 	/*
2703 	 * waitforack should have been called with T_OK_ACK, but the
2704 	 * present implementation needs to be passed T_INFO_ACK to
2705 	 * work correctly.
2706 	 */
2707 	error = waitforack(e, T_INFO_ACK, waitp, nosignal);
2708 	if (error == EINTR)
2709 		interrupted = 1;
2710 	if (zone_status_get(curproc->p_zone) >= ZONE_IS_EMPTY) {
2711 		/*
2712 		 * No time to lose; we essentially have been signaled to
2713 		 * quit.
2714 		 */
2715 		interrupted = 1;
2716 	}
2717 #ifdef RPCDEBUG
2718 	if (error == ETIME)
2719 		RPCLOG0(8, "connmgr_connect: giving up "
2720 		    "on connection attempt; "
2721 		    "clnt_dispatch notifyconn "
2722 		    "diagnostic 'no one waiting for "
2723 		    "connection' should not be "
2724 		    "unexpected\n");
2725 #endif
2726 	if (e->call_prev)
2727 		e->call_prev->call_next = e->call_next;
2728 	else
2729 		clnt_pending = e->call_next;
2730 	if (e->call_next)
2731 		e->call_next->call_prev = e->call_prev;
2732 	mutex_exit(&clnt_pending_lock);
2733 
2734 	if (e->call_status != RPC_SUCCESS || error != 0) {
2735 		if (interrupted)
2736 			e->call_status = RPC_INTR;
2737 		else if (error == ETIME)
2738 			e->call_status = RPC_TIMEDOUT;
2739 		else if (error == EPROTO) {
2740 			e->call_status = RPC_SYSTEMERROR;
2741 			e->call_reason = EPROTO;
2742 		}
2743 
2744 		RPCLOG(8, "connmgr_connect: can't connect, status: "
2745 		    "%s\n", clnt_sperrno(e->call_status));
2746 
2747 		if (e->call_reply) {
2748 			freemsg(e->call_reply);
2749 			e->call_reply = NULL;
2750 		}
2751 
2752 		return (FALSE);
2753 	}
2754 	/*
2755 	 * The result of the "connection accept" is a T_info_ack
2756 	 * in the call_reply field.
2757 	 */
2758 	ASSERT(e->call_reply != NULL);
2759 	mp = e->call_reply;
2760 	e->call_reply = NULL;
2761 	tinfo = (struct T_info_ack *)mp->b_rptr;
2762 
2763 	tidu_size = tinfo->TIDU_size;
2764 	tidu_size -= (tidu_size % BYTES_PER_XDR_UNIT);
2765 	if (tidu_size > COTS_DEFAULT_ALLOCSIZE || (tidu_size <= 0))
2766 		tidu_size = COTS_DEFAULT_ALLOCSIZE;
2767 	*tidu_ptr = tidu_size;
2768 
2769 	freemsg(mp);
2770 
2771 	/*
2772 	 * Set up the pertinent options.  NODELAY is so the transport doesn't
2773 	 * buffer up RPC messages on either end.  This may not be valid for
2774 	 * all transports.  Failure to set this option is not cause to
2775 	 * bail out so we return success anyway.  Note that lack of NODELAY
2776 	 * or some other way to flush the message on both ends will cause
2777 	 * lots of retries and terrible performance.
2778 	 */
2779 	if (addrfmly == AF_INET || addrfmly == AF_INET6) {
2780 		(void) connmgr_setopt(wq, IPPROTO_TCP, TCP_NODELAY, e, cr);
2781 		if (e->call_status == RPC_XPRTFAILED)
2782 			return (FALSE);
2783 	}
2784 
2785 	/*
2786 	 * Since we have a connection, we now need to figure out if
2787 	 * we need to create a kstat. If x_ksp is not NULL then we
2788 	 * are reusing a connection and so we do not need to create
2789 	 * another kstat -- lets just return.
2790 	 */
2791 	if (cm_entry->x_ksp != NULL)
2792 		return (TRUE);
2793 
2794 	/*
2795 	 * We need to increment rpc_kstat_instance atomically to prevent
2796 	 * two kstats being created with the same instance.
2797 	 */
2798 	kstat_instance = atomic_inc_32_nv((uint32_t *)&rpc_kstat_instance);
2799 
2800 	if ((cm_entry->x_ksp = kstat_create_zone("unix", kstat_instance,
2801 	    "rpc_cots_connections", "rpc", KSTAT_TYPE_NAMED,
2802 	    (uint_t)(sizeof (cm_kstat_xprt_t) / sizeof (kstat_named_t)),
2803 	    KSTAT_FLAG_VIRTUAL, cm_entry->x_zoneid)) == NULL) {
2804 		return (TRUE);
2805 	}
2806 
2807 	cm_entry->x_ksp->ks_lock = &connmgr_lock;
2808 	cm_entry->x_ksp->ks_private = cm_entry;
2809 	cm_entry->x_ksp->ks_data_size = ((INET6_ADDRSTRLEN * sizeof (char))
2810 	    + sizeof (cm_kstat_template));
2811 	cm_entry->x_ksp->ks_data = kmem_alloc(cm_entry->x_ksp->ks_data_size,
2812 	    KM_SLEEP);
2813 	bcopy(&cm_kstat_template, cm_entry->x_ksp->ks_data,
2814 	    cm_entry->x_ksp->ks_data_size);
2815 	((struct cm_kstat_xprt *)(cm_entry->x_ksp->ks_data))->
2816 	    x_server.value.str.addr.ptr =
2817 	    kmem_alloc(INET6_ADDRSTRLEN, KM_SLEEP);
2818 
2819 	cm_entry->x_ksp->ks_update = conn_kstat_update;
2820 	kstat_install(cm_entry->x_ksp);
2821 	return (TRUE);
2822 }
2823 
2824 /*
2825  * Verify that the specified offset falls within the mblk and
2826  * that the resulting pointer is aligned.
2827  * Returns NULL if not.
2828  *
2829  * code from fs/sockfs/socksubr.c
2830  */
2831 static void *
2832 connmgr_opt_getoff(mblk_t *mp, t_uscalar_t offset,
2833     t_uscalar_t length, uint_t align_size)
2834 {
2835 	uintptr_t ptr1, ptr2;
2836 
2837 	ASSERT(mp && mp->b_wptr >= mp->b_rptr);
2838 	ptr1 = (uintptr_t)mp->b_rptr + offset;
2839 	ptr2 = (uintptr_t)ptr1 + length;
2840 	if (ptr1 < (uintptr_t)mp->b_rptr || ptr2 > (uintptr_t)mp->b_wptr) {
2841 		return (NULL);
2842 	}
2843 	if ((ptr1 & (align_size - 1)) != 0) {
2844 		return (NULL);
2845 	}
2846 	return ((void *)ptr1);
2847 }
2848 
2849 static bool_t
2850 connmgr_getopt_int(queue_t *wq, int level, int name, int *val,
2851     calllist_t *e, cred_t *cr)
2852 {
2853 	mblk_t *mp;
2854 	struct opthdr *opt, *opt_res;
2855 	struct T_optmgmt_req *tor;
2856 	struct T_optmgmt_ack *opt_ack;
2857 	struct timeval waitp;
2858 	int error;
2859 
2860 	mp = allocb_cred(sizeof (struct T_optmgmt_req) +
2861 	    sizeof (struct opthdr) + sizeof (int), cr, NOPID);
2862 	if (mp == NULL)
2863 		return (FALSE);
2864 
2865 	mp->b_datap->db_type = M_PROTO;
2866 	tor = (struct T_optmgmt_req *)(mp->b_rptr);
2867 	tor->PRIM_type = T_SVR4_OPTMGMT_REQ;
2868 	tor->MGMT_flags = T_CURRENT;
2869 	tor->OPT_length = sizeof (struct opthdr) + sizeof (int);
2870 	tor->OPT_offset = sizeof (struct T_optmgmt_req);
2871 
2872 	opt = (struct opthdr *)(mp->b_rptr + sizeof (struct T_optmgmt_req));
2873 	opt->level = level;
2874 	opt->name = name;
2875 	opt->len = sizeof (int);
2876 	mp->b_wptr += sizeof (struct T_optmgmt_req) + sizeof (struct opthdr) +
2877 	    sizeof (int);
2878 
2879 	/*
2880 	 * We will use this connection regardless
2881 	 * of whether or not the option is readable.
2882 	 */
2883 	if (clnt_dispatch_send(wq, mp, e, 0, 0) != RPC_SUCCESS) {
2884 		DTRACE_PROBE(krpc__e__connmgr__getopt__cantsend);
2885 		freemsg(mp);
2886 		return (FALSE);
2887 	}
2888 
2889 	mutex_enter(&clnt_pending_lock);
2890 
2891 	waitp.tv_sec = clnt_cots_min_conntout;
2892 	waitp.tv_usec = 0;
2893 	error = waitforack(e, T_OPTMGMT_ACK, &waitp, 1);
2894 
2895 	if (e->call_prev)
2896 		e->call_prev->call_next = e->call_next;
2897 	else
2898 		clnt_pending = e->call_next;
2899 	if (e->call_next)
2900 		e->call_next->call_prev = e->call_prev;
2901 	mutex_exit(&clnt_pending_lock);
2902 
2903 	/* get reply message */
2904 	mp = e->call_reply;
2905 	e->call_reply = NULL;
2906 
2907 	if ((!mp) || (e->call_status != RPC_SUCCESS) || (error != 0)) {
2908 
2909 		DTRACE_PROBE4(krpc__e__connmgr_getopt, int, name,
2910 		    int, e->call_status, int, error, mblk_t *, mp);
2911 
2912 		if (mp)
2913 			freemsg(mp);
2914 		return (FALSE);
2915 	}
2916 
2917 	opt_ack = (struct T_optmgmt_ack *)mp->b_rptr;
2918 	opt_res = (struct opthdr *)connmgr_opt_getoff(mp, opt_ack->OPT_offset,
2919 	    opt_ack->OPT_length, __TPI_ALIGN_SIZE);
2920 
2921 	if (!opt_res) {
2922 		DTRACE_PROBE4(krpc__e__connmgr_optres, mblk_t *, mp, int, name,
2923 		    int, opt_ack->OPT_offset, int, opt_ack->OPT_length);
2924 		freemsg(mp);
2925 		return (FALSE);
2926 	}
2927 	*val = *(int *)&opt_res[1];
2928 
2929 	DTRACE_PROBE2(connmgr_getopt__ok, int, name, int, *val);
2930 
2931 	freemsg(mp);
2932 	return (TRUE);
2933 }
2934 
2935 /*
2936  * Called by connmgr_connect to set an option on the new stream.
2937  */
2938 static bool_t
2939 connmgr_setopt_int(queue_t *wq, int level, int name, int val,
2940     calllist_t *e, cred_t *cr)
2941 {
2942 	mblk_t *mp;
2943 	struct opthdr *opt;
2944 	struct T_optmgmt_req *tor;
2945 	struct timeval waitp;
2946 	int error;
2947 
2948 	mp = allocb_cred(sizeof (struct T_optmgmt_req) +
2949 	    sizeof (struct opthdr) + sizeof (int), cr, NOPID);
2950 	if (mp == NULL) {
2951 		RPCLOG0(1, "connmgr_setopt: cannot alloc mp for option "
2952 		    "request\n");
2953 		return (FALSE);
2954 	}
2955 
2956 	mp->b_datap->db_type = M_PROTO;
2957 	tor = (struct T_optmgmt_req *)(mp->b_rptr);
2958 	tor->PRIM_type = T_SVR4_OPTMGMT_REQ;
2959 	tor->MGMT_flags = T_NEGOTIATE;
2960 	tor->OPT_length = sizeof (struct opthdr) + sizeof (int);
2961 	tor->OPT_offset = sizeof (struct T_optmgmt_req);
2962 
2963 	opt = (struct opthdr *)(mp->b_rptr + sizeof (struct T_optmgmt_req));
2964 	opt->level = level;
2965 	opt->name = name;
2966 	opt->len = sizeof (int);
2967 	*(int *)((char *)opt + sizeof (*opt)) = val;
2968 	mp->b_wptr += sizeof (struct T_optmgmt_req) + sizeof (struct opthdr) +
2969 	    sizeof (int);
2970 
2971 	/*
2972 	 * We will use this connection regardless
2973 	 * of whether or not the option is settable.
2974 	 */
2975 	if (clnt_dispatch_send(wq, mp, e, 0, 0) != RPC_SUCCESS) {
2976 		DTRACE_PROBE(krpc__e__connmgr__setopt__cantsend);
2977 		freemsg(mp);
2978 		return (FALSE);
2979 	}
2980 
2981 	mutex_enter(&clnt_pending_lock);
2982 
2983 	waitp.tv_sec = clnt_cots_min_conntout;
2984 	waitp.tv_usec = 0;
2985 	error = waitforack(e, T_OPTMGMT_ACK, &waitp, 1);
2986 
2987 	if (e->call_prev)
2988 		e->call_prev->call_next = e->call_next;
2989 	else
2990 		clnt_pending = e->call_next;
2991 	if (e->call_next)
2992 		e->call_next->call_prev = e->call_prev;
2993 	mutex_exit(&clnt_pending_lock);
2994 
2995 	if (e->call_reply != NULL) {
2996 		freemsg(e->call_reply);
2997 		e->call_reply = NULL;
2998 	}
2999 
3000 	if (e->call_status != RPC_SUCCESS || error != 0) {
3001 		RPCLOG(1, "connmgr_setopt: can't set option: %d\n", name);
3002 		return (FALSE);
3003 	}
3004 	RPCLOG(8, "connmgr_setopt: successfully set option: %d\n", name);
3005 	return (TRUE);
3006 }
3007 
3008 static bool_t
3009 connmgr_setopt(queue_t *wq, int level, int name, calllist_t *e, cred_t *cr)
3010 {
3011 	return (connmgr_setopt_int(wq, level, name, 1, e, cr));
3012 }
3013 
3014 #ifdef	DEBUG
3015 
3016 /*
3017  * This is a knob to let us force code coverage in allocation failure
3018  * case.
3019  */
3020 static int	connmgr_failsnd;
3021 #define	CONN_SND_ALLOC(Size, Pri)	\
3022 	((connmgr_failsnd-- > 0) ? NULL : allocb(Size, Pri))
3023 
3024 #else
3025 
3026 #define	CONN_SND_ALLOC(Size, Pri)	allocb(Size, Pri)
3027 
3028 #endif
3029 
3030 /*
3031  * Sends an orderly release on the specified queue.
3032  * Entered with connmgr_lock. Exited without connmgr_lock
3033  */
3034 static void
3035 connmgr_sndrel(struct cm_xprt *cm_entry)
3036 {
3037 	struct T_ordrel_req *torr;
3038 	mblk_t *mp;
3039 	queue_t *q = cm_entry->x_wq;
3040 	ASSERT(MUTEX_HELD(&connmgr_lock));
3041 	mp = CONN_SND_ALLOC(sizeof (struct T_ordrel_req), BPRI_LO);
3042 	if (mp == NULL) {
3043 		cm_entry->x_needrel = TRUE;
3044 		mutex_exit(&connmgr_lock);
3045 		RPCLOG(1, "connmgr_sndrel: cannot alloc mp for sending ordrel "
3046 		    "to queue %p\n", (void *)q);
3047 		return;
3048 	}
3049 	mutex_exit(&connmgr_lock);
3050 
3051 	mp->b_datap->db_type = M_PROTO;
3052 	torr = (struct T_ordrel_req *)(mp->b_rptr);
3053 	torr->PRIM_type = T_ORDREL_REQ;
3054 	mp->b_wptr = mp->b_rptr + sizeof (struct T_ordrel_req);
3055 
3056 	RPCLOG(8, "connmgr_sndrel: sending ordrel to queue %p\n", (void *)q);
3057 	put(q, mp);
3058 }
3059 
3060 /*
3061  * Sends an disconnect on the specified queue.
3062  * Entered with connmgr_lock. Exited without connmgr_lock
3063  */
3064 static void
3065 connmgr_snddis(struct cm_xprt *cm_entry)
3066 {
3067 	struct T_discon_req *tdis;
3068 	mblk_t *mp;
3069 	queue_t *q = cm_entry->x_wq;
3070 
3071 	ASSERT(MUTEX_HELD(&connmgr_lock));
3072 	mp = CONN_SND_ALLOC(sizeof (*tdis), BPRI_LO);
3073 	if (mp == NULL) {
3074 		cm_entry->x_needdis = TRUE;
3075 		mutex_exit(&connmgr_lock);
3076 		RPCLOG(1, "connmgr_snddis: cannot alloc mp for sending discon "
3077 		    "to queue %p\n", (void *)q);
3078 		return;
3079 	}
3080 	mutex_exit(&connmgr_lock);
3081 
3082 	mp->b_datap->db_type = M_PROTO;
3083 	tdis = (struct T_discon_req *)mp->b_rptr;
3084 	tdis->PRIM_type = T_DISCON_REQ;
3085 	mp->b_wptr = mp->b_rptr + sizeof (*tdis);
3086 
3087 	RPCLOG(8, "connmgr_snddis: sending discon to queue %p\n", (void *)q);
3088 	put(q, mp);
3089 }
3090 
3091 /*
3092  * Sets up the entry for receiving replies, and calls rpcmod's write put proc
3093  * (through put) to send the call.
3094  */
3095 static int
3096 clnt_dispatch_send(queue_t *q, mblk_t *mp, calllist_t *e, uint_t xid,
3097 			uint_t queue_flag)
3098 {
3099 	ASSERT(e != NULL);
3100 
3101 	e->call_status = RPC_TIMEDOUT;	/* optimistic, eh? */
3102 	e->call_reason = 0;
3103 	e->call_wq = q;
3104 	e->call_xid = xid;
3105 	e->call_notified = FALSE;
3106 
3107 	if (!canput(q)) {
3108 		e->call_status = RPC_CANTSEND;
3109 		e->call_reason = ENOBUFS;
3110 		return (RPC_CANTSEND);
3111 	}
3112 
3113 	/*
3114 	 * If queue_flag is set then the calllist_t is already on the hash
3115 	 * queue.  In this case just send the message and return.
3116 	 */
3117 	if (queue_flag) {
3118 		put(q, mp);
3119 		return (RPC_SUCCESS);
3120 
3121 	}
3122 
3123 	/*
3124 	 * Set up calls for RPC requests (with XID != 0) on the hash
3125 	 * queue for fast lookups and place other calls (i.e.
3126 	 * connection management) on the linked list.
3127 	 */
3128 	if (xid != 0) {
3129 		RPCLOG(64, "clnt_dispatch_send: putting xid 0x%x on "
3130 		    "dispatch list\n", xid);
3131 		e->call_hash = call_hash(xid, clnt_cots_hash_size);
3132 		e->call_bucket = &cots_call_ht[e->call_hash];
3133 		call_table_enter(e);
3134 	} else {
3135 		mutex_enter(&clnt_pending_lock);
3136 		if (clnt_pending)
3137 			clnt_pending->call_prev = e;
3138 		e->call_next = clnt_pending;
3139 		e->call_prev = NULL;
3140 		clnt_pending = e;
3141 		mutex_exit(&clnt_pending_lock);
3142 	}
3143 
3144 	put(q, mp);
3145 	return (RPC_SUCCESS);
3146 }
3147 
3148 /*
3149  * Called by rpcmod to notify a client with a clnt_pending call that its reply
3150  * has arrived.  If we can't find a client waiting for this reply, we log
3151  * the error and return.
3152  */
3153 bool_t
3154 clnt_dispatch_notify(mblk_t *mp, zoneid_t zoneid)
3155 {
3156 	calllist_t *e = NULL;
3157 	call_table_t *chtp;
3158 	uint32_t xid;
3159 	uint_t hash;
3160 
3161 	if ((IS_P2ALIGNED(mp->b_rptr, sizeof (uint32_t))) &&
3162 	    (mp->b_wptr - mp->b_rptr) >= sizeof (xid))
3163 		xid = *((uint32_t *)mp->b_rptr);
3164 	else {
3165 		int i = 0;
3166 		unsigned char *p = (unsigned char *)&xid;
3167 		unsigned char *rptr;
3168 		mblk_t *tmp = mp;
3169 
3170 		/*
3171 		 * Copy the xid, byte-by-byte into xid.
3172 		 */
3173 		while (tmp) {
3174 			rptr = tmp->b_rptr;
3175 			while (rptr < tmp->b_wptr) {
3176 				*p++ = *rptr++;
3177 				if (++i >= sizeof (xid))
3178 					goto done_xid_copy;
3179 			}
3180 			tmp = tmp->b_cont;
3181 		}
3182 
3183 		/*
3184 		 * If we got here, we ran out of mblk space before the
3185 		 * xid could be copied.
3186 		 */
3187 		ASSERT(tmp == NULL && i < sizeof (xid));
3188 
3189 		RPCLOG0(1,
3190 		    "clnt_dispatch_notify: message less than size of xid\n");
3191 		return (FALSE);
3192 
3193 	}
3194 done_xid_copy:
3195 
3196 	hash = call_hash(xid, clnt_cots_hash_size);
3197 	chtp = &cots_call_ht[hash];
3198 	/* call_table_find returns with the hash bucket locked */
3199 	call_table_find(chtp, xid, e);
3200 
3201 	if (e != NULL) {
3202 		/*
3203 		 * Found thread waiting for this reply
3204 		 */
3205 		mutex_enter(&e->call_lock);
3206 
3207 		/*
3208 		 * verify that the reply is coming in on
3209 		 * the same zone that it was sent from.
3210 		 */
3211 		if (e->call_zoneid != zoneid) {
3212 			mutex_exit(&e->call_lock);
3213 			mutex_exit(&chtp->ct_lock);
3214 			RPCLOG0(1, "clnt_dispatch_notify: incorrect zoneid\n");
3215 			return (FALSE);
3216 		}
3217 
3218 		if (e->call_reply)
3219 			/*
3220 			 * This can happen under the following scenario:
3221 			 * clnt_cots_kcallit() times out on the response,
3222 			 * rfscall() repeats the CLNT_CALL() with
3223 			 * the same xid, clnt_cots_kcallit() sends the retry,
3224 			 * thereby putting the clnt handle on the pending list,
3225 			 * the first response arrives, signalling the thread
3226 			 * in clnt_cots_kcallit(). Before that thread is
3227 			 * dispatched, the second response arrives as well,
3228 			 * and clnt_dispatch_notify still finds the handle on
3229 			 * the pending list, with call_reply set. So free the
3230 			 * old reply now.
3231 			 *
3232 			 * It is also possible for a response intended for
3233 			 * an RPC call with a different xid to reside here.
3234 			 * This can happen if the thread that owned this
3235 			 * client handle prior to the current owner bailed
3236 			 * out and left its call record on the dispatch
3237 			 * queue.  A window exists where the response can
3238 			 * arrive before the current owner dispatches its
3239 			 * RPC call.
3240 			 *
3241 			 * In any case, this is the very last point where we
3242 			 * can safely check the call_reply field before
3243 			 * placing the new response there.
3244 			 */
3245 			freemsg(e->call_reply);
3246 		e->call_reply = mp;
3247 		e->call_status = RPC_SUCCESS;
3248 		e->call_notified = TRUE;
3249 		cv_signal(&e->call_cv);
3250 		mutex_exit(&e->call_lock);
3251 		mutex_exit(&chtp->ct_lock);
3252 		return (TRUE);
3253 	} else {
3254 		zone_t *zone;
3255 		struct rpcstat *rpcstat;
3256 
3257 		mutex_exit(&chtp->ct_lock);
3258 		RPCLOG(65, "clnt_dispatch_notify: no caller for reply 0x%x\n",
3259 		    xid);
3260 		/*
3261 		 * This is unfortunate, but we need to lookup the zone so we
3262 		 * can increment its "rcbadxids" counter.
3263 		 */
3264 		zone = zone_find_by_id(zoneid);
3265 		if (zone == NULL) {
3266 			/*
3267 			 * The zone went away...
3268 			 */
3269 			return (FALSE);
3270 		}
3271 		rpcstat = zone_getspecific(rpcstat_zone_key, zone);
3272 		if (zone_status_get(zone) >= ZONE_IS_SHUTTING_DOWN) {
3273 			/*
3274 			 * Not interested
3275 			 */
3276 			zone_rele(zone);
3277 			return (FALSE);
3278 		}
3279 		COTSRCSTAT_INCR(rpcstat->rpc_cots_client, rcbadxids);
3280 		zone_rele(zone);
3281 	}
3282 	return (FALSE);
3283 }
3284 
3285 /*
3286  * Called by rpcmod when a non-data indication arrives.  The ones in which we
3287  * are interested are connection indications and options acks.  We dispatch
3288  * based on the queue the indication came in on.  If we are not interested in
3289  * what came in, we return false to rpcmod, who will then pass it upstream.
3290  */
3291 bool_t
3292 clnt_dispatch_notifyconn(queue_t *q, mblk_t *mp)
3293 {
3294 	calllist_t *e;
3295 	int type;
3296 
3297 	ASSERT((q->q_flag & QREADR) == 0);
3298 
3299 	type = ((union T_primitives *)mp->b_rptr)->type;
3300 	RPCLOG(8, "clnt_dispatch_notifyconn: prim type: [%s]\n",
3301 	    rpc_tpiprim2name(type));
3302 	mutex_enter(&clnt_pending_lock);
3303 	for (e = clnt_pending; /* NO CONDITION */; e = e->call_next) {
3304 		if (e == NULL) {
3305 			mutex_exit(&clnt_pending_lock);
3306 			RPCLOG(1, "clnt_dispatch_notifyconn: no one waiting "
3307 			    "for connection on queue 0x%p\n", (void *)q);
3308 			return (FALSE);
3309 		}
3310 		if (e->call_wq == q)
3311 			break;
3312 	}
3313 
3314 	switch (type) {
3315 	case T_CONN_CON:
3316 		/*
3317 		 * The transport is now connected, send a T_INFO_REQ to get
3318 		 * the tidu size.
3319 		 */
3320 		mutex_exit(&clnt_pending_lock);
3321 		ASSERT(mp->b_datap->db_lim - mp->b_datap->db_base >=
3322 		    sizeof (struct T_info_req));
3323 		mp->b_rptr = mp->b_datap->db_base;
3324 		((union T_primitives *)mp->b_rptr)->type = T_INFO_REQ;
3325 		mp->b_wptr = mp->b_rptr + sizeof (struct T_info_req);
3326 		mp->b_datap->db_type = M_PCPROTO;
3327 		put(q, mp);
3328 		return (TRUE);
3329 	case T_INFO_ACK:
3330 	case T_OPTMGMT_ACK:
3331 		e->call_status = RPC_SUCCESS;
3332 		e->call_reply = mp;
3333 		e->call_notified = TRUE;
3334 		cv_signal(&e->call_cv);
3335 		break;
3336 	case T_ERROR_ACK:
3337 		e->call_status = RPC_CANTCONNECT;
3338 		e->call_reply = mp;
3339 		e->call_notified = TRUE;
3340 		cv_signal(&e->call_cv);
3341 		break;
3342 	case T_OK_ACK:
3343 		/*
3344 		 * Great, but we are really waiting for a T_CONN_CON
3345 		 */
3346 		freemsg(mp);
3347 		break;
3348 	default:
3349 		mutex_exit(&clnt_pending_lock);
3350 		RPCLOG(1, "clnt_dispatch_notifyconn: bad type %d\n", type);
3351 		return (FALSE);
3352 	}
3353 
3354 	mutex_exit(&clnt_pending_lock);
3355 	return (TRUE);
3356 }
3357 
3358 /*
3359  * Called by rpcmod when the transport is (or should be) going away.  Informs
3360  * all callers waiting for replies and marks the entry in the connection
3361  * manager's list as unconnected, and either closing (close handshake in
3362  * progress) or dead.
3363  */
3364 void
3365 clnt_dispatch_notifyall(queue_t *q, int32_t msg_type, int32_t reason)
3366 {
3367 	calllist_t *e;
3368 	call_table_t *ctp;
3369 	struct cm_xprt *cm_entry;
3370 	int have_connmgr_lock;
3371 	int i;
3372 
3373 	ASSERT((q->q_flag & QREADR) == 0);
3374 
3375 	RPCLOG(1, "clnt_dispatch_notifyall on queue %p", (void *)q);
3376 	RPCLOG(1, " received a notifcation prim type [%s]",
3377 	    rpc_tpiprim2name(msg_type));
3378 	RPCLOG(1, " and reason %d\n", reason);
3379 
3380 	/*
3381 	 * Find the transport entry in the connection manager's list, close
3382 	 * the transport and delete the entry.  In the case where rpcmod's
3383 	 * idle timer goes off, it sends us a T_ORDREL_REQ, indicating we
3384 	 * should gracefully close the connection.
3385 	 */
3386 	have_connmgr_lock = 1;
3387 	mutex_enter(&connmgr_lock);
3388 	for (cm_entry = cm_hd; cm_entry; cm_entry = cm_entry->x_next) {
3389 		ASSERT(cm_entry != cm_entry->x_next);
3390 		if (cm_entry->x_wq == q) {
3391 			ASSERT(MUTEX_HELD(&connmgr_lock));
3392 			ASSERT(have_connmgr_lock == 1);
3393 			switch (msg_type) {
3394 			case T_ORDREL_REQ:
3395 
3396 				if (cm_entry->x_dead) {
3397 					RPCLOG(1, "idle timeout on dead "
3398 					    "connection: %p\n",
3399 					    (void *)cm_entry);
3400 					if (clnt_stop_idle != NULL)
3401 						(*clnt_stop_idle)(q);
3402 					break;
3403 				}
3404 
3405 				/*
3406 				 * Only mark the connection as dead if it is
3407 				 * connected and idle.
3408 				 * An unconnected connection has probably
3409 				 * gone idle because the server is down,
3410 				 * and when it comes back up there will be
3411 				 * retries that need to use that connection.
3412 				 */
3413 				if (cm_entry->x_connected ||
3414 				    cm_entry->x_doomed) {
3415 					if (cm_entry->x_ordrel) {
3416 						if (cm_entry->x_closing ==
3417 						    TRUE) {
3418 							/*
3419 							 * The connection is
3420 							 * obviously wedged due
3421 							 * to a bug or problem
3422 							 * with the transport.
3423 							 * Mark it as dead.
3424 							 * Otherwise we can
3425 							 * leak connections.
3426 							 */
3427 							cm_entry->x_dead = TRUE;
3428 							mutex_exit(
3429 							    &connmgr_lock);
3430 							have_connmgr_lock = 0;
3431 							if (clnt_stop_idle !=
3432 							    NULL)
3433 							(*clnt_stop_idle)(q);
3434 							break;
3435 						}
3436 						cm_entry->x_closing = TRUE;
3437 						connmgr_sndrel(cm_entry);
3438 						have_connmgr_lock = 0;
3439 					} else {
3440 						cm_entry->x_dead = TRUE;
3441 						mutex_exit(&connmgr_lock);
3442 						have_connmgr_lock = 0;
3443 						if (clnt_stop_idle != NULL)
3444 							(*clnt_stop_idle)(q);
3445 					}
3446 				} else {
3447 					/*
3448 					 * We don't mark the connection
3449 					 * as dead, but we turn off the
3450 					 * idle timer.
3451 					 */
3452 					mutex_exit(&connmgr_lock);
3453 					have_connmgr_lock = 0;
3454 					if (clnt_stop_idle != NULL)
3455 						(*clnt_stop_idle)(q);
3456 					RPCLOG(1, "clnt_dispatch_notifyall:"
3457 					    " ignoring timeout from rpcmod"
3458 					    " (q %p) because we are not "
3459 					    " connected\n", (void *)q);
3460 				}
3461 				break;
3462 			case T_ORDREL_IND:
3463 				/*
3464 				 * If this entry is marked closing, then we are
3465 				 * completing a close handshake, and the
3466 				 * connection is dead.  Otherwise, the server is
3467 				 * trying to close. Since the server will not
3468 				 * be sending any more RPC replies, we abort
3469 				 * the connection, including flushing
3470 				 * any RPC requests that are in-transit.
3471 				 * In either case, mark the entry as dead so
3472 				 * that it can be closed by the connection
3473 				 * manager's garbage collector.
3474 				 */
3475 				cm_entry->x_dead = TRUE;
3476 				if (cm_entry->x_closing) {
3477 					mutex_exit(&connmgr_lock);
3478 					have_connmgr_lock = 0;
3479 					if (clnt_stop_idle != NULL)
3480 						(*clnt_stop_idle)(q);
3481 				} else {
3482 					/*
3483 					 * if we're getting a disconnect
3484 					 * before we've finished our
3485 					 * connect attempt, mark it for
3486 					 * later processing
3487 					 */
3488 					if (cm_entry->x_thread)
3489 						cm_entry->x_early_disc = TRUE;
3490 					else
3491 						cm_entry->x_connected = FALSE;
3492 					cm_entry->x_waitdis = TRUE;
3493 					connmgr_snddis(cm_entry);
3494 					have_connmgr_lock = 0;
3495 				}
3496 				break;
3497 
3498 			case T_ERROR_ACK:
3499 			case T_OK_ACK:
3500 				cm_entry->x_waitdis = FALSE;
3501 				cv_signal(&cm_entry->x_dis_cv);
3502 				mutex_exit(&connmgr_lock);
3503 				return;
3504 
3505 			case T_DISCON_REQ:
3506 				if (cm_entry->x_thread)
3507 					cm_entry->x_early_disc = TRUE;
3508 				else
3509 					cm_entry->x_connected = FALSE;
3510 				cm_entry->x_waitdis = TRUE;
3511 
3512 				connmgr_snddis(cm_entry);
3513 				have_connmgr_lock = 0;
3514 				break;
3515 
3516 			case T_DISCON_IND:
3517 			default:
3518 				/*
3519 				 * if we're getting a disconnect before
3520 				 * we've finished our connect attempt,
3521 				 * mark it for later processing
3522 				 */
3523 				if (cm_entry->x_closing) {
3524 					cm_entry->x_dead = TRUE;
3525 					mutex_exit(&connmgr_lock);
3526 					have_connmgr_lock = 0;
3527 					if (clnt_stop_idle != NULL)
3528 						(*clnt_stop_idle)(q);
3529 				} else {
3530 					if (cm_entry->x_thread) {
3531 						cm_entry->x_early_disc = TRUE;
3532 					} else {
3533 						cm_entry->x_dead = TRUE;
3534 						cm_entry->x_connected = FALSE;
3535 					}
3536 				}
3537 				break;
3538 			}
3539 			break;
3540 		}
3541 	}
3542 
3543 	if (have_connmgr_lock)
3544 		mutex_exit(&connmgr_lock);
3545 
3546 	if (msg_type == T_ERROR_ACK || msg_type == T_OK_ACK) {
3547 		RPCLOG(1, "clnt_dispatch_notifyall: (wq %p) could not find "
3548 		    "connmgr entry for discon ack\n", (void *)q);
3549 		return;
3550 	}
3551 
3552 	/*
3553 	 * Then kick all the clnt_pending calls out of their wait.  There
3554 	 * should be no clnt_pending calls in the case of rpcmod's idle
3555 	 * timer firing.
3556 	 */
3557 	for (i = 0; i < clnt_cots_hash_size; i++) {
3558 		ctp = &cots_call_ht[i];
3559 		mutex_enter(&ctp->ct_lock);
3560 		for (e = ctp->ct_call_next;
3561 		    e != (calllist_t *)ctp;
3562 		    e = e->call_next) {
3563 			if (e->call_wq == q && e->call_notified == FALSE) {
3564 				RPCLOG(1,
3565 				    "clnt_dispatch_notifyall for queue %p ",
3566 				    (void *)q);
3567 				RPCLOG(1, "aborting clnt_pending call %p\n",
3568 				    (void *)e);
3569 
3570 				if (msg_type == T_DISCON_IND)
3571 					e->call_reason = reason;
3572 				e->call_notified = TRUE;
3573 				e->call_status = RPC_XPRTFAILED;
3574 				cv_signal(&e->call_cv);
3575 			}
3576 		}
3577 		mutex_exit(&ctp->ct_lock);
3578 	}
3579 
3580 	mutex_enter(&clnt_pending_lock);
3581 	for (e = clnt_pending; e; e = e->call_next) {
3582 		/*
3583 		 * Only signal those RPC handles that haven't been
3584 		 * signalled yet. Otherwise we can get a bogus call_reason.
3585 		 * This can happen if thread A is making a call over a
3586 		 * connection. If the server is killed, it will cause
3587 		 * reset, and reason will default to EIO as a result of
3588 		 * a T_ORDREL_IND. Thread B then attempts to recreate
3589 		 * the connection but gets a T_DISCON_IND. If we set the
3590 		 * call_reason code for all threads, then if thread A
3591 		 * hasn't been dispatched yet, it will get the wrong
3592 		 * reason. The bogus call_reason can make it harder to
3593 		 * discriminate between calls that fail because the
3594 		 * connection attempt failed versus those where the call
3595 		 * may have been executed on the server.
3596 		 */
3597 		if (e->call_wq == q && e->call_notified == FALSE) {
3598 			RPCLOG(1, "clnt_dispatch_notifyall for queue %p ",
3599 			    (void *)q);
3600 			RPCLOG(1, " aborting clnt_pending call %p\n",
3601 			    (void *)e);
3602 
3603 			if (msg_type == T_DISCON_IND)
3604 				e->call_reason = reason;
3605 			e->call_notified = TRUE;
3606 			/*
3607 			 * Let the caller timeout, else he will retry
3608 			 * immediately.
3609 			 */
3610 			e->call_status = RPC_XPRTFAILED;
3611 
3612 			/*
3613 			 * We used to just signal those threads
3614 			 * waiting for a connection, (call_xid = 0).
3615 			 * That meant that threads waiting for a response
3616 			 * waited till their timeout expired. This
3617 			 * could be a long time if they've specified a
3618 			 * maximum timeout. (2^31 - 1). So we
3619 			 * Signal all threads now.
3620 			 */
3621 			cv_signal(&e->call_cv);
3622 		}
3623 	}
3624 	mutex_exit(&clnt_pending_lock);
3625 }
3626 
3627 
3628 /*ARGSUSED*/
3629 /*
3630  * after resuming a system that's been suspended for longer than the
3631  * NFS server's idle timeout (svc_idle_timeout for Solaris 2), rfscall()
3632  * generates "NFS server X not responding" and "NFS server X ok" messages;
3633  * here we reset inet connections to cause a re-connect and avoid those
3634  * NFS messages.  see 4045054
3635  */
3636 boolean_t
3637 connmgr_cpr_reset(void *arg, int code)
3638 {
3639 	struct cm_xprt *cxp;
3640 
3641 	if (code == CB_CODE_CPR_CHKPT)
3642 		return (B_TRUE);
3643 
3644 	if (mutex_tryenter(&connmgr_lock) == 0)
3645 		return (B_FALSE);
3646 	for (cxp = cm_hd; cxp; cxp = cxp->x_next) {
3647 		if ((cxp->x_family == AF_INET || cxp->x_family == AF_INET6) &&
3648 		    cxp->x_connected == TRUE) {
3649 			if (cxp->x_thread)
3650 				cxp->x_early_disc = TRUE;
3651 			else
3652 				cxp->x_connected = FALSE;
3653 			cxp->x_needdis = TRUE;
3654 		}
3655 	}
3656 	mutex_exit(&connmgr_lock);
3657 	return (B_TRUE);
3658 }
3659 
3660 void
3661 clnt_cots_stats_init(zoneid_t zoneid, struct rpc_cots_client **statsp)
3662 {
3663 
3664 	*statsp = (struct rpc_cots_client *)rpcstat_zone_init_common(zoneid,
3665 	    "unix", "rpc_cots_client", (const kstat_named_t *)&cots_rcstat_tmpl,
3666 	    sizeof (cots_rcstat_tmpl));
3667 }
3668 
3669 void
3670 clnt_cots_stats_fini(zoneid_t zoneid, struct rpc_cots_client **statsp)
3671 {
3672 	rpcstat_zone_fini_common(zoneid, "unix", "rpc_cots_client");
3673 	kmem_free(*statsp, sizeof (cots_rcstat_tmpl));
3674 }
3675 
3676 void
3677 clnt_cots_init(void)
3678 {
3679 	mutex_init(&connmgr_lock, NULL, MUTEX_DEFAULT, NULL);
3680 	mutex_init(&clnt_pending_lock, NULL, MUTEX_DEFAULT, NULL);
3681 
3682 	if (clnt_cots_hash_size < DEFAULT_MIN_HASH_SIZE)
3683 		clnt_cots_hash_size = DEFAULT_MIN_HASH_SIZE;
3684 
3685 	cots_call_ht = call_table_init(clnt_cots_hash_size);
3686 	zone_key_create(&zone_cots_key, NULL, NULL, clnt_zone_destroy);
3687 }
3688 
3689 void
3690 clnt_cots_fini(void)
3691 {
3692 	(void) zone_key_delete(zone_cots_key);
3693 }
3694 
3695 /*
3696  * Wait for TPI ack, returns success only if expected ack is received
3697  * within timeout period.
3698  */
3699 
3700 static int
3701 waitforack(calllist_t *e, t_scalar_t ack_prim, const struct timeval *waitp,
3702     bool_t nosignal)
3703 {
3704 	union T_primitives *tpr;
3705 	clock_t timout;
3706 	int cv_stat = 1;
3707 
3708 	ASSERT(MUTEX_HELD(&clnt_pending_lock));
3709 	while (e->call_reply == NULL) {
3710 		if (waitp != NULL) {
3711 			timout = waitp->tv_sec * drv_usectohz(MICROSEC) +
3712 			    drv_usectohz(waitp->tv_usec);
3713 			if (nosignal)
3714 				cv_stat = cv_reltimedwait(&e->call_cv,
3715 				    &clnt_pending_lock, timout, TR_CLOCK_TICK);
3716 			else
3717 				cv_stat = cv_reltimedwait_sig(&e->call_cv,
3718 				    &clnt_pending_lock, timout, TR_CLOCK_TICK);
3719 		} else {
3720 			if (nosignal)
3721 				cv_wait(&e->call_cv, &clnt_pending_lock);
3722 			else
3723 				cv_stat = cv_wait_sig(&e->call_cv,
3724 				    &clnt_pending_lock);
3725 		}
3726 		if (cv_stat == -1)
3727 			return (ETIME);
3728 		if (cv_stat == 0)
3729 			return (EINTR);
3730 		/*
3731 		 * if we received an error from the server and we know a reply
3732 		 * is not going to be sent, do not wait for the full timeout,
3733 		 * return now.
3734 		 */
3735 		if (e->call_status == RPC_XPRTFAILED)
3736 			return (e->call_reason);
3737 	}
3738 	tpr = (union T_primitives *)e->call_reply->b_rptr;
3739 	if (tpr->type == ack_prim)
3740 		return (0); /* Success */
3741 
3742 	if (tpr->type == T_ERROR_ACK) {
3743 		if (tpr->error_ack.TLI_error == TSYSERR)
3744 			return (tpr->error_ack.UNIX_error);
3745 		else
3746 			return (t_tlitosyserr(tpr->error_ack.TLI_error));
3747 	}
3748 
3749 	return (EPROTO); /* unknown or unexpected primitive */
3750 }
3751