xref: /titanic_50/usr/src/uts/common/fs/nfs/nfs4_recovery.c (revision f22acdfff536d452df49dd85c5ecd42092b8fcad)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License, Version 1.0 only
6  * (the "License").  You may not use this file except in compliance
7  * with the License.
8  *
9  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
10  * or http://www.opensolaris.org/os/licensing.
11  * See the License for the specific language governing permissions
12  * and limitations under the License.
13  *
14  * When distributing Covered Code, include this CDDL HEADER in each
15  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
16  * If applicable, add the following below this CDDL HEADER, with the
17  * fields enclosed by brackets "[]" replaced with your own identifying
18  * information: Portions Copyright [yyyy] [name of copyright owner]
19  *
20  * CDDL HEADER END
21  */
22 /*
23  * Copyright 2005 Sun Microsystems, Inc.  All rights reserved.
24  * Use is subject to license terms.
25  */
26 
27 #pragma ident	"%Z%%M%	%I%	%E% SMI"
28 
29 /*
30  * NFS Version 4 state recovery code.
31  */
32 
33 #include <nfs/nfs4_clnt.h>
34 #include <nfs/nfs4.h>
35 #include <nfs/rnode4.h>
36 #include <sys/cmn_err.h>
37 #include <sys/cred.h>
38 #include <sys/systm.h>
39 #include <sys/flock.h>
40 #include <sys/dnlc.h>
41 #include <sys/ddi.h>
42 #include <sys/disp.h>
43 #include <sys/list.h>
44 #include <sys/sdt.h>
45 
46 extern r4hashq_t *rtable4;
47 
48 /*
49  * Information that describes what needs to be done for recovery.  It is
50  * passed to a client recovery thread as well as passed to various recovery
51  * routines.  rc_mi, rc_vp1, and rc_vp2 refer to the filesystem and
52  * vnode(s) affected by recovery.  rc_vp1 and rc_vp2 are references (use
53  * VN_HOLD) or NULL.  rc_lost_rqst contains information about the lost
54  * lock or open/close request, and it holds reference counts for the
55  * various objects (vnode, etc.).  The recovery thread also uses flags set
56  * in the mntinfo4_t or vnode_t to tell it what to do.  rc_error is used
57  * to save the error that originally triggered the recovery event -- will
58  * later be used to set mi_error if recovery doesn't work.  rc_bseqid_rqst
59  * contains information about the request that got NFS4ERR_BAD_SEQID, and
60  * it holds reference count for the various objects (vnode, open owner,
61  * open stream, lock owner).
62  */
63 
64 typedef struct {
65 	mntinfo4_t *rc_mi;
66 	vnode_t *rc_vp1;
67 	vnode_t *rc_vp2;
68 	nfs4_recov_t rc_action;
69 	stateid4 rc_stateid;
70 	bool_t rc_srv_reboot;		/* server has rebooted */
71 	nfs4_lost_rqst_t *rc_lost_rqst;
72 	nfs4_error_t rc_orig_errors;	/* original errors causing recovery */
73 	int rc_error;
74 	nfs4_bseqid_entry_t *rc_bseqid_rqst;
75 } recov_info_t;
76 
77 /*
78  * How long to wait before trying again if there is an error doing
79  * recovery, in seconds.
80  */
81 
82 static int recov_err_delay = 1;
83 
84 /*
85  * How long to wait when processing NFS4ERR_GRACE or NFS4ERR_DELAY
86  * errors.  Expressed in seconds.  Default is defined as
87  * NFS4ERR_DELAY_TIME and this variable is initialized in nfs4_subr_init()
88  */
89 time_t nfs4err_delay_time = 0;
90 
91 /*
92  * Tuneable to limit how many time "exempt" ops go OTW
93  * after a recovery error.  Exempt op hints are OH_CLOSE,
94  * OH_LOCKU, OH_DELEGRETURN.  These previously always went
95  * OTW even after rnode was "dead" due to recovery errors.
96  *
97  * The tuneable below limits the number of times a start_fop
98  * invocation will retry the exempt hints.  After the limit
99  * is reached, nfs4_start_fop will return an error just like
100  * it would for non-exempt op hints.
101  */
102 int nfs4_max_recov_error_retry = 3;
103 
104 /*
105  * Number of seconds the recovery thread should pause before retry when the
106  * filesystem has been forcibly unmounted.
107  */
108 
109 int nfs4_unmount_delay = 1;
110 
111 #ifdef DEBUG
112 
113 /*
114  * How long to wait (in seconds) between recovery operations on a given
115  * file.  Normally zero, but could be set longer for testing purposes.
116  */
117 static int nfs4_recovdelay = 0;
118 
119 /*
120  * Switch that controls whether to go into the debugger when recovery
121  * fails.
122  */
123 static int nfs4_fail_recov_stop = 0;
124 
125 /*
126  * Tuneables to debug client namespace interaction with server
127  * mount points:
128  *
129  *	nfs4_srvmnt_fail_cnt:
130  *		number of times EACCES returned because client
131  *		attempted to cross server mountpoint
132  *
133  *	nfs4_srvmnt_debug:
134  *		trigger console printf whenever client attempts
135  *		to cross server mountpoint
136  */
137 int nfs4_srvmnt_fail_cnt = 0;
138 int nfs4_srvmnt_debug = 0;
139 #endif
140 
141 /* forward references, in alphabetic order */
142 static void close_after_open_resend(vnode_t *, cred_t *, uint32_t,
143 	nfs4_error_t *);
144 static void errs_to_action(recov_info_t *,
145 	nfs4_server_t *, mntinfo4_t *, stateid4 *, nfs4_lost_rqst_t *, int,
146 	nfs_opnum4, nfs4_bseqid_entry_t *);
147 static void flush_reinstate(nfs4_lost_rqst_t *);
148 static void free_milist(mntinfo4_t **, int);
149 static mntinfo4_t **make_milist(nfs4_server_t *, int *);
150 static int nfs4_check_recov_err(vnode_t *, nfs4_op_hint_t,
151 	nfs4_recov_state_t *, int, char *);
152 static int nfs4_check_srvstub(vnode_t *vp, rnode4_t *rp, nfs4_op_hint_t op);
153 static char *nfs4_getsrvnames(mntinfo4_t *, size_t *);
154 static void nfs4_recov_fh_fail(vnode_t *, int, nfsstat4);
155 static void nfs4_recov_thread(recov_info_t *);
156 static void nfs4_remove_lost_rqsts(mntinfo4_t *, nfs4_server_t *);
157 static void nfs4_resend_lost_rqsts(recov_info_t *, nfs4_server_t *);
158 static cred_t *pid_to_cr(pid_t);
159 static void reclaim_one_lock(vnode_t *, flock64_t *, nfs4_error_t *, int *);
160 static void recov_bad_seqid(recov_info_t *);
161 static void recov_badstate(recov_info_t *, vnode_t *, nfsstat4);
162 static void recov_clientid(recov_info_t *, nfs4_server_t *);
163 static void recov_done(mntinfo4_t *, recov_info_t *);
164 static void recov_filehandle(nfs4_recov_t, mntinfo4_t *, vnode_t *);
165 static void recov_newserver(recov_info_t *, nfs4_server_t **, bool_t *);
166 static void recov_openfiles(recov_info_t *, nfs4_server_t *);
167 static void recov_stale(mntinfo4_t *, vnode_t *);
168 static void nfs4_free_lost_rqst(nfs4_lost_rqst_t *, nfs4_server_t *);
169 static void recov_throttle(recov_info_t *, vnode_t *);
170 static void relock_skip_pid(locklist_t *, pid_t);
171 static void resend_lock(nfs4_lost_rqst_t *, nfs4_error_t *);
172 static void resend_one_op(nfs4_lost_rqst_t *, nfs4_error_t *, mntinfo4_t *,
173 	nfs4_server_t *);
174 static void save_bseqid_rqst(nfs4_bseqid_entry_t *, recov_info_t *);
175 static void start_recovery(recov_info_t *, mntinfo4_t *, vnode_t *, vnode_t *,
176 	nfs4_server_t *);
177 static void start_recovery_action(nfs4_recov_t, bool_t, mntinfo4_t *, vnode_t *,
178 	vnode_t *);
179 static int wait_for_recovery(mntinfo4_t *, nfs4_op_hint_t);
180 
181 /*
182  * Return non-zero if the given errno, status, and rpc status codes
183  * in the nfs4_error_t indicate that client recovery is needed.
184  * "stateful" indicates whether the call that got the error establishes or
185  * removes state on the server (open, close, lock, unlock, delegreturn).
186  */
187 
188 int
189 nfs4_needs_recovery(nfs4_error_t *ep, bool_t stateful, vfs_t *vfsp)
190 {
191 	int recov = 0;
192 	mntinfo4_t *mi;
193 
194 	/*
195 	 * Try failover if the error values justify it and if
196 	 * it's a failover mount.  Don't try if the mount is in
197 	 * progress, failures are handled explicitly by nfs4rootvp.
198 	 */
199 	if (nfs4_try_failover(ep)) {
200 		mi = VFTOMI4(vfsp);
201 		mutex_enter(&mi->mi_lock);
202 		recov = FAILOVER_MOUNT4(mi) && !(mi->mi_flags & MI4_MOUNTING);
203 		mutex_exit(&mi->mi_lock);
204 		if (recov)
205 			return (recov);
206 	}
207 
208 	if (ep->error == EINTR || NFS4_FRC_UNMT_ERR(ep->error, vfsp)) {
209 		/*
210 		 * The server may have gotten the request, so for stateful
211 		 * ops we need to resynchronize and possibly back out the
212 		 * op.
213 		 */
214 		return (stateful);
215 	}
216 	if (ep->error != 0)
217 		return (0);
218 
219 	/* stat values are listed alphabetically */
220 	/*
221 	 * There are two lists here: the errors for which we have code, and
222 	 * the errors for which we plan to have code before FCS.  For the
223 	 * second list, print a warning message but don't attempt recovery.
224 	 */
225 	switch (ep->stat) {
226 	case NFS4ERR_BADHANDLE:
227 	case NFS4ERR_BAD_SEQID:
228 	case NFS4ERR_BAD_STATEID:
229 	case NFS4ERR_DELAY:
230 	case NFS4ERR_EXPIRED:
231 	case NFS4ERR_FHEXPIRED:
232 	case NFS4ERR_GRACE:
233 	case NFS4ERR_OLD_STATEID:
234 	case NFS4ERR_RESOURCE:
235 	case NFS4ERR_STALE_CLIENTID:
236 	case NFS4ERR_STALE_STATEID:
237 	case NFS4ERR_WRONGSEC:
238 	case NFS4ERR_STALE:
239 		recov = 1;
240 		break;
241 #ifdef DEBUG
242 	case NFS4ERR_LEASE_MOVED:
243 	case NFS4ERR_MOVED:
244 		zcmn_err(VFTOMI4(vfsp)->mi_zone->zone_id,
245 		    CE_WARN, "!Can't yet recover from NFS status %d",
246 				ep->stat);
247 		break;
248 #endif
249 	}
250 
251 	return (recov);
252 }
253 
254 /*
255  * Some operations such as DELEGRETURN want to avoid invoking
256  * recovery actions that will only mark the file dead.  If
257  * better handlers are invoked for any of these errors, this
258  * routine should be modified.
259  */
260 int
261 nfs4_recov_marks_dead(nfsstat4 status)
262 {
263 	if (status == NFS4ERR_BAD_SEQID ||
264 	    status == NFS4ERR_EXPIRED ||
265 	    status == NFS4ERR_BAD_STATEID ||
266 	    status == NFS4ERR_OLD_STATEID)
267 		return (1);
268 	return (0);
269 }
270 
271 /*
272  * Transfer the state recovery information in recovp to mi's resend queue,
273  * and mark mi as having a lost state request.
274  */
275 static void
276 nfs4_enqueue_lost_rqst(recov_info_t *recovp, mntinfo4_t *mi)
277 {
278 	nfs4_lost_rqst_t *lrp = recovp->rc_lost_rqst;
279 
280 	ASSERT(nfs_rw_lock_held(&mi->mi_recovlock, RW_READER) ||
281 	    nfs_rw_lock_held(&mi->mi_recovlock, RW_WRITER));
282 
283 	ASSERT(lrp != NULL && lrp->lr_op != 0);
284 
285 	NFS4_DEBUG(nfs4_lost_rqst_debug, (CE_NOTE,
286 		"nfs4_enqueue_lost_rqst %p, op %d",
287 		(void *)lrp, lrp->lr_op));
288 
289 	mutex_enter(&mi->mi_lock);
290 	mi->mi_recovflags |= MI4R_LOST_STATE;
291 	if (lrp->lr_putfirst)
292 		list_insert_head(&mi->mi_lost_state, lrp);
293 	else
294 		list_insert_tail(&mi->mi_lost_state, lrp);
295 	recovp->rc_lost_rqst = NULL;
296 	mutex_exit(&mi->mi_lock);
297 
298 	nfs4_queue_event(RE_LOST_STATE, mi, NULL, lrp->lr_op, lrp->lr_vp,
299 		lrp->lr_dvp, 0, NULL, 0, TAG_NONE, TAG_NONE, 0, 0);
300 }
301 
302 /*
303  * Transfer the bad seqid recovery information in recovp to mi's
304  * bad seqid queue, and mark mi as having a bad seqid request.
305  */
306 void
307 enqueue_bseqid_rqst(recov_info_t *recovp, mntinfo4_t *mi)
308 {
309 	ASSERT(nfs_rw_lock_held(&mi->mi_recovlock, RW_READER) ||
310 	    nfs_rw_lock_held(&mi->mi_recovlock, RW_WRITER));
311 	ASSERT(recovp->rc_bseqid_rqst != NULL);
312 
313 	mutex_enter(&mi->mi_lock);
314 	mi->mi_recovflags |= MI4R_BAD_SEQID;
315 	list_insert_tail(&mi->mi_bseqid_list, recovp->rc_bseqid_rqst);
316 	recovp->rc_bseqid_rqst = NULL;
317 	mutex_exit(&mi->mi_lock);
318 }
319 
320 /*
321  * Initiate recovery.
322  *
323  * The nfs4_error_t contains the return codes that triggered a recovery
324  * attempt.  mi, vp1, and vp2 refer to the filesystem and files that were
325  * being operated on.  vp1 and vp2 may be NULL.
326  *
327  * Multiple calls are okay.  If recovery is already underway, the call
328  * updates the information about what state needs recovery but does not
329  * start a new thread.  The caller should hold mi->mi_recovlock as a reader
330  * for proper synchronization with any recovery thread.
331  *
332  * This will return TRUE if recovery was aborted, and FALSE otherwise.
333  */
334 bool_t
335 nfs4_start_recovery(nfs4_error_t *ep, mntinfo4_t *mi, vnode_t *vp1,
336     vnode_t *vp2, stateid4 *sid, nfs4_lost_rqst_t *lost_rqstp, nfs_opnum4 op,
337     nfs4_bseqid_entry_t *bsep)
338 {
339 	recov_info_t *recovp;
340 	nfs4_server_t *sp;
341 	bool_t abort = FALSE;
342 	bool_t gone = FALSE;
343 
344 	ASSERT(curproc->p_zone == mi->mi_zone);
345 	mutex_enter(&mi->mi_lock);
346 	/*
347 	 * If there is lost state, we need to kick off recovery even if the
348 	 * filesystem has been unmounted or the zone is shutting down.
349 	 */
350 	gone = FS_OR_ZONE_GONE4(mi->mi_vfsp);
351 	if (gone) {
352 		ASSERT(ep->error != EINTR || lost_rqstp != NULL);
353 		if (ep->error == EIO && lost_rqstp == NULL) {
354 			/* failed due to forced unmount, no new lost state */
355 			abort = TRUE;
356 		}
357 		if ((ep->error == 0 || ep->error == ETIMEDOUT) &&
358 		    !(mi->mi_recovflags & MI4R_LOST_STATE)) {
359 			/* some other failure, no existing lost state */
360 			abort = TRUE;
361 		}
362 		if (abort) {
363 			mutex_exit(&mi->mi_lock);
364 			NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
365 				    "nfs4_start_recovery: fs unmounted"));
366 			return (TRUE);
367 		}
368 	}
369 	mi->mi_in_recovery++;
370 	mutex_exit(&mi->mi_lock);
371 
372 	recovp = kmem_alloc(sizeof (recov_info_t), KM_SLEEP);
373 	recovp->rc_orig_errors = *ep;
374 	sp = find_nfs4_server(mi);
375 	errs_to_action(recovp, sp, mi, sid, lost_rqstp,
376 		gone, op, bsep);
377 	if (sp != NULL)
378 		mutex_exit(&sp->s_lock);
379 	start_recovery(recovp, mi, vp1, vp2, sp);
380 	if (sp != NULL)
381 		nfs4_server_rele(sp);
382 	return (FALSE);
383 }
384 
385 /*
386  * Internal version of nfs4_start_recovery.  The difference is that the
387  * caller specifies the recovery action, rather than the errors leading to
388  * recovery.
389  */
390 static void
391 start_recovery_action(nfs4_recov_t what, bool_t reboot, mntinfo4_t *mi,
392 	vnode_t *vp1, vnode_t *vp2)
393 {
394 	recov_info_t *recovp;
395 
396 	ASSERT(curproc->p_zone == mi->mi_zone);
397 	mutex_enter(&mi->mi_lock);
398 	mi->mi_in_recovery++;
399 	mutex_exit(&mi->mi_lock);
400 
401 	recovp = kmem_zalloc(sizeof (recov_info_t), KM_SLEEP);
402 	recovp->rc_action = what;
403 	recovp->rc_srv_reboot = reboot;
404 	recovp->rc_error = EIO;
405 	start_recovery(recovp, mi, vp1, vp2, NULL);
406 }
407 
408 static void
409 start_recovery(recov_info_t *recovp, mntinfo4_t *mi,
410 	vnode_t *vp1, vnode_t *vp2, nfs4_server_t *sp)
411 {
412 	NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
413 		"start_recovery: mi %p, what %s", (void*)mi,
414 		nfs4_recov_action_to_str(recovp->rc_action)));
415 
416 	/*
417 	 * Bump the reference on the vfs so that we can pass it to the
418 	 * recovery thread.
419 	 */
420 	VFS_HOLD(mi->mi_vfsp);
421 
422 again:
423 	switch (recovp->rc_action) {
424 	case NR_FAILOVER:
425 		ASSERT(nfs_rw_lock_held(&mi->mi_recovlock, RW_READER) ||
426 		    nfs_rw_lock_held(&mi->mi_recovlock, RW_WRITER));
427 		if (mi->mi_servers->sv_next == NULL)
428 			goto out_no_thread;
429 		mutex_enter(&mi->mi_lock);
430 		mi->mi_recovflags |= MI4R_NEED_NEW_SERVER;
431 		mutex_exit(&mi->mi_lock);
432 
433 		if (recovp->rc_lost_rqst != NULL)
434 			nfs4_enqueue_lost_rqst(recovp, mi);
435 		break;
436 
437 	case NR_CLIENTID:
438 		/*
439 		 * If the filesystem has been unmounted, punt.
440 		 */
441 		if (sp == NULL)
442 			goto out_no_thread;
443 
444 		/*
445 		 * If nobody else is working on the clientid, mark the
446 		 * clientid as being no longer set.  Then mark the specific
447 		 * filesystem being worked on.
448 		 */
449 		if (!nfs4_server_in_recovery(sp)) {
450 			mutex_enter(&sp->s_lock);
451 			sp->s_flags &= ~N4S_CLIENTID_SET;
452 			mutex_exit(&sp->s_lock);
453 		}
454 		ASSERT(nfs_rw_lock_held(&mi->mi_recovlock, RW_READER) ||
455 		    nfs_rw_lock_held(&mi->mi_recovlock, RW_WRITER));
456 		mutex_enter(&mi->mi_lock);
457 		mi->mi_recovflags |= MI4R_NEED_CLIENTID;
458 		if (recovp->rc_srv_reboot)
459 			mi->mi_recovflags |= MI4R_SRV_REBOOT;
460 		mutex_exit(&mi->mi_lock);
461 		break;
462 
463 	case NR_OPENFILES:
464 		ASSERT(nfs_rw_lock_held(&mi->mi_recovlock, RW_READER) ||
465 		    nfs_rw_lock_held(&mi->mi_recovlock, RW_WRITER));
466 		mutex_enter(&mi->mi_lock);
467 		mi->mi_recovflags |= MI4R_REOPEN_FILES;
468 		if (recovp->rc_srv_reboot)
469 			mi->mi_recovflags |= MI4R_SRV_REBOOT;
470 		mutex_exit(&mi->mi_lock);
471 		break;
472 
473 	case NR_WRONGSEC:
474 		ASSERT(nfs_rw_lock_held(&mi->mi_recovlock, RW_READER) ||
475 		    nfs_rw_lock_held(&mi->mi_recovlock, RW_WRITER));
476 		mutex_enter(&mi->mi_lock);
477 		mi->mi_recovflags |= MI4R_NEED_SECINFO;
478 		mutex_exit(&mi->mi_lock);
479 		break;
480 
481 	case NR_EXPIRED:
482 		if (vp1 != NULL)
483 			recov_badstate(recovp, vp1, NFS4ERR_EXPIRED);
484 		if (vp2 != NULL)
485 			recov_badstate(recovp, vp2, NFS4ERR_EXPIRED);
486 		goto out_no_thread;	/* no further recovery possible */
487 
488 	case NR_BAD_STATEID:
489 		if (vp1 != NULL)
490 			recov_badstate(recovp, vp1, NFS4ERR_BAD_STATEID);
491 		if (vp2 != NULL)
492 			recov_badstate(recovp, vp2, NFS4ERR_BAD_STATEID);
493 		goto out_no_thread;	/* no further recovery possible */
494 
495 	case NR_FHEXPIRED:
496 	case NR_BADHANDLE:
497 		if (vp1 != NULL)
498 			recov_throttle(recovp, vp1);
499 		if (vp2 != NULL)
500 			recov_throttle(recovp, vp2);
501 		/*
502 		 * Recover the filehandle now, rather than using a
503 		 * separate thread.  We can do this because filehandle
504 		 * recovery is independent of any other state, and because
505 		 * we know that we are not competing with the recovery
506 		 * thread at this time.  recov_filehandle will deal with
507 		 * threads that are competing to recover this filehandle.
508 		 */
509 		ASSERT(nfs_rw_lock_held(&mi->mi_recovlock, RW_READER) ||
510 		    nfs_rw_lock_held(&mi->mi_recovlock, RW_WRITER));
511 		if (vp1 != NULL)
512 			recov_filehandle(recovp->rc_action, mi, vp1);
513 		if (vp2 != NULL)
514 			recov_filehandle(recovp->rc_action, mi, vp2);
515 		goto out_no_thread;	/* no further recovery needed */
516 
517 	case NR_STALE:
518 		/*
519 		 * NFS4ERR_STALE handling
520 		 * recov_stale() could set MI4R_NEED_NEW_SERVER to
521 		 * indicate that we can and should failover.
522 		 */
523 		ASSERT(nfs_rw_lock_held(&mi->mi_recovlock, RW_READER) ||
524 		    nfs_rw_lock_held(&mi->mi_recovlock, RW_WRITER));
525 
526 		if (vp1 != NULL)
527 			recov_stale(mi, vp1);
528 		if (vp2 != NULL)
529 			recov_stale(mi, vp2);
530 		mutex_enter(&mi->mi_lock);
531 		if ((mi->mi_recovflags & MI4R_NEED_NEW_SERVER) == 0) {
532 			mutex_exit(&mi->mi_lock);
533 			goto out_no_thread;
534 		}
535 		mutex_exit(&mi->mi_lock);
536 		recovp->rc_action = NR_FAILOVER;
537 		goto again;
538 
539 	case NR_BAD_SEQID:
540 		if (recovp->rc_bseqid_rqst) {
541 			enqueue_bseqid_rqst(recovp, mi);
542 			break;
543 		}
544 
545 		if (vp1 != NULL)
546 			recov_badstate(recovp, vp1, NFS4ERR_BAD_SEQID);
547 		if (vp2 != NULL)
548 			recov_badstate(recovp, vp2, NFS4ERR_BAD_SEQID);
549 		goto out_no_thread; /* no further recovery possible */
550 
551 	case NR_OLDSTATEID:
552 		if (vp1 != NULL)
553 			recov_badstate(recovp, vp1, NFS4ERR_OLD_STATEID);
554 		if (vp2 != NULL)
555 			recov_badstate(recovp, vp2, NFS4ERR_OLD_STATEID);
556 		goto out_no_thread;	/* no further recovery possible */
557 
558 	case NR_GRACE:
559 		nfs4_set_grace_wait(mi);
560 		goto out_no_thread; /* no further action required for GRACE */
561 
562 	case NR_DELAY:
563 		if (vp1)
564 			nfs4_set_delay_wait(vp1);
565 		goto out_no_thread; /* no further action required for DELAY */
566 
567 	case NR_LOST_STATE_RQST:
568 	case NR_LOST_LOCK:
569 		nfs4_enqueue_lost_rqst(recovp, mi);
570 		break;
571 
572 	default:
573 		nfs4_queue_event(RE_UNEXPECTED_ACTION, mi, NULL,
574 		    recovp->rc_action, NULL, NULL, 0, NULL, 0, TAG_NONE,
575 		    TAG_NONE, 0, 0);
576 		goto out_no_thread;
577 	}
578 
579 	/*
580 	 * If either file recently went through the same recovery, wait
581 	 * awhile.  This is in case there is some sort of bug; we might not
582 	 * be able to recover properly, but at least we won't bombard the
583 	 * server with calls, and we won't tie up the client.
584 	 */
585 	if (vp1 != NULL)
586 		recov_throttle(recovp, vp1);
587 	if (vp2 != NULL)
588 		recov_throttle(recovp, vp2);
589 
590 	/*
591 	 * If there's already a recovery thread, don't start another one.
592 	 */
593 
594 	mutex_enter(&mi->mi_lock);
595 	if (mi->mi_flags & MI4_RECOV_ACTIV) {
596 		mutex_exit(&mi->mi_lock);
597 		goto out_no_thread;
598 	}
599 	mi->mi_flags |= MI4_RECOV_ACTIV;
600 	mutex_exit(&mi->mi_lock);
601 	NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
602 		"start_recovery: starting new thread for mi %p", (void*)mi));
603 
604 	recovp->rc_mi = mi;
605 	recovp->rc_vp1 = vp1;
606 	if (vp1 != NULL) {
607 		ASSERT(VTOMI4(vp1) == mi);
608 		VN_HOLD(recovp->rc_vp1);
609 	}
610 	recovp->rc_vp2 = vp2;
611 	if (vp2 != NULL) {
612 		ASSERT(VTOMI4(vp2) == mi);
613 		VN_HOLD(recovp->rc_vp2);
614 	}
615 
616 	(void) zthread_create(NULL, 0, nfs4_recov_thread, recovp, 0,
617 			    minclsyspri);
618 	return;
619 
620 	/* not reached by thread creating call */
621 out_no_thread:
622 	mutex_enter(&mi->mi_lock);
623 	mi->mi_in_recovery--;
624 	cv_broadcast(&mi->mi_cv_in_recov);
625 	mutex_exit(&mi->mi_lock);
626 
627 	VFS_RELE(mi->mi_vfsp);
628 	/*
629 	 * Free up resources that were allocated for us.
630 	 */
631 	kmem_free(recovp, sizeof (recov_info_t));
632 }
633 
634 static int
635 nfs4_check_srvstub(vnode_t *vp, rnode4_t *rp, nfs4_op_hint_t op)
636 {
637 	int err = 0;
638 
639 	/*
640 	 * If tuneable does not allow client to cross srv mountpoints and
641 	 * object is a stub, then check check op hint and return EACCES for
642 	 * any hint other than access, rddir, getattr, lookup.
643 	 */
644 	if (rp->r_flags & R4SRVSTUB && op != OH_ACCESS && op != OH_GETACL &&
645 	    op != OH_GETATTR && op != OH_READDIR && op != OH_LOOKUP) {
646 		err = EACCES;
647 #ifdef DEBUG
648 		NFS4_DEBUG(nfs4_srvmnt_debug, (CE_NOTE,
649 			"nfs4_check_srvstub: op=%d err=%d rp=%p vp=%p\n"
650 			"va_nod=%llx r_mntd_fid=%llx\n"
651 			"sv_fsid=(%llx:%llx) r_srv_fsid=(%llx:%llx)",
652 			op, err, (void *)rp, (void *)vp,
653 			(u_longlong_t)rp->r_attr.va_nodeid,
654 			(u_longlong_t)rp->r_mntd_fid,
655 			(u_longlong_t)rp->r_server->sv_fsid.major,
656 			(u_longlong_t)rp->r_server->sv_fsid.minor,
657 			(u_longlong_t)rp->r_srv_fsid.major,
658 			(u_longlong_t)rp->r_srv_fsid.minor));
659 #endif
660 	}
661 
662 	return (err);
663 }
664 
665 static int
666 nfs4_check_recov_err(vnode_t *vp, nfs4_op_hint_t op,
667 			nfs4_recov_state_t *rsp, int retry_err_cnt, char *str)
668 {
669 	rnode4_t *rp;
670 	int error = 0;
671 	int exempt;
672 
673 	if (vp == NULL)
674 		return (0);
675 
676 	exempt = (op == OH_CLOSE || op == OH_LOCKU || op == OH_DELEGRETURN);
677 	rp = VTOR4(vp);
678 	mutex_enter(&rp->r_statelock);
679 
680 	/*
681 	 * If there was a recovery error, then allow op hints "exempt" from
682 	 * recov errors to retry (currently 3 times).  Either r_error or
683 	 * EIO is returned for non-exempt op hints.
684 	 *
685 	 *	Error heirarchy:
686 	 *	a) check for R4ERECOVERR
687 	 *	b) check for R4SRVSTUB (only if R4RECOVERR is not set).
688 	 */
689 	if (rp->r_flags & R4RECOVERR) {
690 		if (exempt && rsp->rs_num_retry_despite_err <=
691 				nfs4_max_recov_error_retry) {
692 
693 			/*
694 			 * Check to make sure that we haven't already inc'd
695 			 * rs_num_retry_despite_err for current nfs4_start_fop
696 			 * instance.  We don't want to double inc (if we were
697 			 * called with vp2, then the vp1 call could have
698 			 * already incremented.
699 			 */
700 			if (retry_err_cnt == rsp->rs_num_retry_despite_err)
701 				rsp->rs_num_retry_despite_err++;
702 
703 			NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
704 				"nfs4_start_fop: %s %p DEAD, cnt=%d", str,
705 				(void *)vp, rsp->rs_num_retry_despite_err));
706 		} else {
707 			error = (rp->r_error ? rp->r_error : EIO);
708 			/*
709 			 * An ESTALE error on a non-regular file is not
710 			 * "sticky".  Return the ESTALE error once, but
711 			 * clear the condition to allow future operations
712 			 * to go OTW.  This will allow the client to
713 			 * recover if the server has merely unshared then
714 			 * re-shared the file system.  For regular files,
715 			 * the unshare has destroyed the open state at the
716 			 * server and we aren't willing to do a reopen (yet).
717 			 */
718 			if (error == ESTALE && vp->v_type != VREG) {
719 				rp->r_flags &=
720 					~(R4RECOVERR|R4RECOVERRP|R4STALE);
721 				rp->r_error = 0;
722 				error = ESTALE;
723 			}
724 			NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
725 				"nfs4_start_fop: %s %p DEAD, cnt=%d error=%d",
726 				str, (void *)vp,
727 				rsp->rs_num_retry_despite_err, error));
728 		}
729 	} else {
730 		error = nfs4_check_srvstub(vp, rp, op);
731 		NFS4_DEBUG(nfs4_client_recov_stub_debug, (CE_NOTE,
732 			"nfs4_start_fop: %s %p SRVSTUB, error=%d", str,
733 			(void *)vp, error));
734 	}
735 	mutex_exit(&rp->r_statelock);
736 	return (error);
737 }
738 
739 /*
740  * Initial setup code that every operation should call if it might invoke
741  * client recovery.  Can block waiting for recovery to finish on a
742  * filesystem.  Either vnode ptr can be NULL.
743  *
744  * Returns 0 if there are no outstanding errors.  Can return an
745  * errno value under various circumstances (e.g., failed recovery, or
746  * interrupted while waiting for recovery to finish).
747  *
748  * There must be a corresponding call to nfs4_end_op() to free up any locks
749  * or resources allocated by this call (assuming this call succeeded),
750  * using the same rsp that's passed in here.
751  *
752  * The open and lock seqid synchronization must be stopped before calling this
753  * function, as it could lead to deadlock when trying to reopen a file or
754  * reclaim a lock.  The synchronization is obtained with calls to:
755  *   nfs4_start_open_seqid_sync()
756  *   nfs4_start_lock_seqid_sync()
757  *
758  * *startrecovp is set TRUE if the caller should not bother with the
759  * over-the-wire call, and just initiate recovery for the given request.
760  * This is typically used for state-releasing ops if the filesystem has
761  * been forcibly unmounted.  startrecovp may be NULL for
762  * non-state-releasing ops.
763  */
764 
765 int
766 nfs4_start_fop(mntinfo4_t *mi, vnode_t *vp1, vnode_t *vp2, nfs4_op_hint_t op,
767 		nfs4_recov_state_t *rsp, bool_t *startrecovp)
768 {
769 	int error = 0, rerr_cnt;
770 	nfs4_server_t *sp = NULL;
771 	nfs4_server_t *tsp;
772 	nfs4_error_t e = { 0, NFS4_OK, RPC_SUCCESS };
773 	time_t droplock_time;
774 #ifdef DEBUG
775 	void *fop_caller;
776 #endif
777 
778 	ASSERT(vp1 == NULL || vp1->v_vfsp == mi->mi_vfsp);
779 	ASSERT(vp2 == NULL || vp2->v_vfsp == mi->mi_vfsp);
780 
781 #ifdef	DEBUG
782 	if ((fop_caller = tsd_get(nfs4_tsd_key)) != NULL) {
783 		cmn_err(CE_PANIC, "Missing nfs4_end_fop: last caller %p",
784 			fop_caller);
785 	}
786 	(void) tsd_set(nfs4_tsd_key, caller());
787 #endif
788 
789 	rsp->rs_sp = NULL;
790 	rsp->rs_flags &= ~NFS4_RS_RENAME_HELD;
791 	rerr_cnt = rsp->rs_num_retry_despite_err;
792 
793 	/*
794 	 * Process the items that may delay() based on server response
795 	 */
796 	error = nfs4_wait_for_grace(mi, rsp);
797 	if (error)
798 		goto out;
799 
800 	if (vp1 != NULL) {
801 		error = nfs4_wait_for_delay(vp1, rsp);
802 		if (error)
803 			goto out;
804 	}
805 
806 	/* Wait for a delegation recall to complete. */
807 
808 	error = wait_for_recall(vp1, vp2, op, rsp);
809 	if (error)
810 		goto out;
811 
812 	/*
813 	 * Wait for any current recovery actions to finish.  Note that a
814 	 * recovery thread can still start up after wait_for_recovery()
815 	 * finishes.  We don't block out recovery operations until we
816 	 * acquire s_recovlock and mi_recovlock.
817 	 */
818 	error = wait_for_recovery(mi, op);
819 	if (error)
820 		goto out;
821 
822 	/*
823 	 * Check to see if the rnode is already marked with a
824 	 * recovery error.  If so, return it immediately.  But
825 	 * always pass CLOSE, LOCKU, and DELEGRETURN so we can
826 	 * clean up state on the server.
827 	 */
828 
829 	if (vp1 != NULL) {
830 		if (error = nfs4_check_recov_err(vp1, op, rsp, rerr_cnt, "vp1"))
831 			goto out;
832 		nfs4_check_remap(mi, vp1, NFS4_REMAP_CKATTRS, &e);
833 	}
834 
835 	if (vp2 != NULL) {
836 		if (error = nfs4_check_recov_err(vp2, op, rsp, rerr_cnt, "vp2"))
837 			goto out;
838 		nfs4_check_remap(mi, vp2, NFS4_REMAP_CKATTRS, &e);
839 	}
840 
841 	/*
842 	 * The lock order calls for us to acquire s_recovlock before
843 	 * mi_recovlock, but we have to hold mi_recovlock to look up sp (to
844 	 * prevent races with the failover/migration code).  So acquire
845 	 * mi_recovlock, look up sp, drop mi_recovlock, acquire
846 	 * s_recovlock and mi_recovlock, then verify that sp is still the
847 	 * right object.  XXX Can we find a simpler way to deal with this?
848 	 */
849 	if (nfs_rw_enter_sig(&mi->mi_recovlock, RW_READER,
850 	    mi->mi_flags & MI4_INT)) {
851 		error = EINTR;
852 		goto out;
853 	}
854 get_sp:
855 	sp = find_nfs4_server(mi);
856 	if (sp != NULL) {
857 		sp->s_otw_call_count++;
858 		mutex_exit(&sp->s_lock);
859 		droplock_time = gethrestime_sec();
860 	}
861 	nfs_rw_exit(&mi->mi_recovlock);
862 
863 	if (sp != NULL) {
864 		if (nfs_rw_enter_sig(&sp->s_recovlock, RW_READER,
865 			    mi->mi_flags & MI4_INT)) {
866 			error = EINTR;
867 			goto out;
868 		}
869 	}
870 	if (nfs_rw_enter_sig(&mi->mi_recovlock, RW_READER,
871 			    mi->mi_flags & MI4_INT)) {
872 		if (sp != NULL)
873 			nfs_rw_exit(&sp->s_recovlock);
874 		error = EINTR;
875 		goto out;
876 	}
877 	/*
878 	 * If the mntinfo4_t hasn't changed nfs4_sever_ts then
879 	 * there's no point in double checking to make sure it
880 	 * has switched.
881 	 */
882 	if (sp == NULL || droplock_time < mi->mi_srvsettime) {
883 		tsp = find_nfs4_server(mi);
884 		if (tsp != sp) {
885 			/* try again */
886 			if (tsp != NULL) {
887 				mutex_exit(&tsp->s_lock);
888 				nfs4_server_rele(tsp);
889 				tsp = NULL;
890 			}
891 			if (sp != NULL) {
892 				nfs_rw_exit(&sp->s_recovlock);
893 				mutex_enter(&sp->s_lock);
894 				sp->s_otw_call_count--;
895 				mutex_exit(&sp->s_lock);
896 				nfs4_server_rele(sp);
897 				sp = NULL;
898 			}
899 			goto get_sp;
900 		} else {
901 			if (tsp != NULL) {
902 				mutex_exit(&tsp->s_lock);
903 				nfs4_server_rele(tsp);
904 				tsp = NULL;
905 			}
906 		}
907 	}
908 
909 	if (sp != NULL) {
910 		rsp->rs_sp = sp;
911 	}
912 
913 	/*
914 	 * If the fileystem uses volatile filehandles, obtain a lock so
915 	 * that we synchronize with renames.  Exception: mount operations
916 	 * can change mi_fh_expire_type, which could be a problem, since
917 	 * the end_op code needs to be consistent with the start_op code
918 	 * about mi_rename_lock.  Since mounts don't compete with renames,
919 	 * it's simpler to just not acquire the rename lock for mounts.
920 	 */
921 	if (NFS4_VOLATILE_FH(mi) && op != OH_MOUNT) {
922 		if (nfs_rw_enter_sig(&mi->mi_rename_lock,
923 				    op == OH_VFH_RENAME ? RW_WRITER : RW_READER,
924 				    mi->mi_flags & MI4_INT)) {
925 			nfs_rw_exit(&mi->mi_recovlock);
926 			if (sp != NULL)
927 				nfs_rw_exit(&sp->s_recovlock);
928 			error = EINTR;
929 			goto out;
930 		}
931 		rsp->rs_flags |= NFS4_RS_RENAME_HELD;
932 	}
933 
934 	if (OH_IS_STATE_RELE(op)) {
935 		/*
936 		 * For forced unmount, letting the request proceed will
937 		 * almost always delay response to the user, so hand it off
938 		 * to the recovery thread.  For exiting lwp's, we don't
939 		 * have a good way to tell if the request will hang.  We
940 		 * generally want processes to handle their own requests so
941 		 * that they can be done in parallel, but if there is
942 		 * already a recovery thread, hand the request off to it.
943 		 * This will improve user response at no cost to overall
944 		 * system throughput.  For zone shutdown, we'd prefer
945 		 * the recovery thread to handle this as well.
946 		 */
947 		ASSERT(startrecovp != NULL);
948 		mutex_enter(&mi->mi_lock);
949 		if (FS_OR_ZONE_GONE4(mi->mi_vfsp))
950 			*startrecovp = TRUE;
951 		else if ((curthread->t_proc_flag & TP_LWPEXIT) &&
952 		    (mi->mi_flags & MI4_RECOV_ACTIV))
953 			*startrecovp = TRUE;
954 		else
955 			*startrecovp = FALSE;
956 		mutex_exit(&mi->mi_lock);
957 	} else
958 		if (startrecovp != NULL)
959 			*startrecovp = FALSE;
960 
961 	ASSERT(error == 0);
962 	return (error);
963 
964 out:
965 	ASSERT(error != 0);
966 	if (sp != NULL) {
967 		mutex_enter(&sp->s_lock);
968 		sp->s_otw_call_count--;
969 		mutex_exit(&sp->s_lock);
970 		nfs4_server_rele(sp);
971 		rsp->rs_sp = NULL;
972 	}
973 	nfs4_end_op_recall(vp1, vp2, rsp);
974 
975 #ifdef	DEBUG
976 	(void) tsd_set(nfs4_tsd_key, NULL);
977 #endif
978 	return (error);
979 }
980 
981 /*
982  * It is up to the caller to determine if rsp->rs_sp being NULL
983  * is detrimental or not.
984  */
985 int
986 nfs4_start_op(mntinfo4_t *mi, vnode_t *vp1, vnode_t *vp2,
987 	nfs4_recov_state_t *rsp)
988 {
989 	ASSERT(rsp->rs_num_retry_despite_err == 0);
990 	rsp->rs_num_retry_despite_err = 0;
991 	return (nfs4_start_fop(mi, vp1, vp2, OH_OTHER, rsp, NULL));
992 }
993 
994 /*
995  * Release any resources acquired by nfs4_start_op().
996  * 'sp' should be the nfs4_server pointer returned by nfs4_start_op().
997  *
998  * The operation hint is used to avoid a deadlock by bypassing delegation
999  * return logic for writes, which are done while returning a delegation.
1000  */
1001 
1002 void
1003 nfs4_end_fop(mntinfo4_t *mi, vnode_t *vp1, vnode_t *vp2, nfs4_op_hint_t op,
1004 		nfs4_recov_state_t *rsp, bool_t needs_recov)
1005 {
1006 	nfs4_server_t *sp = rsp->rs_sp;
1007 	rnode4_t *rp = NULL;
1008 
1009 #ifdef	lint
1010 	/*
1011 	 * The op hint isn't used any more, but might be in
1012 	 * the future.
1013 	 */
1014 	op = op;
1015 #endif
1016 
1017 #ifdef	DEBUG
1018 	ASSERT(tsd_get(nfs4_tsd_key) != NULL);
1019 	(void) tsd_set(nfs4_tsd_key, NULL);
1020 #endif
1021 
1022 	nfs4_end_op_recall(vp1, vp2, rsp);
1023 
1024 	if (rsp->rs_flags & NFS4_RS_RENAME_HELD)
1025 		nfs_rw_exit(&mi->mi_rename_lock);
1026 
1027 	if (!needs_recov) {
1028 		if (rsp->rs_flags & NFS4_RS_DELAY_MSG) {
1029 			/* may need to clear the delay interval */
1030 			if (vp1 != NULL) {
1031 				rp = VTOR4(vp1);
1032 				mutex_enter(&rp->r_statelock);
1033 				rp->r_delay_interval = 0;
1034 				mutex_exit(&rp->r_statelock);
1035 			}
1036 		}
1037 		rsp->rs_flags &= ~(NFS4_RS_GRACE_MSG|NFS4_RS_DELAY_MSG);
1038 	}
1039 
1040 	/*
1041 	 * If the corresponding nfs4_start_op() found a sp,
1042 	 * then there must still be a sp.
1043 	 */
1044 	if (sp != NULL) {
1045 		nfs_rw_exit(&mi->mi_recovlock);
1046 		nfs_rw_exit(&sp->s_recovlock);
1047 		mutex_enter(&sp->s_lock);
1048 		sp->s_otw_call_count--;
1049 		cv_broadcast(&sp->s_cv_otw_count);
1050 		mutex_exit(&sp->s_lock);
1051 		nfs4_server_rele(sp);
1052 	} else {
1053 		nfs_rw_exit(&mi->mi_recovlock);
1054 	}
1055 }
1056 
1057 void
1058 nfs4_end_op(mntinfo4_t *mi, vnode_t *vp1, vnode_t *vp2,
1059 	    nfs4_recov_state_t *rsp, bool_t needrecov)
1060 {
1061 	nfs4_end_fop(mi, vp1, vp2, OH_OTHER, rsp, needrecov);
1062 }
1063 
1064 /*
1065  * If the filesystem is going through client recovery, block until
1066  * finished.
1067  * Exceptions:
1068  * - state-releasing ops (CLOSE, LOCKU, DELEGRETURN) are allowed to proceed
1069  *   if the filesystem has been forcibly unmounted or the lwp is exiting.
1070  *
1071  * Return value:
1072  * - 0 if no errors
1073  * - EINTR if the call was interrupted
1074  * - EIO if the filesystem has been forcibly unmounted (non-state-releasing
1075  *   op)
1076  * - the errno value from the recovery thread, if recovery failed
1077  */
1078 
1079 static int
1080 wait_for_recovery(mntinfo4_t *mi, nfs4_op_hint_t op_hint)
1081 {
1082 	int error = 0;
1083 
1084 	mutex_enter(&mi->mi_lock);
1085 
1086 	while (mi->mi_recovflags != 0) {
1087 		klwp_t *lwp = ttolwp(curthread);
1088 
1089 		if (mi->mi_flags & MI4_RECOV_FAIL)
1090 			break;
1091 		if (mi->mi_vfsp->vfs_flag & VFS_UNMOUNTED)
1092 			break;
1093 		if (OH_IS_STATE_RELE(op_hint) &&
1094 		    (curthread->t_proc_flag & TP_LWPEXIT))
1095 			break;
1096 
1097 		if (lwp != NULL)
1098 			lwp->lwp_nostop++;
1099 		/* XXX - use different cv? */
1100 		if (cv_wait_sig(&mi->mi_failover_cv, &mi->mi_lock) == 0) {
1101 			error = EINTR;
1102 			if (lwp != NULL)
1103 				lwp->lwp_nostop--;
1104 			break;
1105 		}
1106 		if (lwp != NULL)
1107 			lwp->lwp_nostop--;
1108 	}
1109 
1110 	if (mi->mi_flags & MI4_RECOV_FAIL) {
1111 		NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
1112 			"wait_for_recovery: fail since RECOV FAIL"));
1113 		error = mi->mi_error;
1114 	} else if ((mi->mi_vfsp->vfs_flag & VFS_UNMOUNTED) &&
1115 	    !OH_IS_STATE_RELE(op_hint)) {
1116 		NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
1117 			"wait_for_recovery: forced unmount"));
1118 		error = EIO;
1119 	}
1120 
1121 	mutex_exit(&mi->mi_lock);
1122 
1123 	return (error);
1124 }
1125 
1126 /*
1127  * If the client received NFS4ERR_GRACE for this particular mount,
1128  * the client blocks here until it is time to try again.
1129  *
1130  * Return value:
1131  * - 0 if wait was successful
1132  * - EINTR if the call was interrupted
1133  */
1134 
1135 int
1136 nfs4_wait_for_grace(mntinfo4_t *mi, nfs4_recov_state_t *rsp)
1137 {
1138 	int error = 0;
1139 	time_t curtime, time_to_wait;
1140 
1141 	/* do a unprotected check to reduce mi_lock contention */
1142 	if (mi->mi_grace_wait != 0) {
1143 		mutex_enter(&mi->mi_lock);
1144 
1145 		if (mi->mi_grace_wait != 0) {
1146 			if (!(rsp->rs_flags & NFS4_RS_GRACE_MSG))
1147 				rsp->rs_flags |= NFS4_RS_GRACE_MSG;
1148 
1149 			curtime = gethrestime_sec();
1150 
1151 			if (curtime < mi->mi_grace_wait) {
1152 
1153 				time_to_wait = mi->mi_grace_wait - curtime;
1154 
1155 				mutex_exit(&mi->mi_lock);
1156 
1157 				delay(SEC_TO_TICK(time_to_wait));
1158 
1159 				curtime = gethrestime_sec();
1160 
1161 				mutex_enter(&mi->mi_lock);
1162 
1163 				if (curtime >= mi->mi_grace_wait)
1164 					mi->mi_grace_wait = 0;
1165 			} else {
1166 				mi->mi_grace_wait = 0;
1167 			}
1168 		}
1169 		mutex_exit(&mi->mi_lock);
1170 	}
1171 
1172 	return (error);
1173 }
1174 
1175 /*
1176  * If the client received NFS4ERR_DELAY for an operation on a vnode,
1177  * the client blocks here until it is time to try again.
1178  *
1179  * Return value:
1180  * - 0 if wait was successful
1181  * - EINTR if the call was interrupted
1182  */
1183 
1184 int
1185 nfs4_wait_for_delay(vnode_t *vp, nfs4_recov_state_t *rsp)
1186 {
1187 	int error = 0;
1188 	time_t curtime, time_to_wait;
1189 	rnode4_t *rp;
1190 
1191 	ASSERT(vp != NULL);
1192 
1193 	rp = VTOR4(vp);
1194 
1195 	/* do a unprotected check to reduce r_statelock contention */
1196 	if (rp->r_delay_wait != 0) {
1197 		mutex_enter(&rp->r_statelock);
1198 
1199 		if (rp->r_delay_wait != 0) {
1200 
1201 			if (!(rsp->rs_flags & NFS4_RS_DELAY_MSG)) {
1202 				rsp->rs_flags |= NFS4_RS_DELAY_MSG;
1203 				nfs4_mi_kstat_inc_delay(VTOMI4(vp));
1204 			}
1205 
1206 			curtime = gethrestime_sec();
1207 
1208 			if (curtime < rp->r_delay_wait) {
1209 
1210 				time_to_wait = rp->r_delay_wait - curtime;
1211 
1212 				mutex_exit(&rp->r_statelock);
1213 
1214 				delay(SEC_TO_TICK(time_to_wait));
1215 
1216 				curtime = gethrestime_sec();
1217 
1218 				mutex_enter(&rp->r_statelock);
1219 
1220 				if (curtime >= rp->r_delay_wait)
1221 					rp->r_delay_wait = 0;
1222 			} else {
1223 				rp->r_delay_wait = 0;
1224 			}
1225 		}
1226 		mutex_exit(&rp->r_statelock);
1227 	}
1228 
1229 	return (error);
1230 }
1231 
1232 /*
1233  * The recovery thread.
1234  */
1235 
1236 static void
1237 nfs4_recov_thread(recov_info_t *recovp)
1238 {
1239 	mntinfo4_t *mi = recovp->rc_mi;
1240 	nfs4_server_t *sp;
1241 	int done = 0, error = 0;
1242 	bool_t recov_fail = FALSE;
1243 	callb_cpr_t cpr_info;
1244 	kmutex_t cpr_lock;
1245 
1246 	nfs4_queue_event(RE_START, mi, NULL, mi->mi_recovflags,
1247 	    recovp->rc_vp1, recovp->rc_vp2, 0, NULL, 0, TAG_NONE, TAG_NONE,
1248 	    0, 0);
1249 
1250 	mutex_init(&cpr_lock, NULL, MUTEX_DEFAULT, NULL);
1251 	CALLB_CPR_INIT(&cpr_info, &cpr_lock, callb_generic_cpr, "nfsv4Recov");
1252 
1253 	mutex_enter(&mi->mi_lock);
1254 	mi->mi_recovthread = curthread;
1255 	mutex_exit(&mi->mi_lock);
1256 
1257 	/*
1258 	 * We don't really need protection here against failover or
1259 	 * migration, since the current thread is the one that would make
1260 	 * any changes, but hold mi_recovlock anyway for completeness (and
1261 	 * to satisfy any ASSERTs).
1262 	 */
1263 	(void) nfs_rw_enter_sig(&mi->mi_recovlock, RW_READER, 0);
1264 	sp = find_nfs4_server(mi);
1265 	if (sp != NULL)
1266 		mutex_exit(&sp->s_lock);
1267 	nfs_rw_exit(&mi->mi_recovlock);
1268 
1269 	/*
1270 	 * Do any necessary recovery, based on the information in recovp
1271 	 * and any recovery flags.
1272 	 */
1273 
1274 	do {
1275 		mutex_enter(&mi->mi_lock);
1276 		if (FS_OR_ZONE_GONE4(mi->mi_vfsp)) {
1277 			bool_t activesrv;
1278 
1279 			NFS4_DEBUG(nfs4_client_recov_debug &&
1280 			    mi->mi_vfsp->vfs_flag & VFS_UNMOUNTED, (CE_NOTE,
1281 				"nfs4_recov_thread: file system has been "
1282 				"unmounted"));
1283 			NFS4_DEBUG(nfs4_client_recov_debug &&
1284 			    zone_status_get(curproc->p_zone) >=
1285 			    ZONE_IS_SHUTTING_DOWN, (CE_NOTE,
1286 				"nfs4_recov_thread: zone shutting down"));
1287 			/*
1288 			 * If the server has lost its state for us and
1289 			 * the filesystem is unmounted, then the filesystem
1290 			 * can be tossed, even if there are lost lock or
1291 			 * lost state calls in the recovery queue.
1292 			 */
1293 			if (mi->mi_recovflags &
1294 			    (MI4R_NEED_CLIENTID | MI4R_REOPEN_FILES)) {
1295 				NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
1296 				"nfs4_recov_thread: bailing out"));
1297 				mi->mi_flags |= MI4_RECOV_FAIL;
1298 				mi->mi_error = recovp->rc_error;
1299 				recov_fail = TRUE;
1300 			}
1301 			/*
1302 			 * We don't know if the server has any state for
1303 			 * us, and the filesystem has been unmounted.  If
1304 			 * there are "lost state" recovery items, keep
1305 			 * trying to process them until there are no more
1306 			 * mounted filesystems for the server.  Otherwise,
1307 			 * bail out.  The reason we don't mark the
1308 			 * filesystem as failing recovery is in case we
1309 			 * have to do "lost state" recovery later (e.g., a
1310 			 * user process exits).
1311 			 */
1312 			if (!(mi->mi_recovflags & MI4R_LOST_STATE)) {
1313 				recov_done(mi, recovp);
1314 				mutex_exit(&mi->mi_lock);
1315 				break;
1316 			}
1317 			mutex_exit(&mi->mi_lock);
1318 
1319 			if (sp == NULL)
1320 				activesrv = FALSE;
1321 			else {
1322 				mutex_enter(&sp->s_lock);
1323 				activesrv = nfs4_fs_active(sp);
1324 			}
1325 			if (!activesrv) {
1326 				NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
1327 					"no active fs for server %p",
1328 					(void *)sp));
1329 				mutex_enter(&mi->mi_lock);
1330 				mi->mi_flags |= MI4_RECOV_FAIL;
1331 				mi->mi_error = recovp->rc_error;
1332 				mutex_exit(&mi->mi_lock);
1333 				recov_fail = TRUE;
1334 				if (sp != NULL) {
1335 					/*
1336 					 * Mark the server instance as
1337 					 * dead, so that nobody will attach
1338 					 * a new filesystem.
1339 					 */
1340 					nfs4_mark_srv_dead(sp);
1341 				}
1342 			}
1343 			if (sp != NULL)
1344 				mutex_exit(&sp->s_lock);
1345 		} else {
1346 			mutex_exit(&mi->mi_lock);
1347 		}
1348 
1349 		/*
1350 		 * Check if we need to select a new server for a
1351 		 * failover.  Choosing a new server will force at
1352 		 * least a check of the clientid.
1353 		 */
1354 		mutex_enter(&mi->mi_lock);
1355 		if (!recov_fail &&
1356 		    (mi->mi_recovflags & MI4R_NEED_NEW_SERVER)) {
1357 			mutex_exit(&mi->mi_lock);
1358 			recov_newserver(recovp, &sp, &recov_fail);
1359 		} else
1360 			mutex_exit(&mi->mi_lock);
1361 
1362 		/*
1363 		 * Check if we need to recover the clientid.  This
1364 		 * must be done before file and lock recovery, and it
1365 		 * potentially affects the recovery threads for other
1366 		 * filesystems, so it gets special treatment.
1367 		 */
1368 		if (sp != NULL && recov_fail == FALSE) {
1369 			mutex_enter(&sp->s_lock);
1370 			if (!(sp->s_flags & N4S_CLIENTID_SET)) {
1371 				mutex_exit(&sp->s_lock);
1372 				recov_clientid(recovp, sp);
1373 			} else {
1374 				/*
1375 				 * Unset this flag in case another recovery
1376 				 * thread successfully recovered the clientid
1377 				 * for us already.
1378 				 */
1379 				mutex_enter(&mi->mi_lock);
1380 				mi->mi_recovflags &= ~MI4R_NEED_CLIENTID;
1381 				mutex_exit(&mi->mi_lock);
1382 				mutex_exit(&sp->s_lock);
1383 			}
1384 		}
1385 
1386 		/*
1387 		 * Check if we need to get the security information.
1388 		 */
1389 		mutex_enter(&mi->mi_lock);
1390 		if ((mi->mi_recovflags & MI4R_NEED_SECINFO) &&
1391 		    !(mi->mi_flags & MI4_RECOV_FAIL)) {
1392 			mutex_exit(&mi->mi_lock);
1393 			(void) nfs_rw_enter_sig(&mi->mi_recovlock,
1394 							RW_WRITER, 0);
1395 			error = nfs4_secinfo_recov(recovp->rc_mi,
1396 					recovp->rc_vp1, recovp->rc_vp2);
1397 			/*
1398 			 * If error, nothing more can be done, stop
1399 			 * the recovery.
1400 			 */
1401 			if (error) {
1402 				mutex_enter(&mi->mi_lock);
1403 				mi->mi_flags |= MI4_RECOV_FAIL;
1404 				mi->mi_error = recovp->rc_error;
1405 				mutex_exit(&mi->mi_lock);
1406 				nfs4_queue_event(RE_WRONGSEC, mi, NULL,
1407 				    error, recovp->rc_vp1, recovp->rc_vp2,
1408 				    0, NULL, 0, TAG_NONE, TAG_NONE, 0, 0);
1409 			}
1410 			nfs_rw_exit(&mi->mi_recovlock);
1411 		} else
1412 			mutex_exit(&mi->mi_lock);
1413 
1414 		/*
1415 		 * Check if there's a bad seqid to recover.
1416 		 */
1417 		mutex_enter(&mi->mi_lock);
1418 		if ((mi->mi_recovflags & MI4R_BAD_SEQID) &&
1419 		    !(mi->mi_flags & MI4_RECOV_FAIL)) {
1420 			mutex_exit(&mi->mi_lock);
1421 			(void) nfs_rw_enter_sig(&mi->mi_recovlock,
1422 					RW_WRITER, 0);
1423 			recov_bad_seqid(recovp);
1424 			nfs_rw_exit(&mi->mi_recovlock);
1425 		} else
1426 			mutex_exit(&mi->mi_lock);
1427 
1428 		/*
1429 		 * Next check for recovery that affects the entire
1430 		 * filesystem.
1431 		 */
1432 		if (sp != NULL) {
1433 			mutex_enter(&mi->mi_lock);
1434 			if ((mi->mi_recovflags & MI4R_REOPEN_FILES) &&
1435 			    !(mi->mi_flags & MI4_RECOV_FAIL)) {
1436 				mutex_exit(&mi->mi_lock);
1437 				recov_openfiles(recovp, sp);
1438 			} else
1439 				mutex_exit(&mi->mi_lock);
1440 		}
1441 
1442 		/*
1443 		 * Send any queued state recovery requests.
1444 		 */
1445 		mutex_enter(&mi->mi_lock);
1446 		if (sp != NULL &&
1447 		    (mi->mi_recovflags & MI4R_LOST_STATE) &&
1448 		    !(mi->mi_flags & MI4_RECOV_FAIL)) {
1449 			mutex_exit(&mi->mi_lock);
1450 			(void) nfs_rw_enter_sig(&mi->mi_recovlock,
1451 				    RW_WRITER, 0);
1452 			nfs4_resend_lost_rqsts(recovp, sp);
1453 			if (list_head(&mi->mi_lost_state) == NULL) {
1454 				/* done */
1455 				mutex_enter(&mi->mi_lock);
1456 				mi->mi_recovflags &= ~MI4R_LOST_STATE;
1457 				mutex_exit(&mi->mi_lock);
1458 			}
1459 			nfs_rw_exit(&mi->mi_recovlock);
1460 		} else {
1461 			mutex_exit(&mi->mi_lock);
1462 		}
1463 
1464 		/*
1465 		 * See if there is anything more to do.  If not, announce
1466 		 * that we are done and exit.
1467 		 *
1468 		 * Need mi_recovlock to keep 'sp' valid.  Must grab
1469 		 * mi_recovlock before mi_lock to preserve lock ordering.
1470 		 */
1471 		(void) nfs_rw_enter_sig(&mi->mi_recovlock, RW_READER, 0);
1472 		mutex_enter(&mi->mi_lock);
1473 		if ((mi->mi_recovflags & ~MI4R_SRV_REBOOT) == 0 ||
1474 		    (mi->mi_flags & MI4_RECOV_FAIL)) {
1475 			list_t local_lost_state;
1476 			nfs4_lost_rqst_t *lrp;
1477 
1478 			/*
1479 			 * We need to remove the lost requests before we
1480 			 * unmark the mi as no longer doing recovery to
1481 			 * avoid a race with a new thread putting new lost
1482 			 * requests on the same mi (and the going away
1483 			 * thread would remove the new lost requests).
1484 			 *
1485 			 * Move the lost requests to a local list since
1486 			 * nfs4_remove_lost_rqst() drops mi_lock, and
1487 			 * dropping the mi_lock would make our check to
1488 			 * see if recovery is done no longer valid.
1489 			 */
1490 			list_create(&local_lost_state,
1491 			    sizeof (nfs4_lost_rqst_t),
1492 			    offsetof(nfs4_lost_rqst_t, lr_node));
1493 			list_move_tail(&local_lost_state, &mi->mi_lost_state);
1494 
1495 			done = 1;
1496 			recov_done(mi, recovp);
1497 			mutex_exit(&mi->mi_lock);
1498 			/*
1499 			 * Now officially free the "moved"
1500 			 * lost requests.
1501 			 */
1502 			while ((lrp = list_head(&local_lost_state)) != NULL) {
1503 				list_remove(&local_lost_state, lrp);
1504 				nfs4_free_lost_rqst(lrp, sp);
1505 			}
1506 			list_destroy(&local_lost_state);
1507 		} else
1508 			mutex_exit(&mi->mi_lock);
1509 		nfs_rw_exit(&mi->mi_recovlock);
1510 
1511 		/*
1512 		 * If the filesystem has been forcibly unmounted, there is
1513 		 * probably no point in retrying immediately.  Furthermore,
1514 		 * there might be user processes waiting for a chance to
1515 		 * queue up "lost state" requests, so that they can exit.
1516 		 * So pause here for a moment.  Same logic for zone shutdown.
1517 		 */
1518 		if (!done && FS_OR_ZONE_GONE4(mi->mi_vfsp)) {
1519 			mutex_enter(&mi->mi_lock);
1520 			cv_broadcast(&mi->mi_failover_cv);
1521 			mutex_exit(&mi->mi_lock);
1522 			delay(SEC_TO_TICK(nfs4_unmount_delay));
1523 		}
1524 
1525 	} while (!done);
1526 
1527 	mutex_enter(&mi->mi_lock);
1528 	mi->mi_in_recovery--;
1529 	cv_broadcast(&mi->mi_cv_in_recov);
1530 	mutex_exit(&mi->mi_lock);
1531 
1532 	if (sp != NULL)
1533 		nfs4_server_rele(sp);
1534 
1535 	/*
1536 	 * Return all recalled delegations
1537 	 */
1538 	nfs4_dlistclean();
1539 
1540 	/*
1541 	 * Free up resources that were allocated for us.
1542 	 */
1543 	if (recovp->rc_vp1 != NULL)
1544 		VN_RELE(recovp->rc_vp1);
1545 	if (recovp->rc_vp2 != NULL)
1546 		VN_RELE(recovp->rc_vp2);
1547 	VFS_RELE(mi->mi_vfsp);
1548 	kmem_free(recovp, sizeof (recov_info_t));
1549 	mutex_enter(&cpr_lock);
1550 	CALLB_CPR_EXIT(&cpr_info);
1551 	mutex_destroy(&cpr_lock);
1552 	zthread_exit();
1553 }
1554 
1555 /*
1556  * Log the end of recovery and notify any waiting threads.
1557  */
1558 
1559 static void
1560 recov_done(mntinfo4_t *mi, recov_info_t *recovp)
1561 {
1562 
1563 	ASSERT(MUTEX_HELD(&mi->mi_lock));
1564 
1565 	nfs4_queue_event(RE_END, mi, NULL, 0, recovp->rc_vp1,
1566 		recovp->rc_vp2, 0, NULL, 0, TAG_NONE, TAG_NONE, 0, 0);
1567 	mi->mi_recovthread = NULL;
1568 	mi->mi_flags &= ~MI4_RECOV_ACTIV;
1569 	mi->mi_recovflags &= ~MI4R_SRV_REBOOT;
1570 	cv_broadcast(&mi->mi_failover_cv);
1571 }
1572 
1573 /*
1574  * State-specific recovery routines, by state.
1575  */
1576 
1577 /*
1578  * Failover.
1579  *
1580  * Replaces *spp with a reference to the new server, which must
1581  * eventually be freed.
1582  */
1583 
1584 static void
1585 recov_newserver(recov_info_t *recovp, nfs4_server_t **spp, bool_t *recov_fail)
1586 {
1587 	mntinfo4_t *mi = recovp->rc_mi;
1588 	servinfo4_t *svp = NULL;
1589 	nfs4_server_t *osp = *spp;
1590 	CLIENT *cl;
1591 	enum clnt_stat status;
1592 	struct timeval tv;
1593 	int error;
1594 	int oncethru = 0;
1595 	rnode4_t *rp;
1596 	int index;
1597 	nfs_fh4 fh;
1598 	char *snames;
1599 	size_t len;
1600 
1601 	(void) nfs_rw_enter_sig(&mi->mi_recovlock, RW_WRITER, 0);
1602 
1603 	tv.tv_sec = 2;
1604 	tv.tv_usec = 0;
1605 
1606 #ifdef lint
1607 	/*
1608 	 * Lint can't follow the logic, so thinks that snames and len
1609 	 * can be used before being set.  They can't, but lint can't
1610 	 * figure it out.  To address the lint warning, initialize
1611 	 * snames and len for lint.
1612 	 */
1613 	snames = NULL;
1614 	len = 0;
1615 #endif
1616 
1617 	/*
1618 	 * Ping the null NFS procedure of every server in
1619 	 * the list until one responds.  We always start
1620 	 * at the head of the list and always skip the one
1621 	 * that is current, since it's caused us a problem.
1622 	 */
1623 	while (svp == NULL) {
1624 		for (svp = mi->mi_servers; svp; svp = svp->sv_next) {
1625 
1626 			mutex_enter(&mi->mi_lock);
1627 			if (FS_OR_ZONE_GONE4(mi->mi_vfsp)) {
1628 				mi->mi_flags |= MI4_RECOV_FAIL;
1629 				mutex_exit(&mi->mi_lock);
1630 				(void) nfs_rw_exit(&mi->mi_recovlock);
1631 				*recov_fail = TRUE;
1632 				if (oncethru)
1633 					kmem_free(snames, len);
1634 				return;
1635 			}
1636 			mutex_exit(&mi->mi_lock);
1637 
1638 			(void) nfs_rw_enter_sig(&svp->sv_lock, RW_READER, 0);
1639 			if (svp->sv_flags & SV4_NOTINUSE) {
1640 				nfs_rw_exit(&svp->sv_lock);
1641 				continue;
1642 			}
1643 			nfs_rw_exit(&svp->sv_lock);
1644 
1645 			if (!oncethru && svp == mi->mi_curr_serv)
1646 				continue;
1647 
1648 			error = clnt_tli_kcreate(svp->sv_knconf, &svp->sv_addr,
1649 			    NFS_PROGRAM, NFS_V4, 0, 1, CRED(), &cl);
1650 			if (error)
1651 				continue;
1652 
1653 			if (!(mi->mi_flags & MI4_INT))
1654 				cl->cl_nosignal = TRUE;
1655 			status = CLNT_CALL(cl, RFS_NULL, xdr_void, NULL,
1656 			    xdr_void, NULL, tv);
1657 			if (!(mi->mi_flags & MI4_INT))
1658 				cl->cl_nosignal = FALSE;
1659 			AUTH_DESTROY(cl->cl_auth);
1660 			CLNT_DESTROY(cl);
1661 			if (status == RPC_SUCCESS) {
1662 				nfs4_queue_event(RE_FAILOVER, mi,
1663 				    svp == mi->mi_curr_serv ? NULL :
1664 				    svp->sv_hostname, 0, NULL, NULL, 0,
1665 				    NULL, 0, TAG_NONE, TAG_NONE, 0, 0);
1666 				break;
1667 			}
1668 		}
1669 
1670 		if (svp == NULL) {
1671 			if (!oncethru) {
1672 				snames = nfs4_getsrvnames(mi, &len);
1673 				nfs4_queue_fact(RF_SRVS_NOT_RESPOND, mi,
1674 				    0, 0, 0, FALSE, snames, 0, NULL);
1675 				oncethru = 1;
1676 			}
1677 			delay(hz);
1678 		}
1679 	}
1680 
1681 	if (oncethru) {
1682 		nfs4_queue_fact(RF_SRVS_OK, mi, 0, 0, 0, FALSE, snames,
1683 		    0, NULL);
1684 		kmem_free(snames, len);
1685 	}
1686 
1687 #if DEBUG
1688 	(void) nfs_rw_enter_sig(&svp->sv_lock, RW_READER, 0);
1689 	ASSERT((svp->sv_flags & SV4_NOTINUSE) == 0);
1690 	nfs_rw_exit(&svp->sv_lock);
1691 #endif
1692 
1693 	mutex_enter(&mi->mi_lock);
1694 	mi->mi_recovflags &= ~MI4R_NEED_NEW_SERVER;
1695 	if (svp != mi->mi_curr_serv) {
1696 		servinfo4_t *osvp = mi->mi_curr_serv;
1697 
1698 		mutex_exit(&mi->mi_lock);
1699 
1700 		/*
1701 		 * Update server-dependent fields in the root vnode.
1702 		 */
1703 		index = rtable4hash(mi->mi_rootfh);
1704 		rw_enter(&rtable4[index].r_lock, RW_WRITER);
1705 
1706 		rp = r4find(&rtable4[index], mi->mi_rootfh, mi->mi_vfsp);
1707 		if (rp != NULL) {
1708 			NFS4_DEBUG(nfs4_client_failover_debug, (CE_NOTE,
1709 			    "recov_newserver: remapping %s", rnode4info(rp)));
1710 			mutex_enter(&rp->r_statelock);
1711 			rp->r_server = svp;
1712 			PURGE_ATTRCACHE4_LOCKED(rp);
1713 			mutex_exit(&rp->r_statelock);
1714 			(void) nfs4_free_data_reclaim(rp);
1715 			nfs4_purge_rddir_cache(RTOV4(rp));
1716 			rw_exit(&rtable4[index].r_lock);
1717 			NFS4_DEBUG(nfs4_client_failover_debug, (CE_NOTE,
1718 			    "recov_newserver: done with %s",
1719 			    rnode4info(rp)));
1720 			VN_RELE(RTOV4(rp));
1721 		} else
1722 			rw_exit(&rtable4[index].r_lock);
1723 		(void) dnlc_purge_vfsp(mi->mi_vfsp, 0);
1724 
1725 		mutex_enter(&mi->mi_lock);
1726 		mi->mi_recovflags |= MI4R_REOPEN_FILES | MI4R_REMAP_FILES;
1727 		if (recovp->rc_srv_reboot)
1728 			mi->mi_recovflags |= MI4R_SRV_REBOOT;
1729 		mi->mi_curr_serv = svp;
1730 		mi->mi_failover++;
1731 		mi->mi_flags &= ~MI4_BADOWNER_DEBUG;
1732 		mutex_exit(&mi->mi_lock);
1733 
1734 		(void) nfs_rw_enter_sig(&svp->sv_lock, RW_READER, 0);
1735 		fh.nfs_fh4_len = svp->sv_fhandle.fh_len;
1736 		fh.nfs_fh4_val = svp->sv_fhandle.fh_buf;
1737 		sfh4_update(mi->mi_rootfh, &fh);
1738 		fh.nfs_fh4_len = svp->sv_pfhandle.fh_len;
1739 		fh.nfs_fh4_val = svp->sv_pfhandle.fh_buf;
1740 		sfh4_update(mi->mi_srvparentfh, &fh);
1741 		nfs_rw_exit(&svp->sv_lock);
1742 
1743 		*spp = nfs4_move_mi(mi, osvp, svp);
1744 		if (osp != NULL)
1745 			nfs4_server_rele(osp);
1746 	} else
1747 		mutex_exit(&mi->mi_lock);
1748 	(void) nfs_rw_exit(&mi->mi_recovlock);
1749 }
1750 
1751 /*
1752  * Clientid.
1753  */
1754 
1755 static void
1756 recov_clientid(recov_info_t *recovp, nfs4_server_t *sp)
1757 {
1758 	mntinfo4_t *mi = recovp->rc_mi;
1759 	int error = 0;
1760 	int still_stale;
1761 	int need_new_s;
1762 
1763 	ASSERT(sp != NULL);
1764 
1765 	/*
1766 	 * Acquire the recovery lock and then verify that the clientid
1767 	 * still needs to be recovered.  (Note that s_recovlock is supposed
1768 	 * to be acquired before s_lock.)  Since the thread holds the
1769 	 * recovery lock, no other thread will recover the clientid.
1770 	 */
1771 	(void) nfs_rw_enter_sig(&sp->s_recovlock, RW_WRITER, 0);
1772 	(void) nfs_rw_enter_sig(&mi->mi_recovlock, RW_WRITER, 0);
1773 	mutex_enter(&sp->s_lock);
1774 	still_stale = ((sp->s_flags & N4S_CLIENTID_SET) == 0);
1775 	mutex_exit(&sp->s_lock);
1776 
1777 	if (still_stale) {
1778 		nfs4_error_t n4e;
1779 
1780 		nfs4_error_zinit(&n4e);
1781 		nfs4setclientid(mi, kcred, TRUE, &n4e);
1782 		error = n4e.error;
1783 		if (error != 0) {
1784 
1785 			/*
1786 			 * nfs4setclientid may have set MI4R_NEED_NEW_SERVER,
1787 			 * if so, just return and let recov_thread drive
1788 			 * failover.
1789 			 */
1790 			mutex_enter(&mi->mi_lock);
1791 			need_new_s = mi->mi_recovflags & MI4R_NEED_NEW_SERVER;
1792 			mutex_exit(&mi->mi_lock);
1793 
1794 			if (need_new_s) {
1795 				nfs_rw_exit(&mi->mi_recovlock);
1796 				nfs_rw_exit(&sp->s_recovlock);
1797 				return;
1798 			}
1799 
1800 			nfs4_queue_event(RE_CLIENTID, mi, NULL, n4e.error, NULL,
1801 			    NULL, n4e.stat, NULL, 0, TAG_NONE, TAG_NONE, 0, 0);
1802 			mutex_enter(&mi->mi_lock);
1803 			mi->mi_flags |= MI4_RECOV_FAIL;
1804 			mi->mi_error = recovp->rc_error;
1805 			mutex_exit(&mi->mi_lock);
1806 			/* don't destroy the nfs4_server, let umount do it */
1807 		}
1808 	}
1809 
1810 	if (error == 0) {
1811 		mutex_enter(&mi->mi_lock);
1812 		mi->mi_recovflags &= ~MI4R_NEED_CLIENTID;
1813 		/*
1814 		 * If still_stale isn't true, then another thread already
1815 		 * recovered the clientid.  And that thread that set the
1816 		 * clientid will have initiated reopening files on all the
1817 		 * filesystems for the server, so we should not initiate
1818 		 * reopening for this filesystem here.
1819 		 */
1820 		if (still_stale) {
1821 			mi->mi_recovflags |= MI4R_REOPEN_FILES;
1822 			if (recovp->rc_srv_reboot)
1823 				mi->mi_recovflags |= MI4R_SRV_REBOOT;
1824 		}
1825 		mutex_exit(&mi->mi_lock);
1826 	}
1827 
1828 	nfs_rw_exit(&mi->mi_recovlock);
1829 
1830 	if (error != 0) {
1831 		nfs_rw_exit(&sp->s_recovlock);
1832 		mutex_enter(&mi->mi_lock);
1833 		if ((mi->mi_flags & MI4_RECOV_FAIL) == 0)
1834 			delay(SEC_TO_TICK(recov_err_delay));
1835 		mutex_exit(&mi->mi_lock);
1836 	} else {
1837 		mntinfo4_t **milist;
1838 		mntinfo4_t *tmi;
1839 		int nummi, i;
1840 
1841 		/*
1842 		 * Initiate recovery of open files for other filesystems.
1843 		 * We create an array of filesystems, rather than just
1844 		 * walking the filesystem list, to avoid deadlock issues
1845 		 * with s_lock and mi_recovlock.
1846 		 */
1847 		milist = make_milist(sp, &nummi);
1848 		for (i = 0; i < nummi; i++) {
1849 			tmi = milist[i];
1850 			if (tmi != mi) {
1851 				(void) nfs_rw_enter_sig(&tmi->mi_recovlock,
1852 							RW_READER, 0);
1853 				start_recovery_action(NR_OPENFILES, TRUE, tmi,
1854 					NULL, NULL);
1855 				nfs_rw_exit(&tmi->mi_recovlock);
1856 			}
1857 		}
1858 		free_milist(milist, nummi);
1859 
1860 		nfs_rw_exit(&sp->s_recovlock);
1861 	}
1862 }
1863 
1864 /*
1865  * Return an array of filesystems associated with the given server.  The
1866  * caller should call free_milist() to free the references and memory.
1867  */
1868 
1869 static mntinfo4_t **
1870 make_milist(nfs4_server_t *sp, int *nummip)
1871 {
1872 	int nummi, i;
1873 	mntinfo4_t **milist;
1874 	mntinfo4_t *tmi;
1875 
1876 	mutex_enter(&sp->s_lock);
1877 	nummi = 0;
1878 	for (tmi = sp->mntinfo4_list; tmi != NULL; tmi = tmi->mi_clientid_next)
1879 		nummi++;
1880 
1881 	milist = kmem_alloc(nummi * sizeof (mntinfo4_t *), KM_NOSLEEP);
1882 
1883 	for (i = 0, tmi = sp->mntinfo4_list; tmi != NULL; i++,
1884 	    tmi = tmi->mi_clientid_next) {
1885 		milist[i] = tmi;
1886 		VFS_HOLD(tmi->mi_vfsp);
1887 	}
1888 	mutex_exit(&sp->s_lock);
1889 
1890 	*nummip = nummi;
1891 	return (milist);
1892 }
1893 
1894 /*
1895  * Free the filesystem list created by make_milist().
1896  */
1897 
1898 static void
1899 free_milist(mntinfo4_t **milist, int nummi)
1900 {
1901 	mntinfo4_t *tmi;
1902 	int i;
1903 
1904 	for (i = 0; i < nummi; i++) {
1905 		tmi = milist[i];
1906 		VFS_RELE(tmi->mi_vfsp);
1907 	}
1908 	kmem_free(milist, nummi * sizeof (mntinfo4_t *));
1909 }
1910 
1911 /*
1912  * Filehandle
1913  */
1914 
1915 /*
1916  * Lookup the filehandle for the given vnode and update the rnode if it has
1917  * changed.
1918  *
1919  * Errors:
1920  * - if the filehandle could not be updated because of an error that
1921  *   requires further recovery, initiate that recovery and return.
1922  * - if the filehandle could not be updated because of a signal, pretend we
1923  *   succeeded and let someone else deal with it.
1924  * - if the filehandle could not be updated and the filesystem has been
1925  *   forcibly unmounted, pretend we succeeded, and let the caller deal with
1926  *   the forced unmount (to retry or not to retry, that is the question).
1927  * - if the filehandle could not be updated because of some other error,
1928  *   mark the rnode bad and return.
1929  */
1930 static void
1931 recov_filehandle(nfs4_recov_t action, mntinfo4_t *mi, vnode_t *vp)
1932 {
1933 	rnode4_t *rp = VTOR4(vp);
1934 	nfs4_error_t e = { 0, NFS4_OK, RPC_SUCCESS };
1935 	bool_t needrecov;
1936 
1937 	mutex_enter(&rp->r_statelock);
1938 
1939 	if (rp->r_flags & R4RECOVERR) {
1940 		mutex_exit(&rp->r_statelock);
1941 		return;
1942 	}
1943 
1944 	/*
1945 	 * If someone else is updating the filehandle, wait for them to
1946 	 * finish and then let our caller retry.
1947 	 */
1948 	if (rp->r_flags & R4RECEXPFH) {
1949 		while (rp->r_flags & R4RECEXPFH) {
1950 			cv_wait(&rp->r_cv, &rp->r_statelock);
1951 		}
1952 		mutex_exit(&rp->r_statelock);
1953 		return;
1954 	}
1955 	rp->r_flags |= R4RECEXPFH;
1956 	mutex_exit(&rp->r_statelock);
1957 
1958 	if (action == NR_BADHANDLE) {
1959 		/* shouldn't happen */
1960 		nfs4_queue_event(RE_BADHANDLE, mi, NULL, 0,
1961 		    vp, NULL, 0, NULL, 0, TAG_NONE, TAG_NONE, 0, 0);
1962 	}
1963 
1964 	nfs4_remap_file(mi, vp, 0, &e);
1965 	needrecov = nfs4_needs_recovery(&e, FALSE, mi->mi_vfsp);
1966 
1967 	/*
1968 	 * If we get BADHANDLE or FHEXPIRED in their handler, something is
1969 	 * broken.  Don't try to recover, just mark the file dead.
1970 	 */
1971 	if (needrecov && e.error == 0 &&
1972 	    (e.stat == NFS4ERR_BADHANDLE || e.stat == NFS4ERR_FHEXPIRED))
1973 		needrecov = FALSE;
1974 	if (needrecov) {
1975 		(void) nfs4_start_recovery(&e, mi, vp,
1976 				NULL, NULL, NULL, OP_LOOKUP, NULL);
1977 	} else if (e.error != EINTR &&
1978 	    !NFS4_FRC_UNMT_ERR(e.error, mi->mi_vfsp) &&
1979 	    (e.error != 0 || e.stat != NFS4_OK)) {
1980 		nfs4_recov_fh_fail(vp, e.error, e.stat);
1981 		/*
1982 		 * Don't set r_error to ESTALE.  Higher-level code (e.g.,
1983 		 * cstatat_getvp()) retries on ESTALE, which would cause
1984 		 * an infinite loop.
1985 		 */
1986 	}
1987 
1988 	mutex_enter(&rp->r_statelock);
1989 	rp->r_flags &= ~R4RECEXPFH;
1990 	cv_broadcast(&rp->r_cv);
1991 	mutex_exit(&rp->r_statelock);
1992 }
1993 
1994 /*
1995  * Stale Filehandle
1996  */
1997 
1998 /*
1999  * A stale filehandle can happen when an individual file has
2000  * been removed, or when an entire filesystem has been taken
2001  * offline.  To distinguish these cases, we do this:
2002  * - if a GETATTR with the current filehandle is okay, we do
2003  *   nothing (this can happen with two-filehandle ops)
2004  * - if the GETATTR fails, but a GETATTR of the root filehandle
2005  *   succeeds, mark the rnode with R4STALE, which will stop use
2006  * - if the GETATTR fails, and a GETATTR of the root filehandle
2007  *   also fails, we consider the problem filesystem-wide, so:
2008  *   - if we can failover, we should
2009  *   - if we can't failover, we should mark both the original
2010  *     vnode and the root bad
2011  */
2012 static void
2013 recov_stale(mntinfo4_t *mi, vnode_t *vp)
2014 {
2015 	rnode4_t *rp = VTOR4(vp);
2016 	vnode_t *rootvp = NULL;
2017 	nfs4_error_t e = { 0, NFS4_OK, RPC_SUCCESS };
2018 	nfs4_ga_res_t gar;
2019 	char *fail_msg = "failed to recover from NFS4ERR_STALE";
2020 	bool_t needrecov;
2021 
2022 	mutex_enter(&rp->r_statelock);
2023 
2024 	if (rp->r_flags & R4RECOVERR) {
2025 		mutex_exit(&rp->r_statelock);
2026 		NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
2027 		    "recov_stale: already marked dead, rp %s",
2028 		    rnode4info(rp)));
2029 		return;
2030 	}
2031 
2032 	if (rp->r_flags & R4STALE) {
2033 		mutex_exit(&rp->r_statelock);
2034 		NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
2035 		    "recov_stale: already marked stale, rp %s",
2036 		    rnode4info(rp)));
2037 		return;
2038 	}
2039 
2040 	mutex_exit(&rp->r_statelock);
2041 
2042 	/* Try a GETATTR on this vnode */
2043 	nfs4_getattr_otw_norecovery(vp, &gar, &e, CRED(), 0);
2044 
2045 	/*
2046 	 * Handle non-STALE recoverable errors
2047 	 */
2048 	needrecov = nfs4_needs_recovery(&e, FALSE, vp->v_vfsp);
2049 	if (needrecov && (e.error != 0 || e.stat != NFS4ERR_STALE)) {
2050 		(void) nfs4_start_recovery(&e, mi, vp,
2051 				NULL, NULL, NULL, OP_GETATTR, NULL);
2052 		NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
2053 		    "recov_stale: error=%d, stat=%d seen on rp %s",
2054 		    e.error, e.stat, rnode4info(rp)));
2055 		goto out;
2056 	}
2057 
2058 	/* Are things OK for this vnode? */
2059 	if (!e.error && e.stat == NFS4_OK) {
2060 		NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
2061 		    "recov_stale: file appears fine, rp %s",
2062 		    rnode4info(rp)));
2063 		goto out;
2064 	}
2065 
2066 	/* Did we get an unrelated non-recoverable error? */
2067 	if (e.error || e.stat != NFS4ERR_STALE) {
2068 		nfs4_fail_recov(vp, fail_msg, e.error, e.stat);
2069 		NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
2070 		    "recov_stale: unrelated fatal error, rp %s",
2071 		    rnode4info(rp)));
2072 		goto out;
2073 	}
2074 
2075 	/*
2076 	 * If we don't appear to be dealing with the root node, find it.
2077 	 */
2078 	if ((vp->v_flag & VROOT) == 0) {
2079 		nfs4_error_zinit(&e);
2080 		e.error = VFS_ROOT(vp->v_vfsp, &rootvp);
2081 		if (e.error) {
2082 			nfs4_fail_recov(vp, fail_msg, 0, NFS4ERR_STALE);
2083 			NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
2084 			    "recov_stale: can't find root node for rp %s",
2085 			    rnode4info(rp)));
2086 			goto out;
2087 		}
2088 	}
2089 
2090 	/* Try a GETATTR on the root vnode */
2091 	if (rootvp != NULL) {
2092 		nfs4_error_zinit(&e);
2093 		nfs4_getattr_otw_norecovery(rootvp, &gar, &e, CRED(), 0);
2094 
2095 		/* Try recovery? */
2096 		if (e.error != 0 || e.stat != NFS4ERR_STALE) {
2097 			needrecov = nfs4_needs_recovery(&e, FALSE, vp->v_vfsp);
2098 			if (needrecov) {
2099 				(void) nfs4_start_recovery(&e,
2100 					mi, rootvp, NULL, NULL, NULL,
2101 					OP_GETATTR, NULL);
2102 				NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
2103 				    "recov_stale: error=%d, stat=%d seen "
2104 				    "on rp %s", e.error, e.stat,
2105 				    rnode4info(rp)));
2106 			}
2107 		}
2108 
2109 		/*
2110 		 * Check to see if a failover attempt is warranted
2111 		 * NB: nfs4_try_failover doesn't check for STALE
2112 		 * because recov_stale gets a shot first.  Now that
2113 		 * recov_stale has failed, go ahead and try failover.
2114 		 *
2115 		 * If the getattr on the root filehandle was successful,
2116 		 * then mark recovery as failed for 'vp' and exit.
2117 		 */
2118 		if (nfs4_try_failover(&e) == 0 && e.stat != NFS4ERR_STALE) {
2119 			/*
2120 			 * pass the original error to fail_recov, not
2121 			 * the one from trying the root vnode.
2122 			 */
2123 			nfs4_fail_recov(vp, fail_msg, 0, NFS4ERR_STALE);
2124 			NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
2125 			    "recov_stale: root node OK, marking "
2126 			    "dead rp %s", rnode4info(rp)));
2127 			goto out;
2128 		}
2129 	}
2130 
2131 	/*
2132 	 * Here, we know that both the original file and the
2133 	 * root filehandle (which may be the same) are stale.
2134 	 * We want to fail over if we can, and if we can't, we
2135 	 * want to mark everything in sight bad.
2136 	 */
2137 	if (FAILOVER_MOUNT4(mi)) {
2138 		mutex_enter(&mi->mi_lock);
2139 		mi->mi_recovflags |= MI4R_NEED_NEW_SERVER;
2140 		NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
2141 		    "recov_stale: failing over due to rp %s",
2142 		    rnode4info(rp)));
2143 		mutex_exit(&mi->mi_lock);
2144 	} else {
2145 		rnode4_t *rootrp;
2146 		servinfo4_t *svp;
2147 
2148 		/*
2149 		 * Can't fail over, so mark things dead.
2150 		 *
2151 		 * If rootvp is set, we know we have a distinct
2152 		 * non-root vnode which can be marked dead in
2153 		 * the usual way.
2154 		 *
2155 		 * Then we want to mark the root vnode dead.
2156 		 * Note that if rootvp wasn't set, our vp is
2157 		 * actually the root vnode.
2158 		 */
2159 		if (rootvp != NULL) {
2160 			NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
2161 			    "recov_stale: can't fail over, marking dead rp %s",
2162 			    rnode4info(rp)));
2163 			nfs4_fail_recov(vp, fail_msg, 0, NFS4ERR_STALE);
2164 		} else {
2165 			rootvp = vp;
2166 			VN_HOLD(rootvp);
2167 		}
2168 
2169 		/*
2170 		 * Mark root dead, but quietly - since
2171 		 * the root rnode is frequently recreated,
2172 		 * we can encounter this at every access.
2173 		 * Also mark recovery as failed on this VFS.
2174 		 */
2175 		rootrp = VTOR4(rootvp);
2176 		NFS4_DEBUG(nfs4_client_recov_debug, (CE_CONT,
2177 		    "recov_stale: marking dead root rp %s",
2178 		    rnode4info(rootrp)));
2179 		mutex_enter(&rootrp->r_statelock);
2180 		rootrp->r_flags |= (R4RECOVERR | R4STALE);
2181 		rootrp->r_error = ESTALE;
2182 		mutex_exit(&rootrp->r_statelock);
2183 		mutex_enter(&mi->mi_lock);
2184 		mi->mi_error = ESTALE;
2185 		mutex_exit(&mi->mi_lock);
2186 
2187 		svp = mi->mi_curr_serv;
2188 		(void) nfs_rw_enter_sig(&svp->sv_lock, RW_WRITER, 0);
2189 		svp->sv_flags |= SV4_ROOT_STALE;
2190 		nfs_rw_exit(&svp->sv_lock);
2191 	}
2192 
2193 out:
2194 	if (rootvp)
2195 		VN_RELE(rootvp);
2196 }
2197 
2198 /*
2199  * Locks.
2200  */
2201 
2202 /*
2203  * Reclaim all the active (acquired) locks for the given file.
2204  * If a process lost a lock, the process is sent a SIGLOST.  This is not
2205  * considered an error.
2206  *
2207  * Return values:
2208  * Errors and status are returned via the nfs4_error_t parameter
2209  * If an error indicates that recovery is needed, the caller is responsible
2210  * for dealing with it.
2211  */
2212 
2213 static void
2214 relock_file(vnode_t *vp, mntinfo4_t *mi, nfs4_error_t *ep,
2215     fattr4_change pre_change)
2216 {
2217 	locklist_t *locks, *llp;
2218 	rnode4_t *rp;
2219 
2220 	ASSERT(ep != NULL);
2221 	nfs4_error_zinit(ep);
2222 
2223 	if (VTOMI4(vp)->mi_flags & MI4_LLOCK)
2224 		return;
2225 
2226 	nfs4_flush_lock_owners(VTOR4(vp));
2227 
2228 	/*
2229 	 * If we get an error that requires recovery actions, just bail out
2230 	 * and let the top-level recovery code handle it.
2231 	 *
2232 	 * If we get some other error, kill the process that owned the lock
2233 	 * and mark its remaining locks (if any) as belonging to NOPID, so
2234 	 * that we don't make any more reclaim requests for that process.
2235 	 */
2236 
2237 	rp = VTOR4(vp);
2238 	locks = flk_active_locks_for_vp(vp);
2239 	for (llp = locks; llp != NULL; llp = llp->ll_next) {
2240 		int did_reclaim = 1;
2241 
2242 		ASSERT(llp->ll_vp == vp);
2243 		if (llp->ll_flock.l_pid == NOPID)
2244 			continue;
2245 		reclaim_one_lock(vp, &llp->ll_flock, ep, &did_reclaim);
2246 		/*
2247 		 * If we need to restart recovery, stop processing the
2248 		 * list.  Some errors would be recoverable under other
2249 		 * circumstances, but if they happen here we just give up
2250 		 * on the lock.
2251 		 */
2252 		if (nfs4_needs_recovery(ep, TRUE, vp->v_vfsp)) {
2253 			if (ep->error != 0)
2254 				break;
2255 			if (!nfs4_recov_marks_dead(ep->stat))
2256 				break;
2257 		}
2258 		/*
2259 		 *   In case the server isn't offering us a grace period, or
2260 		 * if we missed it, we might have opened & locked from scratch,
2261 		 * rather than reopened/reclaimed.
2262 		 *   We need to ensure that the object hadn't been otherwise
2263 		 * changed during this time, by comparing the changeinfo.
2264 		 *   We get passed the changeinfo from before the reopen by our
2265 		 * caller, in pre_change.
2266 		 *   The changeinfo from after the reopen is in rp->r_change,
2267 		 * courtesy of the GETATTR in the reopen.
2268 		 *   If they're different, then the file has changed, and we
2269 		 * have to SIGLOST the app.
2270 		 */
2271 		if (ep->error == 0 && ep->stat == NFS4_OK && !did_reclaim) {
2272 			mutex_enter(&rp->r_statelock);
2273 			if (pre_change != rp->r_change)
2274 				ep->stat = NFS4ERR_NO_GRACE;
2275 			mutex_exit(&rp->r_statelock);
2276 		}
2277 		if (ep->error != 0 || ep->stat != NFS4_OK) {
2278 			if (ep->error != 0)
2279 				nfs4_queue_event(RE_FAIL_RELOCK, mi,
2280 				    NULL, ep->error, vp, NULL, 0, NULL,
2281 				    llp->ll_flock.l_pid, TAG_NONE, TAG_NONE,
2282 				    0, 0);
2283 			else
2284 				nfs4_queue_event(RE_FAIL_RELOCK, mi,
2285 				    NULL, 0, vp, NULL, ep->stat, NULL,
2286 				    llp->ll_flock.l_pid, TAG_NONE, TAG_NONE,
2287 				    0, 0);
2288 			nfs4_send_siglost(llp->ll_flock.l_pid, mi, vp, TRUE,
2289 			    ep->error, ep->stat);
2290 			relock_skip_pid(llp, llp->ll_flock.l_pid);
2291 
2292 			/* Reinitialize the nfs4_error and continue */
2293 			nfs4_error_zinit(ep);
2294 		}
2295 	}
2296 
2297 	if (locks != NULL)
2298 		flk_free_locklist(locks);
2299 }
2300 
2301 /*
2302  * Reclaim the given lock.
2303  * If the lock can't be reclaimed, the process is sent SIGLOST, but this is
2304  * not considered an error.
2305  *
2306  * Errors are returned via the nfs4_error_t parameter.
2307  */
2308 static void
2309 reclaim_one_lock(vnode_t *vp, flock64_t *flk, nfs4_error_t *ep,
2310 	int *did_reclaimp)
2311 {
2312 	cred_t *cr;
2313 	rnode4_t *rp = VTOR4(vp);
2314 
2315 	cr = pid_to_cr(flk->l_pid);
2316 	if (cr == NULL) {
2317 		nfs4_error_zinit(ep);
2318 		ep->error = ESRCH;
2319 		return;
2320 	}
2321 
2322 	do {
2323 		mutex_enter(&rp->r_statelock);
2324 		if (rp->r_flags & R4RECOVERR) {
2325 			/*
2326 			 * This shouldn't affect other reclaims, so don't
2327 			 * return an error.
2328 			 */
2329 			mutex_exit(&rp->r_statelock);
2330 			break;
2331 		}
2332 		mutex_exit(&rp->r_statelock);
2333 
2334 		nfs4frlock(NFS4_LCK_CTYPE_RECLAIM, vp, F_SETLK, flk,
2335 				FREAD|FWRITE, 0, cr, ep, NULL, did_reclaimp);
2336 		if (ep->error == 0 && ep->stat == NFS4ERR_FHEXPIRED)
2337 			start_recovery_action(NR_FHEXPIRED, TRUE, VTOMI4(vp),
2338 					    vp, NULL);
2339 	} while (ep->error == 0 && ep->stat == NFS4ERR_FHEXPIRED);
2340 
2341 	crfree(cr);
2342 }
2343 
2344 /*
2345  * Open files.
2346  */
2347 
2348 /*
2349  * Verifies if the nfsstat4 is a valid error for marking this vnode dead.
2350  * Returns 1 if the error is valid; 0 otherwise.
2351  */
2352 static int
2353 nfs4_valid_recov_err_for_vp(vnode_t *vp, nfsstat4 stat)
2354 {
2355 	/*
2356 	 * We should not be marking non-regular files as dead,
2357 	 * except in very rare cases (eg: BADHANDLE or NFS4ERR_BADNAME).
2358 	 */
2359 	if (vp->v_type != VREG && stat != NFS4ERR_BADHANDLE &&
2360 	    stat != NFS4ERR_BADNAME)
2361 		return (0);
2362 
2363 	return (1);
2364 }
2365 
2366 /*
2367  * Failed attempting to recover a filehandle.  If 'stat' is valid for 'vp',
2368  * then mark the object dead.  Since we've had to do a lookup for
2369  * filehandle recovery, we will mark the object dead if we got NOENT.
2370  */
2371 static void
2372 nfs4_recov_fh_fail(vnode_t *vp, int error, nfsstat4 stat)
2373 {
2374 	ASSERT(vp != NULL);
2375 
2376 	if ((error == 0) && (stat != NFS4ERR_NOENT) &&
2377 	    (!nfs4_valid_recov_err_for_vp(vp, stat)))
2378 		return;
2379 
2380 	nfs4_fail_recov(vp, "can't recover filehandle", error, stat);
2381 }
2382 
2383 /*
2384  * Recovery from a "shouldn't happen" error.  In the long term, we'd like
2385  * to mark only the data structure(s) that provided the bad value as being
2386  * bad.  But for now we'll just mark the entire file.
2387  */
2388 
2389 static void
2390 recov_badstate(recov_info_t *recovp, vnode_t *vp, nfsstat4 stat)
2391 {
2392 	ASSERT(vp != NULL);
2393 	recov_throttle(recovp, vp);
2394 
2395 	if (!nfs4_valid_recov_err_for_vp(vp, stat))
2396 		return;
2397 
2398 	nfs4_fail_recov(vp, "", 0, stat);
2399 }
2400 
2401 /*
2402  * Free up the information saved for a lost state request.
2403  */
2404 static void
2405 nfs4_free_lost_rqst(nfs4_lost_rqst_t *lrp, nfs4_server_t *sp)
2406 {
2407 	component4 *filep;
2408 	nfs4_open_stream_t *osp;
2409 	int have_sync_lock;
2410 
2411 	NFS4_DEBUG(nfs4_lost_rqst_debug,
2412 		(CE_NOTE, "nfs4_free_lost_rqst:"));
2413 
2414 	switch (lrp->lr_op) {
2415 	case OP_OPEN:
2416 		filep = &lrp->lr_ofile;
2417 		if (filep->utf8string_val) {
2418 			kmem_free(filep->utf8string_val, filep->utf8string_len);
2419 			filep->utf8string_val = NULL;
2420 		}
2421 		break;
2422 	case OP_DELEGRETURN:
2423 		nfs4delegreturn_cleanup(VTOR4(lrp->lr_vp), sp);
2424 		break;
2425 	case OP_CLOSE:
2426 		osp = lrp->lr_osp;
2427 		ASSERT(osp != NULL);
2428 		mutex_enter(&osp->os_sync_lock);
2429 		have_sync_lock = 1;
2430 		if (osp->os_pending_close) {
2431 			/* clean up the open file state. */
2432 			osp->os_pending_close = 0;
2433 			nfs4close_notw(lrp->lr_vp, osp, &have_sync_lock);
2434 		}
2435 		if (have_sync_lock)
2436 			mutex_exit(&osp->os_sync_lock);
2437 		break;
2438 	}
2439 
2440 	lrp->lr_op = 0;
2441 	if (lrp->lr_oop != NULL) {
2442 		open_owner_rele(lrp->lr_oop);
2443 		lrp->lr_oop = NULL;
2444 	}
2445 	if (lrp->lr_osp != NULL) {
2446 		open_stream_rele(lrp->lr_osp, VTOR4(lrp->lr_vp));
2447 		lrp->lr_osp = NULL;
2448 	}
2449 	if (lrp->lr_lop != NULL) {
2450 		lock_owner_rele(lrp->lr_lop);
2451 		lrp->lr_lop = NULL;
2452 	}
2453 	if (lrp->lr_flk != NULL) {
2454 		kmem_free(lrp->lr_flk, sizeof (flock64_t));
2455 		lrp->lr_flk = NULL;
2456 	}
2457 	if (lrp->lr_vp != NULL) {
2458 		VN_RELE(lrp->lr_vp);
2459 		lrp->lr_vp = NULL;
2460 	}
2461 	if (lrp->lr_dvp != NULL) {
2462 		VN_RELE(lrp->lr_dvp);
2463 		lrp->lr_dvp = NULL;
2464 	}
2465 	if (lrp->lr_cr != NULL) {
2466 		crfree(lrp->lr_cr);
2467 		lrp->lr_cr = NULL;
2468 	}
2469 
2470 	kmem_free(lrp, sizeof (nfs4_lost_rqst_t));
2471 }
2472 
2473 /*
2474  * Remove any lost state requests and free them.
2475  */
2476 static void
2477 nfs4_remove_lost_rqsts(mntinfo4_t *mi, nfs4_server_t *sp)
2478 {
2479 	nfs4_lost_rqst_t *lrp;
2480 
2481 	mutex_enter(&mi->mi_lock);
2482 	while ((lrp = list_head(&mi->mi_lost_state)) != NULL) {
2483 		list_remove(&mi->mi_lost_state, lrp);
2484 		mutex_exit(&mi->mi_lock);
2485 		nfs4_free_lost_rqst(lrp, sp);
2486 		mutex_enter(&mi->mi_lock);
2487 	}
2488 	mutex_exit(&mi->mi_lock);
2489 }
2490 
2491 /*
2492  * Reopen all the files for the given filesystem and reclaim any locks.
2493  */
2494 
2495 static void
2496 recov_openfiles(recov_info_t *recovp, nfs4_server_t *sp)
2497 {
2498 	mntinfo4_t *mi = recovp->rc_mi;
2499 	nfs4_opinst_t *reopenlist = NULL, *rep;
2500 	nfs4_error_t e = { 0, NFS4_OK, RPC_SUCCESS };
2501 	open_claim_type4 claim;
2502 	int remap;
2503 	char *fail_msg = "No such file or directory on replica";
2504 	rnode4_t *rp;
2505 	fattr4_change pre_change;
2506 
2507 	ASSERT(sp != NULL);
2508 
2509 	/*
2510 	 * This check is to allow a 10ms pause before we reopen files
2511 	 * it should allow the server time to have received the CB_NULL
2512 	 * reply and update its internal structures such that (if
2513 	 * applicable) we are granted a delegation on reopened files.
2514 	 */
2515 	mutex_enter(&sp->s_lock);
2516 	if ((sp->s_flags & (N4S_CB_PINGED | N4S_CB_WAITER)) == 0) {
2517 		sp->s_flags |= N4S_CB_WAITER;
2518 		(void) cv_timedwait(&sp->wait_cb_null, &sp->s_lock,
2519 			(lbolt+drv_usectohz(N4S_CB_PAUSE_TIME)));
2520 	}
2521 	mutex_exit(&sp->s_lock);
2522 
2523 	(void) nfs_rw_enter_sig(&sp->s_recovlock, RW_READER, 0);
2524 	(void) nfs_rw_enter_sig(&mi->mi_recovlock, RW_WRITER, 0);
2525 
2526 	if (NFS4_VOLATILE_FH(mi)) {
2527 		nfs4_remap_root(mi, &e, 0);
2528 		if (nfs4_needs_recovery(&e, FALSE, mi->mi_vfsp)) {
2529 			(void) nfs4_start_recovery(&e, mi, NULL,
2530 					NULL, NULL, NULL, OP_LOOKUP, NULL);
2531 		}
2532 	}
2533 
2534 	mutex_enter(&mi->mi_lock);
2535 	if (recovp->rc_srv_reboot || (mi->mi_recovflags & MI4R_SRV_REBOOT))
2536 		claim = CLAIM_PREVIOUS;
2537 	else
2538 		claim = CLAIM_NULL;
2539 	mutex_exit(&mi->mi_lock);
2540 
2541 	if (e.error == 0 && e.stat == NFS4_OK) {
2542 		/*
2543 		 * Get a snapshot of open files in the filesystem.  Note
2544 		 * that new opens will stall until the server's grace
2545 		 * period is done.
2546 		 */
2547 		reopenlist = r4mkopenlist(mi);
2548 
2549 		mutex_enter(&mi->mi_lock);
2550 		remap = mi->mi_recovflags & MI4R_REMAP_FILES;
2551 		mutex_exit(&mi->mi_lock);
2552 		/*
2553 		 * Since we are re-establishing state on the
2554 		 * server, its ok to blow away the saved lost
2555 		 * requests since we don't need to reissue it.
2556 		 */
2557 		nfs4_remove_lost_rqsts(mi, sp);
2558 
2559 		for (rep = reopenlist; rep; rep = rep->re_next) {
2560 
2561 			if (remap) {
2562 				nfs4_remap_file(mi, rep->re_vp,
2563 					NFS4_REMAP_CKATTRS, &e);
2564 			}
2565 			if (e.error == ENOENT || e.stat == NFS4ERR_NOENT) {
2566 				/*
2567 				 * The current server does not have the file
2568 				 * that is to be remapped.  This is most
2569 				 * likely due to an improperly maintained
2570 				 * replica.   The files that are missing from
2571 				 * the server will be marked dead and logged
2572 				 * in order to make sys admins aware of the
2573 				 * problem.
2574 				 */
2575 				nfs4_fail_recov(rep->re_vp,
2576 					fail_msg, e.error, e.stat);
2577 				/*
2578 				 * We've already handled the error so clear it.
2579 				 */
2580 				nfs4_error_zinit(&e);
2581 				continue;
2582 			} else if (e.error == 0 && e.stat == NFS4_OK) {
2583 				int j;
2584 
2585 				rp = VTOR4(rep->re_vp);
2586 				mutex_enter(&rp->r_statelock);
2587 				pre_change = rp->r_change;
2588 				mutex_exit(&rp->r_statelock);
2589 
2590 				for (j = 0; j < rep->re_numosp; j++) {
2591 					nfs4_reopen(rep->re_vp, rep->re_osp[j],
2592 						&e, claim, FALSE, TRUE);
2593 					if (e.error != 0 || e.stat != NFS4_OK)
2594 						break;
2595 				}
2596 				if (nfs4_needs_recovery(&e, TRUE,
2597 				    mi->mi_vfsp)) {
2598 					(void) nfs4_start_recovery(&e, mi,
2599 						rep->re_vp, NULL, NULL, NULL,
2600 						OP_OPEN, NULL);
2601 					break;
2602 				}
2603 			}
2604 #ifdef DEBUG
2605 			if (nfs4_recovdelay > 0)
2606 				delay(MSEC_TO_TICK(nfs4_recovdelay * 1000));
2607 #endif
2608 			if (e.error == 0 && e.stat == NFS4_OK)
2609 				relock_file(rep->re_vp, mi, &e, pre_change);
2610 
2611 			if (nfs4_needs_recovery(&e, TRUE, mi->mi_vfsp))
2612 				(void) nfs4_start_recovery(&e, mi,
2613 					rep->re_vp, NULL, NULL, NULL, OP_LOCK,
2614 					NULL);
2615 			if (e.error != 0 || e.stat != NFS4_OK)
2616 				break;
2617 		}
2618 
2619 		/*
2620 		 * Check to see if we need to remap files passed in
2621 		 * via the recovery arguments; this will have been
2622 		 * done for open files.  A failure here is not fatal.
2623 		 */
2624 		if (remap) {
2625 			nfs4_error_t ignore;
2626 			nfs4_check_remap(mi, recovp->rc_vp1, NFS4_REMAP_CKATTRS,
2627 				&ignore);
2628 			nfs4_check_remap(mi, recovp->rc_vp2, NFS4_REMAP_CKATTRS,
2629 				&ignore);
2630 		}
2631 	}
2632 
2633 	if (e.error == 0 && e.stat == NFS4_OK) {
2634 		mutex_enter(&mi->mi_lock);
2635 		mi->mi_recovflags &= ~(MI4R_REOPEN_FILES | MI4R_REMAP_FILES);
2636 		mutex_exit(&mi->mi_lock);
2637 	}
2638 
2639 	nfs_rw_exit(&mi->mi_recovlock);
2640 	nfs_rw_exit(&sp->s_recovlock);
2641 
2642 	if (reopenlist != NULL)
2643 		r4releopenlist(reopenlist);
2644 }
2645 
2646 /*
2647  * Resend the queued state recovery requests in "rqsts".
2648  */
2649 
2650 static void
2651 nfs4_resend_lost_rqsts(recov_info_t *recovp, nfs4_server_t *sp)
2652 {
2653 	nfs4_lost_rqst_t	*lrp, *tlrp;
2654 	mntinfo4_t		*mi = recovp->rc_mi;
2655 	nfs4_error_t		n4e;
2656 #ifdef NOTYET
2657 	uint32_t		deny_bits = 0;
2658 #endif
2659 
2660 	NFS4_DEBUG(nfs4_lost_rqst_debug, (CE_NOTE, "nfs4_resend_lost_rqsts"));
2661 
2662 	ASSERT(mi != NULL);
2663 	ASSERT(nfs_rw_lock_held(&mi->mi_recovlock, RW_WRITER));
2664 
2665 	mutex_enter(&mi->mi_lock);
2666 	lrp = list_head(&mi->mi_lost_state);
2667 	mutex_exit(&mi->mi_lock);
2668 	while (lrp != NULL) {
2669 		nfs4_error_zinit(&n4e);
2670 		resend_one_op(lrp, &n4e, mi, sp);
2671 		NFS4_DEBUG(nfs4_lost_rqst_debug, (CE_NOTE,
2672 		    "nfs4_resend_lost_rqsts: resend request: for vp %p got "
2673 		    "error %d stat %d", (void *)lrp->lr_vp, n4e.error,
2674 		    n4e.stat));
2675 
2676 		/*
2677 		 * If we get a recovery error that we can actually
2678 		 * recover from (such as ETIMEDOUT, FHEXPIRED), we
2679 		 * return and let the recovery thread redrive the call.
2680 		 * Don't requeue unless the zone is still healthy.
2681 		 */
2682 		if (zone_status_get(curproc->p_zone) < ZONE_IS_SHUTTING_DOWN &&
2683 		    nfs4_needs_recovery(&n4e, TRUE, mi->mi_vfsp) &&
2684 		    (nfs4_try_failover(&n4e) ||
2685 		    NFS4_FRC_UNMT_ERR(n4e.error, mi->mi_vfsp) ||
2686 		    (n4e.error == 0 && n4e.stat != NFS4ERR_BADHANDLE &&
2687 		    !nfs4_recov_marks_dead(n4e.stat)))) {
2688 			/*
2689 			 * For these three errors, we want to delay a bit
2690 			 * instead of pounding the server into submission.
2691 			 * We have to do this manually; the normal
2692 			 * processing for these errors only works for
2693 			 * non-recovery requests.
2694 			 */
2695 			if ((n4e.error == 0 && n4e.stat == NFS4ERR_DELAY) ||
2696 			    (n4e.error == 0 && n4e.stat == NFS4ERR_GRACE) ||
2697 			    (n4e.error == 0 && n4e.stat == NFS4ERR_RESOURCE) ||
2698 			    NFS4_FRC_UNMT_ERR(n4e.error, mi->mi_vfsp)) {
2699 				delay(SEC_TO_TICK(nfs4err_delay_time));
2700 			} else {
2701 				(void) nfs4_start_recovery(&n4e,
2702 					mi, lrp->lr_dvp, lrp->lr_vp, NULL, NULL,
2703 					lrp->lr_op, NULL);
2704 			}
2705 			return;
2706 		}
2707 
2708 		mutex_enter(&mi->mi_lock);
2709 		list_remove(&mi->mi_lost_state, lrp);
2710 		tlrp = lrp;
2711 		lrp = list_head(&mi->mi_lost_state);
2712 		mutex_exit(&mi->mi_lock);
2713 		nfs4_free_lost_rqst(tlrp, sp);
2714 	}
2715 }
2716 
2717 /*
2718  * Resend the given op, and issue any necessary undo call.
2719  * errors are returned via the nfs4_error_t parameter.
2720  */
2721 
2722 static void
2723 resend_one_op(nfs4_lost_rqst_t *lrp, nfs4_error_t *ep,
2724 	mntinfo4_t *mi, nfs4_server_t *sp)
2725 {
2726 	vnode_t *vp;
2727 	nfs4_open_stream_t *osp;
2728 	cred_t *cr;
2729 	uint32_t acc_bits;
2730 
2731 	vp = lrp->lr_vp;
2732 	NFS4_DEBUG(nfs4_lost_rqst_debug, (CE_NOTE, "resend_one_op: "
2733 	    "have a lost open/close request for vp %p", (void *)vp));
2734 
2735 	switch (lrp->lr_op) {
2736 	case OP_OPEN:
2737 		nfs4_resend_open_otw(&vp, lrp, ep);
2738 		break;
2739 	case OP_OPEN_DOWNGRADE:
2740 		ASSERT(lrp->lr_oop != NULL);
2741 		ep->error = nfs4_start_open_seqid_sync(lrp->lr_oop, mi);
2742 		ASSERT(!ep->error);	/* recov thread always succeeds */
2743 		ASSERT(lrp->lr_osp != NULL);
2744 		mutex_enter(&lrp->lr_osp->os_sync_lock);
2745 		nfs4_open_downgrade(lrp->lr_dg_acc, lrp->lr_dg_deny,
2746 			    lrp->lr_oop, lrp->lr_osp, vp, lrp->lr_cr, lrp,
2747 			    ep, NULL, NULL);
2748 		mutex_exit(&lrp->lr_osp->os_sync_lock);
2749 		nfs4_end_open_seqid_sync(lrp->lr_oop);
2750 		break;
2751 	case OP_CLOSE:
2752 		osp = lrp->lr_osp;
2753 		cr = lrp->lr_cr;
2754 		acc_bits = 0;
2755 		mutex_enter(&osp->os_sync_lock);
2756 		if (osp->os_share_acc_read)
2757 			acc_bits |= OPEN4_SHARE_ACCESS_READ;
2758 		if (osp->os_share_acc_write)
2759 			acc_bits |= OPEN4_SHARE_ACCESS_WRITE;
2760 		mutex_exit(&osp->os_sync_lock);
2761 		nfs4close_one(vp, osp, cr, acc_bits, lrp, ep,
2762 				CLOSE_RESEND, 0, 0, 0);
2763 		break;
2764 	case OP_LOCK:
2765 	case OP_LOCKU:
2766 		resend_lock(lrp, ep);
2767 		goto done;
2768 	case OP_DELEGRETURN:
2769 		nfs4_resend_delegreturn(lrp, ep, sp);
2770 		goto done;
2771 	default:
2772 #ifdef DEBUG
2773 		cmn_err(CE_PANIC, "resend_one_op: unexpected op: %d",
2774 			lrp->lr_op);
2775 #endif
2776 		nfs4_queue_event(RE_LOST_STATE_BAD_OP, mi, NULL,
2777 		    lrp->lr_op, lrp->lr_vp, lrp->lr_dvp, NFS4_OK, NULL, 0,
2778 		    TAG_NONE, TAG_NONE, 0, 0);
2779 		nfs4_error_init(ep, EINVAL);
2780 		return;
2781 	}
2782 
2783 	/*
2784 	 * No need to retry nor send an "undo" CLOSE in the
2785 	 * event the server rebooted.
2786 	 */
2787 	if (ep->error == 0 && (ep->stat == NFS4ERR_STALE_CLIENTID ||
2788 	    ep->stat == NFS4ERR_STALE_STATEID || ep->stat == NFS4ERR_EXPIRED))
2789 		goto done;
2790 
2791 	/*
2792 	 * If we resent a CLOSE or OPEN_DOWNGRADE, there's nothing
2793 	 * to undo.  Undoing locking operations was handled by
2794 	 * resend_lock().
2795 	 */
2796 	if (lrp->lr_op == OP_OPEN_DOWNGRADE || lrp->lr_op == OP_CLOSE)
2797 		goto done;
2798 
2799 	/*
2800 	 * If we get any other error for OPEN, then don't attempt
2801 	 * to undo the resend of the open (since it was never
2802 	 * successful!).
2803 	 */
2804 	ASSERT(lrp->lr_op == OP_OPEN);
2805 	if (ep->error || ep->stat != NFS4_OK)
2806 		goto done;
2807 
2808 	/*
2809 	 * Now let's undo our OPEN.
2810 	 */
2811 	nfs4_error_zinit(ep);
2812 	close_after_open_resend(vp, lrp->lr_cr, lrp->lr_oacc, ep);
2813 	NFS4_DEBUG(nfs4_lost_rqst_debug, (CE_NOTE, "resend_one_op: "
2814 	    "nfs4close_one: for vp %p got error %d stat %d",
2815 	    (void *)vp, ep->error, ep->stat));
2816 
2817 done:
2818 	if (vp != lrp->lr_vp)
2819 		VN_RELE(vp);
2820 }
2821 
2822 /*
2823  * Close a file that was opened via a resent OPEN.
2824  * Most errors are passed back to the caller (via the return value and
2825  * *statp), except for FHEXPIRED, which is retried.
2826  *
2827  * It might be conceptually cleaner to push the CLOSE request onto the
2828  * front of the resend queue, rather than sending it here.  That would
2829  * match the way we undo lost lock requests.  On the other
2830  * hand, we've already got something that works, and there's no reason to
2831  * change it at this time.
2832  */
2833 
2834 static void
2835 close_after_open_resend(vnode_t *vp, cred_t *cr, uint32_t acc_bits,
2836 			nfs4_error_t *ep)
2837 {
2838 
2839 	for (;;) {
2840 		nfs4close_one(vp, NULL, cr, acc_bits, NULL, ep,
2841 				CLOSE_AFTER_RESEND, 0, 0, 0);
2842 		if (ep->error == 0 && ep->stat == NFS4_OK)
2843 			break;		/* success; done */
2844 		if (ep->error != 0 || ep->stat != NFS4ERR_FHEXPIRED)
2845 			break;
2846 		/* else retry FHEXPIRED */
2847 	}
2848 
2849 }
2850 
2851 /*
2852  * Resend the given lost lock request.  Return an errno value.  If zero,
2853  * *statp is set to the NFS status code for the call.
2854  *
2855  * Issue a SIGLOST and mark the rnode dead if we get a non-recovery error or
2856  * a recovery error that we don't actually recover from yet (eg: BAD_SEQID).
2857  * Let the recovery thread redrive the call if we get a recovery error that
2858  * we can actually recover from.
2859  */
2860 static void
2861 resend_lock(nfs4_lost_rqst_t *lrp, nfs4_error_t *ep)
2862 {
2863 	bool_t		send_siglost = FALSE;
2864 	vnode_t		*vp = lrp->lr_vp;
2865 
2866 	NFS4_DEBUG(nfs4_lost_rqst_debug, (CE_NOTE, "resend_lock:"));
2867 	ASSERT(lrp->lr_ctype == NFS4_LCK_CTYPE_REINSTATE ||
2868 	    lrp->lr_ctype == NFS4_LCK_CTYPE_RESEND);
2869 
2870 	nfs4frlock(lrp->lr_ctype, vp, F_SETLK,
2871 		    lrp->lr_flk, FREAD|FWRITE, 0, lrp->lr_cr, ep, lrp, NULL);
2872 
2873 	NFS4_DEBUG(nfs4_lost_rqst_debug, (CE_NOTE, "resend_lock: "
2874 	    "nfs4frlock for vp %p returned error %d, stat %d",
2875 	    (void *)vp, ep->error, ep->stat));
2876 
2877 	if (ep->error == 0 && ep->stat == 0)
2878 		goto done;
2879 	if (ep->error == 0 && ep->stat == NFS4ERR_DENIED &&
2880 	    lrp->lr_ctype == NFS4_LCK_CTYPE_RESEND)
2881 		goto done;
2882 
2883 	/*
2884 	 * If we failed with a non-recovery error, send SIGLOST and
2885 	 * mark the file dead.
2886 	 */
2887 	if (!nfs4_needs_recovery(ep, TRUE, vp->v_vfsp))
2888 		send_siglost = TRUE;
2889 	else {
2890 		/*
2891 		 * Done with recovering LOST LOCK in the event the
2892 		 * server rebooted or we've lost the lease.
2893 		 */
2894 		if (ep->error == 0 && (ep->stat == NFS4ERR_STALE_CLIENTID ||
2895 		    ep->stat == NFS4ERR_STALE_STATEID ||
2896 		    ep->stat == NFS4ERR_EXPIRED)) {
2897 			goto done;
2898 		}
2899 
2900 		/*
2901 		 * BAD_STATEID on an unlock indicates that the server has
2902 		 * forgotten about the lock anyway, so act like the call
2903 		 * was successful.
2904 		 */
2905 		if (ep->error == 0 && ep->stat == NFS4ERR_BAD_STATEID &&
2906 		    lrp->lr_op == OP_LOCKU)
2907 			goto done;
2908 
2909 		/*
2910 		 * If we got a recovery error that we don't actually
2911 		 * recover from, send SIGLOST.  If the filesystem was
2912 		 * forcibly unmounted, we skip the SIGLOST because (a) it's
2913 		 * unnecessary noise, and (b) there could be a new process
2914 		 * with the same pid as the one that had generated the lost
2915 		 * state request.
2916 		 */
2917 		if (ep->error == 0 && (ep->stat == NFS4ERR_BADHANDLE ||
2918 		    nfs4_recov_marks_dead(ep->stat))) {
2919 			if (!(vp->v_vfsp->vfs_flag & VFS_UNMOUNTED))
2920 				send_siglost = TRUE;
2921 			goto done;
2922 		}
2923 
2924 		/*
2925 		 * If the filesystem was forcibly unmounted, we
2926 		 * still need to synchronize with the server and
2927 		 * release state.  Try again later.
2928 		 */
2929 		if (NFS4_FRC_UNMT_ERR(ep->error, vp->v_vfsp))
2930 			goto done;
2931 
2932 		/*
2933 		 * If we get a recovery error that we can actually
2934 		 * recover from (such as ETIMEDOUT, FHEXPIRED),
2935 		 * return and let the recovery thread redrive the call.
2936 		 *
2937 		 * For the three errors below, we want to delay a bit
2938 		 * instead of pounding the server into submission.
2939 		 */
2940 		if ((ep->error == 0 && ep->stat == NFS4ERR_DELAY) ||
2941 		    (ep->error == 0 && ep->stat == NFS4ERR_GRACE) ||
2942 		    (ep->error == 0 && ep->stat == NFS4ERR_RESOURCE))
2943 			delay(SEC_TO_TICK(recov_err_delay));
2944 		goto done;
2945 	}
2946 
2947 done:
2948 	if (send_siglost) {
2949 		cred_t *sv_cred;
2950 
2951 		/*
2952 		 * Must be root or the actual thread being issued the
2953 		 * SIGLOST for this to work, so just become root.
2954 		 */
2955 		sv_cred = curthread->t_cred;
2956 		curthread->t_cred = kcred;
2957 		nfs4_send_siglost(lrp->lr_flk->l_pid, VTOMI4(vp), vp, FALSE,
2958 		    ep->error, ep->stat);
2959 		curthread->t_cred = sv_cred;
2960 
2961 		/*
2962 		 * Flush any additional reinstantiation requests for
2963 		 * this operation.  Sending multiple SIGLOSTs to the user
2964 		 * process is unlikely to help and may cause trouble.
2965 		 */
2966 		if (lrp->lr_ctype == NFS4_LCK_CTYPE_REINSTATE)
2967 			flush_reinstate(lrp);
2968 	}
2969 }
2970 
2971 /*
2972  * Remove any lock reinstantiation requests that correspond to the given
2973  * lost request.  We only remove items that follow lrp in the queue,
2974  * assuming that lrp will be removed by the generic lost state code.
2975  */
2976 
2977 static void
2978 flush_reinstate(nfs4_lost_rqst_t *lrp)
2979 {
2980 	vnode_t *vp;
2981 	pid_t pid;
2982 	mntinfo4_t *mi;
2983 	nfs4_lost_rqst_t *nlrp;
2984 
2985 	vp = lrp->lr_vp;
2986 	mi = VTOMI4(vp);
2987 	pid = lrp->lr_flk->l_pid;
2988 
2989 	/*
2990 	 * If there are any more reinstantation requests to get rid of,
2991 	 * they should all be clustered at the front of the lost state
2992 	 * queue.
2993 	 */
2994 	mutex_enter(&mi->mi_lock);
2995 	for (lrp = list_next(&mi->mi_lost_state, lrp); lrp != NULL;
2996 	    lrp = nlrp) {
2997 		nlrp = list_next(&mi->mi_lost_state, lrp);
2998 		if (lrp->lr_op != OP_LOCK && lrp->lr_op != OP_LOCKU)
2999 			break;
3000 		if (lrp->lr_ctype != NFS4_LCK_CTYPE_REINSTATE)
3001 			break;
3002 		ASSERT(lrp->lr_vp == vp);
3003 		ASSERT(lrp->lr_flk->l_pid == pid);
3004 		NFS4_DEBUG(nfs4_lost_rqst_debug, (CE_NOTE,
3005 				"remove reinstantiation %p", (void *)lrp));
3006 		list_remove(&mi->mi_lost_state, lrp);
3007 		nfs4_free_lost_rqst(lrp, NULL);
3008 	}
3009 	mutex_exit(&mi->mi_lock);
3010 }
3011 
3012 /*
3013  * End of state-specific recovery routines.
3014  */
3015 
3016 /*
3017  * Allocate a lost request struct, initialize it from lost_rqstp (including
3018  * bumping the reference counts for the referenced vnode, etc.), and hang
3019  * it off of recovp.
3020  */
3021 
3022 static void
3023 nfs4_save_lost_rqst(nfs4_lost_rqst_t *lost_rqstp, recov_info_t *recovp,
3024 	nfs4_recov_t *action, mntinfo4_t *mi)
3025 {
3026 	nfs4_lost_rqst_t *destp;
3027 
3028 	ASSERT(recovp->rc_lost_rqst == NULL);
3029 
3030 	destp = kmem_alloc(sizeof (nfs4_lost_rqst_t), KM_SLEEP);
3031 	recovp->rc_lost_rqst = destp;
3032 
3033 	if (lost_rqstp->lr_op == OP_LOCK ||
3034 	    lost_rqstp->lr_op == OP_LOCKU) {
3035 		ASSERT(lost_rqstp->lr_lop);
3036 		*action = NR_LOST_LOCK;
3037 		destp->lr_ctype = lost_rqstp->lr_ctype;
3038 		destp->lr_locktype = lost_rqstp->lr_locktype;
3039 	} else if (lost_rqstp->lr_op == OP_OPEN) {
3040 		component4 *srcfp, *destfp;
3041 
3042 		destp->lr_oacc = lost_rqstp->lr_oacc;
3043 		destp->lr_odeny = lost_rqstp->lr_odeny;
3044 		destp->lr_oclaim = lost_rqstp->lr_oclaim;
3045 		if (lost_rqstp->lr_oclaim == CLAIM_DELEGATE_CUR)
3046 			destp->lr_ostateid = lost_rqstp->lr_ostateid;
3047 
3048 		srcfp = &lost_rqstp->lr_ofile;
3049 		destfp = &destp->lr_ofile;
3050 		/*
3051 		 * Consume caller's utf8string
3052 		 */
3053 		destfp->utf8string_len = srcfp->utf8string_len;
3054 		destfp->utf8string_val = srcfp->utf8string_val;
3055 		srcfp->utf8string_len = 0;
3056 		srcfp->utf8string_val = NULL;	/* make sure not reused */
3057 
3058 		*action = NR_LOST_STATE_RQST;
3059 	} else if (lost_rqstp->lr_op == OP_OPEN_DOWNGRADE) {
3060 		destp->lr_dg_acc = lost_rqstp->lr_dg_acc;
3061 		destp->lr_dg_deny = lost_rqstp->lr_dg_deny;
3062 
3063 		*action = NR_LOST_STATE_RQST;
3064 	} else if (lost_rqstp->lr_op == OP_CLOSE) {
3065 		ASSERT(lost_rqstp->lr_oop);
3066 		*action = NR_LOST_STATE_RQST;
3067 	} else if (lost_rqstp->lr_op == OP_DELEGRETURN) {
3068 		*action = NR_LOST_STATE_RQST;
3069 	} else {
3070 #ifdef DEBUG
3071 		cmn_err(CE_PANIC, "nfs4_save_lost_rqst: bad op %d",
3072 			lost_rqstp->lr_op);
3073 #endif
3074 		nfs4_queue_event(RE_LOST_STATE_BAD_OP, mi, NULL,
3075 		    lost_rqstp->lr_op, lost_rqstp->lr_vp, lost_rqstp->lr_dvp,
3076 		    NFS4_OK, NULL, curproc->p_pid, TAG_NONE, TAG_NONE, 0, 0);
3077 		*action = NR_UNUSED;
3078 		recovp->rc_lost_rqst = NULL;
3079 		kmem_free(destp, sizeof (nfs4_lost_rqst_t));
3080 		return;
3081 	}
3082 
3083 	destp->lr_op = lost_rqstp->lr_op;
3084 	destp->lr_vp = lost_rqstp->lr_vp;
3085 	if (destp->lr_vp)
3086 		VN_HOLD(destp->lr_vp);
3087 	destp->lr_dvp = lost_rqstp->lr_dvp;
3088 	if (destp->lr_dvp)
3089 		VN_HOLD(destp->lr_dvp);
3090 	destp->lr_oop = lost_rqstp->lr_oop;
3091 	if (destp->lr_oop)
3092 		open_owner_hold(destp->lr_oop);
3093 	destp->lr_osp = lost_rqstp->lr_osp;
3094 	if (destp->lr_osp)
3095 		open_stream_hold(destp->lr_osp);
3096 	destp->lr_lop = lost_rqstp->lr_lop;
3097 	if (destp->lr_lop)
3098 		lock_owner_hold(destp->lr_lop);
3099 	destp->lr_cr = lost_rqstp->lr_cr;
3100 	if (destp->lr_cr)
3101 		crhold(destp->lr_cr);
3102 	if (lost_rqstp->lr_flk == NULL)
3103 		destp->lr_flk = NULL;
3104 	else {
3105 		destp->lr_flk = kmem_alloc(sizeof (flock64_t), KM_SLEEP);
3106 		*destp->lr_flk = *lost_rqstp->lr_flk;
3107 	}
3108 	destp->lr_putfirst = lost_rqstp->lr_putfirst;
3109 }
3110 
3111 /*
3112  * Map the given return values (errno and nfs4 status code) to a recovery
3113  * action and fill in the following fields of recovp: rc_action,
3114  * rc_srv_reboot, rc_stateid, rc_lost_rqst.
3115  */
3116 
3117 void
3118 errs_to_action(recov_info_t *recovp,
3119 	nfs4_server_t *sp, mntinfo4_t *mi, stateid4 *sidp,
3120 	nfs4_lost_rqst_t *lost_rqstp, int unmounted, nfs_opnum4 op,
3121 	nfs4_bseqid_entry_t *bsep)
3122 {
3123 	nfs4_recov_t action = NR_UNUSED;
3124 	bool_t reboot = FALSE;
3125 	int try_f;
3126 	int error = recovp->rc_orig_errors.error;
3127 	nfsstat4 stat = recovp->rc_orig_errors.stat;
3128 
3129 	bzero(&recovp->rc_stateid, sizeof (stateid4));
3130 	recovp->rc_lost_rqst = NULL;
3131 	recovp->rc_bseqid_rqst = NULL;
3132 
3133 	try_f = nfs4_try_failover(&recovp->rc_orig_errors) &&
3134 			FAILOVER_MOUNT4(mi);
3135 
3136 	/*
3137 	 * We start recovery for EINTR only in the lost lock
3138 	 * or lost open/close case.
3139 	 */
3140 
3141 	if (try_f || error == EINTR || (error == EIO && unmounted)) {
3142 		recovp->rc_error = (error != 0 ? error : geterrno4(stat));
3143 		if (lost_rqstp) {
3144 			ASSERT(lost_rqstp->lr_op != 0);
3145 			nfs4_save_lost_rqst(lost_rqstp, recovp, &action, mi);
3146 		}
3147 		if (try_f)
3148 			action = NR_FAILOVER;
3149 	} else if (error != 0) {
3150 		recovp->rc_error = error;
3151 		nfs4_queue_event(RE_UNEXPECTED_ERRNO, mi, NULL, error, NULL,
3152 		    NULL, 0, NULL, 0, TAG_NONE, TAG_NONE, 0, 0);
3153 		action = NR_CLIENTID;
3154 	} else {
3155 		recovp->rc_error = geterrno4(stat);
3156 		switch (stat) {
3157 #ifdef notyet
3158 		case NFS4ERR_LEASE_MOVED:
3159 			action = xxx;
3160 			break;
3161 		case NFS4ERR_MOVED:
3162 			action = xxx;
3163 			break;
3164 #endif
3165 		case NFS4ERR_BADHANDLE:
3166 			action = NR_BADHANDLE;
3167 			break;
3168 		case NFS4ERR_BAD_SEQID:
3169 			if (bsep)
3170 				save_bseqid_rqst(bsep, recovp);
3171 			action = NR_BAD_SEQID;
3172 			break;
3173 		case NFS4ERR_OLD_STATEID:
3174 			action = NR_OLDSTATEID;
3175 			break;
3176 		case NFS4ERR_WRONGSEC:
3177 			action = NR_WRONGSEC;
3178 			break;
3179 		case NFS4ERR_FHEXPIRED:
3180 			action = NR_FHEXPIRED;
3181 			break;
3182 		case NFS4ERR_BAD_STATEID:
3183 			if (sp == NULL || (sp != NULL && inlease(sp))) {
3184 
3185 				action = NR_BAD_STATEID;
3186 				if (sidp)
3187 					recovp->rc_stateid = *sidp;
3188 			} else
3189 				action = NR_CLIENTID;
3190 			break;
3191 		case NFS4ERR_EXPIRED:
3192 			/*
3193 			 * The client's lease has expired, either due
3194 			 * to a network partition or perhaps a client
3195 			 * error.  In either case, try an NR_CLIENTID
3196 			 * style recovery.  reboot remains false, since
3197 			 * there is no evidence the server has rebooted.
3198 			 * This will cause CLAIM_NULL opens and lock
3199 			 * requests without the reclaim bit.
3200 			 */
3201 			action = NR_CLIENTID;
3202 
3203 			DTRACE_PROBE4(nfs4__expired,
3204 					nfs4_server_t *, sp,
3205 					mntinfo4_t *, mi,
3206 					stateid4 *, sidp, int, op);
3207 
3208 			break;
3209 		case NFS4ERR_STALE_CLIENTID:
3210 		case NFS4ERR_STALE_STATEID:
3211 			action = NR_CLIENTID;
3212 			reboot = TRUE;
3213 			break;
3214 		case NFS4ERR_RESOURCE:
3215 			/*
3216 			 * If this had been a FAILOVER mount, then
3217 			 * we'd have tried failover.  Since it's not,
3218 			 * just delay a while and retry.
3219 			 */
3220 			action = NR_DELAY;
3221 			break;
3222 		case NFS4ERR_GRACE:
3223 			action = NR_GRACE;
3224 			break;
3225 		case NFS4ERR_DELAY:
3226 			action = NR_DELAY;
3227 			break;
3228 		case NFS4ERR_STALE:
3229 			action = NR_STALE;
3230 			break;
3231 		default:
3232 			nfs4_queue_event(RE_UNEXPECTED_STATUS, mi, NULL, 0,
3233 			    NULL, NULL, stat, NULL, 0, TAG_NONE, TAG_NONE,
3234 			    0, 0);
3235 			action = NR_CLIENTID;
3236 			break;
3237 		}
3238 	}
3239 
3240 	/* make sure action got set */
3241 	ASSERT(action != NR_UNUSED);
3242 	recovp->rc_srv_reboot = reboot;
3243 	recovp->rc_action = action;
3244 	nfs4_queue_fact(RF_ERR, mi, stat, action, op, reboot, NULL, error,
3245 		NULL);
3246 }
3247 
3248 /*
3249  * Return the (held) credential for the process with the given pid.
3250  * May return NULL (e.g., process not found).
3251  */
3252 
3253 static cred_t *
3254 pid_to_cr(pid_t pid)
3255 {
3256 	proc_t *p;
3257 	cred_t *cr;
3258 
3259 	mutex_enter(&pidlock);
3260 	if ((p = prfind(pid)) == NULL) {
3261 		mutex_exit(&pidlock);
3262 		return (NULL);
3263 	}
3264 
3265 	mutex_enter(&p->p_crlock);
3266 	crhold(cr = p->p_cred);
3267 	mutex_exit(&p->p_crlock);
3268 	mutex_exit(&pidlock);
3269 
3270 	return (cr);
3271 }
3272 
3273 /*
3274  * Send SIGLOST to the given process and queue the event.
3275  *
3276  * The 'dump' boolean tells us whether this action should dump the
3277  * in-kernel queue of recovery messages or not.
3278  */
3279 
3280 void
3281 nfs4_send_siglost(pid_t pid, mntinfo4_t *mi, vnode_t *vp, bool_t dump,
3282     int error, nfsstat4 stat)
3283 {
3284 	proc_t *p;
3285 
3286 	mutex_enter(&pidlock);
3287 	p = prfind(pid);
3288 	if (p)
3289 		psignal(p, SIGLOST);
3290 	mutex_exit(&pidlock);
3291 	nfs4_queue_event(dump ? RE_SIGLOST : RE_SIGLOST_NO_DUMP, mi,
3292 	    NULL, error, vp, NULL, stat, NULL, pid, TAG_NONE, TAG_NONE, 0, 0);
3293 }
3294 
3295 /*
3296  * Scan the lock list for entries that match the given pid.  Change the
3297  * pid in those that do to NOPID.
3298  */
3299 
3300 static void
3301 relock_skip_pid(locklist_t *llp, pid_t pid)
3302 {
3303 	for (; llp != NULL; llp = llp->ll_next) {
3304 		if (llp->ll_flock.l_pid == pid)
3305 			llp->ll_flock.l_pid = NOPID;
3306 	}
3307 }
3308 
3309 /*
3310  * Mark a file as having failed recovery, after making a last-ditch effort
3311  * to return any delegation.
3312  *
3313  * Sets r_error to EIO or ESTALE for the given vnode.
3314  */
3315 void
3316 nfs4_fail_recov(vnode_t *vp, char *why, int error, nfsstat4 stat)
3317 {
3318 	rnode4_t *rp = VTOR4(vp);
3319 
3320 #ifdef DEBUG
3321 	if (nfs4_fail_recov_stop)
3322 		debug_enter("nfs4_fail_recov");
3323 #endif
3324 
3325 	mutex_enter(&rp->r_statelock);
3326 	if (rp->r_flags & (R4RECOVERR|R4RECOVERRP)) {
3327 		mutex_exit(&rp->r_statelock);
3328 		return;
3329 	}
3330 
3331 	/*
3332 	 * Set R4RECOVERRP to indicate that a recovery error is in
3333 	 * progress.  This will shut down reads and writes at the top
3334 	 * half.  Don't set R4RECOVERR until after we've returned the
3335 	 * delegation, otherwise it will fail.
3336 	 */
3337 
3338 	rp->r_flags |= R4RECOVERRP;
3339 	mutex_exit(&rp->r_statelock);
3340 
3341 	nfs4delegabandon(rp);
3342 
3343 	mutex_enter(&rp->r_statelock);
3344 	rp->r_flags |= (R4RECOVERR | R4STALE);
3345 	rp->r_error = (error == 0 && stat == NFS4ERR_STALE) ? ESTALE : EIO;
3346 	PURGE_ATTRCACHE4_LOCKED(rp);
3347 	if (!(vp->v_vfsp->vfs_flag & VFS_UNMOUNTED))
3348 		nfs4_queue_event(RE_DEAD_FILE, VTOMI4(vp), NULL, error,
3349 		    vp, NULL, stat, why, 0, TAG_NONE, TAG_NONE, 0, 0);
3350 	mutex_exit(&rp->r_statelock);
3351 
3352 	dnlc_purge_vp(vp);
3353 }
3354 
3355 /*
3356  * recov_throttle: if the file had the same recovery action within the
3357  * throttle interval, wait for the throttle interval to finish before
3358  * proceeding.
3359  *
3360  * Side effects: updates the rnode with the current recovery information.
3361  */
3362 
3363 static void
3364 recov_throttle(recov_info_t *recovp, vnode_t *vp)
3365 {
3366 	time_t curtime, time_to_wait;
3367 	rnode4_t *rp = VTOR4(vp);
3368 
3369 	curtime = gethrestime_sec();
3370 
3371 	mutex_enter(&rp->r_statelock);
3372 	NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
3373 		"recov_throttle: now: (%d, %ld), last: (%d, %ld)",
3374 		recovp->rc_action, curtime,
3375 		rp->r_recov_act, rp->r_last_recov));
3376 	if (recovp->rc_action == rp->r_recov_act &&
3377 	    rp->r_last_recov + recov_err_delay > curtime) {
3378 		time_to_wait = rp->r_last_recov + recov_err_delay - curtime;
3379 		mutex_exit(&rp->r_statelock);
3380 		delay(SEC_TO_TICK(time_to_wait));
3381 		curtime = gethrestime_sec();
3382 		mutex_enter(&rp->r_statelock);
3383 	}
3384 
3385 	rp->r_last_recov = curtime;
3386 	rp->r_recov_act = recovp->rc_action;
3387 	mutex_exit(&rp->r_statelock);
3388 }
3389 
3390 /*
3391  * React to NFS4ERR_GRACE by setting the time we'll permit
3392  * the next call to this filesystem.
3393  */
3394 void
3395 nfs4_set_grace_wait(mntinfo4_t *mi)
3396 {
3397 	mutex_enter(&mi->mi_lock);
3398 	/* Mark the time for the future */
3399 	mi->mi_grace_wait = gethrestime_sec() + nfs4err_delay_time;
3400 	mutex_exit(&mi->mi_lock);
3401 }
3402 
3403 /*
3404  * React to MFS4ERR_DELAY by setting the time we'll permit
3405  * the next call to this vnode.
3406  */
3407 void
3408 nfs4_set_delay_wait(vnode_t *vp)
3409 {
3410 	rnode4_t *rp = VTOR4(vp);
3411 
3412 	mutex_enter(&rp->r_statelock);
3413 	/*
3414 	 * Calculate amount we should delay, initial
3415 	 * delay will be short and then we will back off.
3416 	 */
3417 	if (rp->r_delay_interval == 0)
3418 		rp->r_delay_interval = NFS4_INITIAL_DELAY_INTERVAL;
3419 	else
3420 		/* calculate next interval value */
3421 		rp->r_delay_interval =
3422 		    MIN(NFS4_MAX_DELAY_INTERVAL, (rp->r_delay_interval << 1));
3423 	rp->r_delay_wait = gethrestime_sec() + rp->r_delay_interval;
3424 	mutex_exit(&rp->r_statelock);
3425 }
3426 
3427 /*
3428  * The caller is responsible for freeing the returned string.
3429  */
3430 static char *
3431 nfs4_getsrvnames(mntinfo4_t *mi, size_t *len)
3432 {
3433 	servinfo4_t *svp;
3434 	char *srvnames;
3435 	char *namep;
3436 	size_t length;
3437 
3438 	/*
3439 	 * Calculate the length of the string required to hold all
3440 	 * of the server names plus either a comma or a null
3441 	 * character following each individual one.
3442 	 */
3443 	length = 0;
3444 	for (svp = mi->mi_servers; svp != NULL; svp = svp->sv_next) {
3445 		(void) nfs_rw_enter_sig(&svp->sv_lock, RW_READER, 0);
3446 		if (svp->sv_flags & SV4_NOTINUSE) {
3447 			nfs_rw_exit(&svp->sv_lock);
3448 			continue;
3449 		}
3450 		nfs_rw_exit(&svp->sv_lock);
3451 		length += svp->sv_hostnamelen;
3452 	}
3453 
3454 	srvnames = kmem_alloc(length, KM_SLEEP);
3455 
3456 	namep = srvnames;
3457 	for (svp = mi->mi_servers; svp != NULL; svp = svp->sv_next) {
3458 		(void) nfs_rw_enter_sig(&svp->sv_lock, RW_READER, 0);
3459 		if (svp->sv_flags & SV4_NOTINUSE) {
3460 			nfs_rw_exit(&svp->sv_lock);
3461 			continue;
3462 		}
3463 		nfs_rw_exit(&svp->sv_lock);
3464 		(void) strcpy(namep, svp->sv_hostname);
3465 		namep += svp->sv_hostnamelen - 1;
3466 		*namep++ = ',';
3467 	}
3468 	*--namep = '\0';
3469 
3470 	*len = length;
3471 
3472 	return (srvnames);
3473 }
3474 
3475 static void
3476 save_bseqid_rqst(nfs4_bseqid_entry_t *bsep, recov_info_t *recovp)
3477 {
3478 	nfs4_bseqid_entry_t *destp;
3479 
3480 	destp = kmem_alloc(sizeof (nfs4_bseqid_entry_t), KM_SLEEP);
3481 	recovp->rc_bseqid_rqst = destp;
3482 
3483 	if (bsep->bs_oop)
3484 		open_owner_hold(bsep->bs_oop);
3485 	destp->bs_oop = bsep->bs_oop;
3486 	if (bsep->bs_lop)
3487 		lock_owner_hold(bsep->bs_lop);
3488 	destp->bs_lop = bsep->bs_lop;
3489 	if (bsep->bs_vp)
3490 		VN_HOLD(bsep->bs_vp);
3491 	destp->bs_vp = bsep->bs_vp;
3492 	destp->bs_pid = bsep->bs_pid;
3493 	destp->bs_tag = bsep->bs_tag;
3494 	destp->bs_seqid = bsep->bs_seqid;
3495 }
3496 
3497 static void
3498 free_bseqid_rqst(nfs4_bseqid_entry_t *bsep)
3499 {
3500 	if (bsep->bs_oop)
3501 		open_owner_rele(bsep->bs_oop);
3502 	if (bsep->bs_lop)
3503 		lock_owner_rele(bsep->bs_lop);
3504 	if (bsep->bs_vp)
3505 		VN_RELE(bsep->bs_vp);
3506 	kmem_free(bsep, sizeof (nfs4_bseqid_entry_t));
3507 }
3508 
3509 /*
3510  * We don't actually fully recover from NFS4ERR_BAD_SEQID.  We
3511  * simply mark the open owner and open stream (if provided) as "bad".
3512  * Then future uses of these data structures will be limited to basically
3513  * just cleaning up the internal client state (no going OTW).
3514  *
3515  * The result of this is to return errors back to the app/usr when
3516  * we receive NFS4ERR_BAD_SEQID, but also allow future/new calls to
3517  * succeed so progress can be made.
3518  */
3519 void
3520 recov_bad_seqid(recov_info_t *recovp)
3521 {
3522 	mntinfo4_t		*mi = recovp->rc_mi;
3523 	nfs4_open_owner_t	*bad_oop;
3524 	nfs4_lock_owner_t	*bad_lop;
3525 	vnode_t			*vp;
3526 	rnode4_t		*rp = NULL;
3527 	pid_t			pid;
3528 	nfs4_bseqid_entry_t	*bsep, *tbsep;
3529 	int			error;
3530 
3531 	ASSERT(mi != NULL);
3532 	ASSERT(nfs_rw_lock_held(&mi->mi_recovlock, RW_WRITER));
3533 
3534 	mutex_enter(&mi->mi_lock);
3535 	bsep = list_head(&mi->mi_bseqid_list);
3536 	mutex_exit(&mi->mi_lock);
3537 
3538 	/*
3539 	 * Handle all the bad seqid entries on mi's list.
3540 	 */
3541 	while (bsep != NULL) {
3542 		bad_oop = bsep->bs_oop;
3543 		bad_lop = bsep->bs_lop;
3544 		vp = bsep->bs_vp;
3545 		pid = bsep->bs_pid;
3546 
3547 		NFS4_DEBUG(nfs4_client_recov_debug, (CE_NOTE,
3548 		    "recov_bad_seqid: mark oop %p lop %p as bad for "
3549 		    "vp %p tag %s pid %d: last good seqid %d for tag %s",
3550 		    (void *)bad_oop, (void *)bad_lop, (void *)vp,
3551 		    nfs4_ctags[bsep->bs_tag].ct_str, pid,
3552 		    bad_oop ?  bad_oop->oo_last_good_seqid : 0,
3553 		    bad_oop ? nfs4_ctags[bad_oop->oo_last_good_op].ct_str :
3554 		    nfs4_ctags[TAG_NONE].ct_str));
3555 
3556 		nfs4_queue_event(RE_BAD_SEQID, mi, NULL,
3557 		    0, vp, NULL, NFS4ERR_BAD_SEQID, NULL, pid, bsep->bs_tag,
3558 		    bad_oop ? bad_oop->oo_last_good_op : TAG_NONE,
3559 		    bsep->bs_seqid, bad_oop ? bad_oop->oo_last_good_seqid : 0);
3560 
3561 		if (bad_oop) {
3562 			/* essentially reset the open owner */
3563 			error = nfs4_start_open_seqid_sync(bad_oop, mi);
3564 			ASSERT(!error);	/* recov thread always succeeds */
3565 			bad_oop->oo_name = nfs4_get_new_oo_name();
3566 			bad_oop->oo_seqid = 0;
3567 			nfs4_end_open_seqid_sync(bad_oop);
3568 		}
3569 
3570 		if (bad_lop) {
3571 			mutex_enter(&bad_lop->lo_lock);
3572 			bad_lop->lo_flags |= NFS4_BAD_SEQID_LOCK;
3573 			mutex_exit(&bad_lop->lo_lock);
3574 
3575 			ASSERT(vp != NULL);
3576 			rp = VTOR4(vp);
3577 			mutex_enter(&rp->r_statelock);
3578 			rp->r_flags |= R4LODANGLERS;
3579 			mutex_exit(&rp->r_statelock);
3580 
3581 			nfs4_send_siglost(pid, mi, vp, TRUE,
3582 			    0, NFS4ERR_BAD_SEQID);
3583 		}
3584 
3585 		mutex_enter(&mi->mi_lock);
3586 		list_remove(&mi->mi_bseqid_list, bsep);
3587 		tbsep = bsep;
3588 		bsep = list_head(&mi->mi_bseqid_list);
3589 		mutex_exit(&mi->mi_lock);
3590 		free_bseqid_rqst(tbsep);
3591 	}
3592 
3593 	mutex_enter(&mi->mi_lock);
3594 	mi->mi_recovflags &= ~MI4R_BAD_SEQID;
3595 	mutex_exit(&mi->mi_lock);
3596 }
3597