xref: /titanic_52/usr/src/uts/common/fs/ufs/ufs_vfsops.c (revision a1b5e537933659371285214eae1db2603e6364b4)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License, Version 1.0 only
6  * (the "License").  You may not use this file except in compliance
7  * with the License.
8  *
9  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
10  * or http://www.opensolaris.org/os/licensing.
11  * See the License for the specific language governing permissions
12  * and limitations under the License.
13  *
14  * When distributing Covered Code, include this CDDL HEADER in each
15  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
16  * If applicable, add the following below this CDDL HEADER, with the
17  * fields enclosed by brackets "[]" replaced with your own identifying
18  * information: Portions Copyright [yyyy] [name of copyright owner]
19  *
20  * CDDL HEADER END
21  */
22 /*
23  * Copyright 2005 Sun Microsystems, Inc.  All rights reserved.
24  * Use is subject to license terms.
25  */
26 
27 /*	Copyright (c) 1983, 1984, 1985, 1986, 1987, 1988, 1989 AT&T	*/
28 /*	  All Rights Reserved  	*/
29 
30 /*
31  * University Copyright- Copyright (c) 1982, 1986, 1988
32  * The Regents of the University of California
33  * All Rights Reserved
34  *
35  * University Acknowledgment- Portions of this document are derived from
36  * software developed by the University of California, Berkeley, and its
37  * contributors.
38  */
39 
40 
41 #pragma ident	"%Z%%M%	%I%	%E% SMI"
42 
43 #include <sys/types.h>
44 #include <sys/t_lock.h>
45 #include <sys/param.h>
46 #include <sys/systm.h>
47 #include <sys/bitmap.h>
48 #include <sys/sysmacros.h>
49 #include <sys/kmem.h>
50 #include <sys/signal.h>
51 #include <sys/user.h>
52 #include <sys/proc.h>
53 #include <sys/disp.h>
54 #include <sys/buf.h>
55 #include <sys/pathname.h>
56 #include <sys/vfs.h>
57 #include <sys/vnode.h>
58 #include <sys/file.h>
59 #include <sys/atomic.h>
60 #include <sys/uio.h>
61 #include <sys/dkio.h>
62 #include <sys/cred.h>
63 #include <sys/conf.h>
64 #include <sys/dnlc.h>
65 #include <sys/kstat.h>
66 #include <sys/acl.h>
67 #include <sys/fs/ufs_fsdir.h>
68 #include <sys/fs/ufs_fs.h>
69 #include <sys/fs/ufs_inode.h>
70 #include <sys/fs/ufs_mount.h>
71 #include <sys/fs/ufs_acl.h>
72 #include <sys/fs/ufs_panic.h>
73 #include <sys/fs/ufs_bio.h>
74 #include <sys/fs/ufs_quota.h>
75 #include <sys/fs/ufs_log.h>
76 #undef NFS
77 #include <sys/statvfs.h>
78 #include <sys/mount.h>
79 #include <sys/mntent.h>
80 #include <sys/swap.h>
81 #include <sys/errno.h>
82 #include <sys/debug.h>
83 #include "fs/fs_subr.h"
84 #include <sys/cmn_err.h>
85 #include <sys/dnlc.h>
86 #include <sys/fssnap_if.h>
87 #include <sys/sunddi.h>
88 #include <sys/bootconf.h>
89 #include <sys/policy.h>
90 #include <sys/zone.h>
91 
92 /*
93  * This is the loadable module wrapper.
94  */
95 #include <sys/modctl.h>
96 
97 int			ufsfstype;
98 vfsops_t		*ufs_vfsops;
99 static int		ufsinit(int, char *);
100 static int		mountfs();
101 extern int		highbit();
102 extern struct instats	ins;
103 extern struct vnode *common_specvp(struct vnode *vp);
104 extern vfs_t		EIO_vfs;
105 
106 struct  dquot *dquot, *dquotNDQUOT;
107 
108 /*
109  * Cylinder group summary information handling tunable.
110  * This defines when these deltas get logged.
111  * If the number of cylinders in the file system is over the
112  * tunable then we log csum updates. Otherwise the updates are only
113  * done for performance on unmount. After a panic they can be
114  * quickly constructed during mounting. See ufs_construct_si()
115  * called from ufs_getsummaryinfo().
116  *
117  * This performance feature can of course be disabled by setting
118  * ufs_ncg_log to 0, and fully enabled by setting it to 0xffffffff.
119  */
120 #define	UFS_LOG_NCG_DEFAULT 10000
121 uint32_t ufs_ncg_log = UFS_LOG_NCG_DEFAULT;
122 
123 /*
124  * ufs_clean_root indicates whether the root fs went down cleanly
125  */
126 static int ufs_clean_root = 0;
127 
128 /*
129  * UFS Mount options table
130  */
131 static char *intr_cancel[] = { MNTOPT_NOINTR, NULL };
132 static char *nointr_cancel[] = { MNTOPT_INTR, NULL };
133 static char *forcedirectio_cancel[] = { MNTOPT_NOFORCEDIRECTIO, NULL };
134 static char *noforcedirectio_cancel[] = { MNTOPT_FORCEDIRECTIO, NULL };
135 static char *largefiles_cancel[] = { MNTOPT_NOLARGEFILES, NULL };
136 static char *nolargefiles_cancel[] = { MNTOPT_LARGEFILES, NULL };
137 static char *logging_cancel[] = { MNTOPT_NOLOGGING, NULL };
138 static char *nologging_cancel[] = { MNTOPT_LOGGING, NULL };
139 static char *xattr_cancel[] = { MNTOPT_NOXATTR, NULL };
140 static char *noxattr_cancel[] = { MNTOPT_XATTR, NULL };
141 static char *quota_cancel[] = { MNTOPT_NOQUOTA, NULL };
142 static char *noquota_cancel[] = { MNTOPT_QUOTA, NULL };
143 static char *dfratime_cancel[] = { MNTOPT_NODFRATIME, NULL };
144 static char *nodfratime_cancel[] = { MNTOPT_DFRATIME, NULL };
145 
146 static mntopt_t mntopts[] = {
147 /*
148  *	option name		cancel option	default arg	flags
149  *		ufs arg flag
150  */
151 	{ MNTOPT_INTR,		intr_cancel,	NULL,		MO_DEFAULT,
152 		(void *)0 },
153 	{ MNTOPT_NOINTR,	nointr_cancel,	NULL,		0,
154 		(void *)UFSMNT_NOINTR },
155 	{ MNTOPT_SYNCDIR,	NULL,		NULL,		0,
156 		(void *)UFSMNT_SYNCDIR },
157 	{ MNTOPT_FORCEDIRECTIO,	forcedirectio_cancel, NULL,	0,
158 		(void *)UFSMNT_FORCEDIRECTIO },
159 	{ MNTOPT_NOFORCEDIRECTIO, noforcedirectio_cancel, NULL, 0,
160 		(void *)UFSMNT_NOFORCEDIRECTIO },
161 	{ MNTOPT_NOSETSEC,	NULL,		NULL,		0,
162 		(void *)UFSMNT_NOSETSEC },
163 	{ MNTOPT_LARGEFILES,	largefiles_cancel, NULL,	MO_DEFAULT,
164 		(void *)UFSMNT_LARGEFILES },
165 	{ MNTOPT_NOLARGEFILES,	nolargefiles_cancel, NULL,	0,
166 		(void *)0 },
167 	{ MNTOPT_LOGGING,	logging_cancel, NULL,		MO_TAG,
168 		(void *)UFSMNT_LOGGING },
169 	{ MNTOPT_NOLOGGING,	nologging_cancel, NULL,
170 		MO_NODISPLAY|MO_DEFAULT|MO_TAG, (void *)0 },
171 	{ MNTOPT_QUOTA,		quota_cancel, NULL,		MO_IGNORE,
172 		(void *)0 },
173 	{ MNTOPT_NOQUOTA,	noquota_cancel,	NULL,
174 		MO_NODISPLAY|MO_DEFAULT, (void *)0 },
175 	{ MNTOPT_GLOBAL,	NULL,		NULL,		0,
176 		(void *)0 },
177 	{ MNTOPT_XATTR,	xattr_cancel,		NULL,		MO_DEFAULT,
178 		(void *)0 },
179 	{ MNTOPT_NOXATTR,	noxattr_cancel,		NULL,		0,
180 		(void *)0 },
181 	{ MNTOPT_NOATIME,	NULL,		NULL,		0,
182 		(void *)UFSMNT_NOATIME },
183 	{ MNTOPT_DFRATIME,	dfratime_cancel, NULL,		0,
184 		(void *)0 },
185 	{ MNTOPT_NODFRATIME,	nodfratime_cancel, NULL,
186 		MO_NODISPLAY|MO_DEFAULT, (void *)UFSMNT_NODFRATIME },
187 	{ MNTOPT_ONERROR,	NULL,		UFSMNT_ONERROR_PANIC_STR,
188 		MO_DEFAULT|MO_HASVALUE,	(void *)0 },
189 };
190 
191 static mntopts_t ufs_mntopts = {
192 	sizeof (mntopts) / sizeof (mntopt_t),
193 	mntopts
194 };
195 
196 static vfsdef_t vfw = {
197 	VFSDEF_VERSION,
198 	"ufs",
199 	ufsinit,
200 	VSW_HASPROTO|VSW_CANREMOUNT,
201 	&ufs_mntopts
202 };
203 
204 /*
205  * Module linkage information for the kernel.
206  */
207 extern struct mod_ops mod_fsops;
208 
209 static struct modlfs modlfs = {
210 	&mod_fsops, "filesystem for ufs", &vfw
211 };
212 
213 static struct modlinkage modlinkage = {
214 	MODREV_1, (void *)&modlfs, NULL
215 };
216 
217 /*
218  * An attempt has been made to make this module unloadable.  In order to
219  * test it, we need a system in which the root fs is NOT ufs.  THIS HAS NOT
220  * BEEN DONE
221  */
222 
223 extern kstat_t *ufs_inode_kstat;
224 extern uint_t ufs_lockfs_key;
225 extern void ufs_lockfs_tsd_destructor(void *);
226 extern uint_t bypass_snapshot_throttle_key;
227 
228 int
229 _init(void)
230 {
231 	/*
232 	 * Create an index into the per thread array so that any thread doing
233 	 * VOP will have a lockfs mark on it.
234 	 */
235 	tsd_create(&ufs_lockfs_key, ufs_lockfs_tsd_destructor);
236 	tsd_create(&bypass_snapshot_throttle_key, NULL);
237 	return (mod_install(&modlinkage));
238 }
239 
240 int
241 _fini(void)
242 {
243 	return (EBUSY);
244 }
245 
246 int
247 _info(struct modinfo *modinfop)
248 {
249 	return (mod_info(&modlinkage, modinfop));
250 }
251 
252 extern struct vnode *makespecvp(dev_t dev, vtype_t type);
253 
254 extern kmutex_t	ufs_scan_lock;
255 
256 static int mountfs(struct vfs *, enum whymountroot, struct vnode *, char *,
257 		struct cred *, int, void *, int);
258 
259 
260 static int
261 ufs_mount(struct vfs *vfsp, struct vnode *mvp, struct mounta *uap,
262 	struct cred *cr)
263 
264 {
265 	char *data = uap->dataptr;
266 	int datalen = uap->datalen;
267 	dev_t dev;
268 	struct vnode *bvp;
269 	struct pathname dpn;
270 	int error;
271 	enum whymountroot why = ROOT_INIT;
272 	struct ufs_args args;
273 	int oflag, aflag;
274 	int fromspace = (uap->flags & MS_SYSSPACE) ?
275 	    UIO_SYSSPACE : UIO_USERSPACE;
276 
277 	if ((error = secpolicy_fs_mount(cr, mvp, vfsp)) != 0)
278 		return (error);
279 
280 	if (mvp->v_type != VDIR)
281 		return (ENOTDIR);
282 
283 	mutex_enter(&mvp->v_lock);
284 	if ((uap->flags & MS_REMOUNT) == 0 &&
285 	    (uap->flags & MS_OVERLAY) == 0 &&
286 	    (mvp->v_count != 1 || (mvp->v_flag & VROOT))) {
287 		mutex_exit(&mvp->v_lock);
288 		return (EBUSY);
289 	}
290 	mutex_exit(&mvp->v_lock);
291 
292 	/*
293 	 * Get arguments
294 	 */
295 	bzero(&args, sizeof (args));
296 	if ((uap->flags & MS_DATA) && data != NULL && datalen != 0) {
297 		int copy_result = 0;
298 
299 		if (datalen > sizeof (args))
300 			return (EINVAL);
301 		if (uap->flags & MS_SYSSPACE)
302 			bcopy(data, &args, datalen);
303 		else
304 			copy_result = copyin(data, &args, datalen);
305 		if (copy_result)
306 			return (EFAULT);
307 		datalen = sizeof (struct ufs_args);
308 	} else {
309 		datalen = 0;
310 	}
311 	/*
312 	 * Read in the mount point pathname
313 	 * (so we can record the directory the file system was last mounted on).
314 	 */
315 	if (error = pn_get(uap->dir, fromspace, &dpn))
316 		return (error);
317 
318 	/*
319 	 * Resolve path name of special file being mounted.
320 	 */
321 	if (error = lookupname(uap->spec, fromspace, FOLLOW, NULL, &bvp)) {
322 		pn_free(&dpn);
323 		return (error);
324 	}
325 	if (bvp->v_type != VBLK) {
326 		VN_RELE(bvp);
327 		pn_free(&dpn);
328 		return (ENOTBLK);
329 	}
330 	dev = bvp->v_rdev;
331 	if (getmajor(dev) >= devcnt) {
332 		pn_free(&dpn);
333 		VN_RELE(bvp);
334 		return (ENXIO);
335 	}
336 	if (uap->flags & MS_REMOUNT)
337 		why = ROOT_REMOUNT;
338 
339 	/*
340 	 * In SunCluster, requests to a global device are satisfied by
341 	 * a local device. We substitute the global pxfs node with a
342 	 * local spec node here.
343 	 */
344 	if (IS_PXFSVP(bvp)) {
345 		VN_RELE(bvp);
346 		bvp = makespecvp(dev, VBLK);
347 	}
348 
349 	/*
350 	 * Open block device mounted on.  We need this to
351 	 * check whether the caller has sufficient rights to
352 	 * access the device in question.
353 	 * When bio is fixed for vnodes this can all be vnode
354 	 * operations.
355 	 */
356 	if ((vfsp->vfs_flag & VFS_RDONLY) != 0 ||
357 	    (uap->flags & MS_RDONLY) != 0) {
358 		oflag = FREAD;
359 		aflag = VREAD;
360 	} else {
361 		oflag = FREAD | FWRITE;
362 		aflag = VREAD | VWRITE;
363 	}
364 	if ((error = VOP_ACCESS(bvp, aflag, 0, cr)) != 0 ||
365 	    (error = secpolicy_spec_open(cr, bvp, oflag)) != 0) {
366 		pn_free(&dpn);
367 		VN_RELE(bvp);
368 		return (error);
369 	}
370 
371 	/*
372 	 * Ensure that this device isn't already mounted or in progress on a
373 	 * mount unless this is a REMOUNT request or we are told to suppress
374 	 * mount checks. Global mounts require special handling.
375 	 */
376 	if ((uap->flags & MS_NOCHECK) == 0) {
377 		if ((uap->flags & MS_GLOBAL) == 0 &&
378 		    vfs_devmounting(dev, vfsp)) {
379 			pn_free(&dpn);
380 			VN_RELE(bvp);
381 			return (EBUSY);
382 		}
383 		if (vfs_devismounted(dev)) {
384 			if ((uap->flags & MS_REMOUNT) == 0) {
385 				pn_free(&dpn);
386 				VN_RELE(bvp);
387 				return (EBUSY);
388 			}
389 		}
390 	}
391 
392 	/*
393 	 * If the device is a tape, mount it read only
394 	 */
395 	if (devopsp[getmajor(dev)]->devo_cb_ops->cb_flag & D_TAPE) {
396 		vfsp->vfs_flag |= VFS_RDONLY;
397 		vfs_setmntopt(vfsp, MNTOPT_RO, NULL, 0);
398 	}
399 	if (uap->flags & MS_RDONLY)
400 		vfsp->vfs_flag |= VFS_RDONLY;
401 
402 	/*
403 	 * Mount the filesystem, free the device vnode on error.
404 	 */
405 	error = mountfs(vfsp, why, bvp, dpn.pn_path, cr, 0, &args, datalen);
406 	pn_free(&dpn);
407 	if (error) {
408 		VN_RELE(bvp);
409 	}
410 	return (error);
411 }
412 /*
413  * Mount root file system.
414  * "why" is ROOT_INIT on initial call ROOT_REMOUNT if called to
415  * remount the root file system, and ROOT_UNMOUNT if called to
416  * unmount the root (e.g., as part of a system shutdown).
417  *
418  * XXX - this may be partially machine-dependent; it, along with the VFS_SWAPVP
419  * operation, goes along with auto-configuration.  A mechanism should be
420  * provided by which machine-INdependent code in the kernel can say "get me the
421  * right root file system" and "get me the right initial swap area", and have
422  * that done in what may well be a machine-dependent fashion.
423  * Unfortunately, it is also file-system-type dependent (NFS gets it via
424  * bootparams calls, UFS gets it from various and sundry machine-dependent
425  * mechanisms, as SPECFS does for swap).
426  */
427 static int
428 ufs_mountroot(struct vfs *vfsp, enum whymountroot why)
429 {
430 	struct fs *fsp;
431 	int error;
432 	static int ufsrootdone = 0;
433 	dev_t rootdev;
434 	struct vnode *vp;
435 	struct vnode *devvp = 0;
436 	int ovflags;
437 	int doclkset;
438 	ufsvfs_t *ufsvfsp;
439 
440 	if (why == ROOT_INIT) {
441 		if (ufsrootdone++)
442 			return (EBUSY);
443 		rootdev = getrootdev();
444 		if (rootdev == (dev_t)NODEV)
445 			return (ENODEV);
446 		vfsp->vfs_dev = rootdev;
447 		vfsp->vfs_flag |= VFS_RDONLY;
448 	} else if (why == ROOT_REMOUNT) {
449 		vp = ((struct ufsvfs *)vfsp->vfs_data)->vfs_devvp;
450 		(void) dnlc_purge_vfsp(vfsp, 0);
451 		vp = common_specvp(vp);
452 		(void) VOP_PUTPAGE(vp, (offset_t)0, (size_t)0, B_INVAL, CRED());
453 		(void) bfinval(vfsp->vfs_dev, 0);
454 		fsp = getfs(vfsp);
455 
456 		ovflags = vfsp->vfs_flag;
457 		vfsp->vfs_flag &= ~VFS_RDONLY;
458 		vfsp->vfs_flag |= VFS_REMOUNT;
459 		rootdev = vfsp->vfs_dev;
460 	} else if (why == ROOT_UNMOUNT) {
461 		if (vfs_lock(vfsp) == 0) {
462 			(void) ufs_flush(vfsp);
463 			/*
464 			 * Mark the log as fully rolled
465 			 */
466 			ufsvfsp = (ufsvfs_t *)vfsp->vfs_data;
467 			fsp = ufsvfsp->vfs_fs;
468 			if (TRANS_ISTRANS(ufsvfsp) &&
469 			    !TRANS_ISERROR(ufsvfsp) &&
470 			    (fsp->fs_rolled == FS_NEED_ROLL)) {
471 				ml_unit_t *ul = ufsvfsp->vfs_log;
472 
473 				error = ufs_putsummaryinfo(ul->un_dev,
474 				    ufsvfsp, fsp);
475 				if (error == 0) {
476 					fsp->fs_rolled = FS_ALL_ROLLED;
477 					UFS_BWRITE2(NULL, ufsvfsp->vfs_bufp);
478 				}
479 			}
480 			vfs_unlock(vfsp);
481 		} else {
482 			ufs_update(0);
483 		}
484 
485 		vp = ((struct ufsvfs *)vfsp->vfs_data)->vfs_devvp;
486 		(void) VOP_CLOSE(vp, FREAD|FWRITE, 1,
487 			(offset_t)0, CRED());
488 		return (0);
489 	}
490 	error = vfs_lock(vfsp);
491 	if (error)
492 		return (error);
493 
494 	devvp = makespecvp(rootdev, VBLK);
495 
496 	/* If RO media, don't call clkset() (see below) */
497 	doclkset = 1;
498 	if (why == ROOT_INIT) {
499 		error = VOP_OPEN(&devvp, FREAD|FWRITE, CRED());
500 		if (error == 0) {
501 			(void) VOP_CLOSE(devvp, FREAD|FWRITE, 1,
502 				(offset_t)0, CRED());
503 		} else {
504 			doclkset = 0;
505 		}
506 	}
507 
508 	error = mountfs(vfsp, why, devvp, "/", CRED(), 1, NULL, 0);
509 	/*
510 	 * XXX - assumes root device is not indirect, because we don't set
511 	 * rootvp.  Is rootvp used for anything?  If so, make another arg
512 	 * to mountfs.
513 	 */
514 	if (error) {
515 		vfs_unlock(vfsp);
516 		if (why == ROOT_REMOUNT)
517 			vfsp->vfs_flag = ovflags;
518 		if (rootvp) {
519 			VN_RELE(rootvp);
520 			rootvp = (struct vnode *)0;
521 		}
522 		VN_RELE(devvp);
523 		return (error);
524 	}
525 	if (why == ROOT_INIT)
526 		vfs_add((struct vnode *)0, vfsp,
527 		    (vfsp->vfs_flag & VFS_RDONLY) ? MS_RDONLY : 0);
528 	vfs_unlock(vfsp);
529 	fsp = getfs(vfsp);
530 	clkset(doclkset ? fsp->fs_time : -1);
531 	ufsvfsp = (ufsvfs_t *)vfsp->vfs_data;
532 	if (ufsvfsp->vfs_log) {
533 		vfs_setmntopt(vfsp, MNTOPT_LOGGING, NULL, 0);
534 	}
535 	return (0);
536 }
537 
538 static int
539 remountfs(struct vfs *vfsp, dev_t dev, void *raw_argsp, int args_len)
540 {
541 	struct ufsvfs *ufsvfsp = (struct ufsvfs *)vfsp->vfs_data;
542 	struct ulockfs *ulp = &ufsvfsp->vfs_ulockfs;
543 	struct buf *bp = ufsvfsp->vfs_bufp;
544 	struct fs *fsp = (struct fs *)bp->b_un.b_addr;
545 	struct fs *fspt;
546 	struct buf *tpt = 0;
547 	int error = 0;
548 	int flags = 0;
549 
550 	if (args_len == sizeof (struct ufs_args) && raw_argsp)
551 		flags = ((struct ufs_args *)raw_argsp)->flags;
552 
553 	/* cannot remount to RDONLY */
554 	if (vfsp->vfs_flag & VFS_RDONLY)
555 		return (EINVAL);
556 
557 	/* whoops, wrong dev */
558 	if (vfsp->vfs_dev != dev)
559 		return (EINVAL);
560 
561 	/*
562 	 * synchronize w/ufs ioctls
563 	 */
564 	mutex_enter(&ulp->ul_lock);
565 	atomic_add_long(&ufs_quiesce_pend, 1);
566 
567 	/*
568 	 * reset options
569 	 */
570 	ufsvfsp->vfs_nointr  = flags & UFSMNT_NOINTR;
571 	ufsvfsp->vfs_syncdir = flags & UFSMNT_SYNCDIR;
572 	ufsvfsp->vfs_nosetsec = flags & UFSMNT_NOSETSEC;
573 	ufsvfsp->vfs_noatime = flags & UFSMNT_NOATIME;
574 	if ((flags & UFSMNT_NODFRATIME) || ufsvfsp->vfs_noatime)
575 		ufsvfsp->vfs_dfritime &= ~UFS_DFRATIME;
576 	else	/* dfratime, default behavior */
577 		ufsvfsp->vfs_dfritime |= UFS_DFRATIME;
578 	if (flags & UFSMNT_FORCEDIRECTIO)
579 		ufsvfsp->vfs_forcedirectio = 1;
580 	else	/* default is no direct I/O */
581 		ufsvfsp->vfs_forcedirectio = 0;
582 	ufsvfsp->vfs_iotstamp = lbolt;
583 
584 	/*
585 	 * set largefiles flag in ufsvfs equal to the
586 	 * value passed in by the mount command. If
587 	 * it is "nolargefiles", and the flag is set
588 	 * in the superblock, the mount fails.
589 	 */
590 	if (!(flags & UFSMNT_LARGEFILES)) {  /* "nolargefiles" */
591 		if (fsp->fs_flags & FSLARGEFILES) {
592 			error = EFBIG;
593 			goto remounterr;
594 		}
595 		ufsvfsp->vfs_lfflags &= ~UFS_LARGEFILES;
596 	} else	/* "largefiles" */
597 		ufsvfsp->vfs_lfflags |= UFS_LARGEFILES;
598 	/*
599 	 * read/write to read/write; all done
600 	 */
601 	if (fsp->fs_ronly == 0)
602 		goto remounterr;
603 
604 	/*
605 	 * fix-on-panic assumes RO->RW remount implies system-critical fs
606 	 * if it is shortly after boot; so, don't attempt to lock and fix
607 	 * (unless the user explicitly asked for another action on error)
608 	 * XXX UFSMNT_ONERROR_RDONLY rather than UFSMNT_ONERROR_PANIC
609 	 */
610 #define	BOOT_TIME_LIMIT	(180*hz)
611 	if (!(flags & UFSMNT_ONERROR_FLGMASK) && lbolt < BOOT_TIME_LIMIT) {
612 		cmn_err(CE_WARN, "%s is required to be mounted onerror=%s",
613 			ufsvfsp->vfs_fs->fs_fsmnt, UFSMNT_ONERROR_PANIC_STR);
614 		flags |= UFSMNT_ONERROR_PANIC;
615 	}
616 
617 	if ((error = ufsfx_mount(ufsvfsp, flags)) != 0)
618 		goto remounterr;
619 
620 	/*
621 	 * Lock the file system and flush stuff from memory
622 	 */
623 	error = ufs_quiesce(ulp);
624 	if (error)
625 		goto remounterr;
626 
627 	/*
628 	 * We don't need to call the expensive ufs_flush when going from
629 	 * read only to read/write, except if the root fs didn't come
630 	 * down cleanly.
631 	 */
632 	if ((ufsvfsp->vfs_devvp == rootvp) && !ufs_clean_root) {
633 		error = ufs_flush(vfsp);
634 		if (error) {
635 			goto remounterr;
636 		}
637 	}
638 
639 	tpt = UFS_BREAD(ufsvfsp, ufsvfsp->vfs_dev, SBLOCK, SBSIZE);
640 	if (tpt->b_flags & B_ERROR) {
641 		error = EIO;
642 		goto remounterr;
643 	}
644 	fspt = (struct fs *)tpt->b_un.b_addr;
645 	if (((fspt->fs_magic != FS_MAGIC) &&
646 	    (fspt->fs_magic != MTB_UFS_MAGIC)) ||
647 	    (fspt->fs_magic == MTB_UFS_MAGIC &&
648 		(fspt->fs_version > MTB_UFS_VERSION_1 ||
649 		fspt->fs_version < MTB_UFS_VERSION_MIN)) ||
650 	    fspt->fs_bsize > MAXBSIZE || fspt->fs_frag > MAXFRAG ||
651 	    fspt->fs_bsize < sizeof (struct fs) || fspt->fs_bsize < PAGESIZE) {
652 		tpt->b_flags |= B_STALE | B_AGE;
653 		error = EINVAL;
654 		goto remounterr;
655 	}
656 
657 	if (ufsvfsp->vfs_log && (ufsvfsp->vfs_log->un_flags & LDL_NOROLL)) {
658 		ufsvfsp->vfs_log->un_flags &= ~LDL_NOROLL;
659 		logmap_start_roll(ufsvfsp->vfs_log);
660 	}
661 
662 	if (TRANS_ISERROR(ufsvfsp))
663 		goto remounterr;
664 	TRANS_DOMATAMAP(ufsvfsp);
665 
666 	if ((fspt->fs_state + fspt->fs_time == FSOKAY) &&
667 	    fspt->fs_clean == FSLOG && !TRANS_ISTRANS(ufsvfsp)) {
668 		ufsvfsp->vfs_log = NULL;
669 		ufsvfsp->vfs_domatamap = 0;
670 		error = ENOSPC;
671 		goto remounterr;
672 	}
673 
674 	if (fspt->fs_state + fspt->fs_time == FSOKAY &&
675 	    (fspt->fs_clean == FSCLEAN ||
676 	    fspt->fs_clean == FSSTABLE ||
677 	    fspt->fs_clean == FSLOG)) {
678 
679 		/*
680 		 * Ensure that ufs_getsummaryinfo doesn't reconstruct
681 		 * the summary info.
682 		 */
683 		error = ufs_getsummaryinfo(vfsp->vfs_dev, ufsvfsp, fspt);
684 		if (error)
685 			goto remounterr;
686 
687 		/* preserve mount name */
688 		(void) strncpy(fspt->fs_fsmnt, fsp->fs_fsmnt, MAXMNTLEN);
689 		/* free the old cg space */
690 		kmem_free(fsp->fs_u.fs_csp, fsp->fs_cssize);
691 		/* switch in the new superblock */
692 		fspt->fs_rolled = FS_NEED_ROLL;
693 		bcopy(tpt->b_un.b_addr, bp->b_un.b_addr, fspt->fs_sbsize);
694 
695 		fsp->fs_clean = FSSTABLE;
696 	} /* superblock updated in memory */
697 	tpt->b_flags |= B_STALE | B_AGE;
698 	brelse(tpt);
699 	tpt = 0;
700 
701 	if (fsp->fs_clean != FSSTABLE) {
702 		error = ENOSPC;
703 		goto remounterr;
704 	}
705 
706 
707 	if (TRANS_ISTRANS(ufsvfsp)) {
708 		fsp->fs_clean = FSLOG;
709 		ufsvfsp->vfs_dio = 0;
710 	} else
711 		if (ufsvfsp->vfs_dio)
712 			fsp->fs_clean = FSSUSPEND;
713 
714 	TRANS_MATA_MOUNT(ufsvfsp);
715 
716 	fsp->fs_fmod = 0;
717 	fsp->fs_ronly = 0;
718 
719 	atomic_add_long(&ufs_quiesce_pend, -1);
720 	cv_broadcast(&ulp->ul_cv);
721 	mutex_exit(&ulp->ul_lock);
722 
723 	if (TRANS_ISTRANS(ufsvfsp)) {
724 
725 		/*
726 		 * start the delete thread
727 		 */
728 		ufs_thread_start(&ufsvfsp->vfs_delete, ufs_thread_delete, vfsp);
729 
730 		/*
731 		 * start the reclaim thread
732 		 */
733 		if (fsp->fs_reclaim & (FS_RECLAIM|FS_RECLAIMING)) {
734 			fsp->fs_reclaim &= ~FS_RECLAIM;
735 			fsp->fs_reclaim |=  FS_RECLAIMING;
736 			ufs_thread_start(&ufsvfsp->vfs_reclaim,
737 				ufs_thread_reclaim, vfsp);
738 		}
739 	}
740 
741 	TRANS_SBWRITE(ufsvfsp, TOP_MOUNT);
742 
743 	return (0);
744 
745 remounterr:
746 	if (tpt)
747 		brelse(tpt);
748 	atomic_add_long(&ufs_quiesce_pend, -1);
749 	cv_broadcast(&ulp->ul_cv);
750 	mutex_exit(&ulp->ul_lock);
751 	return (error);
752 }
753 
754 /*
755  * If the device maxtransfer size is not available, we use ufs_maxmaxphys
756  * along with the system value for maxphys to determine the value for
757  * maxtransfer.
758  */
759 int ufs_maxmaxphys = (1024 * 1024);
760 
761 #include <sys/ddi.h>		/* for delay(9f) */
762 
763 int ufs_mount_error_delay = 20;	/* default to 20ms */
764 int ufs_mount_timeout = 60;	/* default to 1 minute */
765 
766 static int
767 mountfs(struct vfs *vfsp, enum whymountroot why, struct vnode *devvp,
768 	char *path, cred_t *cr, int isroot, void *raw_argsp, int args_len)
769 {
770 	dev_t dev = devvp->v_rdev;
771 	struct fs *fsp;
772 	struct ufsvfs *ufsvfsp = 0;
773 	struct buf *bp = 0;
774 	struct buf *tp = 0;
775 	struct dk_cinfo ci;
776 	int error = 0;
777 	size_t len;
778 	int needclose = 0;
779 	int needtrans = 0;
780 	struct inode *rip;
781 	struct vnode *rvp = NULL;
782 	int flags = 0;
783 	kmutex_t *ihm;
784 	int elapsed;
785 	int status;
786 	extern	int	maxphys;
787 
788 	if (args_len == sizeof (struct ufs_args) && raw_argsp)
789 		flags = ((struct ufs_args *)raw_argsp)->flags;
790 
791 	ASSERT(vfs_lock_held(vfsp));
792 
793 	if (why == ROOT_INIT) {
794 		/*
795 		 * Open block device mounted on.
796 		 * When bio is fixed for vnodes this can all be vnode
797 		 * operations.
798 		 */
799 		error = VOP_OPEN(&devvp,
800 		    (vfsp->vfs_flag & VFS_RDONLY) ? FREAD : FREAD|FWRITE, cr);
801 		if (error)
802 			goto out;
803 		needclose = 1;
804 
805 		/*
806 		 * Refuse to go any further if this
807 		 * device is being used for swapping.
808 		 */
809 		if (IS_SWAPVP(devvp)) {
810 			error = EBUSY;
811 			goto out;
812 		}
813 	}
814 
815 	/*
816 	 * check for dev already mounted on
817 	 */
818 	if (vfsp->vfs_flag & VFS_REMOUNT) {
819 		error = remountfs(vfsp, dev, raw_argsp, args_len);
820 		if (error == 0)
821 			VN_RELE(devvp);
822 		return (error);
823 	}
824 
825 	ASSERT(devvp != 0);
826 
827 	/*
828 	 * Flush back any dirty pages on the block device to
829 	 * try and keep the buffer cache in sync with the page
830 	 * cache if someone is trying to use block devices when
831 	 * they really should be using the raw device.
832 	 */
833 	(void) VOP_PUTPAGE(common_specvp(devvp), (offset_t)0,
834 	    (size_t)0, B_INVAL, cr);
835 
836 	/*
837 	 * read in superblock
838 	 */
839 	ufsvfsp = kmem_zalloc(sizeof (struct ufsvfs), KM_SLEEP);
840 	tp = UFS_BREAD(ufsvfsp, dev, SBLOCK, SBSIZE);
841 	if (tp->b_flags & B_ERROR)
842 		goto out;
843 	fsp = (struct fs *)tp->b_un.b_addr;
844 #ifdef _LP64
845 	if ((fsp->fs_magic != FS_MAGIC) && (fsp->fs_magic != MTB_UFS_MAGIC)) {
846 		if (why == ROOT_INIT) {
847 			cmn_err(CE_NOTE,
848 			    "mount: not a UFS magic number (0x%x)",
849 			    fsp->fs_magic);
850 		}
851 		error = EINVAL;
852 		goto out;
853 	}
854 
855 	if ((fsp->fs_magic == MTB_UFS_MAGIC) &&
856 	    (fsp->fs_version > MTB_UFS_VERSION_1 ||
857 	    fsp->fs_version < MTB_UFS_VERSION_MIN)) {
858 		cmn_err(CE_NOTE,
859 		    "mount: unrecognized version of UFS on-disk format: %d",
860 		    fsp->fs_version);
861 		error = EINVAL;
862 		goto out;
863 	}
864 
865 #else
866 	if (fsp->fs_magic == MTB_UFS_MAGIC) {
867 		if (fsp->fs_version > MTB_UFS_VERSION_1 ||
868 		    fsp->fs_version < MTB_UFS_VERSION_MIN) {
869 			cmn_err(CE_NOTE,
870 			    "mount: unrecognized version of UFS"
871 			    " on-disk format: %d", fsp->fs_version);
872 			error = EINVAL;
873 			goto out;
874 		}
875 
876 		/*
877 		 * Find the size of the device in sectors.  If the
878 		 * the size in sectors is greater than INT_MAX, it's
879 		 * a multi-terabyte file system, which can't be
880 		 * mounted by a 32-bit kernel.  We can't use the
881 		 * fsbtodb() macro in the next line because the macro
882 		 * casts the intermediate values to daddr_t, which is
883 		 * a 32-bit quantity in a 32-bit kernel.  Here we
884 		 * really do need the intermediate values to be held
885 		 * in 64-bit quantities because we're checking for
886 		 * overflow of a 32-bit field.
887 		 */
888 		if ((((diskaddr_t)(fsp->fs_size)) << fsp->fs_fsbtodb)
889 		    > INT_MAX) {
890 			cmn_err(CE_NOTE,
891 			    "mount: multi-terabyte UFS cannot be"
892 			    " mounted by a 32-bit kernel");
893 			error = EINVAL;
894 			goto out;
895 		}
896 
897 	} else if (fsp->fs_magic != FS_MAGIC) {
898 		cmn_err(CE_NOTE,
899 		    "mount: not a UFS magic number (0x%x)", fsp->fs_magic);
900 		error = EINVAL;
901 		goto out;
902 	}
903 #endif /* _LP64 */
904 
905 	if (fsp->fs_bsize > MAXBSIZE || fsp->fs_frag > MAXFRAG ||
906 	    fsp->fs_bsize < sizeof (struct fs) || fsp->fs_bsize < PAGESIZE) {
907 		error = EINVAL;	/* also needs translation */
908 		goto out;
909 	}
910 
911 	/*
912 	 * Allocate VFS private data.
913 	 */
914 	vfsp->vfs_bcount = 0;
915 	vfsp->vfs_data = (caddr_t)ufsvfsp;
916 	vfsp->vfs_fstype = ufsfstype;
917 	vfsp->vfs_dev = dev;
918 	vfsp->vfs_flag |= VFS_NOTRUNC;
919 	vfs_make_fsid(&vfsp->vfs_fsid, dev, ufsfstype);
920 	ufsvfsp->vfs_devvp = devvp;
921 
922 	/*
923 	 * Cross-link with vfs and add to instance list.
924 	 */
925 	ufsvfsp->vfs_vfs = vfsp;
926 	ufs_vfs_add(ufsvfsp);
927 
928 	ufsvfsp->vfs_dev = dev;
929 	ufsvfsp->vfs_bufp = tp;
930 
931 	ufsvfsp->vfs_dirsize = INODESIZE + (4 * ALLOCSIZE) + fsp->fs_fsize;
932 	ufsvfsp->vfs_minfrags = (int)((int64_t)fsp->fs_dsize *
933 							fsp->fs_minfree / 100);
934 	/*
935 	 * if mount allows largefiles, indicate so in ufsvfs
936 	 */
937 	if (flags & UFSMNT_LARGEFILES)
938 		ufsvfsp->vfs_lfflags |= UFS_LARGEFILES;
939 	/*
940 	 * Initialize threads
941 	 */
942 	ufs_thread_init(&ufsvfsp->vfs_delete, 1);
943 	ufs_thread_init(&ufsvfsp->vfs_reclaim, 0);
944 
945 	/*
946 	 * Chicken and egg problem. The superblock may have deltas
947 	 * in the log.  So after the log is scanned we reread the
948 	 * superblock. We guarantee that the fields needed to
949 	 * scan the log will not be in the log.
950 	 */
951 	if (fsp->fs_logbno && fsp->fs_clean == FSLOG &&
952 	    (fsp->fs_state + fsp->fs_time == FSOKAY)) {
953 		error = lufs_snarf(ufsvfsp, fsp, (vfsp->vfs_flag & VFS_RDONLY));
954 		if (error) {
955 			/*
956 			 * Allow a ro mount to continue even if the
957 			 * log cannot be processed - yet.
958 			 */
959 			if (!(vfsp->vfs_flag & VFS_RDONLY)) {
960 				cmn_err(CE_WARN, "Error accessing ufs "
961 					"log for %s; Please run fsck(1M)",
962 					path);
963 				goto out;
964 			}
965 		}
966 		tp->b_flags |= (B_AGE | B_STALE);
967 		brelse(tp);
968 		tp = UFS_BREAD(ufsvfsp, dev, SBLOCK, SBSIZE);
969 		fsp = (struct fs *)tp->b_un.b_addr;
970 		ufsvfsp->vfs_bufp = tp;
971 		if (tp->b_flags & B_ERROR)
972 			goto out;
973 	}
974 
975 	/*
976 	 * Set logging mounted flag used by lockfs
977 	 */
978 	ufsvfsp->vfs_validfs = UT_MOUNTED;
979 
980 	/*
981 	 * Copy the super block into a buffer in its native size.
982 	 * Use ngeteblk to allocate the buffer
983 	 */
984 	bp = ngeteblk(fsp->fs_bsize);
985 	ufsvfsp->vfs_bufp = bp;
986 	bp->b_edev = dev;
987 	bp->b_dev = cmpdev(dev);
988 	bp->b_blkno = SBLOCK;
989 	bp->b_bcount = fsp->fs_sbsize;
990 	bcopy(tp->b_un.b_addr, bp->b_un.b_addr, fsp->fs_sbsize);
991 	tp->b_flags |= B_STALE | B_AGE;
992 	brelse(tp);
993 	tp = 0;
994 
995 	fsp = (struct fs *)bp->b_un.b_addr;
996 	/*
997 	 * Mount fails if superblock flag indicates presence of large
998 	 * files and filesystem is attempted to be mounted 'nolargefiles'.
999 	 * The exception is for a read only mount of root, which we
1000 	 * always want to succeed, so fsck can fix potential problems.
1001 	 * The assumption is that we will remount root at some point,
1002 	 * and the remount will enforce the mount option.
1003 	 */
1004 	if (!(isroot & (vfsp->vfs_flag & VFS_RDONLY)) &&
1005 	    (fsp->fs_flags & FSLARGEFILES) &&
1006 	    !(flags & UFSMNT_LARGEFILES)) {
1007 		error = EFBIG;
1008 		goto out;
1009 	}
1010 
1011 	if (vfsp->vfs_flag & VFS_RDONLY) {
1012 		fsp->fs_ronly = 1;
1013 		fsp->fs_fmod = 0;
1014 		if (((fsp->fs_state + fsp->fs_time) == FSOKAY) &&
1015 		    ((fsp->fs_clean == FSCLEAN) ||
1016 		    (fsp->fs_clean == FSSTABLE) ||
1017 		    (fsp->fs_clean == FSLOG))) {
1018 			if (isroot) {
1019 				if (fsp->fs_clean == FSLOG) {
1020 					if (fsp->fs_rolled == FS_ALL_ROLLED) {
1021 						ufs_clean_root = 1;
1022 					}
1023 				} else {
1024 					ufs_clean_root = 1;
1025 				}
1026 			}
1027 			fsp->fs_clean = FSSTABLE;
1028 		} else {
1029 			fsp->fs_clean = FSBAD;
1030 		}
1031 	} else {
1032 
1033 		fsp->fs_fmod = 0;
1034 		fsp->fs_ronly = 0;
1035 
1036 		TRANS_DOMATAMAP(ufsvfsp);
1037 
1038 		if ((TRANS_ISERROR(ufsvfsp)) ||
1039 		    (((fsp->fs_state + fsp->fs_time) == FSOKAY) &&
1040 			fsp->fs_clean == FSLOG && !TRANS_ISTRANS(ufsvfsp))) {
1041 			ufsvfsp->vfs_log = NULL;
1042 			ufsvfsp->vfs_domatamap = 0;
1043 			error = ENOSPC;
1044 			goto out;
1045 		}
1046 
1047 		if (((fsp->fs_state + fsp->fs_time) == FSOKAY) &&
1048 		    (fsp->fs_clean == FSCLEAN ||
1049 			fsp->fs_clean == FSSTABLE ||
1050 			fsp->fs_clean == FSLOG))
1051 			fsp->fs_clean = FSSTABLE;
1052 		else {
1053 			if (isroot) {
1054 				/*
1055 				 * allow root partition to be mounted even
1056 				 * when fs_state is not ok
1057 				 * will be fixed later by a remount root
1058 				 */
1059 				fsp->fs_clean = FSBAD;
1060 				ufsvfsp->vfs_log = NULL;
1061 				ufsvfsp->vfs_domatamap = 0;
1062 			} else {
1063 				error = ENOSPC;
1064 				goto out;
1065 			}
1066 		}
1067 
1068 		if (fsp->fs_clean == FSSTABLE && TRANS_ISTRANS(ufsvfsp))
1069 			fsp->fs_clean = FSLOG;
1070 	}
1071 	TRANS_MATA_MOUNT(ufsvfsp);
1072 	needtrans = 1;
1073 
1074 	vfsp->vfs_bsize = fsp->fs_bsize;
1075 
1076 	/*
1077 	 * Read in summary info
1078 	 */
1079 	if (error = ufs_getsummaryinfo(dev, ufsvfsp, fsp))
1080 		goto out;
1081 
1082 	/*
1083 	 * lastwhinetime is set to zero rather than lbolt, so that after
1084 	 * mounting if the filesystem is found to be full, then immediately the
1085 	 * "file system message" will be logged.
1086 	 */
1087 	ufsvfsp->vfs_lastwhinetime = 0L;
1088 
1089 
1090 	mutex_init(&ufsvfsp->vfs_lock, NULL, MUTEX_DEFAULT, NULL);
1091 	(void) copystr(path, fsp->fs_fsmnt, sizeof (fsp->fs_fsmnt) - 1, &len);
1092 	bzero(fsp->fs_fsmnt + len, sizeof (fsp->fs_fsmnt) - len);
1093 
1094 	/*
1095 	 * Sanity checks for old file systems
1096 	 */
1097 	if (fsp->fs_postblformat == FS_42POSTBLFMT)
1098 		ufsvfsp->vfs_nrpos = 8;
1099 	else
1100 		ufsvfsp->vfs_nrpos = fsp->fs_nrpos;
1101 
1102 	/*
1103 	 * Initialize lockfs structure to support file system locking
1104 	 */
1105 	bzero(&ufsvfsp->vfs_ulockfs.ul_lockfs,
1106 	    sizeof (struct lockfs));
1107 	ufsvfsp->vfs_ulockfs.ul_fs_lock = ULOCKFS_ULOCK;
1108 	mutex_init(&ufsvfsp->vfs_ulockfs.ul_lock, NULL,
1109 	    MUTEX_DEFAULT, NULL);
1110 	cv_init(&ufsvfsp->vfs_ulockfs.ul_cv, NULL, CV_DEFAULT, NULL);
1111 
1112 	/*
1113 	 * We don't need to grab vfs_dqrwlock for this ufs_iget() call.
1114 	 * We are in the process of mounting the file system so there
1115 	 * is no need to grab the quota lock. If a quota applies to the
1116 	 * root inode, then it will be updated when quotas are enabled.
1117 	 *
1118 	 * However, we have an ASSERT(RW_LOCK_HELD(&ufsvfsp->vfs_dqrwlock))
1119 	 * in getinoquota() that we want to keep so grab it anyway.
1120 	 */
1121 	rw_enter(&ufsvfsp->vfs_dqrwlock, RW_READER);
1122 
1123 	error = ufs_iget_alloced(vfsp, UFSROOTINO, &rip, cr);
1124 
1125 	rw_exit(&ufsvfsp->vfs_dqrwlock);
1126 
1127 	if (error)
1128 		goto out;
1129 
1130 	/*
1131 	 * make sure root inode is a directory.  Returning ENOTDIR might
1132 	 * be confused with the mount point not being a directory, so
1133 	 * we use EIO instead.
1134 	 */
1135 	if ((rip->i_mode & IFMT) != IFDIR) {
1136 		/*
1137 		 * Mark this inode as subject for cleanup
1138 		 * to avoid stray inodes in the cache.
1139 		 */
1140 		rvp = ITOV(rip);
1141 		error = EIO;
1142 		goto out;
1143 	}
1144 
1145 	rvp = ITOV(rip);
1146 	mutex_enter(&rvp->v_lock);
1147 	rvp->v_flag |= VROOT;
1148 	mutex_exit(&rvp->v_lock);
1149 	ufsvfsp->vfs_root = rvp;
1150 	/* The buffer for the root inode does not contain a valid b_vp */
1151 	(void) bfinval(dev, 0);
1152 
1153 	/* options */
1154 	ufsvfsp->vfs_nosetsec = flags & UFSMNT_NOSETSEC;
1155 	ufsvfsp->vfs_nointr  = flags & UFSMNT_NOINTR;
1156 	ufsvfsp->vfs_syncdir = flags & UFSMNT_SYNCDIR;
1157 	ufsvfsp->vfs_noatime = flags & UFSMNT_NOATIME;
1158 	if ((flags & UFSMNT_NODFRATIME) || ufsvfsp->vfs_noatime)
1159 		ufsvfsp->vfs_dfritime &= ~UFS_DFRATIME;
1160 	else	/* dfratime, default behavior */
1161 		ufsvfsp->vfs_dfritime |= UFS_DFRATIME;
1162 	if (flags & UFSMNT_FORCEDIRECTIO)
1163 		ufsvfsp->vfs_forcedirectio = 1;
1164 	else if (flags & UFSMNT_NOFORCEDIRECTIO)
1165 		ufsvfsp->vfs_forcedirectio = 0;
1166 	ufsvfsp->vfs_iotstamp = lbolt;
1167 
1168 	ufsvfsp->vfs_nindiroffset = fsp->fs_nindir - 1;
1169 	ufsvfsp->vfs_nindirshift = highbit(ufsvfsp->vfs_nindiroffset);
1170 	ufsvfsp->vfs_ioclustsz = fsp->fs_bsize * fsp->fs_maxcontig;
1171 
1172 	if (cdev_ioctl(dev, DKIOCINFO, (intptr_t)&ci,
1173 	    FKIOCTL|FNATIVE|FREAD, CRED(), &status) == 0) {
1174 		ufsvfsp->vfs_iotransz = ci.dki_maxtransfer * DEV_BSIZE;
1175 	} else {
1176 		ufsvfsp->vfs_iotransz = MIN(maxphys, ufs_maxmaxphys);
1177 	}
1178 
1179 	if (ufsvfsp->vfs_iotransz <= 0) {
1180 		ufsvfsp->vfs_iotransz = MIN(maxphys, ufs_maxmaxphys);
1181 	}
1182 
1183 	/*
1184 	 * When logging, used to reserve log space for writes and truncs
1185 	 */
1186 	ufsvfsp->vfs_avgbfree = fsp->fs_cstotal.cs_nbfree / fsp->fs_ncg;
1187 
1188 	/*
1189 	 * Determine whether to log cylinder group summary info.
1190 	 */
1191 	ufsvfsp->vfs_nolog_si = (fsp->fs_ncg < ufs_ncg_log);
1192 
1193 	if (TRANS_ISTRANS(ufsvfsp)) {
1194 		/*
1195 		 * start the delete thread
1196 		 */
1197 		ufs_thread_start(&ufsvfsp->vfs_delete, ufs_thread_delete, vfsp);
1198 
1199 		/*
1200 		 * start reclaim thread if the filesystem was not mounted
1201 		 * read only.
1202 		 */
1203 		if (!fsp->fs_ronly && (fsp->fs_reclaim &
1204 			(FS_RECLAIM|FS_RECLAIMING))) {
1205 			fsp->fs_reclaim &= ~FS_RECLAIM;
1206 			fsp->fs_reclaim |=  FS_RECLAIMING;
1207 			ufs_thread_start(&ufsvfsp->vfs_reclaim,
1208 			    ufs_thread_reclaim, vfsp);
1209 		}
1210 
1211 		/* Mark the fs as unrolled */
1212 		fsp->fs_rolled = FS_NEED_ROLL;
1213 	} else if (!fsp->fs_ronly && (fsp->fs_reclaim &
1214 	    (FS_RECLAIM|FS_RECLAIMING))) {
1215 		/*
1216 		 * If a file system that is mounted nologging, after
1217 		 * having previously been mounted logging, becomes
1218 		 * unmounted whilst the reclaim thread is in the throes
1219 		 * of reclaiming open/deleted inodes, a subsequent mount
1220 		 * of such a file system with logging disabled could lead
1221 		 * to inodes becoming lost.  So, start reclaim now, even
1222 		 * though logging was disabled for the previous mount, to
1223 		 * tidy things up.
1224 		 */
1225 		fsp->fs_reclaim &= ~FS_RECLAIM;
1226 		fsp->fs_reclaim |=  FS_RECLAIMING;
1227 		ufs_thread_start(&ufsvfsp->vfs_reclaim,
1228 		    ufs_thread_reclaim, vfsp);
1229 	}
1230 
1231 	if (!fsp->fs_ronly) {
1232 		TRANS_SBWRITE(ufsvfsp, TOP_MOUNT);
1233 		if (error = geterror(ufsvfsp->vfs_bufp))
1234 			goto out;
1235 	}
1236 
1237 	/* fix-on-panic initialization */
1238 	if (isroot && !(flags & UFSMNT_ONERROR_FLGMASK))
1239 		flags |= UFSMNT_ONERROR_PANIC;	/* XXX ..._RDONLY */
1240 
1241 	if ((error = ufsfx_mount(ufsvfsp, flags)) != 0)
1242 		goto out;
1243 
1244 	if (why == ROOT_INIT && isroot)
1245 		rootvp = devvp;
1246 
1247 	return (0);
1248 out:
1249 	if (error == 0)
1250 		error = EIO;
1251 	if (rvp) {
1252 		/* the following sequence is similar to ufs_unmount() */
1253 
1254 		/*
1255 		 * There's a problem that ufs_iget() puts inodes into
1256 		 * the inode cache before it returns them.  If someone
1257 		 * traverses that cache and gets a reference to our
1258 		 * inode, there's a chance they'll still be using it
1259 		 * after we've destroyed it.  This is a hard race to
1260 		 * hit, but it's happened (putting in a medium delay
1261 		 * here, and a large delay in ufs_scan_inodes() for
1262 		 * inodes on the device we're bailing out on, makes
1263 		 * the race easy to demonstrate).  The symptom is some
1264 		 * other part of UFS faulting on bad inode contents,
1265 		 * or when grabbing one of the locks inside the inode,
1266 		 * etc.  The usual victim is ufs_scan_inodes() or
1267 		 * someone called by it.
1268 		 */
1269 
1270 		/*
1271 		 * First, isolate it so that no new references can be
1272 		 * gotten via the inode cache.
1273 		 */
1274 		ihm = &ih_lock[INOHASH(UFSROOTINO)];
1275 		mutex_enter(ihm);
1276 		remque(rip);
1277 		mutex_exit(ihm);
1278 
1279 		/*
1280 		 * Now wait for all outstanding references except our
1281 		 * own to drain.  This could, in theory, take forever,
1282 		 * so don't wait *too* long.  If we time out, mark
1283 		 * it stale and leak it, so we don't hit the problem
1284 		 * described above.
1285 		 *
1286 		 * Note that v_count is an int, which means we can read
1287 		 * it in one operation.  Thus, there's no need to lock
1288 		 * around our tests.
1289 		 */
1290 		elapsed = 0;
1291 		while ((rvp->v_count > 1) && (elapsed < ufs_mount_timeout)) {
1292 			delay(ufs_mount_error_delay * drv_usectohz(1000));
1293 			elapsed += ufs_mount_error_delay;
1294 		}
1295 
1296 		if (rvp->v_count > 1) {
1297 			mutex_enter(&rip->i_tlock);
1298 			rip->i_flag |= ISTALE;
1299 			mutex_exit(&rip->i_tlock);
1300 			cmn_err(CE_WARN,
1301 		"Timed out while cleaning up after failed mount of %s",
1302 				path);
1303 		} else {
1304 
1305 			/*
1306 			 * Now we're the only one with a handle left, so tear
1307 			 * it down the rest of the way.
1308 			 */
1309 			if (ufs_rmidle(rip))
1310 				VN_RELE(rvp);
1311 			ufs_si_del(rip);
1312 			rip->i_ufsvfs = NULL;
1313 			rvp->v_vfsp = NULL;
1314 			rvp->v_type = VBAD;
1315 			VN_RELE(rvp);
1316 		}
1317 	}
1318 	if (needtrans) {
1319 		TRANS_MATA_UMOUNT(ufsvfsp);
1320 	}
1321 	if (ufsvfsp) {
1322 		ufs_vfs_remove(ufsvfsp);
1323 		ufs_thread_exit(&ufsvfsp->vfs_delete);
1324 		ufs_thread_exit(&ufsvfsp->vfs_reclaim);
1325 		mutex_destroy(&ufsvfsp->vfs_lock);
1326 		if (ufsvfsp->vfs_log) {
1327 			lufs_unsnarf(ufsvfsp);
1328 		}
1329 		kmem_free(ufsvfsp, sizeof (struct ufsvfs));
1330 	}
1331 	if (bp) {
1332 		bp->b_flags |= (B_STALE|B_AGE);
1333 		brelse(bp);
1334 	}
1335 	if (tp) {
1336 		tp->b_flags |= (B_STALE|B_AGE);
1337 		brelse(tp);
1338 	}
1339 	if (needclose) {
1340 		(void) VOP_CLOSE(devvp, (vfsp->vfs_flag & VFS_RDONLY) ?
1341 		    FREAD : FREAD|FWRITE, 1, (offset_t)0, cr);
1342 		bflush(dev);
1343 		(void) bfinval(dev, 1);
1344 	}
1345 	return (error);
1346 }
1347 
1348 /*
1349  * vfs operations
1350  */
1351 static int
1352 ufs_unmount(struct vfs *vfsp, int fflag, struct cred *cr)
1353 {
1354 	dev_t 		dev		= vfsp->vfs_dev;
1355 	struct ufsvfs	*ufsvfsp	= (struct ufsvfs *)vfsp->vfs_data;
1356 	struct fs	*fs		= ufsvfsp->vfs_fs;
1357 	struct ulockfs	*ulp		= &ufsvfsp->vfs_ulockfs;
1358 	struct vnode 	*bvp, *vp;
1359 	struct buf	*bp;
1360 	struct inode	*ip, *inext, *rip;
1361 	union ihead	*ih;
1362 	int 		error, flag, i;
1363 	struct lockfs	lockfs;
1364 	int		poll_events = POLLPRI;
1365 	extern struct pollhead ufs_pollhd;
1366 	refstr_t	*mountpoint;
1367 
1368 	ASSERT(vfs_lock_held(vfsp));
1369 
1370 	if (secpolicy_fs_unmount(cr, vfsp) != 0)
1371 		return (EPERM);
1372 	/*
1373 	 * Forced unmount is now supported through the
1374 	 * lockfs protocol.
1375 	 */
1376 	if (fflag & MS_FORCE) {
1377 		/*
1378 		 * Mark the filesystem as being unmounted now in
1379 		 * case of a forcible umount before we take any
1380 		 * locks inside UFS to prevent racing with a VFS_VGET()
1381 		 * request. Throw these VFS_VGET() requests away for
1382 		 * the duration of the forcible umount so they won't
1383 		 * use stale or even freed data later on when we're done.
1384 		 * It may happen that the VFS has had a additional hold
1385 		 * placed on it by someone other than UFS and thus will
1386 		 * not get freed immediately once we're done with the
1387 		 * umount by dounmount() - use VFS_UNMOUNTED to inform
1388 		 * users of this still-alive VFS that its corresponding
1389 		 * filesystem being gone so they can detect that and error
1390 		 * out.
1391 		 */
1392 		vfsp->vfs_flag |= VFS_UNMOUNTED;
1393 
1394 		ufs_thread_suspend(&ufsvfsp->vfs_delete);
1395 		mutex_enter(&ulp->ul_lock);
1396 		/*
1397 		 * If file system is already hard locked,
1398 		 * unmount the file system, otherwise
1399 		 * hard lock it before unmounting.
1400 		 */
1401 		if (!ULOCKFS_IS_HLOCK(ulp)) {
1402 			atomic_add_long(&ufs_quiesce_pend, 1);
1403 			lockfs.lf_lock = LOCKFS_HLOCK;
1404 			lockfs.lf_flags = 0;
1405 			lockfs.lf_key = ulp->ul_lockfs.lf_key + 1;
1406 			lockfs.lf_comlen = 0;
1407 			lockfs.lf_comment = NULL;
1408 			ufs_freeze(ulp, &lockfs);
1409 			ULOCKFS_SET_BUSY(ulp);
1410 			LOCKFS_SET_BUSY(&ulp->ul_lockfs);
1411 			(void) ufs_quiesce(ulp);
1412 			(void) ufs_flush(vfsp);
1413 			(void) ufs_thaw(vfsp, ufsvfsp, ulp);
1414 			atomic_add_long(&ufs_quiesce_pend, -1);
1415 			ULOCKFS_CLR_BUSY(ulp);
1416 			LOCKFS_CLR_BUSY(&ulp->ul_lockfs);
1417 			poll_events |= POLLERR;
1418 			pollwakeup(&ufs_pollhd, poll_events);
1419 		}
1420 		ufs_thread_continue(&ufsvfsp->vfs_delete);
1421 		mutex_exit(&ulp->ul_lock);
1422 	}
1423 
1424 	/* let all types of writes go through */
1425 	ufsvfsp->vfs_iotstamp = lbolt;
1426 
1427 	/* coordinate with global hlock thread */
1428 	if (TRANS_ISTRANS(ufsvfsp) && (ufsvfsp->vfs_validfs == UT_HLOCKING)) {
1429 		/*
1430 		 * last possibility for a forced umount to fail hence clear
1431 		 * VFS_UNMOUNTED if appropriate.
1432 		 */
1433 		if (fflag & MS_FORCE)
1434 			vfsp->vfs_flag &= ~VFS_UNMOUNTED;
1435 		return (EAGAIN);
1436 	}
1437 
1438 	ufsvfsp->vfs_validfs = UT_UNMOUNTED;
1439 
1440 	/* kill the reclaim thread */
1441 	ufs_thread_exit(&ufsvfsp->vfs_reclaim);
1442 
1443 	/* suspend the delete thread */
1444 	ufs_thread_suspend(&ufsvfsp->vfs_delete);
1445 
1446 	/*
1447 	 * drain the delete and idle queues
1448 	 */
1449 	ufs_delete_drain(vfsp, -1, 1);
1450 	ufs_idle_drain(vfsp);
1451 
1452 	/*
1453 	 * use the lockfs protocol to prevent new ops from starting
1454 	 * a forcible umount can not fail beyond this point as
1455 	 * we hard-locked the filesystem and drained all current consumers
1456 	 * before.
1457 	 */
1458 	mutex_enter(&ulp->ul_lock);
1459 
1460 	/*
1461 	 * if the file system is busy; return EBUSY
1462 	 */
1463 	if (ulp->ul_vnops_cnt || ULOCKFS_IS_SLOCK(ulp)) {
1464 		error = EBUSY;
1465 		goto out;
1466 	}
1467 
1468 	/*
1469 	 * if this is not a forced unmount (!hard/error locked), then
1470 	 * get rid of every inode except the root and quota inodes
1471 	 * also, commit any outstanding transactions
1472 	 */
1473 	if (!ULOCKFS_IS_HLOCK(ulp) && !ULOCKFS_IS_ELOCK(ulp))
1474 		if (error = ufs_flush(vfsp))
1475 			goto out;
1476 
1477 	/*
1478 	 * ignore inodes in the cache if fs is hard locked or error locked
1479 	 */
1480 	rip = VTOI(ufsvfsp->vfs_root);
1481 	if (!ULOCKFS_IS_HLOCK(ulp) && !ULOCKFS_IS_ELOCK(ulp)) {
1482 		/*
1483 		 * Otherwise, only the quota and root inodes are in the cache.
1484 		 *
1485 		 * Avoid racing with ufs_update() and ufs_sync().
1486 		 */
1487 		mutex_enter(&ufs_scan_lock);
1488 
1489 		for (i = 0, ih = ihead; i < inohsz; i++, ih++) {
1490 			mutex_enter(&ih_lock[i]);
1491 			for (ip = ih->ih_chain[0];
1492 					ip != (struct inode *)ih;
1493 					ip = ip->i_forw) {
1494 				if (ip->i_ufsvfs != ufsvfsp)
1495 					continue;
1496 				if (ip == ufsvfsp->vfs_qinod)
1497 					continue;
1498 				if (ip == rip && ITOV(ip)->v_count == 1)
1499 					continue;
1500 				mutex_exit(&ih_lock[i]);
1501 				mutex_exit(&ufs_scan_lock);
1502 				error = EBUSY;
1503 				goto out;
1504 			}
1505 			mutex_exit(&ih_lock[i]);
1506 		}
1507 		mutex_exit(&ufs_scan_lock);
1508 	}
1509 
1510 	/*
1511 	 * if a snapshot exists and this is a forced unmount, then delete
1512 	 * the snapshot.  Otherwise return EBUSY.  This will insure the
1513 	 * snapshot always belongs to a valid file system.
1514 	 */
1515 	if (ufsvfsp->vfs_snapshot) {
1516 		if (ULOCKFS_IS_HLOCK(ulp) || ULOCKFS_IS_ELOCK(ulp)) {
1517 			(void) fssnap_delete(&ufsvfsp->vfs_snapshot);
1518 		} else {
1519 			error = EBUSY;
1520 			goto out;
1521 		}
1522 	}
1523 
1524 	/*
1525 	 * Close the quota file and invalidate anything left in the quota
1526 	 * cache for this file system.  Pass kcred to allow all quota
1527 	 * manipulations.
1528 	 */
1529 	(void) closedq(ufsvfsp, kcred);
1530 	invalidatedq(ufsvfsp);
1531 	/*
1532 	 * drain the delete and idle queues
1533 	 */
1534 	ufs_delete_drain(vfsp, -1, 0);
1535 	ufs_idle_drain(vfsp);
1536 
1537 	/*
1538 	 * discard the inodes for this fs (including root, shadow, and quota)
1539 	 */
1540 	for (i = 0, ih = ihead; i < inohsz; i++, ih++) {
1541 		mutex_enter(&ih_lock[i]);
1542 		for (inext = 0, ip = ih->ih_chain[0];
1543 				ip != (struct inode *)ih;
1544 				ip = inext) {
1545 			inext = ip->i_forw;
1546 			if (ip->i_ufsvfs != ufsvfsp)
1547 				continue;
1548 			vp = ITOV(ip);
1549 			VN_HOLD(vp)
1550 			remque(ip);
1551 			if (ufs_rmidle(ip))
1552 				VN_RELE(vp);
1553 			ufs_si_del(ip);
1554 			/*
1555 			 * rip->i_ufsvfsp is needed by bflush()
1556 			 */
1557 			if (ip != rip)
1558 				ip->i_ufsvfs = NULL;
1559 			/*
1560 			 * Set vnode's vfsops to dummy ops, which return
1561 			 * EIO. This is needed to forced unmounts to work
1562 			 * with lofs/nfs properly.
1563 			 */
1564 			if (ULOCKFS_IS_HLOCK(ulp) || ULOCKFS_IS_ELOCK(ulp))
1565 				vp->v_vfsp = &EIO_vfs;
1566 			else
1567 				vp->v_vfsp = NULL;
1568 			vp->v_type = VBAD;
1569 			VN_RELE(vp);
1570 		}
1571 		mutex_exit(&ih_lock[i]);
1572 	}
1573 	ufs_si_cache_flush(dev);
1574 
1575 	/*
1576 	 * kill the delete thread and drain the idle queue
1577 	 */
1578 	ufs_thread_exit(&ufsvfsp->vfs_delete);
1579 	ufs_idle_drain(vfsp);
1580 
1581 	bp = ufsvfsp->vfs_bufp;
1582 	bvp = ufsvfsp->vfs_devvp;
1583 	flag = !fs->fs_ronly;
1584 	if (flag) {
1585 		bflush(dev);
1586 		if (fs->fs_clean != FSBAD) {
1587 			if (fs->fs_clean == FSSTABLE)
1588 				fs->fs_clean = FSCLEAN;
1589 			fs->fs_reclaim &= ~FS_RECLAIM;
1590 		}
1591 		if (TRANS_ISTRANS(ufsvfsp) &&
1592 		    !TRANS_ISERROR(ufsvfsp) &&
1593 		    !ULOCKFS_IS_HLOCK(ulp) &&
1594 		    (fs->fs_rolled == FS_NEED_ROLL)) {
1595 			/*
1596 			 * ufs_flush() above has flushed the last Moby.
1597 			 * This is needed to ensure the following superblock
1598 			 * update really is the last metadata update
1599 			 */
1600 			error = ufs_putsummaryinfo(dev, ufsvfsp, fs);
1601 			if (error == 0) {
1602 				fs->fs_rolled = FS_ALL_ROLLED;
1603 			}
1604 		}
1605 		TRANS_SBUPDATE(ufsvfsp, vfsp, TOP_SBUPDATE_UNMOUNT);
1606 		/*
1607 		 * push this last transaction
1608 		 */
1609 		curthread->t_flag |= T_DONTBLOCK;
1610 		TRANS_BEGIN_SYNC(ufsvfsp, TOP_COMMIT_UNMOUNT, TOP_COMMIT_SIZE,
1611 		    error);
1612 		if (!error)
1613 			TRANS_END_SYNC(ufsvfsp, error, TOP_COMMIT_UNMOUNT,
1614 			    TOP_COMMIT_SIZE);
1615 		curthread->t_flag &= ~T_DONTBLOCK;
1616 	}
1617 
1618 	TRANS_MATA_UMOUNT(ufsvfsp);
1619 	lufs_unsnarf(ufsvfsp);		/* Release the in-memory structs */
1620 	ufsfx_unmount(ufsvfsp);		/* fix-on-panic bookkeeping */
1621 	kmem_free(fs->fs_u.fs_csp, fs->fs_cssize);
1622 
1623 	bp->b_flags |= B_STALE|B_AGE;
1624 	ufsvfsp->vfs_bufp = NULL;	/* don't point at freed buf */
1625 	brelse(bp);			/* free the superblock buf */
1626 
1627 	(void) VOP_PUTPAGE(common_specvp(bvp), (offset_t)0, (size_t)0,
1628 	    B_INVAL, cr);
1629 	(void) VOP_CLOSE(bvp, flag, 1, (offset_t)0, cr);
1630 	bflush(dev);
1631 	(void) bfinval(dev, 1);
1632 	VN_RELE(bvp);
1633 
1634 	/*
1635 	 * It is now safe to NULL out the ufsvfs pointer and discard
1636 	 * the root inode.
1637 	 */
1638 	rip->i_ufsvfs = NULL;
1639 	VN_RELE(ITOV(rip));
1640 
1641 	/* free up lockfs comment structure, if any */
1642 	if (ulp->ul_lockfs.lf_comlen && ulp->ul_lockfs.lf_comment)
1643 		kmem_free(ulp->ul_lockfs.lf_comment, ulp->ul_lockfs.lf_comlen);
1644 
1645 	/*
1646 	 * Remove from instance list.
1647 	 */
1648 	ufs_vfs_remove(ufsvfsp);
1649 
1650 	/*
1651 	 * For a forcible unmount, threads may be asleep in
1652 	 * ufs_lockfs_begin/ufs_check_lockfs.  These threads will need
1653 	 * the ufsvfs structure so we don't free it, yet.  ufs_update
1654 	 * will free it up after awhile.
1655 	 */
1656 	if (ULOCKFS_IS_HLOCK(ulp) || ULOCKFS_IS_ELOCK(ulp)) {
1657 		extern kmutex_t		ufsvfs_mutex;
1658 		extern struct ufsvfs	*ufsvfslist;
1659 
1660 		mutex_enter(&ufsvfs_mutex);
1661 		ufsvfsp->vfs_dontblock = 1;
1662 		ufsvfsp->vfs_next = ufsvfslist;
1663 		ufsvfslist = ufsvfsp;
1664 		mutex_exit(&ufsvfs_mutex);
1665 		/* wakeup any suspended threads */
1666 		cv_broadcast(&ulp->ul_cv);
1667 		mutex_exit(&ulp->ul_lock);
1668 	} else {
1669 		mutex_destroy(&ufsvfsp->vfs_lock);
1670 		kmem_free(ufsvfsp, sizeof (struct ufsvfs));
1671 	}
1672 
1673 	/*
1674 	 * Now mark the filesystem as unmounted since we're done with it.
1675 	 */
1676 	vfsp->vfs_flag |= VFS_UNMOUNTED;
1677 
1678 	return (0);
1679 out:
1680 	/* open the fs to new ops */
1681 	cv_broadcast(&ulp->ul_cv);
1682 	mutex_exit(&ulp->ul_lock);
1683 
1684 	if (TRANS_ISTRANS(ufsvfsp)) {
1685 		/* allow the delete thread to continue */
1686 		ufs_thread_continue(&ufsvfsp->vfs_delete);
1687 		/* restart the reclaim thread */
1688 		ufs_thread_start(&ufsvfsp->vfs_reclaim, ufs_thread_reclaim,
1689 				vfsp);
1690 		/* coordinate with global hlock thread */
1691 		ufsvfsp->vfs_validfs = UT_MOUNTED;
1692 		/* check for trans errors during umount */
1693 		ufs_trans_onerror();
1694 
1695 		/*
1696 		 * if we have a seperate /usr it will never unmount
1697 		 * when halting. In order to not re-read all the
1698 		 * cylinder group summary info on mounting after
1699 		 * reboot the logging of summary info is re-enabled
1700 		 * and the super block written out.
1701 		 */
1702 		mountpoint = vfs_getmntpoint(vfsp);
1703 		if ((fs->fs_si == FS_SI_OK) &&
1704 		    (strcmp("/usr", refstr_value(mountpoint)) == 0)) {
1705 			ufsvfsp->vfs_nolog_si = 0;
1706 			UFS_BWRITE2(NULL, ufsvfsp->vfs_bufp);
1707 		}
1708 		refstr_rele(mountpoint);
1709 	}
1710 
1711 	return (error);
1712 }
1713 
1714 static int
1715 ufs_root(struct vfs *vfsp, struct vnode **vpp)
1716 {
1717 	struct ufsvfs *ufsvfsp;
1718 	struct vnode *vp;
1719 
1720 	if (!vfsp)
1721 		return (EIO);
1722 
1723 	ufsvfsp = (struct ufsvfs *)vfsp->vfs_data;
1724 	if (!ufsvfsp || !ufsvfsp->vfs_root)
1725 		return (EIO);	/* forced unmount */
1726 
1727 	vp = ufsvfsp->vfs_root;
1728 	VN_HOLD(vp);
1729 	*vpp = vp;
1730 	return (0);
1731 }
1732 
1733 /*
1734  * Get file system statistics.
1735  */
1736 static int
1737 ufs_statvfs(struct vfs *vfsp, struct statvfs64 *sp)
1738 {
1739 	struct fs *fsp;
1740 	struct ufsvfs *ufsvfsp;
1741 	int blk, i;
1742 	long max_avail, used;
1743 	dev32_t d32;
1744 	struct ufs_q *delq;
1745 
1746 	if (vfsp->vfs_flag & VFS_UNMOUNTED)
1747 		return (EIO);
1748 
1749 	ufsvfsp = (struct ufsvfs *)vfsp->vfs_data;
1750 	fsp = ufsvfsp->vfs_fs;
1751 	if ((fsp->fs_magic != FS_MAGIC) && (fsp->fs_magic != MTB_UFS_MAGIC))
1752 		return (EINVAL);
1753 	if (fsp->fs_magic == MTB_UFS_MAGIC &&
1754 	    (fsp->fs_version > MTB_UFS_VERSION_1 ||
1755 	    fsp->fs_version < MTB_UFS_VERSION_MIN))
1756 		return (EINVAL);
1757 
1758 	/*
1759 	 * Can't get a self-consistent result with the delete thread running
1760 	 * or if others come in and jumble the queue up for us.  This is
1761 	 * a no-op if there is no delete thread.
1762 	 */
1763 	delq = &ufsvfsp->vfs_delete;
1764 	ufs_thread_suspend(delq);
1765 
1766 	/*
1767 	 * get the basic numbers
1768 	 */
1769 	(void) bzero(sp, sizeof (*sp));
1770 
1771 	sp->f_bsize = fsp->fs_bsize;
1772 	sp->f_frsize = fsp->fs_fsize;
1773 	sp->f_blocks = (fsblkcnt64_t)fsp->fs_dsize;
1774 	sp->f_bfree = (fsblkcnt64_t)fsp->fs_cstotal.cs_nbfree * fsp->fs_frag +
1775 	    fsp->fs_cstotal.cs_nffree;
1776 
1777 	sp->f_files = (fsfilcnt64_t)fsp->fs_ncg * fsp->fs_ipg;
1778 	sp->f_ffree = (fsfilcnt64_t)fsp->fs_cstotal.cs_nifree;
1779 
1780 	/*
1781 	 * Adjust the numbers based on things waiting to be deleted.
1782 	 * modifies f_bfree and f_ffree.  Afterwards, everything we
1783 	 * come up with will be self-consistent.  By definition, this
1784 	 * is a point-in-time snapshot, so the fact that the delete
1785 	 * thread's probably already invalidated the results is not a
1786 	 * problem.
1787 	 */
1788 	ufs_delete_adjust_stats(ufsvfsp, sp);
1789 	ufs_thread_continue(delq);
1790 
1791 	/*
1792 	 * avail = MAX(max_avail - used, 0)
1793 	 */
1794 	max_avail = fsp->fs_dsize - ufsvfsp->vfs_minfrags;
1795 
1796 	used = (fsp->fs_dsize - sp->f_bfree);
1797 
1798 	if (max_avail > used)
1799 		sp->f_bavail = (fsblkcnt64_t)max_avail - used;
1800 	else
1801 		sp->f_bavail = (fsblkcnt64_t)0;
1802 
1803 	sp->f_favail = sp->f_ffree;
1804 	(void) cmpldev(&d32, vfsp->vfs_dev);
1805 	sp->f_fsid = d32;
1806 	(void) strcpy(sp->f_basetype, vfssw[vfsp->vfs_fstype].vsw_name);
1807 	sp->f_flag = vf_to_stf(vfsp->vfs_flag);
1808 
1809 	/* keep coordinated with ufs_l_pathconf() */
1810 	sp->f_namemax = MAXNAMLEN;
1811 
1812 	if (fsp->fs_cpc == 0) {
1813 		bzero(sp->f_fstr, 14);
1814 		return (0);
1815 	}
1816 	blk = fsp->fs_spc * fsp->fs_cpc / NSPF(fsp);
1817 	for (i = 0; i < blk; i += fsp->fs_frag) /* CSTYLED */
1818 		/* void */;
1819 	i -= fsp->fs_frag;
1820 	blk = i / fsp->fs_frag;
1821 	bcopy(&(fs_rotbl(fsp)[blk]), sp->f_fstr, 14);
1822 	return (0);
1823 }
1824 
1825 /*
1826  * Flush any pending I/O to file system vfsp.
1827  * The ufs_update() routine will only flush *all* ufs files.
1828  * If vfsp is non-NULL, only sync this ufs (in preparation
1829  * for a umount).
1830  */
1831 /*ARGSUSED*/
1832 static int
1833 ufs_sync(struct vfs *vfsp, short flag, struct cred *cr)
1834 {
1835 	struct ufsvfs *ufsvfsp;
1836 	struct fs *fs;
1837 	int cheap = flag & SYNC_ATTR;
1838 	int error;
1839 
1840 	/*
1841 	 * SYNC_CLOSE means we're rebooting.  Toss everything
1842 	 * on the idle queue so we don't have to slog through
1843 	 * a bunch of uninteresting inodes over and over again.
1844 	 */
1845 	if (flag & SYNC_CLOSE)
1846 		ufs_idle_drain(NULL);
1847 
1848 	if (vfsp == NULL) {
1849 		ufs_update(flag);
1850 		return (0);
1851 	}
1852 
1853 	/* Flush a single ufs */
1854 	if (!vfs_matchops(vfsp, ufs_vfsops) || vfs_lock(vfsp) != 0)
1855 		return (0);
1856 
1857 	ufsvfsp = (struct ufsvfs *)vfsp->vfs_data;
1858 	if (!ufsvfsp)
1859 		return (EIO);
1860 	fs = ufsvfsp->vfs_fs;
1861 	mutex_enter(&ufsvfsp->vfs_lock);
1862 
1863 	if (ufsvfsp->vfs_dio &&
1864 	    fs->fs_ronly == 0 &&
1865 	    fs->fs_clean != FSBAD &&
1866 	    fs->fs_clean != FSLOG) {
1867 		/* turn off fast-io on unmount, so no fsck needed (4029401) */
1868 		ufsvfsp->vfs_dio = 0;
1869 		fs->fs_clean = FSACTIVE;
1870 		fs->fs_fmod = 1;
1871 	}
1872 
1873 	/* Write back modified superblock */
1874 	if (fs->fs_fmod == 0) {
1875 		mutex_exit(&ufsvfsp->vfs_lock);
1876 	} else {
1877 		if (fs->fs_ronly != 0) {
1878 			mutex_exit(&ufsvfsp->vfs_lock);
1879 			vfs_unlock(vfsp);
1880 			return (ufs_fault(ufsvfsp->vfs_root,
1881 					    "fs = %s update: ro fs mod\n",
1882 					    fs->fs_fsmnt));
1883 		}
1884 		fs->fs_fmod = 0;
1885 		mutex_exit(&ufsvfsp->vfs_lock);
1886 
1887 		TRANS_SBUPDATE(ufsvfsp, vfsp, TOP_SBUPDATE_UPDATE);
1888 	}
1889 	vfs_unlock(vfsp);
1890 
1891 	/*
1892 	 * Avoid racing with ufs_update() and ufs_unmount().
1893 	 *
1894 	 */
1895 	mutex_enter(&ufs_scan_lock);
1896 
1897 	(void) ufs_scan_inodes(1, ufs_sync_inode,
1898 	    (void *)(uintptr_t)cheap, ufsvfsp);
1899 
1900 	mutex_exit(&ufs_scan_lock);
1901 
1902 	bflush((dev_t)vfsp->vfs_dev);
1903 
1904 	/*
1905 	 * commit any outstanding async transactions
1906 	 */
1907 	curthread->t_flag |= T_DONTBLOCK;
1908 	TRANS_BEGIN_SYNC(ufsvfsp, TOP_COMMIT_UPDATE, TOP_COMMIT_SIZE, error);
1909 	if (!error) {
1910 		TRANS_END_SYNC(ufsvfsp, error, TOP_COMMIT_UPDATE,
1911 		    TOP_COMMIT_SIZE);
1912 	}
1913 	curthread->t_flag &= ~T_DONTBLOCK;
1914 
1915 	return (0);
1916 }
1917 
1918 
1919 void
1920 sbupdate(struct vfs *vfsp)
1921 {
1922 	struct ufsvfs *ufsvfsp = (struct ufsvfs *)vfsp->vfs_data;
1923 	struct fs *fs = ufsvfsp->vfs_fs;
1924 	struct buf *bp;
1925 	int blks;
1926 	caddr_t space;
1927 	int i;
1928 	size_t size;
1929 
1930 	/*
1931 	 * for ulockfs processing, limit the superblock writes
1932 	 */
1933 	if ((ufsvfsp->vfs_ulockfs.ul_sbowner) &&
1934 	    (curthread != ufsvfsp->vfs_ulockfs.ul_sbowner)) {
1935 		/* process later */
1936 		fs->fs_fmod = 1;
1937 		return;
1938 	}
1939 	ULOCKFS_SET_MOD((&ufsvfsp->vfs_ulockfs));
1940 
1941 	if (TRANS_ISTRANS(ufsvfsp)) {
1942 		mutex_enter(&ufsvfsp->vfs_lock);
1943 		ufs_sbwrite(ufsvfsp);
1944 		mutex_exit(&ufsvfsp->vfs_lock);
1945 		return;
1946 	}
1947 
1948 	blks = howmany(fs->fs_cssize, fs->fs_fsize);
1949 	space = (caddr_t)fs->fs_u.fs_csp;
1950 	for (i = 0; i < blks; i += fs->fs_frag) {
1951 		size = fs->fs_bsize;
1952 		if (i + fs->fs_frag > blks)
1953 			size = (blks - i) * fs->fs_fsize;
1954 		bp = UFS_GETBLK(ufsvfsp, ufsvfsp->vfs_dev,
1955 			(daddr_t)(fsbtodb(fs, fs->fs_csaddr + i)),
1956 			fs->fs_bsize);
1957 		bcopy(space, bp->b_un.b_addr, size);
1958 		space += size;
1959 		bp->b_bcount = size;
1960 		UFS_BRWRITE(ufsvfsp, bp);
1961 	}
1962 	mutex_enter(&ufsvfsp->vfs_lock);
1963 	ufs_sbwrite(ufsvfsp);
1964 	mutex_exit(&ufsvfsp->vfs_lock);
1965 }
1966 
1967 int ufs_vget_idle_count = 2;	/* Number of inodes to idle each time */
1968 static int
1969 ufs_vget(struct vfs *vfsp, struct vnode **vpp, struct fid *fidp)
1970 {
1971 	int error = 0;
1972 	struct ufid *ufid;
1973 	struct inode *ip;
1974 	struct ufsvfs *ufsvfsp = (struct ufsvfs *)vfsp->vfs_data;
1975 	struct ulockfs *ulp;
1976 
1977 	/*
1978 	 * Check for unmounted filesystem.
1979 	 */
1980 	if (vfsp->vfs_flag & VFS_UNMOUNTED) {
1981 		error = EIO;
1982 		goto errout;
1983 	}
1984 
1985 	/*
1986 	 * Keep the idle queue from getting too long by
1987 	 * idling an inode before attempting to allocate another.
1988 	 *    This operation must be performed before entering
1989 	 *    lockfs or a transaction.
1990 	 */
1991 	if (ufs_idle_q.uq_ne > ufs_idle_q.uq_hiwat)
1992 		if ((curthread->t_flag & T_DONTBLOCK) == 0) {
1993 			ins.in_vidles.value.ul += ufs_vget_idle_count;
1994 			ufs_idle_some(ufs_vget_idle_count);
1995 		}
1996 
1997 	ufid = (struct ufid *)fidp;
1998 
1999 	if (error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_VGET_MASK))
2000 		goto errout;
2001 
2002 	rw_enter(&ufsvfsp->vfs_dqrwlock, RW_READER);
2003 
2004 	error = ufs_iget(vfsp, ufid->ufid_ino, &ip, CRED());
2005 
2006 	rw_exit(&ufsvfsp->vfs_dqrwlock);
2007 
2008 	ufs_lockfs_end(ulp);
2009 
2010 	if (error)
2011 		goto errout;
2012 
2013 	/*
2014 	 * Check if the inode has been deleted or freed or is in transient state
2015 	 * since the last VFS_VGET() request for it, release it and don't return
2016 	 * it to the caller, presumably NFS, as it's no longer valid.
2017 	 */
2018 	if (ip->i_gen != ufid->ufid_gen || ip->i_mode == 0 ||
2019 	    (ip->i_flag & IDEL)) {
2020 		VN_RELE(ITOV(ip));
2021 		error = EINVAL;
2022 		goto errout;
2023 	}
2024 
2025 	*vpp = ITOV(ip);
2026 	return (0);
2027 
2028 errout:
2029 	*vpp = NULL;
2030 	return (error);
2031 }
2032 
2033 static int
2034 ufsinit(int fstype, char *name)
2035 {
2036 	static const fs_operation_def_t ufs_vfsops_template[] = {
2037 		VFSNAME_MOUNT, ufs_mount,
2038 		VFSNAME_UNMOUNT, ufs_unmount,
2039 		VFSNAME_ROOT, ufs_root,
2040 		VFSNAME_STATVFS, ufs_statvfs,
2041 		VFSNAME_SYNC, (fs_generic_func_p) ufs_sync,
2042 		VFSNAME_VGET, ufs_vget,
2043 		VFSNAME_MOUNTROOT, ufs_mountroot,
2044 		NULL, NULL
2045 	};
2046 	int error;
2047 
2048 	ufsfstype = fstype;
2049 
2050 	error = vfs_setfsops(fstype, ufs_vfsops_template, &ufs_vfsops);
2051 	if (error != 0) {
2052 		cmn_err(CE_WARN, "ufsinit: bad vfs ops template");
2053 		return (error);
2054 	}
2055 
2056 	error = vn_make_ops(name, ufs_vnodeops_template, &ufs_vnodeops);
2057 	if (error != 0) {
2058 		(void) vfs_freevfsops_by_type(fstype);
2059 		cmn_err(CE_WARN, "ufsinit: bad vnode ops template");
2060 		return (error);
2061 	}
2062 
2063 	ufs_iinit();
2064 	return (0);
2065 }
2066 
2067 #ifdef __sparc
2068 
2069 /*
2070  * Mounting a mirrored SVM volume is only supported on ufs,
2071  * this is special-case boot code to support that configuration.
2072  * At this point, we have booted and mounted root on a
2073  * single component of the mirror.  Complete the boot
2074  * by configuring SVM and converting the root to the
2075  * dev_t of the mirrored root device.  This dev_t conversion
2076  * only works because the underlying device doesn't change.
2077  */
2078 int
2079 ufs_remountroot(struct vfs *vfsp)
2080 {
2081 	struct ufsvfs *ufsvfsp;
2082 	struct ulockfs *ulp;
2083 	dev_t new_rootdev;
2084 	dev_t old_rootdev;
2085 	struct vnode *old_rootvp;
2086 	struct vnode *new_rootvp;
2087 	int error, sberror = 0;
2088 	struct inode	*ip;
2089 	union ihead	*ih;
2090 	struct buf	*bp;
2091 	int i;
2092 
2093 	old_rootdev = rootdev;
2094 	old_rootvp = rootvp;
2095 
2096 	new_rootdev = getrootdev();
2097 	if (new_rootdev == (dev_t)NODEV) {
2098 		return (ENODEV);
2099 	}
2100 
2101 	new_rootvp = makespecvp(new_rootdev, VBLK);
2102 
2103 	error = VOP_OPEN(&new_rootvp,
2104 	    (vfsp->vfs_flag & VFS_RDONLY) ? FREAD : FREAD|FWRITE, CRED());
2105 	if (error) {
2106 		cmn_err(CE_CONT,
2107 		    "Cannot open mirrored root device, error %d\n", error);
2108 		return (error);
2109 	}
2110 
2111 	if (vfs_lock(vfsp) != 0) {
2112 		return (EBUSY);
2113 	}
2114 
2115 	ufsvfsp = (struct ufsvfs *)vfsp->vfs_data;
2116 	ulp = &ufsvfsp->vfs_ulockfs;
2117 
2118 	mutex_enter(&ulp->ul_lock);
2119 	atomic_add_long(&ufs_quiesce_pend, 1);
2120 
2121 	(void) ufs_quiesce(ulp);
2122 	(void) ufs_flush(vfsp);
2123 
2124 	/*
2125 	 * Convert root vfs to new dev_t, including vfs hash
2126 	 * table and fs id.
2127 	 */
2128 	vfs_root_redev(vfsp, new_rootdev, ufsfstype);
2129 
2130 	ufsvfsp->vfs_devvp = new_rootvp;
2131 	ufsvfsp->vfs_dev = new_rootdev;
2132 
2133 	bp = ufsvfsp->vfs_bufp;
2134 	bp->b_edev = new_rootdev;
2135 	bp->b_dev = cmpdev(new_rootdev);
2136 
2137 	/*
2138 	 * The buffer for the root inode does not contain a valid b_vp
2139 	 */
2140 	(void) bfinval(new_rootdev, 0);
2141 
2142 	/*
2143 	 * Here we hand-craft inodes with old root device
2144 	 * references to refer to the new device instead.
2145 	 */
2146 	mutex_enter(&ufs_scan_lock);
2147 
2148 	for (i = 0, ih = ihead; i < inohsz; i++, ih++) {
2149 		mutex_enter(&ih_lock[i]);
2150 		for (ip = ih->ih_chain[0];
2151 				ip != (struct inode *)ih;
2152 				ip = ip->i_forw) {
2153 			if (ip->i_ufsvfs != ufsvfsp)
2154 				continue;
2155 			if (ip == ufsvfsp->vfs_qinod)
2156 				continue;
2157 			if (ip->i_dev == old_rootdev) {
2158 				ip->i_dev = new_rootdev;
2159 			}
2160 
2161 			if (ip->i_devvp == old_rootvp) {
2162 				ip->i_devvp = new_rootvp;
2163 			}
2164 		}
2165 		mutex_exit(&ih_lock[i]);
2166 	}
2167 
2168 	mutex_exit(&ufs_scan_lock);
2169 
2170 	/*
2171 	 * Make Sure logging structures are using the new device
2172 	 * if logging is enabled.  Also start any logging thread that
2173 	 * needs to write to the device and couldn't earlier.
2174 	 */
2175 	if (ufsvfsp->vfs_log) {
2176 		buf_t		*bp, *tbp;
2177 		ml_unit_t	*ul = ufsvfsp->vfs_log;
2178 		struct fs	*fsp = ufsvfsp->vfs_fs;
2179 
2180 		/*
2181 		 * Update the main logging structure.
2182 		 */
2183 		ul->un_dev = new_rootdev;
2184 
2185 		/*
2186 		 * Get a new bp for the on disk structures.
2187 		 */
2188 		bp = ul->un_bp;
2189 		tbp = ngeteblk(dbtob(LS_SECTORS));
2190 		tbp->b_edev = new_rootdev;
2191 		tbp->b_dev = cmpdev(new_rootdev);
2192 		tbp->b_blkno = bp->b_blkno;
2193 		bcopy(bp->b_un.b_addr, tbp->b_un.b_addr, DEV_BSIZE);
2194 		bcopy(bp->b_un.b_addr, tbp->b_un.b_addr + DEV_BSIZE, DEV_BSIZE);
2195 		bp->b_flags |= (B_STALE | B_AGE);
2196 		brelse(bp);
2197 		ul->un_bp = tbp;
2198 
2199 		/*
2200 		 * Allocate new circular buffers.
2201 		 */
2202 		alloc_rdbuf(&ul->un_rdbuf, MAPBLOCKSIZE, MAPBLOCKSIZE);
2203 		alloc_wrbuf(&ul->un_wrbuf, ldl_bufsize(ul));
2204 
2205 		/*
2206 		 * Clear the noroll bit which indicates that logging
2207 		 * can't roll the log yet.
2208 		 */
2209 		ASSERT(ul->un_flags & LDL_NOROLL);
2210 		ul->un_flags &= ~LDL_NOROLL;
2211 
2212 		/*
2213 		 * Start the logmap roll thread.
2214 		 */
2215 		logmap_start_roll(ul);
2216 
2217 		/*
2218 		 * Start the reclaim thread if needed.
2219 		 */
2220 		if (!fsp->fs_ronly && (fsp->fs_reclaim &
2221 			(FS_RECLAIM|FS_RECLAIMING))) {
2222 			fsp->fs_reclaim &= ~FS_RECLAIM;
2223 			fsp->fs_reclaim |= FS_RECLAIMING;
2224 			ufs_thread_start(&ufsvfsp->vfs_reclaim,
2225 				ufs_thread_reclaim, vfsp);
2226 			TRANS_SBWRITE(ufsvfsp, TOP_SBUPDATE_UPDATE);
2227 			if (sberror = geterror(ufsvfsp->vfs_bufp)) {
2228 				refstr_t	*mntpt;
2229 				mntpt = vfs_getmntpoint(vfsp);
2230 				cmn_err(CE_WARN,
2231 					"Remountroot failed to update Reclaim"
2232 					"state for filesystem %s "
2233 					"Error writing SuperBlock %d",
2234 					refstr_value(mntpt), error);
2235 				refstr_rele(mntpt);
2236 			}
2237 		}
2238 	}
2239 
2240 	rootdev = new_rootdev;
2241 	rootvp = new_rootvp;
2242 
2243 	atomic_add_long(&ufs_quiesce_pend, -1);
2244 	cv_broadcast(&ulp->ul_cv);
2245 	mutex_exit(&ulp->ul_lock);
2246 
2247 	vfs_unlock(vfsp);
2248 
2249 	error = VOP_CLOSE(old_rootvp, FREAD, 1, (offset_t)0, CRED());
2250 	if (error) {
2251 		cmn_err(CE_CONT,
2252 		    "close of root device component failed, error %d\n",
2253 		    error);
2254 	}
2255 	VN_RELE(old_rootvp);
2256 
2257 	return (sberror ? sberror : error);
2258 }
2259 
2260 #endif	/* __sparc */
2261