1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License, Version 1.0 only 6 * (the "License"). You may not use this file except in compliance 7 * with the License. 8 * 9 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 10 * or http://www.opensolaris.org/os/licensing. 11 * See the License for the specific language governing permissions 12 * and limitations under the License. 13 * 14 * When distributing Covered Code, include this CDDL HEADER in each 15 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 16 * If applicable, add the following below this CDDL HEADER, with the 17 * fields enclosed by brackets "[]" replaced with your own identifying 18 * information: Portions Copyright [yyyy] [name of copyright owner] 19 * 20 * CDDL HEADER END 21 */ 22 /* 23 * Copyright 2005 Sun Microsystems, Inc. All rights reserved. 24 * Use is subject to license terms. 25 */ 26 27 /* Copyright (c) 1983, 1984, 1985, 1986, 1987, 1988, 1989 AT&T */ 28 /* All Rights Reserved */ 29 30 /* 31 * University Copyright- Copyright (c) 1982, 1986, 1988 32 * The Regents of the University of California 33 * All Rights Reserved 34 * 35 * University Acknowledgment- Portions of this document are derived from 36 * software developed by the University of California, Berkeley, and its 37 * contributors. 38 */ 39 40 41 #pragma ident "%Z%%M% %I% %E% SMI" 42 43 #include <sys/types.h> 44 #include <sys/t_lock.h> 45 #include <sys/param.h> 46 #include <sys/systm.h> 47 #include <sys/bitmap.h> 48 #include <sys/sysmacros.h> 49 #include <sys/kmem.h> 50 #include <sys/signal.h> 51 #include <sys/user.h> 52 #include <sys/proc.h> 53 #include <sys/disp.h> 54 #include <sys/buf.h> 55 #include <sys/pathname.h> 56 #include <sys/vfs.h> 57 #include <sys/vnode.h> 58 #include <sys/file.h> 59 #include <sys/atomic.h> 60 #include <sys/uio.h> 61 #include <sys/dkio.h> 62 #include <sys/cred.h> 63 #include <sys/conf.h> 64 #include <sys/dnlc.h> 65 #include <sys/kstat.h> 66 #include <sys/acl.h> 67 #include <sys/fs/ufs_fsdir.h> 68 #include <sys/fs/ufs_fs.h> 69 #include <sys/fs/ufs_inode.h> 70 #include <sys/fs/ufs_mount.h> 71 #include <sys/fs/ufs_acl.h> 72 #include <sys/fs/ufs_panic.h> 73 #include <sys/fs/ufs_bio.h> 74 #include <sys/fs/ufs_quota.h> 75 #include <sys/fs/ufs_log.h> 76 #undef NFS 77 #include <sys/statvfs.h> 78 #include <sys/mount.h> 79 #include <sys/mntent.h> 80 #include <sys/swap.h> 81 #include <sys/errno.h> 82 #include <sys/debug.h> 83 #include "fs/fs_subr.h" 84 #include <sys/cmn_err.h> 85 #include <sys/dnlc.h> 86 #include <sys/fssnap_if.h> 87 #include <sys/sunddi.h> 88 #include <sys/bootconf.h> 89 #include <sys/policy.h> 90 #include <sys/zone.h> 91 92 /* 93 * This is the loadable module wrapper. 94 */ 95 #include <sys/modctl.h> 96 97 int ufsfstype; 98 vfsops_t *ufs_vfsops; 99 static int ufsinit(int, char *); 100 static int mountfs(); 101 extern int highbit(); 102 extern struct instats ins; 103 extern struct vnode *common_specvp(struct vnode *vp); 104 extern vfs_t EIO_vfs; 105 106 struct dquot *dquot, *dquotNDQUOT; 107 108 /* 109 * Cylinder group summary information handling tunable. 110 * This defines when these deltas get logged. 111 * If the number of cylinders in the file system is over the 112 * tunable then we log csum updates. Otherwise the updates are only 113 * done for performance on unmount. After a panic they can be 114 * quickly constructed during mounting. See ufs_construct_si() 115 * called from ufs_getsummaryinfo(). 116 * 117 * This performance feature can of course be disabled by setting 118 * ufs_ncg_log to 0, and fully enabled by setting it to 0xffffffff. 119 */ 120 #define UFS_LOG_NCG_DEFAULT 10000 121 uint32_t ufs_ncg_log = UFS_LOG_NCG_DEFAULT; 122 123 /* 124 * ufs_clean_root indicates whether the root fs went down cleanly 125 */ 126 static int ufs_clean_root = 0; 127 128 /* 129 * UFS Mount options table 130 */ 131 static char *intr_cancel[] = { MNTOPT_NOINTR, NULL }; 132 static char *nointr_cancel[] = { MNTOPT_INTR, NULL }; 133 static char *forcedirectio_cancel[] = { MNTOPT_NOFORCEDIRECTIO, NULL }; 134 static char *noforcedirectio_cancel[] = { MNTOPT_FORCEDIRECTIO, NULL }; 135 static char *largefiles_cancel[] = { MNTOPT_NOLARGEFILES, NULL }; 136 static char *nolargefiles_cancel[] = { MNTOPT_LARGEFILES, NULL }; 137 static char *logging_cancel[] = { MNTOPT_NOLOGGING, NULL }; 138 static char *nologging_cancel[] = { MNTOPT_LOGGING, NULL }; 139 static char *xattr_cancel[] = { MNTOPT_NOXATTR, NULL }; 140 static char *noxattr_cancel[] = { MNTOPT_XATTR, NULL }; 141 static char *quota_cancel[] = { MNTOPT_NOQUOTA, NULL }; 142 static char *noquota_cancel[] = { MNTOPT_QUOTA, NULL }; 143 static char *dfratime_cancel[] = { MNTOPT_NODFRATIME, NULL }; 144 static char *nodfratime_cancel[] = { MNTOPT_DFRATIME, NULL }; 145 146 static mntopt_t mntopts[] = { 147 /* 148 * option name cancel option default arg flags 149 * ufs arg flag 150 */ 151 { MNTOPT_INTR, intr_cancel, NULL, MO_DEFAULT, 152 (void *)0 }, 153 { MNTOPT_NOINTR, nointr_cancel, NULL, 0, 154 (void *)UFSMNT_NOINTR }, 155 { MNTOPT_SYNCDIR, NULL, NULL, 0, 156 (void *)UFSMNT_SYNCDIR }, 157 { MNTOPT_FORCEDIRECTIO, forcedirectio_cancel, NULL, 0, 158 (void *)UFSMNT_FORCEDIRECTIO }, 159 { MNTOPT_NOFORCEDIRECTIO, noforcedirectio_cancel, NULL, 0, 160 (void *)UFSMNT_NOFORCEDIRECTIO }, 161 { MNTOPT_NOSETSEC, NULL, NULL, 0, 162 (void *)UFSMNT_NOSETSEC }, 163 { MNTOPT_LARGEFILES, largefiles_cancel, NULL, MO_DEFAULT, 164 (void *)UFSMNT_LARGEFILES }, 165 { MNTOPT_NOLARGEFILES, nolargefiles_cancel, NULL, 0, 166 (void *)0 }, 167 { MNTOPT_LOGGING, logging_cancel, NULL, MO_TAG, 168 (void *)UFSMNT_LOGGING }, 169 { MNTOPT_NOLOGGING, nologging_cancel, NULL, 170 MO_NODISPLAY|MO_DEFAULT|MO_TAG, (void *)0 }, 171 { MNTOPT_QUOTA, quota_cancel, NULL, MO_IGNORE, 172 (void *)0 }, 173 { MNTOPT_NOQUOTA, noquota_cancel, NULL, 174 MO_NODISPLAY|MO_DEFAULT, (void *)0 }, 175 { MNTOPT_GLOBAL, NULL, NULL, 0, 176 (void *)0 }, 177 { MNTOPT_XATTR, xattr_cancel, NULL, MO_DEFAULT, 178 (void *)0 }, 179 { MNTOPT_NOXATTR, noxattr_cancel, NULL, 0, 180 (void *)0 }, 181 { MNTOPT_NOATIME, NULL, NULL, 0, 182 (void *)UFSMNT_NOATIME }, 183 { MNTOPT_DFRATIME, dfratime_cancel, NULL, 0, 184 (void *)0 }, 185 { MNTOPT_NODFRATIME, nodfratime_cancel, NULL, 186 MO_NODISPLAY|MO_DEFAULT, (void *)UFSMNT_NODFRATIME }, 187 { MNTOPT_ONERROR, NULL, UFSMNT_ONERROR_PANIC_STR, 188 MO_DEFAULT|MO_HASVALUE, (void *)0 }, 189 }; 190 191 static mntopts_t ufs_mntopts = { 192 sizeof (mntopts) / sizeof (mntopt_t), 193 mntopts 194 }; 195 196 static vfsdef_t vfw = { 197 VFSDEF_VERSION, 198 "ufs", 199 ufsinit, 200 VSW_HASPROTO|VSW_CANREMOUNT, 201 &ufs_mntopts 202 }; 203 204 /* 205 * Module linkage information for the kernel. 206 */ 207 extern struct mod_ops mod_fsops; 208 209 static struct modlfs modlfs = { 210 &mod_fsops, "filesystem for ufs", &vfw 211 }; 212 213 static struct modlinkage modlinkage = { 214 MODREV_1, (void *)&modlfs, NULL 215 }; 216 217 /* 218 * An attempt has been made to make this module unloadable. In order to 219 * test it, we need a system in which the root fs is NOT ufs. THIS HAS NOT 220 * BEEN DONE 221 */ 222 223 extern kstat_t *ufs_inode_kstat; 224 extern uint_t ufs_lockfs_key; 225 extern void ufs_lockfs_tsd_destructor(void *); 226 extern uint_t bypass_snapshot_throttle_key; 227 228 int 229 _init(void) 230 { 231 /* 232 * Create an index into the per thread array so that any thread doing 233 * VOP will have a lockfs mark on it. 234 */ 235 tsd_create(&ufs_lockfs_key, ufs_lockfs_tsd_destructor); 236 tsd_create(&bypass_snapshot_throttle_key, NULL); 237 return (mod_install(&modlinkage)); 238 } 239 240 int 241 _fini(void) 242 { 243 return (EBUSY); 244 } 245 246 int 247 _info(struct modinfo *modinfop) 248 { 249 return (mod_info(&modlinkage, modinfop)); 250 } 251 252 extern struct vnode *makespecvp(dev_t dev, vtype_t type); 253 254 extern kmutex_t ufs_scan_lock; 255 256 static int mountfs(struct vfs *, enum whymountroot, struct vnode *, char *, 257 struct cred *, int, void *, int); 258 259 260 static int 261 ufs_mount(struct vfs *vfsp, struct vnode *mvp, struct mounta *uap, 262 struct cred *cr) 263 264 { 265 char *data = uap->dataptr; 266 int datalen = uap->datalen; 267 dev_t dev; 268 struct vnode *bvp; 269 struct pathname dpn; 270 int error; 271 enum whymountroot why = ROOT_INIT; 272 struct ufs_args args; 273 int oflag, aflag; 274 int fromspace = (uap->flags & MS_SYSSPACE) ? 275 UIO_SYSSPACE : UIO_USERSPACE; 276 277 if ((error = secpolicy_fs_mount(cr, mvp, vfsp)) != 0) 278 return (error); 279 280 if (mvp->v_type != VDIR) 281 return (ENOTDIR); 282 283 mutex_enter(&mvp->v_lock); 284 if ((uap->flags & MS_REMOUNT) == 0 && 285 (uap->flags & MS_OVERLAY) == 0 && 286 (mvp->v_count != 1 || (mvp->v_flag & VROOT))) { 287 mutex_exit(&mvp->v_lock); 288 return (EBUSY); 289 } 290 mutex_exit(&mvp->v_lock); 291 292 /* 293 * Get arguments 294 */ 295 bzero(&args, sizeof (args)); 296 if ((uap->flags & MS_DATA) && data != NULL && datalen != 0) { 297 int copy_result = 0; 298 299 if (datalen > sizeof (args)) 300 return (EINVAL); 301 if (uap->flags & MS_SYSSPACE) 302 bcopy(data, &args, datalen); 303 else 304 copy_result = copyin(data, &args, datalen); 305 if (copy_result) 306 return (EFAULT); 307 datalen = sizeof (struct ufs_args); 308 } else { 309 datalen = 0; 310 } 311 /* 312 * Read in the mount point pathname 313 * (so we can record the directory the file system was last mounted on). 314 */ 315 if (error = pn_get(uap->dir, fromspace, &dpn)) 316 return (error); 317 318 /* 319 * Resolve path name of special file being mounted. 320 */ 321 if (error = lookupname(uap->spec, fromspace, FOLLOW, NULL, &bvp)) { 322 pn_free(&dpn); 323 return (error); 324 } 325 if (bvp->v_type != VBLK) { 326 VN_RELE(bvp); 327 pn_free(&dpn); 328 return (ENOTBLK); 329 } 330 dev = bvp->v_rdev; 331 if (getmajor(dev) >= devcnt) { 332 pn_free(&dpn); 333 VN_RELE(bvp); 334 return (ENXIO); 335 } 336 if (uap->flags & MS_REMOUNT) 337 why = ROOT_REMOUNT; 338 339 /* 340 * In SunCluster, requests to a global device are satisfied by 341 * a local device. We substitute the global pxfs node with a 342 * local spec node here. 343 */ 344 if (IS_PXFSVP(bvp)) { 345 VN_RELE(bvp); 346 bvp = makespecvp(dev, VBLK); 347 } 348 349 /* 350 * Open block device mounted on. We need this to 351 * check whether the caller has sufficient rights to 352 * access the device in question. 353 * When bio is fixed for vnodes this can all be vnode 354 * operations. 355 */ 356 if ((vfsp->vfs_flag & VFS_RDONLY) != 0 || 357 (uap->flags & MS_RDONLY) != 0) { 358 oflag = FREAD; 359 aflag = VREAD; 360 } else { 361 oflag = FREAD | FWRITE; 362 aflag = VREAD | VWRITE; 363 } 364 if ((error = VOP_ACCESS(bvp, aflag, 0, cr)) != 0 || 365 (error = secpolicy_spec_open(cr, bvp, oflag)) != 0) { 366 pn_free(&dpn); 367 VN_RELE(bvp); 368 return (error); 369 } 370 371 /* 372 * Ensure that this device isn't already mounted or in progress on a 373 * mount unless this is a REMOUNT request or we are told to suppress 374 * mount checks. Global mounts require special handling. 375 */ 376 if ((uap->flags & MS_NOCHECK) == 0) { 377 if ((uap->flags & MS_GLOBAL) == 0 && 378 vfs_devmounting(dev, vfsp)) { 379 pn_free(&dpn); 380 VN_RELE(bvp); 381 return (EBUSY); 382 } 383 if (vfs_devismounted(dev)) { 384 if ((uap->flags & MS_REMOUNT) == 0) { 385 pn_free(&dpn); 386 VN_RELE(bvp); 387 return (EBUSY); 388 } 389 } 390 } 391 392 /* 393 * If the device is a tape, mount it read only 394 */ 395 if (devopsp[getmajor(dev)]->devo_cb_ops->cb_flag & D_TAPE) { 396 vfsp->vfs_flag |= VFS_RDONLY; 397 vfs_setmntopt(vfsp, MNTOPT_RO, NULL, 0); 398 } 399 if (uap->flags & MS_RDONLY) 400 vfsp->vfs_flag |= VFS_RDONLY; 401 402 /* 403 * Mount the filesystem, free the device vnode on error. 404 */ 405 error = mountfs(vfsp, why, bvp, dpn.pn_path, cr, 0, &args, datalen); 406 pn_free(&dpn); 407 if (error) { 408 VN_RELE(bvp); 409 } 410 return (error); 411 } 412 /* 413 * Mount root file system. 414 * "why" is ROOT_INIT on initial call ROOT_REMOUNT if called to 415 * remount the root file system, and ROOT_UNMOUNT if called to 416 * unmount the root (e.g., as part of a system shutdown). 417 * 418 * XXX - this may be partially machine-dependent; it, along with the VFS_SWAPVP 419 * operation, goes along with auto-configuration. A mechanism should be 420 * provided by which machine-INdependent code in the kernel can say "get me the 421 * right root file system" and "get me the right initial swap area", and have 422 * that done in what may well be a machine-dependent fashion. 423 * Unfortunately, it is also file-system-type dependent (NFS gets it via 424 * bootparams calls, UFS gets it from various and sundry machine-dependent 425 * mechanisms, as SPECFS does for swap). 426 */ 427 static int 428 ufs_mountroot(struct vfs *vfsp, enum whymountroot why) 429 { 430 struct fs *fsp; 431 int error; 432 static int ufsrootdone = 0; 433 dev_t rootdev; 434 struct vnode *vp; 435 struct vnode *devvp = 0; 436 int ovflags; 437 int doclkset; 438 ufsvfs_t *ufsvfsp; 439 440 if (why == ROOT_INIT) { 441 if (ufsrootdone++) 442 return (EBUSY); 443 rootdev = getrootdev(); 444 if (rootdev == (dev_t)NODEV) 445 return (ENODEV); 446 vfsp->vfs_dev = rootdev; 447 vfsp->vfs_flag |= VFS_RDONLY; 448 } else if (why == ROOT_REMOUNT) { 449 vp = ((struct ufsvfs *)vfsp->vfs_data)->vfs_devvp; 450 (void) dnlc_purge_vfsp(vfsp, 0); 451 vp = common_specvp(vp); 452 (void) VOP_PUTPAGE(vp, (offset_t)0, (size_t)0, B_INVAL, CRED()); 453 (void) bfinval(vfsp->vfs_dev, 0); 454 fsp = getfs(vfsp); 455 456 ovflags = vfsp->vfs_flag; 457 vfsp->vfs_flag &= ~VFS_RDONLY; 458 vfsp->vfs_flag |= VFS_REMOUNT; 459 rootdev = vfsp->vfs_dev; 460 } else if (why == ROOT_UNMOUNT) { 461 if (vfs_lock(vfsp) == 0) { 462 (void) ufs_flush(vfsp); 463 /* 464 * Mark the log as fully rolled 465 */ 466 ufsvfsp = (ufsvfs_t *)vfsp->vfs_data; 467 fsp = ufsvfsp->vfs_fs; 468 if (TRANS_ISTRANS(ufsvfsp) && 469 !TRANS_ISERROR(ufsvfsp) && 470 (fsp->fs_rolled == FS_NEED_ROLL)) { 471 ml_unit_t *ul = ufsvfsp->vfs_log; 472 473 error = ufs_putsummaryinfo(ul->un_dev, 474 ufsvfsp, fsp); 475 if (error == 0) { 476 fsp->fs_rolled = FS_ALL_ROLLED; 477 UFS_BWRITE2(NULL, ufsvfsp->vfs_bufp); 478 } 479 } 480 vfs_unlock(vfsp); 481 } else { 482 ufs_update(0); 483 } 484 485 vp = ((struct ufsvfs *)vfsp->vfs_data)->vfs_devvp; 486 (void) VOP_CLOSE(vp, FREAD|FWRITE, 1, 487 (offset_t)0, CRED()); 488 return (0); 489 } 490 error = vfs_lock(vfsp); 491 if (error) 492 return (error); 493 494 devvp = makespecvp(rootdev, VBLK); 495 496 /* If RO media, don't call clkset() (see below) */ 497 doclkset = 1; 498 if (why == ROOT_INIT) { 499 error = VOP_OPEN(&devvp, FREAD|FWRITE, CRED()); 500 if (error == 0) { 501 (void) VOP_CLOSE(devvp, FREAD|FWRITE, 1, 502 (offset_t)0, CRED()); 503 } else { 504 doclkset = 0; 505 } 506 } 507 508 error = mountfs(vfsp, why, devvp, "/", CRED(), 1, NULL, 0); 509 /* 510 * XXX - assumes root device is not indirect, because we don't set 511 * rootvp. Is rootvp used for anything? If so, make another arg 512 * to mountfs. 513 */ 514 if (error) { 515 vfs_unlock(vfsp); 516 if (why == ROOT_REMOUNT) 517 vfsp->vfs_flag = ovflags; 518 if (rootvp) { 519 VN_RELE(rootvp); 520 rootvp = (struct vnode *)0; 521 } 522 VN_RELE(devvp); 523 return (error); 524 } 525 if (why == ROOT_INIT) 526 vfs_add((struct vnode *)0, vfsp, 527 (vfsp->vfs_flag & VFS_RDONLY) ? MS_RDONLY : 0); 528 vfs_unlock(vfsp); 529 fsp = getfs(vfsp); 530 clkset(doclkset ? fsp->fs_time : -1); 531 ufsvfsp = (ufsvfs_t *)vfsp->vfs_data; 532 if (ufsvfsp->vfs_log) { 533 vfs_setmntopt(vfsp, MNTOPT_LOGGING, NULL, 0); 534 } 535 return (0); 536 } 537 538 static int 539 remountfs(struct vfs *vfsp, dev_t dev, void *raw_argsp, int args_len) 540 { 541 struct ufsvfs *ufsvfsp = (struct ufsvfs *)vfsp->vfs_data; 542 struct ulockfs *ulp = &ufsvfsp->vfs_ulockfs; 543 struct buf *bp = ufsvfsp->vfs_bufp; 544 struct fs *fsp = (struct fs *)bp->b_un.b_addr; 545 struct fs *fspt; 546 struct buf *tpt = 0; 547 int error = 0; 548 int flags = 0; 549 550 if (args_len == sizeof (struct ufs_args) && raw_argsp) 551 flags = ((struct ufs_args *)raw_argsp)->flags; 552 553 /* cannot remount to RDONLY */ 554 if (vfsp->vfs_flag & VFS_RDONLY) 555 return (EINVAL); 556 557 /* whoops, wrong dev */ 558 if (vfsp->vfs_dev != dev) 559 return (EINVAL); 560 561 /* 562 * synchronize w/ufs ioctls 563 */ 564 mutex_enter(&ulp->ul_lock); 565 atomic_add_long(&ufs_quiesce_pend, 1); 566 567 /* 568 * reset options 569 */ 570 ufsvfsp->vfs_nointr = flags & UFSMNT_NOINTR; 571 ufsvfsp->vfs_syncdir = flags & UFSMNT_SYNCDIR; 572 ufsvfsp->vfs_nosetsec = flags & UFSMNT_NOSETSEC; 573 ufsvfsp->vfs_noatime = flags & UFSMNT_NOATIME; 574 if ((flags & UFSMNT_NODFRATIME) || ufsvfsp->vfs_noatime) 575 ufsvfsp->vfs_dfritime &= ~UFS_DFRATIME; 576 else /* dfratime, default behavior */ 577 ufsvfsp->vfs_dfritime |= UFS_DFRATIME; 578 if (flags & UFSMNT_FORCEDIRECTIO) 579 ufsvfsp->vfs_forcedirectio = 1; 580 else /* default is no direct I/O */ 581 ufsvfsp->vfs_forcedirectio = 0; 582 ufsvfsp->vfs_iotstamp = lbolt; 583 584 /* 585 * set largefiles flag in ufsvfs equal to the 586 * value passed in by the mount command. If 587 * it is "nolargefiles", and the flag is set 588 * in the superblock, the mount fails. 589 */ 590 if (!(flags & UFSMNT_LARGEFILES)) { /* "nolargefiles" */ 591 if (fsp->fs_flags & FSLARGEFILES) { 592 error = EFBIG; 593 goto remounterr; 594 } 595 ufsvfsp->vfs_lfflags &= ~UFS_LARGEFILES; 596 } else /* "largefiles" */ 597 ufsvfsp->vfs_lfflags |= UFS_LARGEFILES; 598 /* 599 * read/write to read/write; all done 600 */ 601 if (fsp->fs_ronly == 0) 602 goto remounterr; 603 604 /* 605 * fix-on-panic assumes RO->RW remount implies system-critical fs 606 * if it is shortly after boot; so, don't attempt to lock and fix 607 * (unless the user explicitly asked for another action on error) 608 * XXX UFSMNT_ONERROR_RDONLY rather than UFSMNT_ONERROR_PANIC 609 */ 610 #define BOOT_TIME_LIMIT (180*hz) 611 if (!(flags & UFSMNT_ONERROR_FLGMASK) && lbolt < BOOT_TIME_LIMIT) { 612 cmn_err(CE_WARN, "%s is required to be mounted onerror=%s", 613 ufsvfsp->vfs_fs->fs_fsmnt, UFSMNT_ONERROR_PANIC_STR); 614 flags |= UFSMNT_ONERROR_PANIC; 615 } 616 617 if ((error = ufsfx_mount(ufsvfsp, flags)) != 0) 618 goto remounterr; 619 620 /* 621 * Lock the file system and flush stuff from memory 622 */ 623 error = ufs_quiesce(ulp); 624 if (error) 625 goto remounterr; 626 627 /* 628 * We don't need to call the expensive ufs_flush when going from 629 * read only to read/write, except if the root fs didn't come 630 * down cleanly. 631 */ 632 if ((ufsvfsp->vfs_devvp == rootvp) && !ufs_clean_root) { 633 error = ufs_flush(vfsp); 634 if (error) { 635 goto remounterr; 636 } 637 } 638 639 tpt = UFS_BREAD(ufsvfsp, ufsvfsp->vfs_dev, SBLOCK, SBSIZE); 640 if (tpt->b_flags & B_ERROR) { 641 error = EIO; 642 goto remounterr; 643 } 644 fspt = (struct fs *)tpt->b_un.b_addr; 645 if (((fspt->fs_magic != FS_MAGIC) && 646 (fspt->fs_magic != MTB_UFS_MAGIC)) || 647 (fspt->fs_magic == MTB_UFS_MAGIC && 648 (fspt->fs_version > MTB_UFS_VERSION_1 || 649 fspt->fs_version < MTB_UFS_VERSION_MIN)) || 650 fspt->fs_bsize > MAXBSIZE || fspt->fs_frag > MAXFRAG || 651 fspt->fs_bsize < sizeof (struct fs) || fspt->fs_bsize < PAGESIZE) { 652 tpt->b_flags |= B_STALE | B_AGE; 653 error = EINVAL; 654 goto remounterr; 655 } 656 657 if (ufsvfsp->vfs_log && (ufsvfsp->vfs_log->un_flags & LDL_NOROLL)) { 658 ufsvfsp->vfs_log->un_flags &= ~LDL_NOROLL; 659 logmap_start_roll(ufsvfsp->vfs_log); 660 } 661 662 if (TRANS_ISERROR(ufsvfsp)) 663 goto remounterr; 664 TRANS_DOMATAMAP(ufsvfsp); 665 666 if ((fspt->fs_state + fspt->fs_time == FSOKAY) && 667 fspt->fs_clean == FSLOG && !TRANS_ISTRANS(ufsvfsp)) { 668 ufsvfsp->vfs_log = NULL; 669 ufsvfsp->vfs_domatamap = 0; 670 error = ENOSPC; 671 goto remounterr; 672 } 673 674 if (fspt->fs_state + fspt->fs_time == FSOKAY && 675 (fspt->fs_clean == FSCLEAN || 676 fspt->fs_clean == FSSTABLE || 677 fspt->fs_clean == FSLOG)) { 678 679 /* 680 * Ensure that ufs_getsummaryinfo doesn't reconstruct 681 * the summary info. 682 */ 683 error = ufs_getsummaryinfo(vfsp->vfs_dev, ufsvfsp, fspt); 684 if (error) 685 goto remounterr; 686 687 /* preserve mount name */ 688 (void) strncpy(fspt->fs_fsmnt, fsp->fs_fsmnt, MAXMNTLEN); 689 /* free the old cg space */ 690 kmem_free(fsp->fs_u.fs_csp, fsp->fs_cssize); 691 /* switch in the new superblock */ 692 fspt->fs_rolled = FS_NEED_ROLL; 693 bcopy(tpt->b_un.b_addr, bp->b_un.b_addr, fspt->fs_sbsize); 694 695 fsp->fs_clean = FSSTABLE; 696 } /* superblock updated in memory */ 697 tpt->b_flags |= B_STALE | B_AGE; 698 brelse(tpt); 699 tpt = 0; 700 701 if (fsp->fs_clean != FSSTABLE) { 702 error = ENOSPC; 703 goto remounterr; 704 } 705 706 707 if (TRANS_ISTRANS(ufsvfsp)) { 708 fsp->fs_clean = FSLOG; 709 ufsvfsp->vfs_dio = 0; 710 } else 711 if (ufsvfsp->vfs_dio) 712 fsp->fs_clean = FSSUSPEND; 713 714 TRANS_MATA_MOUNT(ufsvfsp); 715 716 fsp->fs_fmod = 0; 717 fsp->fs_ronly = 0; 718 719 atomic_add_long(&ufs_quiesce_pend, -1); 720 cv_broadcast(&ulp->ul_cv); 721 mutex_exit(&ulp->ul_lock); 722 723 if (TRANS_ISTRANS(ufsvfsp)) { 724 725 /* 726 * start the delete thread 727 */ 728 ufs_thread_start(&ufsvfsp->vfs_delete, ufs_thread_delete, vfsp); 729 730 /* 731 * start the reclaim thread 732 */ 733 if (fsp->fs_reclaim & (FS_RECLAIM|FS_RECLAIMING)) { 734 fsp->fs_reclaim &= ~FS_RECLAIM; 735 fsp->fs_reclaim |= FS_RECLAIMING; 736 ufs_thread_start(&ufsvfsp->vfs_reclaim, 737 ufs_thread_reclaim, vfsp); 738 } 739 } 740 741 TRANS_SBWRITE(ufsvfsp, TOP_MOUNT); 742 743 return (0); 744 745 remounterr: 746 if (tpt) 747 brelse(tpt); 748 atomic_add_long(&ufs_quiesce_pend, -1); 749 cv_broadcast(&ulp->ul_cv); 750 mutex_exit(&ulp->ul_lock); 751 return (error); 752 } 753 754 /* 755 * If the device maxtransfer size is not available, we use ufs_maxmaxphys 756 * along with the system value for maxphys to determine the value for 757 * maxtransfer. 758 */ 759 int ufs_maxmaxphys = (1024 * 1024); 760 761 #include <sys/ddi.h> /* for delay(9f) */ 762 763 int ufs_mount_error_delay = 20; /* default to 20ms */ 764 int ufs_mount_timeout = 60; /* default to 1 minute */ 765 766 static int 767 mountfs(struct vfs *vfsp, enum whymountroot why, struct vnode *devvp, 768 char *path, cred_t *cr, int isroot, void *raw_argsp, int args_len) 769 { 770 dev_t dev = devvp->v_rdev; 771 struct fs *fsp; 772 struct ufsvfs *ufsvfsp = 0; 773 struct buf *bp = 0; 774 struct buf *tp = 0; 775 struct dk_cinfo ci; 776 int error = 0; 777 size_t len; 778 int needclose = 0; 779 int needtrans = 0; 780 struct inode *rip; 781 struct vnode *rvp = NULL; 782 int flags = 0; 783 kmutex_t *ihm; 784 int elapsed; 785 int status; 786 extern int maxphys; 787 788 if (args_len == sizeof (struct ufs_args) && raw_argsp) 789 flags = ((struct ufs_args *)raw_argsp)->flags; 790 791 ASSERT(vfs_lock_held(vfsp)); 792 793 if (why == ROOT_INIT) { 794 /* 795 * Open block device mounted on. 796 * When bio is fixed for vnodes this can all be vnode 797 * operations. 798 */ 799 error = VOP_OPEN(&devvp, 800 (vfsp->vfs_flag & VFS_RDONLY) ? FREAD : FREAD|FWRITE, cr); 801 if (error) 802 goto out; 803 needclose = 1; 804 805 /* 806 * Refuse to go any further if this 807 * device is being used for swapping. 808 */ 809 if (IS_SWAPVP(devvp)) { 810 error = EBUSY; 811 goto out; 812 } 813 } 814 815 /* 816 * check for dev already mounted on 817 */ 818 if (vfsp->vfs_flag & VFS_REMOUNT) { 819 error = remountfs(vfsp, dev, raw_argsp, args_len); 820 if (error == 0) 821 VN_RELE(devvp); 822 return (error); 823 } 824 825 ASSERT(devvp != 0); 826 827 /* 828 * Flush back any dirty pages on the block device to 829 * try and keep the buffer cache in sync with the page 830 * cache if someone is trying to use block devices when 831 * they really should be using the raw device. 832 */ 833 (void) VOP_PUTPAGE(common_specvp(devvp), (offset_t)0, 834 (size_t)0, B_INVAL, cr); 835 836 /* 837 * read in superblock 838 */ 839 ufsvfsp = kmem_zalloc(sizeof (struct ufsvfs), KM_SLEEP); 840 tp = UFS_BREAD(ufsvfsp, dev, SBLOCK, SBSIZE); 841 if (tp->b_flags & B_ERROR) 842 goto out; 843 fsp = (struct fs *)tp->b_un.b_addr; 844 #ifdef _LP64 845 if ((fsp->fs_magic != FS_MAGIC) && (fsp->fs_magic != MTB_UFS_MAGIC)) { 846 if (why == ROOT_INIT) { 847 cmn_err(CE_NOTE, 848 "mount: not a UFS magic number (0x%x)", 849 fsp->fs_magic); 850 } 851 error = EINVAL; 852 goto out; 853 } 854 855 if ((fsp->fs_magic == MTB_UFS_MAGIC) && 856 (fsp->fs_version > MTB_UFS_VERSION_1 || 857 fsp->fs_version < MTB_UFS_VERSION_MIN)) { 858 cmn_err(CE_NOTE, 859 "mount: unrecognized version of UFS on-disk format: %d", 860 fsp->fs_version); 861 error = EINVAL; 862 goto out; 863 } 864 865 #else 866 if (fsp->fs_magic == MTB_UFS_MAGIC) { 867 if (fsp->fs_version > MTB_UFS_VERSION_1 || 868 fsp->fs_version < MTB_UFS_VERSION_MIN) { 869 cmn_err(CE_NOTE, 870 "mount: unrecognized version of UFS" 871 " on-disk format: %d", fsp->fs_version); 872 error = EINVAL; 873 goto out; 874 } 875 876 /* 877 * Find the size of the device in sectors. If the 878 * the size in sectors is greater than INT_MAX, it's 879 * a multi-terabyte file system, which can't be 880 * mounted by a 32-bit kernel. We can't use the 881 * fsbtodb() macro in the next line because the macro 882 * casts the intermediate values to daddr_t, which is 883 * a 32-bit quantity in a 32-bit kernel. Here we 884 * really do need the intermediate values to be held 885 * in 64-bit quantities because we're checking for 886 * overflow of a 32-bit field. 887 */ 888 if ((((diskaddr_t)(fsp->fs_size)) << fsp->fs_fsbtodb) 889 > INT_MAX) { 890 cmn_err(CE_NOTE, 891 "mount: multi-terabyte UFS cannot be" 892 " mounted by a 32-bit kernel"); 893 error = EINVAL; 894 goto out; 895 } 896 897 } else if (fsp->fs_magic != FS_MAGIC) { 898 cmn_err(CE_NOTE, 899 "mount: not a UFS magic number (0x%x)", fsp->fs_magic); 900 error = EINVAL; 901 goto out; 902 } 903 #endif /* _LP64 */ 904 905 if (fsp->fs_bsize > MAXBSIZE || fsp->fs_frag > MAXFRAG || 906 fsp->fs_bsize < sizeof (struct fs) || fsp->fs_bsize < PAGESIZE) { 907 error = EINVAL; /* also needs translation */ 908 goto out; 909 } 910 911 /* 912 * Allocate VFS private data. 913 */ 914 vfsp->vfs_bcount = 0; 915 vfsp->vfs_data = (caddr_t)ufsvfsp; 916 vfsp->vfs_fstype = ufsfstype; 917 vfsp->vfs_dev = dev; 918 vfsp->vfs_flag |= VFS_NOTRUNC; 919 vfs_make_fsid(&vfsp->vfs_fsid, dev, ufsfstype); 920 ufsvfsp->vfs_devvp = devvp; 921 922 /* 923 * Cross-link with vfs and add to instance list. 924 */ 925 ufsvfsp->vfs_vfs = vfsp; 926 ufs_vfs_add(ufsvfsp); 927 928 ufsvfsp->vfs_dev = dev; 929 ufsvfsp->vfs_bufp = tp; 930 931 ufsvfsp->vfs_dirsize = INODESIZE + (4 * ALLOCSIZE) + fsp->fs_fsize; 932 ufsvfsp->vfs_minfrags = (int)((int64_t)fsp->fs_dsize * 933 fsp->fs_minfree / 100); 934 /* 935 * if mount allows largefiles, indicate so in ufsvfs 936 */ 937 if (flags & UFSMNT_LARGEFILES) 938 ufsvfsp->vfs_lfflags |= UFS_LARGEFILES; 939 /* 940 * Initialize threads 941 */ 942 ufs_thread_init(&ufsvfsp->vfs_delete, 1); 943 ufs_thread_init(&ufsvfsp->vfs_reclaim, 0); 944 945 /* 946 * Chicken and egg problem. The superblock may have deltas 947 * in the log. So after the log is scanned we reread the 948 * superblock. We guarantee that the fields needed to 949 * scan the log will not be in the log. 950 */ 951 if (fsp->fs_logbno && fsp->fs_clean == FSLOG && 952 (fsp->fs_state + fsp->fs_time == FSOKAY)) { 953 error = lufs_snarf(ufsvfsp, fsp, (vfsp->vfs_flag & VFS_RDONLY)); 954 if (error) { 955 /* 956 * Allow a ro mount to continue even if the 957 * log cannot be processed - yet. 958 */ 959 if (!(vfsp->vfs_flag & VFS_RDONLY)) { 960 cmn_err(CE_WARN, "Error accessing ufs " 961 "log for %s; Please run fsck(1M)", 962 path); 963 goto out; 964 } 965 } 966 tp->b_flags |= (B_AGE | B_STALE); 967 brelse(tp); 968 tp = UFS_BREAD(ufsvfsp, dev, SBLOCK, SBSIZE); 969 fsp = (struct fs *)tp->b_un.b_addr; 970 ufsvfsp->vfs_bufp = tp; 971 if (tp->b_flags & B_ERROR) 972 goto out; 973 } 974 975 /* 976 * Set logging mounted flag used by lockfs 977 */ 978 ufsvfsp->vfs_validfs = UT_MOUNTED; 979 980 /* 981 * Copy the super block into a buffer in its native size. 982 * Use ngeteblk to allocate the buffer 983 */ 984 bp = ngeteblk(fsp->fs_bsize); 985 ufsvfsp->vfs_bufp = bp; 986 bp->b_edev = dev; 987 bp->b_dev = cmpdev(dev); 988 bp->b_blkno = SBLOCK; 989 bp->b_bcount = fsp->fs_sbsize; 990 bcopy(tp->b_un.b_addr, bp->b_un.b_addr, fsp->fs_sbsize); 991 tp->b_flags |= B_STALE | B_AGE; 992 brelse(tp); 993 tp = 0; 994 995 fsp = (struct fs *)bp->b_un.b_addr; 996 /* 997 * Mount fails if superblock flag indicates presence of large 998 * files and filesystem is attempted to be mounted 'nolargefiles'. 999 * The exception is for a read only mount of root, which we 1000 * always want to succeed, so fsck can fix potential problems. 1001 * The assumption is that we will remount root at some point, 1002 * and the remount will enforce the mount option. 1003 */ 1004 if (!(isroot & (vfsp->vfs_flag & VFS_RDONLY)) && 1005 (fsp->fs_flags & FSLARGEFILES) && 1006 !(flags & UFSMNT_LARGEFILES)) { 1007 error = EFBIG; 1008 goto out; 1009 } 1010 1011 if (vfsp->vfs_flag & VFS_RDONLY) { 1012 fsp->fs_ronly = 1; 1013 fsp->fs_fmod = 0; 1014 if (((fsp->fs_state + fsp->fs_time) == FSOKAY) && 1015 ((fsp->fs_clean == FSCLEAN) || 1016 (fsp->fs_clean == FSSTABLE) || 1017 (fsp->fs_clean == FSLOG))) { 1018 if (isroot) { 1019 if (fsp->fs_clean == FSLOG) { 1020 if (fsp->fs_rolled == FS_ALL_ROLLED) { 1021 ufs_clean_root = 1; 1022 } 1023 } else { 1024 ufs_clean_root = 1; 1025 } 1026 } 1027 fsp->fs_clean = FSSTABLE; 1028 } else { 1029 fsp->fs_clean = FSBAD; 1030 } 1031 } else { 1032 1033 fsp->fs_fmod = 0; 1034 fsp->fs_ronly = 0; 1035 1036 TRANS_DOMATAMAP(ufsvfsp); 1037 1038 if ((TRANS_ISERROR(ufsvfsp)) || 1039 (((fsp->fs_state + fsp->fs_time) == FSOKAY) && 1040 fsp->fs_clean == FSLOG && !TRANS_ISTRANS(ufsvfsp))) { 1041 ufsvfsp->vfs_log = NULL; 1042 ufsvfsp->vfs_domatamap = 0; 1043 error = ENOSPC; 1044 goto out; 1045 } 1046 1047 if (((fsp->fs_state + fsp->fs_time) == FSOKAY) && 1048 (fsp->fs_clean == FSCLEAN || 1049 fsp->fs_clean == FSSTABLE || 1050 fsp->fs_clean == FSLOG)) 1051 fsp->fs_clean = FSSTABLE; 1052 else { 1053 if (isroot) { 1054 /* 1055 * allow root partition to be mounted even 1056 * when fs_state is not ok 1057 * will be fixed later by a remount root 1058 */ 1059 fsp->fs_clean = FSBAD; 1060 ufsvfsp->vfs_log = NULL; 1061 ufsvfsp->vfs_domatamap = 0; 1062 } else { 1063 error = ENOSPC; 1064 goto out; 1065 } 1066 } 1067 1068 if (fsp->fs_clean == FSSTABLE && TRANS_ISTRANS(ufsvfsp)) 1069 fsp->fs_clean = FSLOG; 1070 } 1071 TRANS_MATA_MOUNT(ufsvfsp); 1072 needtrans = 1; 1073 1074 vfsp->vfs_bsize = fsp->fs_bsize; 1075 1076 /* 1077 * Read in summary info 1078 */ 1079 if (error = ufs_getsummaryinfo(dev, ufsvfsp, fsp)) 1080 goto out; 1081 1082 /* 1083 * lastwhinetime is set to zero rather than lbolt, so that after 1084 * mounting if the filesystem is found to be full, then immediately the 1085 * "file system message" will be logged. 1086 */ 1087 ufsvfsp->vfs_lastwhinetime = 0L; 1088 1089 1090 mutex_init(&ufsvfsp->vfs_lock, NULL, MUTEX_DEFAULT, NULL); 1091 (void) copystr(path, fsp->fs_fsmnt, sizeof (fsp->fs_fsmnt) - 1, &len); 1092 bzero(fsp->fs_fsmnt + len, sizeof (fsp->fs_fsmnt) - len); 1093 1094 /* 1095 * Sanity checks for old file systems 1096 */ 1097 if (fsp->fs_postblformat == FS_42POSTBLFMT) 1098 ufsvfsp->vfs_nrpos = 8; 1099 else 1100 ufsvfsp->vfs_nrpos = fsp->fs_nrpos; 1101 1102 /* 1103 * Initialize lockfs structure to support file system locking 1104 */ 1105 bzero(&ufsvfsp->vfs_ulockfs.ul_lockfs, 1106 sizeof (struct lockfs)); 1107 ufsvfsp->vfs_ulockfs.ul_fs_lock = ULOCKFS_ULOCK; 1108 mutex_init(&ufsvfsp->vfs_ulockfs.ul_lock, NULL, 1109 MUTEX_DEFAULT, NULL); 1110 cv_init(&ufsvfsp->vfs_ulockfs.ul_cv, NULL, CV_DEFAULT, NULL); 1111 1112 /* 1113 * We don't need to grab vfs_dqrwlock for this ufs_iget() call. 1114 * We are in the process of mounting the file system so there 1115 * is no need to grab the quota lock. If a quota applies to the 1116 * root inode, then it will be updated when quotas are enabled. 1117 * 1118 * However, we have an ASSERT(RW_LOCK_HELD(&ufsvfsp->vfs_dqrwlock)) 1119 * in getinoquota() that we want to keep so grab it anyway. 1120 */ 1121 rw_enter(&ufsvfsp->vfs_dqrwlock, RW_READER); 1122 1123 error = ufs_iget_alloced(vfsp, UFSROOTINO, &rip, cr); 1124 1125 rw_exit(&ufsvfsp->vfs_dqrwlock); 1126 1127 if (error) 1128 goto out; 1129 1130 /* 1131 * make sure root inode is a directory. Returning ENOTDIR might 1132 * be confused with the mount point not being a directory, so 1133 * we use EIO instead. 1134 */ 1135 if ((rip->i_mode & IFMT) != IFDIR) { 1136 /* 1137 * Mark this inode as subject for cleanup 1138 * to avoid stray inodes in the cache. 1139 */ 1140 rvp = ITOV(rip); 1141 error = EIO; 1142 goto out; 1143 } 1144 1145 rvp = ITOV(rip); 1146 mutex_enter(&rvp->v_lock); 1147 rvp->v_flag |= VROOT; 1148 mutex_exit(&rvp->v_lock); 1149 ufsvfsp->vfs_root = rvp; 1150 /* The buffer for the root inode does not contain a valid b_vp */ 1151 (void) bfinval(dev, 0); 1152 1153 /* options */ 1154 ufsvfsp->vfs_nosetsec = flags & UFSMNT_NOSETSEC; 1155 ufsvfsp->vfs_nointr = flags & UFSMNT_NOINTR; 1156 ufsvfsp->vfs_syncdir = flags & UFSMNT_SYNCDIR; 1157 ufsvfsp->vfs_noatime = flags & UFSMNT_NOATIME; 1158 if ((flags & UFSMNT_NODFRATIME) || ufsvfsp->vfs_noatime) 1159 ufsvfsp->vfs_dfritime &= ~UFS_DFRATIME; 1160 else /* dfratime, default behavior */ 1161 ufsvfsp->vfs_dfritime |= UFS_DFRATIME; 1162 if (flags & UFSMNT_FORCEDIRECTIO) 1163 ufsvfsp->vfs_forcedirectio = 1; 1164 else if (flags & UFSMNT_NOFORCEDIRECTIO) 1165 ufsvfsp->vfs_forcedirectio = 0; 1166 ufsvfsp->vfs_iotstamp = lbolt; 1167 1168 ufsvfsp->vfs_nindiroffset = fsp->fs_nindir - 1; 1169 ufsvfsp->vfs_nindirshift = highbit(ufsvfsp->vfs_nindiroffset); 1170 ufsvfsp->vfs_ioclustsz = fsp->fs_bsize * fsp->fs_maxcontig; 1171 1172 if (cdev_ioctl(dev, DKIOCINFO, (intptr_t)&ci, 1173 FKIOCTL|FNATIVE|FREAD, CRED(), &status) == 0) { 1174 ufsvfsp->vfs_iotransz = ci.dki_maxtransfer * DEV_BSIZE; 1175 } else { 1176 ufsvfsp->vfs_iotransz = MIN(maxphys, ufs_maxmaxphys); 1177 } 1178 1179 if (ufsvfsp->vfs_iotransz <= 0) { 1180 ufsvfsp->vfs_iotransz = MIN(maxphys, ufs_maxmaxphys); 1181 } 1182 1183 /* 1184 * When logging, used to reserve log space for writes and truncs 1185 */ 1186 ufsvfsp->vfs_avgbfree = fsp->fs_cstotal.cs_nbfree / fsp->fs_ncg; 1187 1188 /* 1189 * Determine whether to log cylinder group summary info. 1190 */ 1191 ufsvfsp->vfs_nolog_si = (fsp->fs_ncg < ufs_ncg_log); 1192 1193 if (TRANS_ISTRANS(ufsvfsp)) { 1194 /* 1195 * start the delete thread 1196 */ 1197 ufs_thread_start(&ufsvfsp->vfs_delete, ufs_thread_delete, vfsp); 1198 1199 /* 1200 * start reclaim thread if the filesystem was not mounted 1201 * read only. 1202 */ 1203 if (!fsp->fs_ronly && (fsp->fs_reclaim & 1204 (FS_RECLAIM|FS_RECLAIMING))) { 1205 fsp->fs_reclaim &= ~FS_RECLAIM; 1206 fsp->fs_reclaim |= FS_RECLAIMING; 1207 ufs_thread_start(&ufsvfsp->vfs_reclaim, 1208 ufs_thread_reclaim, vfsp); 1209 } 1210 1211 /* Mark the fs as unrolled */ 1212 fsp->fs_rolled = FS_NEED_ROLL; 1213 } else if (!fsp->fs_ronly && (fsp->fs_reclaim & 1214 (FS_RECLAIM|FS_RECLAIMING))) { 1215 /* 1216 * If a file system that is mounted nologging, after 1217 * having previously been mounted logging, becomes 1218 * unmounted whilst the reclaim thread is in the throes 1219 * of reclaiming open/deleted inodes, a subsequent mount 1220 * of such a file system with logging disabled could lead 1221 * to inodes becoming lost. So, start reclaim now, even 1222 * though logging was disabled for the previous mount, to 1223 * tidy things up. 1224 */ 1225 fsp->fs_reclaim &= ~FS_RECLAIM; 1226 fsp->fs_reclaim |= FS_RECLAIMING; 1227 ufs_thread_start(&ufsvfsp->vfs_reclaim, 1228 ufs_thread_reclaim, vfsp); 1229 } 1230 1231 if (!fsp->fs_ronly) { 1232 TRANS_SBWRITE(ufsvfsp, TOP_MOUNT); 1233 if (error = geterror(ufsvfsp->vfs_bufp)) 1234 goto out; 1235 } 1236 1237 /* fix-on-panic initialization */ 1238 if (isroot && !(flags & UFSMNT_ONERROR_FLGMASK)) 1239 flags |= UFSMNT_ONERROR_PANIC; /* XXX ..._RDONLY */ 1240 1241 if ((error = ufsfx_mount(ufsvfsp, flags)) != 0) 1242 goto out; 1243 1244 if (why == ROOT_INIT && isroot) 1245 rootvp = devvp; 1246 1247 return (0); 1248 out: 1249 if (error == 0) 1250 error = EIO; 1251 if (rvp) { 1252 /* the following sequence is similar to ufs_unmount() */ 1253 1254 /* 1255 * There's a problem that ufs_iget() puts inodes into 1256 * the inode cache before it returns them. If someone 1257 * traverses that cache and gets a reference to our 1258 * inode, there's a chance they'll still be using it 1259 * after we've destroyed it. This is a hard race to 1260 * hit, but it's happened (putting in a medium delay 1261 * here, and a large delay in ufs_scan_inodes() for 1262 * inodes on the device we're bailing out on, makes 1263 * the race easy to demonstrate). The symptom is some 1264 * other part of UFS faulting on bad inode contents, 1265 * or when grabbing one of the locks inside the inode, 1266 * etc. The usual victim is ufs_scan_inodes() or 1267 * someone called by it. 1268 */ 1269 1270 /* 1271 * First, isolate it so that no new references can be 1272 * gotten via the inode cache. 1273 */ 1274 ihm = &ih_lock[INOHASH(UFSROOTINO)]; 1275 mutex_enter(ihm); 1276 remque(rip); 1277 mutex_exit(ihm); 1278 1279 /* 1280 * Now wait for all outstanding references except our 1281 * own to drain. This could, in theory, take forever, 1282 * so don't wait *too* long. If we time out, mark 1283 * it stale and leak it, so we don't hit the problem 1284 * described above. 1285 * 1286 * Note that v_count is an int, which means we can read 1287 * it in one operation. Thus, there's no need to lock 1288 * around our tests. 1289 */ 1290 elapsed = 0; 1291 while ((rvp->v_count > 1) && (elapsed < ufs_mount_timeout)) { 1292 delay(ufs_mount_error_delay * drv_usectohz(1000)); 1293 elapsed += ufs_mount_error_delay; 1294 } 1295 1296 if (rvp->v_count > 1) { 1297 mutex_enter(&rip->i_tlock); 1298 rip->i_flag |= ISTALE; 1299 mutex_exit(&rip->i_tlock); 1300 cmn_err(CE_WARN, 1301 "Timed out while cleaning up after failed mount of %s", 1302 path); 1303 } else { 1304 1305 /* 1306 * Now we're the only one with a handle left, so tear 1307 * it down the rest of the way. 1308 */ 1309 if (ufs_rmidle(rip)) 1310 VN_RELE(rvp); 1311 ufs_si_del(rip); 1312 rip->i_ufsvfs = NULL; 1313 rvp->v_vfsp = NULL; 1314 rvp->v_type = VBAD; 1315 VN_RELE(rvp); 1316 } 1317 } 1318 if (needtrans) { 1319 TRANS_MATA_UMOUNT(ufsvfsp); 1320 } 1321 if (ufsvfsp) { 1322 ufs_vfs_remove(ufsvfsp); 1323 ufs_thread_exit(&ufsvfsp->vfs_delete); 1324 ufs_thread_exit(&ufsvfsp->vfs_reclaim); 1325 mutex_destroy(&ufsvfsp->vfs_lock); 1326 if (ufsvfsp->vfs_log) { 1327 lufs_unsnarf(ufsvfsp); 1328 } 1329 kmem_free(ufsvfsp, sizeof (struct ufsvfs)); 1330 } 1331 if (bp) { 1332 bp->b_flags |= (B_STALE|B_AGE); 1333 brelse(bp); 1334 } 1335 if (tp) { 1336 tp->b_flags |= (B_STALE|B_AGE); 1337 brelse(tp); 1338 } 1339 if (needclose) { 1340 (void) VOP_CLOSE(devvp, (vfsp->vfs_flag & VFS_RDONLY) ? 1341 FREAD : FREAD|FWRITE, 1, (offset_t)0, cr); 1342 bflush(dev); 1343 (void) bfinval(dev, 1); 1344 } 1345 return (error); 1346 } 1347 1348 /* 1349 * vfs operations 1350 */ 1351 static int 1352 ufs_unmount(struct vfs *vfsp, int fflag, struct cred *cr) 1353 { 1354 dev_t dev = vfsp->vfs_dev; 1355 struct ufsvfs *ufsvfsp = (struct ufsvfs *)vfsp->vfs_data; 1356 struct fs *fs = ufsvfsp->vfs_fs; 1357 struct ulockfs *ulp = &ufsvfsp->vfs_ulockfs; 1358 struct vnode *bvp, *vp; 1359 struct buf *bp; 1360 struct inode *ip, *inext, *rip; 1361 union ihead *ih; 1362 int error, flag, i; 1363 struct lockfs lockfs; 1364 int poll_events = POLLPRI; 1365 extern struct pollhead ufs_pollhd; 1366 refstr_t *mountpoint; 1367 1368 ASSERT(vfs_lock_held(vfsp)); 1369 1370 if (secpolicy_fs_unmount(cr, vfsp) != 0) 1371 return (EPERM); 1372 /* 1373 * Forced unmount is now supported through the 1374 * lockfs protocol. 1375 */ 1376 if (fflag & MS_FORCE) { 1377 /* 1378 * Mark the filesystem as being unmounted now in 1379 * case of a forcible umount before we take any 1380 * locks inside UFS to prevent racing with a VFS_VGET() 1381 * request. Throw these VFS_VGET() requests away for 1382 * the duration of the forcible umount so they won't 1383 * use stale or even freed data later on when we're done. 1384 * It may happen that the VFS has had a additional hold 1385 * placed on it by someone other than UFS and thus will 1386 * not get freed immediately once we're done with the 1387 * umount by dounmount() - use VFS_UNMOUNTED to inform 1388 * users of this still-alive VFS that its corresponding 1389 * filesystem being gone so they can detect that and error 1390 * out. 1391 */ 1392 vfsp->vfs_flag |= VFS_UNMOUNTED; 1393 1394 ufs_thread_suspend(&ufsvfsp->vfs_delete); 1395 mutex_enter(&ulp->ul_lock); 1396 /* 1397 * If file system is already hard locked, 1398 * unmount the file system, otherwise 1399 * hard lock it before unmounting. 1400 */ 1401 if (!ULOCKFS_IS_HLOCK(ulp)) { 1402 atomic_add_long(&ufs_quiesce_pend, 1); 1403 lockfs.lf_lock = LOCKFS_HLOCK; 1404 lockfs.lf_flags = 0; 1405 lockfs.lf_key = ulp->ul_lockfs.lf_key + 1; 1406 lockfs.lf_comlen = 0; 1407 lockfs.lf_comment = NULL; 1408 ufs_freeze(ulp, &lockfs); 1409 ULOCKFS_SET_BUSY(ulp); 1410 LOCKFS_SET_BUSY(&ulp->ul_lockfs); 1411 (void) ufs_quiesce(ulp); 1412 (void) ufs_flush(vfsp); 1413 (void) ufs_thaw(vfsp, ufsvfsp, ulp); 1414 atomic_add_long(&ufs_quiesce_pend, -1); 1415 ULOCKFS_CLR_BUSY(ulp); 1416 LOCKFS_CLR_BUSY(&ulp->ul_lockfs); 1417 poll_events |= POLLERR; 1418 pollwakeup(&ufs_pollhd, poll_events); 1419 } 1420 ufs_thread_continue(&ufsvfsp->vfs_delete); 1421 mutex_exit(&ulp->ul_lock); 1422 } 1423 1424 /* let all types of writes go through */ 1425 ufsvfsp->vfs_iotstamp = lbolt; 1426 1427 /* coordinate with global hlock thread */ 1428 if (TRANS_ISTRANS(ufsvfsp) && (ufsvfsp->vfs_validfs == UT_HLOCKING)) { 1429 /* 1430 * last possibility for a forced umount to fail hence clear 1431 * VFS_UNMOUNTED if appropriate. 1432 */ 1433 if (fflag & MS_FORCE) 1434 vfsp->vfs_flag &= ~VFS_UNMOUNTED; 1435 return (EAGAIN); 1436 } 1437 1438 ufsvfsp->vfs_validfs = UT_UNMOUNTED; 1439 1440 /* kill the reclaim thread */ 1441 ufs_thread_exit(&ufsvfsp->vfs_reclaim); 1442 1443 /* suspend the delete thread */ 1444 ufs_thread_suspend(&ufsvfsp->vfs_delete); 1445 1446 /* 1447 * drain the delete and idle queues 1448 */ 1449 ufs_delete_drain(vfsp, -1, 1); 1450 ufs_idle_drain(vfsp); 1451 1452 /* 1453 * use the lockfs protocol to prevent new ops from starting 1454 * a forcible umount can not fail beyond this point as 1455 * we hard-locked the filesystem and drained all current consumers 1456 * before. 1457 */ 1458 mutex_enter(&ulp->ul_lock); 1459 1460 /* 1461 * if the file system is busy; return EBUSY 1462 */ 1463 if (ulp->ul_vnops_cnt || ULOCKFS_IS_SLOCK(ulp)) { 1464 error = EBUSY; 1465 goto out; 1466 } 1467 1468 /* 1469 * if this is not a forced unmount (!hard/error locked), then 1470 * get rid of every inode except the root and quota inodes 1471 * also, commit any outstanding transactions 1472 */ 1473 if (!ULOCKFS_IS_HLOCK(ulp) && !ULOCKFS_IS_ELOCK(ulp)) 1474 if (error = ufs_flush(vfsp)) 1475 goto out; 1476 1477 /* 1478 * ignore inodes in the cache if fs is hard locked or error locked 1479 */ 1480 rip = VTOI(ufsvfsp->vfs_root); 1481 if (!ULOCKFS_IS_HLOCK(ulp) && !ULOCKFS_IS_ELOCK(ulp)) { 1482 /* 1483 * Otherwise, only the quota and root inodes are in the cache. 1484 * 1485 * Avoid racing with ufs_update() and ufs_sync(). 1486 */ 1487 mutex_enter(&ufs_scan_lock); 1488 1489 for (i = 0, ih = ihead; i < inohsz; i++, ih++) { 1490 mutex_enter(&ih_lock[i]); 1491 for (ip = ih->ih_chain[0]; 1492 ip != (struct inode *)ih; 1493 ip = ip->i_forw) { 1494 if (ip->i_ufsvfs != ufsvfsp) 1495 continue; 1496 if (ip == ufsvfsp->vfs_qinod) 1497 continue; 1498 if (ip == rip && ITOV(ip)->v_count == 1) 1499 continue; 1500 mutex_exit(&ih_lock[i]); 1501 mutex_exit(&ufs_scan_lock); 1502 error = EBUSY; 1503 goto out; 1504 } 1505 mutex_exit(&ih_lock[i]); 1506 } 1507 mutex_exit(&ufs_scan_lock); 1508 } 1509 1510 /* 1511 * if a snapshot exists and this is a forced unmount, then delete 1512 * the snapshot. Otherwise return EBUSY. This will insure the 1513 * snapshot always belongs to a valid file system. 1514 */ 1515 if (ufsvfsp->vfs_snapshot) { 1516 if (ULOCKFS_IS_HLOCK(ulp) || ULOCKFS_IS_ELOCK(ulp)) { 1517 (void) fssnap_delete(&ufsvfsp->vfs_snapshot); 1518 } else { 1519 error = EBUSY; 1520 goto out; 1521 } 1522 } 1523 1524 /* 1525 * Close the quota file and invalidate anything left in the quota 1526 * cache for this file system. Pass kcred to allow all quota 1527 * manipulations. 1528 */ 1529 (void) closedq(ufsvfsp, kcred); 1530 invalidatedq(ufsvfsp); 1531 /* 1532 * drain the delete and idle queues 1533 */ 1534 ufs_delete_drain(vfsp, -1, 0); 1535 ufs_idle_drain(vfsp); 1536 1537 /* 1538 * discard the inodes for this fs (including root, shadow, and quota) 1539 */ 1540 for (i = 0, ih = ihead; i < inohsz; i++, ih++) { 1541 mutex_enter(&ih_lock[i]); 1542 for (inext = 0, ip = ih->ih_chain[0]; 1543 ip != (struct inode *)ih; 1544 ip = inext) { 1545 inext = ip->i_forw; 1546 if (ip->i_ufsvfs != ufsvfsp) 1547 continue; 1548 vp = ITOV(ip); 1549 VN_HOLD(vp) 1550 remque(ip); 1551 if (ufs_rmidle(ip)) 1552 VN_RELE(vp); 1553 ufs_si_del(ip); 1554 /* 1555 * rip->i_ufsvfsp is needed by bflush() 1556 */ 1557 if (ip != rip) 1558 ip->i_ufsvfs = NULL; 1559 /* 1560 * Set vnode's vfsops to dummy ops, which return 1561 * EIO. This is needed to forced unmounts to work 1562 * with lofs/nfs properly. 1563 */ 1564 if (ULOCKFS_IS_HLOCK(ulp) || ULOCKFS_IS_ELOCK(ulp)) 1565 vp->v_vfsp = &EIO_vfs; 1566 else 1567 vp->v_vfsp = NULL; 1568 vp->v_type = VBAD; 1569 VN_RELE(vp); 1570 } 1571 mutex_exit(&ih_lock[i]); 1572 } 1573 ufs_si_cache_flush(dev); 1574 1575 /* 1576 * kill the delete thread and drain the idle queue 1577 */ 1578 ufs_thread_exit(&ufsvfsp->vfs_delete); 1579 ufs_idle_drain(vfsp); 1580 1581 bp = ufsvfsp->vfs_bufp; 1582 bvp = ufsvfsp->vfs_devvp; 1583 flag = !fs->fs_ronly; 1584 if (flag) { 1585 bflush(dev); 1586 if (fs->fs_clean != FSBAD) { 1587 if (fs->fs_clean == FSSTABLE) 1588 fs->fs_clean = FSCLEAN; 1589 fs->fs_reclaim &= ~FS_RECLAIM; 1590 } 1591 if (TRANS_ISTRANS(ufsvfsp) && 1592 !TRANS_ISERROR(ufsvfsp) && 1593 !ULOCKFS_IS_HLOCK(ulp) && 1594 (fs->fs_rolled == FS_NEED_ROLL)) { 1595 /* 1596 * ufs_flush() above has flushed the last Moby. 1597 * This is needed to ensure the following superblock 1598 * update really is the last metadata update 1599 */ 1600 error = ufs_putsummaryinfo(dev, ufsvfsp, fs); 1601 if (error == 0) { 1602 fs->fs_rolled = FS_ALL_ROLLED; 1603 } 1604 } 1605 TRANS_SBUPDATE(ufsvfsp, vfsp, TOP_SBUPDATE_UNMOUNT); 1606 /* 1607 * push this last transaction 1608 */ 1609 curthread->t_flag |= T_DONTBLOCK; 1610 TRANS_BEGIN_SYNC(ufsvfsp, TOP_COMMIT_UNMOUNT, TOP_COMMIT_SIZE, 1611 error); 1612 if (!error) 1613 TRANS_END_SYNC(ufsvfsp, error, TOP_COMMIT_UNMOUNT, 1614 TOP_COMMIT_SIZE); 1615 curthread->t_flag &= ~T_DONTBLOCK; 1616 } 1617 1618 TRANS_MATA_UMOUNT(ufsvfsp); 1619 lufs_unsnarf(ufsvfsp); /* Release the in-memory structs */ 1620 ufsfx_unmount(ufsvfsp); /* fix-on-panic bookkeeping */ 1621 kmem_free(fs->fs_u.fs_csp, fs->fs_cssize); 1622 1623 bp->b_flags |= B_STALE|B_AGE; 1624 ufsvfsp->vfs_bufp = NULL; /* don't point at freed buf */ 1625 brelse(bp); /* free the superblock buf */ 1626 1627 (void) VOP_PUTPAGE(common_specvp(bvp), (offset_t)0, (size_t)0, 1628 B_INVAL, cr); 1629 (void) VOP_CLOSE(bvp, flag, 1, (offset_t)0, cr); 1630 bflush(dev); 1631 (void) bfinval(dev, 1); 1632 VN_RELE(bvp); 1633 1634 /* 1635 * It is now safe to NULL out the ufsvfs pointer and discard 1636 * the root inode. 1637 */ 1638 rip->i_ufsvfs = NULL; 1639 VN_RELE(ITOV(rip)); 1640 1641 /* free up lockfs comment structure, if any */ 1642 if (ulp->ul_lockfs.lf_comlen && ulp->ul_lockfs.lf_comment) 1643 kmem_free(ulp->ul_lockfs.lf_comment, ulp->ul_lockfs.lf_comlen); 1644 1645 /* 1646 * Remove from instance list. 1647 */ 1648 ufs_vfs_remove(ufsvfsp); 1649 1650 /* 1651 * For a forcible unmount, threads may be asleep in 1652 * ufs_lockfs_begin/ufs_check_lockfs. These threads will need 1653 * the ufsvfs structure so we don't free it, yet. ufs_update 1654 * will free it up after awhile. 1655 */ 1656 if (ULOCKFS_IS_HLOCK(ulp) || ULOCKFS_IS_ELOCK(ulp)) { 1657 extern kmutex_t ufsvfs_mutex; 1658 extern struct ufsvfs *ufsvfslist; 1659 1660 mutex_enter(&ufsvfs_mutex); 1661 ufsvfsp->vfs_dontblock = 1; 1662 ufsvfsp->vfs_next = ufsvfslist; 1663 ufsvfslist = ufsvfsp; 1664 mutex_exit(&ufsvfs_mutex); 1665 /* wakeup any suspended threads */ 1666 cv_broadcast(&ulp->ul_cv); 1667 mutex_exit(&ulp->ul_lock); 1668 } else { 1669 mutex_destroy(&ufsvfsp->vfs_lock); 1670 kmem_free(ufsvfsp, sizeof (struct ufsvfs)); 1671 } 1672 1673 /* 1674 * Now mark the filesystem as unmounted since we're done with it. 1675 */ 1676 vfsp->vfs_flag |= VFS_UNMOUNTED; 1677 1678 return (0); 1679 out: 1680 /* open the fs to new ops */ 1681 cv_broadcast(&ulp->ul_cv); 1682 mutex_exit(&ulp->ul_lock); 1683 1684 if (TRANS_ISTRANS(ufsvfsp)) { 1685 /* allow the delete thread to continue */ 1686 ufs_thread_continue(&ufsvfsp->vfs_delete); 1687 /* restart the reclaim thread */ 1688 ufs_thread_start(&ufsvfsp->vfs_reclaim, ufs_thread_reclaim, 1689 vfsp); 1690 /* coordinate with global hlock thread */ 1691 ufsvfsp->vfs_validfs = UT_MOUNTED; 1692 /* check for trans errors during umount */ 1693 ufs_trans_onerror(); 1694 1695 /* 1696 * if we have a seperate /usr it will never unmount 1697 * when halting. In order to not re-read all the 1698 * cylinder group summary info on mounting after 1699 * reboot the logging of summary info is re-enabled 1700 * and the super block written out. 1701 */ 1702 mountpoint = vfs_getmntpoint(vfsp); 1703 if ((fs->fs_si == FS_SI_OK) && 1704 (strcmp("/usr", refstr_value(mountpoint)) == 0)) { 1705 ufsvfsp->vfs_nolog_si = 0; 1706 UFS_BWRITE2(NULL, ufsvfsp->vfs_bufp); 1707 } 1708 refstr_rele(mountpoint); 1709 } 1710 1711 return (error); 1712 } 1713 1714 static int 1715 ufs_root(struct vfs *vfsp, struct vnode **vpp) 1716 { 1717 struct ufsvfs *ufsvfsp; 1718 struct vnode *vp; 1719 1720 if (!vfsp) 1721 return (EIO); 1722 1723 ufsvfsp = (struct ufsvfs *)vfsp->vfs_data; 1724 if (!ufsvfsp || !ufsvfsp->vfs_root) 1725 return (EIO); /* forced unmount */ 1726 1727 vp = ufsvfsp->vfs_root; 1728 VN_HOLD(vp); 1729 *vpp = vp; 1730 return (0); 1731 } 1732 1733 /* 1734 * Get file system statistics. 1735 */ 1736 static int 1737 ufs_statvfs(struct vfs *vfsp, struct statvfs64 *sp) 1738 { 1739 struct fs *fsp; 1740 struct ufsvfs *ufsvfsp; 1741 int blk, i; 1742 long max_avail, used; 1743 dev32_t d32; 1744 struct ufs_q *delq; 1745 1746 if (vfsp->vfs_flag & VFS_UNMOUNTED) 1747 return (EIO); 1748 1749 ufsvfsp = (struct ufsvfs *)vfsp->vfs_data; 1750 fsp = ufsvfsp->vfs_fs; 1751 if ((fsp->fs_magic != FS_MAGIC) && (fsp->fs_magic != MTB_UFS_MAGIC)) 1752 return (EINVAL); 1753 if (fsp->fs_magic == MTB_UFS_MAGIC && 1754 (fsp->fs_version > MTB_UFS_VERSION_1 || 1755 fsp->fs_version < MTB_UFS_VERSION_MIN)) 1756 return (EINVAL); 1757 1758 /* 1759 * Can't get a self-consistent result with the delete thread running 1760 * or if others come in and jumble the queue up for us. This is 1761 * a no-op if there is no delete thread. 1762 */ 1763 delq = &ufsvfsp->vfs_delete; 1764 ufs_thread_suspend(delq); 1765 1766 /* 1767 * get the basic numbers 1768 */ 1769 (void) bzero(sp, sizeof (*sp)); 1770 1771 sp->f_bsize = fsp->fs_bsize; 1772 sp->f_frsize = fsp->fs_fsize; 1773 sp->f_blocks = (fsblkcnt64_t)fsp->fs_dsize; 1774 sp->f_bfree = (fsblkcnt64_t)fsp->fs_cstotal.cs_nbfree * fsp->fs_frag + 1775 fsp->fs_cstotal.cs_nffree; 1776 1777 sp->f_files = (fsfilcnt64_t)fsp->fs_ncg * fsp->fs_ipg; 1778 sp->f_ffree = (fsfilcnt64_t)fsp->fs_cstotal.cs_nifree; 1779 1780 /* 1781 * Adjust the numbers based on things waiting to be deleted. 1782 * modifies f_bfree and f_ffree. Afterwards, everything we 1783 * come up with will be self-consistent. By definition, this 1784 * is a point-in-time snapshot, so the fact that the delete 1785 * thread's probably already invalidated the results is not a 1786 * problem. 1787 */ 1788 ufs_delete_adjust_stats(ufsvfsp, sp); 1789 ufs_thread_continue(delq); 1790 1791 /* 1792 * avail = MAX(max_avail - used, 0) 1793 */ 1794 max_avail = fsp->fs_dsize - ufsvfsp->vfs_minfrags; 1795 1796 used = (fsp->fs_dsize - sp->f_bfree); 1797 1798 if (max_avail > used) 1799 sp->f_bavail = (fsblkcnt64_t)max_avail - used; 1800 else 1801 sp->f_bavail = (fsblkcnt64_t)0; 1802 1803 sp->f_favail = sp->f_ffree; 1804 (void) cmpldev(&d32, vfsp->vfs_dev); 1805 sp->f_fsid = d32; 1806 (void) strcpy(sp->f_basetype, vfssw[vfsp->vfs_fstype].vsw_name); 1807 sp->f_flag = vf_to_stf(vfsp->vfs_flag); 1808 1809 /* keep coordinated with ufs_l_pathconf() */ 1810 sp->f_namemax = MAXNAMLEN; 1811 1812 if (fsp->fs_cpc == 0) { 1813 bzero(sp->f_fstr, 14); 1814 return (0); 1815 } 1816 blk = fsp->fs_spc * fsp->fs_cpc / NSPF(fsp); 1817 for (i = 0; i < blk; i += fsp->fs_frag) /* CSTYLED */ 1818 /* void */; 1819 i -= fsp->fs_frag; 1820 blk = i / fsp->fs_frag; 1821 bcopy(&(fs_rotbl(fsp)[blk]), sp->f_fstr, 14); 1822 return (0); 1823 } 1824 1825 /* 1826 * Flush any pending I/O to file system vfsp. 1827 * The ufs_update() routine will only flush *all* ufs files. 1828 * If vfsp is non-NULL, only sync this ufs (in preparation 1829 * for a umount). 1830 */ 1831 /*ARGSUSED*/ 1832 static int 1833 ufs_sync(struct vfs *vfsp, short flag, struct cred *cr) 1834 { 1835 struct ufsvfs *ufsvfsp; 1836 struct fs *fs; 1837 int cheap = flag & SYNC_ATTR; 1838 int error; 1839 1840 /* 1841 * SYNC_CLOSE means we're rebooting. Toss everything 1842 * on the idle queue so we don't have to slog through 1843 * a bunch of uninteresting inodes over and over again. 1844 */ 1845 if (flag & SYNC_CLOSE) 1846 ufs_idle_drain(NULL); 1847 1848 if (vfsp == NULL) { 1849 ufs_update(flag); 1850 return (0); 1851 } 1852 1853 /* Flush a single ufs */ 1854 if (!vfs_matchops(vfsp, ufs_vfsops) || vfs_lock(vfsp) != 0) 1855 return (0); 1856 1857 ufsvfsp = (struct ufsvfs *)vfsp->vfs_data; 1858 if (!ufsvfsp) 1859 return (EIO); 1860 fs = ufsvfsp->vfs_fs; 1861 mutex_enter(&ufsvfsp->vfs_lock); 1862 1863 if (ufsvfsp->vfs_dio && 1864 fs->fs_ronly == 0 && 1865 fs->fs_clean != FSBAD && 1866 fs->fs_clean != FSLOG) { 1867 /* turn off fast-io on unmount, so no fsck needed (4029401) */ 1868 ufsvfsp->vfs_dio = 0; 1869 fs->fs_clean = FSACTIVE; 1870 fs->fs_fmod = 1; 1871 } 1872 1873 /* Write back modified superblock */ 1874 if (fs->fs_fmod == 0) { 1875 mutex_exit(&ufsvfsp->vfs_lock); 1876 } else { 1877 if (fs->fs_ronly != 0) { 1878 mutex_exit(&ufsvfsp->vfs_lock); 1879 vfs_unlock(vfsp); 1880 return (ufs_fault(ufsvfsp->vfs_root, 1881 "fs = %s update: ro fs mod\n", 1882 fs->fs_fsmnt)); 1883 } 1884 fs->fs_fmod = 0; 1885 mutex_exit(&ufsvfsp->vfs_lock); 1886 1887 TRANS_SBUPDATE(ufsvfsp, vfsp, TOP_SBUPDATE_UPDATE); 1888 } 1889 vfs_unlock(vfsp); 1890 1891 /* 1892 * Avoid racing with ufs_update() and ufs_unmount(). 1893 * 1894 */ 1895 mutex_enter(&ufs_scan_lock); 1896 1897 (void) ufs_scan_inodes(1, ufs_sync_inode, 1898 (void *)(uintptr_t)cheap, ufsvfsp); 1899 1900 mutex_exit(&ufs_scan_lock); 1901 1902 bflush((dev_t)vfsp->vfs_dev); 1903 1904 /* 1905 * commit any outstanding async transactions 1906 */ 1907 curthread->t_flag |= T_DONTBLOCK; 1908 TRANS_BEGIN_SYNC(ufsvfsp, TOP_COMMIT_UPDATE, TOP_COMMIT_SIZE, error); 1909 if (!error) { 1910 TRANS_END_SYNC(ufsvfsp, error, TOP_COMMIT_UPDATE, 1911 TOP_COMMIT_SIZE); 1912 } 1913 curthread->t_flag &= ~T_DONTBLOCK; 1914 1915 return (0); 1916 } 1917 1918 1919 void 1920 sbupdate(struct vfs *vfsp) 1921 { 1922 struct ufsvfs *ufsvfsp = (struct ufsvfs *)vfsp->vfs_data; 1923 struct fs *fs = ufsvfsp->vfs_fs; 1924 struct buf *bp; 1925 int blks; 1926 caddr_t space; 1927 int i; 1928 size_t size; 1929 1930 /* 1931 * for ulockfs processing, limit the superblock writes 1932 */ 1933 if ((ufsvfsp->vfs_ulockfs.ul_sbowner) && 1934 (curthread != ufsvfsp->vfs_ulockfs.ul_sbowner)) { 1935 /* process later */ 1936 fs->fs_fmod = 1; 1937 return; 1938 } 1939 ULOCKFS_SET_MOD((&ufsvfsp->vfs_ulockfs)); 1940 1941 if (TRANS_ISTRANS(ufsvfsp)) { 1942 mutex_enter(&ufsvfsp->vfs_lock); 1943 ufs_sbwrite(ufsvfsp); 1944 mutex_exit(&ufsvfsp->vfs_lock); 1945 return; 1946 } 1947 1948 blks = howmany(fs->fs_cssize, fs->fs_fsize); 1949 space = (caddr_t)fs->fs_u.fs_csp; 1950 for (i = 0; i < blks; i += fs->fs_frag) { 1951 size = fs->fs_bsize; 1952 if (i + fs->fs_frag > blks) 1953 size = (blks - i) * fs->fs_fsize; 1954 bp = UFS_GETBLK(ufsvfsp, ufsvfsp->vfs_dev, 1955 (daddr_t)(fsbtodb(fs, fs->fs_csaddr + i)), 1956 fs->fs_bsize); 1957 bcopy(space, bp->b_un.b_addr, size); 1958 space += size; 1959 bp->b_bcount = size; 1960 UFS_BRWRITE(ufsvfsp, bp); 1961 } 1962 mutex_enter(&ufsvfsp->vfs_lock); 1963 ufs_sbwrite(ufsvfsp); 1964 mutex_exit(&ufsvfsp->vfs_lock); 1965 } 1966 1967 int ufs_vget_idle_count = 2; /* Number of inodes to idle each time */ 1968 static int 1969 ufs_vget(struct vfs *vfsp, struct vnode **vpp, struct fid *fidp) 1970 { 1971 int error = 0; 1972 struct ufid *ufid; 1973 struct inode *ip; 1974 struct ufsvfs *ufsvfsp = (struct ufsvfs *)vfsp->vfs_data; 1975 struct ulockfs *ulp; 1976 1977 /* 1978 * Check for unmounted filesystem. 1979 */ 1980 if (vfsp->vfs_flag & VFS_UNMOUNTED) { 1981 error = EIO; 1982 goto errout; 1983 } 1984 1985 /* 1986 * Keep the idle queue from getting too long by 1987 * idling an inode before attempting to allocate another. 1988 * This operation must be performed before entering 1989 * lockfs or a transaction. 1990 */ 1991 if (ufs_idle_q.uq_ne > ufs_idle_q.uq_hiwat) 1992 if ((curthread->t_flag & T_DONTBLOCK) == 0) { 1993 ins.in_vidles.value.ul += ufs_vget_idle_count; 1994 ufs_idle_some(ufs_vget_idle_count); 1995 } 1996 1997 ufid = (struct ufid *)fidp; 1998 1999 if (error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_VGET_MASK)) 2000 goto errout; 2001 2002 rw_enter(&ufsvfsp->vfs_dqrwlock, RW_READER); 2003 2004 error = ufs_iget(vfsp, ufid->ufid_ino, &ip, CRED()); 2005 2006 rw_exit(&ufsvfsp->vfs_dqrwlock); 2007 2008 ufs_lockfs_end(ulp); 2009 2010 if (error) 2011 goto errout; 2012 2013 /* 2014 * Check if the inode has been deleted or freed or is in transient state 2015 * since the last VFS_VGET() request for it, release it and don't return 2016 * it to the caller, presumably NFS, as it's no longer valid. 2017 */ 2018 if (ip->i_gen != ufid->ufid_gen || ip->i_mode == 0 || 2019 (ip->i_flag & IDEL)) { 2020 VN_RELE(ITOV(ip)); 2021 error = EINVAL; 2022 goto errout; 2023 } 2024 2025 *vpp = ITOV(ip); 2026 return (0); 2027 2028 errout: 2029 *vpp = NULL; 2030 return (error); 2031 } 2032 2033 static int 2034 ufsinit(int fstype, char *name) 2035 { 2036 static const fs_operation_def_t ufs_vfsops_template[] = { 2037 VFSNAME_MOUNT, ufs_mount, 2038 VFSNAME_UNMOUNT, ufs_unmount, 2039 VFSNAME_ROOT, ufs_root, 2040 VFSNAME_STATVFS, ufs_statvfs, 2041 VFSNAME_SYNC, (fs_generic_func_p) ufs_sync, 2042 VFSNAME_VGET, ufs_vget, 2043 VFSNAME_MOUNTROOT, ufs_mountroot, 2044 NULL, NULL 2045 }; 2046 int error; 2047 2048 ufsfstype = fstype; 2049 2050 error = vfs_setfsops(fstype, ufs_vfsops_template, &ufs_vfsops); 2051 if (error != 0) { 2052 cmn_err(CE_WARN, "ufsinit: bad vfs ops template"); 2053 return (error); 2054 } 2055 2056 error = vn_make_ops(name, ufs_vnodeops_template, &ufs_vnodeops); 2057 if (error != 0) { 2058 (void) vfs_freevfsops_by_type(fstype); 2059 cmn_err(CE_WARN, "ufsinit: bad vnode ops template"); 2060 return (error); 2061 } 2062 2063 ufs_iinit(); 2064 return (0); 2065 } 2066 2067 #ifdef __sparc 2068 2069 /* 2070 * Mounting a mirrored SVM volume is only supported on ufs, 2071 * this is special-case boot code to support that configuration. 2072 * At this point, we have booted and mounted root on a 2073 * single component of the mirror. Complete the boot 2074 * by configuring SVM and converting the root to the 2075 * dev_t of the mirrored root device. This dev_t conversion 2076 * only works because the underlying device doesn't change. 2077 */ 2078 int 2079 ufs_remountroot(struct vfs *vfsp) 2080 { 2081 struct ufsvfs *ufsvfsp; 2082 struct ulockfs *ulp; 2083 dev_t new_rootdev; 2084 dev_t old_rootdev; 2085 struct vnode *old_rootvp; 2086 struct vnode *new_rootvp; 2087 int error, sberror = 0; 2088 struct inode *ip; 2089 union ihead *ih; 2090 struct buf *bp; 2091 int i; 2092 2093 old_rootdev = rootdev; 2094 old_rootvp = rootvp; 2095 2096 new_rootdev = getrootdev(); 2097 if (new_rootdev == (dev_t)NODEV) { 2098 return (ENODEV); 2099 } 2100 2101 new_rootvp = makespecvp(new_rootdev, VBLK); 2102 2103 error = VOP_OPEN(&new_rootvp, 2104 (vfsp->vfs_flag & VFS_RDONLY) ? FREAD : FREAD|FWRITE, CRED()); 2105 if (error) { 2106 cmn_err(CE_CONT, 2107 "Cannot open mirrored root device, error %d\n", error); 2108 return (error); 2109 } 2110 2111 if (vfs_lock(vfsp) != 0) { 2112 return (EBUSY); 2113 } 2114 2115 ufsvfsp = (struct ufsvfs *)vfsp->vfs_data; 2116 ulp = &ufsvfsp->vfs_ulockfs; 2117 2118 mutex_enter(&ulp->ul_lock); 2119 atomic_add_long(&ufs_quiesce_pend, 1); 2120 2121 (void) ufs_quiesce(ulp); 2122 (void) ufs_flush(vfsp); 2123 2124 /* 2125 * Convert root vfs to new dev_t, including vfs hash 2126 * table and fs id. 2127 */ 2128 vfs_root_redev(vfsp, new_rootdev, ufsfstype); 2129 2130 ufsvfsp->vfs_devvp = new_rootvp; 2131 ufsvfsp->vfs_dev = new_rootdev; 2132 2133 bp = ufsvfsp->vfs_bufp; 2134 bp->b_edev = new_rootdev; 2135 bp->b_dev = cmpdev(new_rootdev); 2136 2137 /* 2138 * The buffer for the root inode does not contain a valid b_vp 2139 */ 2140 (void) bfinval(new_rootdev, 0); 2141 2142 /* 2143 * Here we hand-craft inodes with old root device 2144 * references to refer to the new device instead. 2145 */ 2146 mutex_enter(&ufs_scan_lock); 2147 2148 for (i = 0, ih = ihead; i < inohsz; i++, ih++) { 2149 mutex_enter(&ih_lock[i]); 2150 for (ip = ih->ih_chain[0]; 2151 ip != (struct inode *)ih; 2152 ip = ip->i_forw) { 2153 if (ip->i_ufsvfs != ufsvfsp) 2154 continue; 2155 if (ip == ufsvfsp->vfs_qinod) 2156 continue; 2157 if (ip->i_dev == old_rootdev) { 2158 ip->i_dev = new_rootdev; 2159 } 2160 2161 if (ip->i_devvp == old_rootvp) { 2162 ip->i_devvp = new_rootvp; 2163 } 2164 } 2165 mutex_exit(&ih_lock[i]); 2166 } 2167 2168 mutex_exit(&ufs_scan_lock); 2169 2170 /* 2171 * Make Sure logging structures are using the new device 2172 * if logging is enabled. Also start any logging thread that 2173 * needs to write to the device and couldn't earlier. 2174 */ 2175 if (ufsvfsp->vfs_log) { 2176 buf_t *bp, *tbp; 2177 ml_unit_t *ul = ufsvfsp->vfs_log; 2178 struct fs *fsp = ufsvfsp->vfs_fs; 2179 2180 /* 2181 * Update the main logging structure. 2182 */ 2183 ul->un_dev = new_rootdev; 2184 2185 /* 2186 * Get a new bp for the on disk structures. 2187 */ 2188 bp = ul->un_bp; 2189 tbp = ngeteblk(dbtob(LS_SECTORS)); 2190 tbp->b_edev = new_rootdev; 2191 tbp->b_dev = cmpdev(new_rootdev); 2192 tbp->b_blkno = bp->b_blkno; 2193 bcopy(bp->b_un.b_addr, tbp->b_un.b_addr, DEV_BSIZE); 2194 bcopy(bp->b_un.b_addr, tbp->b_un.b_addr + DEV_BSIZE, DEV_BSIZE); 2195 bp->b_flags |= (B_STALE | B_AGE); 2196 brelse(bp); 2197 ul->un_bp = tbp; 2198 2199 /* 2200 * Allocate new circular buffers. 2201 */ 2202 alloc_rdbuf(&ul->un_rdbuf, MAPBLOCKSIZE, MAPBLOCKSIZE); 2203 alloc_wrbuf(&ul->un_wrbuf, ldl_bufsize(ul)); 2204 2205 /* 2206 * Clear the noroll bit which indicates that logging 2207 * can't roll the log yet. 2208 */ 2209 ASSERT(ul->un_flags & LDL_NOROLL); 2210 ul->un_flags &= ~LDL_NOROLL; 2211 2212 /* 2213 * Start the logmap roll thread. 2214 */ 2215 logmap_start_roll(ul); 2216 2217 /* 2218 * Start the reclaim thread if needed. 2219 */ 2220 if (!fsp->fs_ronly && (fsp->fs_reclaim & 2221 (FS_RECLAIM|FS_RECLAIMING))) { 2222 fsp->fs_reclaim &= ~FS_RECLAIM; 2223 fsp->fs_reclaim |= FS_RECLAIMING; 2224 ufs_thread_start(&ufsvfsp->vfs_reclaim, 2225 ufs_thread_reclaim, vfsp); 2226 TRANS_SBWRITE(ufsvfsp, TOP_SBUPDATE_UPDATE); 2227 if (sberror = geterror(ufsvfsp->vfs_bufp)) { 2228 refstr_t *mntpt; 2229 mntpt = vfs_getmntpoint(vfsp); 2230 cmn_err(CE_WARN, 2231 "Remountroot failed to update Reclaim" 2232 "state for filesystem %s " 2233 "Error writing SuperBlock %d", 2234 refstr_value(mntpt), error); 2235 refstr_rele(mntpt); 2236 } 2237 } 2238 } 2239 2240 rootdev = new_rootdev; 2241 rootvp = new_rootvp; 2242 2243 atomic_add_long(&ufs_quiesce_pend, -1); 2244 cv_broadcast(&ulp->ul_cv); 2245 mutex_exit(&ulp->ul_lock); 2246 2247 vfs_unlock(vfsp); 2248 2249 error = VOP_CLOSE(old_rootvp, FREAD, 1, (offset_t)0, CRED()); 2250 if (error) { 2251 cmn_err(CE_CONT, 2252 "close of root device component failed, error %d\n", 2253 error); 2254 } 2255 VN_RELE(old_rootvp); 2256 2257 return (sberror ? sberror : error); 2258 } 2259 2260 #endif /* __sparc */ 2261