17c478bd9Sstevel@tonic-gate /* 27c478bd9Sstevel@tonic-gate * CDDL HEADER START 37c478bd9Sstevel@tonic-gate * 47c478bd9Sstevel@tonic-gate * The contents of this file are subject to the terms of the 55b024a5bSbatschul * Common Development and Distribution License (the "License"). 65b024a5bSbatschul * You may not use this file except in compliance with the License. 77c478bd9Sstevel@tonic-gate * 87c478bd9Sstevel@tonic-gate * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 97c478bd9Sstevel@tonic-gate * or http://www.opensolaris.org/os/licensing. 107c478bd9Sstevel@tonic-gate * See the License for the specific language governing permissions 117c478bd9Sstevel@tonic-gate * and limitations under the License. 127c478bd9Sstevel@tonic-gate * 137c478bd9Sstevel@tonic-gate * When distributing Covered Code, include this CDDL HEADER in each 147c478bd9Sstevel@tonic-gate * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 157c478bd9Sstevel@tonic-gate * If applicable, add the following below this CDDL HEADER, with the 167c478bd9Sstevel@tonic-gate * fields enclosed by brackets "[]" replaced with your own identifying 177c478bd9Sstevel@tonic-gate * information: Portions Copyright [yyyy] [name of copyright owner] 187c478bd9Sstevel@tonic-gate * 197c478bd9Sstevel@tonic-gate * CDDL HEADER END 207c478bd9Sstevel@tonic-gate */ 21342440ecSPrasad Singamsetty 227c478bd9Sstevel@tonic-gate /* 239b5097eeSOwen Roberts * Copyright (c) 1984, 2010, Oracle and/or its affiliates. All rights reserved. 2472102e74SBryan Cantrill * Copyright (c) 2013, Joyent, Inc. All rights reserved. 257c478bd9Sstevel@tonic-gate */ 267c478bd9Sstevel@tonic-gate 277c478bd9Sstevel@tonic-gate /* Copyright (c) 1983, 1984, 1985, 1986, 1987, 1988, 1989 AT&T */ 287c478bd9Sstevel@tonic-gate /* All Rights Reserved */ 297c478bd9Sstevel@tonic-gate 307c478bd9Sstevel@tonic-gate /* 317c478bd9Sstevel@tonic-gate * Portions of this source code were derived from Berkeley 4.3 BSD 327c478bd9Sstevel@tonic-gate * under license from the Regents of the University of California. 337c478bd9Sstevel@tonic-gate */ 347c478bd9Sstevel@tonic-gate 357c478bd9Sstevel@tonic-gate #include <sys/types.h> 367c478bd9Sstevel@tonic-gate #include <sys/t_lock.h> 377c478bd9Sstevel@tonic-gate #include <sys/ksynch.h> 387c478bd9Sstevel@tonic-gate #include <sys/param.h> 397c478bd9Sstevel@tonic-gate #include <sys/time.h> 407c478bd9Sstevel@tonic-gate #include <sys/systm.h> 417c478bd9Sstevel@tonic-gate #include <sys/sysmacros.h> 427c478bd9Sstevel@tonic-gate #include <sys/resource.h> 437c478bd9Sstevel@tonic-gate #include <sys/signal.h> 447c478bd9Sstevel@tonic-gate #include <sys/cred.h> 457c478bd9Sstevel@tonic-gate #include <sys/user.h> 467c478bd9Sstevel@tonic-gate #include <sys/buf.h> 477c478bd9Sstevel@tonic-gate #include <sys/vfs.h> 48aa59c4cbSrsb #include <sys/vfs_opreg.h> 497c478bd9Sstevel@tonic-gate #include <sys/vnode.h> 507c478bd9Sstevel@tonic-gate #include <sys/proc.h> 517c478bd9Sstevel@tonic-gate #include <sys/disp.h> 527c478bd9Sstevel@tonic-gate #include <sys/file.h> 537c478bd9Sstevel@tonic-gate #include <sys/fcntl.h> 547c478bd9Sstevel@tonic-gate #include <sys/flock.h> 55bc69f433Saguzovsk #include <sys/atomic.h> 567c478bd9Sstevel@tonic-gate #include <sys/kmem.h> 577c478bd9Sstevel@tonic-gate #include <sys/uio.h> 587c478bd9Sstevel@tonic-gate #include <sys/dnlc.h> 597c478bd9Sstevel@tonic-gate #include <sys/conf.h> 607c478bd9Sstevel@tonic-gate #include <sys/mman.h> 617c478bd9Sstevel@tonic-gate #include <sys/pathname.h> 627c478bd9Sstevel@tonic-gate #include <sys/debug.h> 637c478bd9Sstevel@tonic-gate #include <sys/vmsystm.h> 647c478bd9Sstevel@tonic-gate #include <sys/cmn_err.h> 657c478bd9Sstevel@tonic-gate #include <sys/filio.h> 667c478bd9Sstevel@tonic-gate #include <sys/policy.h> 677c478bd9Sstevel@tonic-gate 687c478bd9Sstevel@tonic-gate #include <sys/fs/ufs_fs.h> 697c478bd9Sstevel@tonic-gate #include <sys/fs/ufs_lockfs.h> 707c478bd9Sstevel@tonic-gate #include <sys/fs/ufs_filio.h> 717c478bd9Sstevel@tonic-gate #include <sys/fs/ufs_inode.h> 727c478bd9Sstevel@tonic-gate #include <sys/fs/ufs_fsdir.h> 737c478bd9Sstevel@tonic-gate #include <sys/fs/ufs_quota.h> 747c478bd9Sstevel@tonic-gate #include <sys/fs/ufs_log.h> 757c478bd9Sstevel@tonic-gate #include <sys/fs/ufs_snap.h> 767c478bd9Sstevel@tonic-gate #include <sys/fs/ufs_trans.h> 777c478bd9Sstevel@tonic-gate #include <sys/fs/ufs_panic.h> 787c478bd9Sstevel@tonic-gate #include <sys/fs/ufs_bio.h> 797c478bd9Sstevel@tonic-gate #include <sys/dirent.h> /* must be AFTER <sys/fs/fsdir.h>! */ 807c478bd9Sstevel@tonic-gate #include <sys/errno.h> 817c478bd9Sstevel@tonic-gate #include <sys/fssnap_if.h> 827c478bd9Sstevel@tonic-gate #include <sys/unistd.h> 837c478bd9Sstevel@tonic-gate #include <sys/sunddi.h> 847c478bd9Sstevel@tonic-gate 857c478bd9Sstevel@tonic-gate #include <sys/filio.h> /* _FIOIO */ 867c478bd9Sstevel@tonic-gate 877c478bd9Sstevel@tonic-gate #include <vm/hat.h> 887c478bd9Sstevel@tonic-gate #include <vm/page.h> 897c478bd9Sstevel@tonic-gate #include <vm/pvn.h> 907c478bd9Sstevel@tonic-gate #include <vm/as.h> 917c478bd9Sstevel@tonic-gate #include <vm/seg.h> 927c478bd9Sstevel@tonic-gate #include <vm/seg_map.h> 937c478bd9Sstevel@tonic-gate #include <vm/seg_vn.h> 947c478bd9Sstevel@tonic-gate #include <vm/seg_kmem.h> 957c478bd9Sstevel@tonic-gate #include <vm/rm.h> 967c478bd9Sstevel@tonic-gate #include <sys/swap.h> 977c478bd9Sstevel@tonic-gate 987c478bd9Sstevel@tonic-gate #include <fs/fs_subr.h> 997c478bd9Sstevel@tonic-gate 100986fd29aSsetje #include <sys/fs/decomp.h> 101986fd29aSsetje 1027c478bd9Sstevel@tonic-gate static struct instats ins; 1037c478bd9Sstevel@tonic-gate 1047c478bd9Sstevel@tonic-gate static int ufs_getpage_ra(struct vnode *, u_offset_t, struct seg *, caddr_t); 1057c478bd9Sstevel@tonic-gate static int ufs_getpage_miss(struct vnode *, u_offset_t, size_t, struct seg *, 1067c478bd9Sstevel@tonic-gate caddr_t, struct page **, size_t, enum seg_rw, int); 107da6c28aaSamw static int ufs_open(struct vnode **, int, struct cred *, caller_context_t *); 108da6c28aaSamw static int ufs_close(struct vnode *, int, int, offset_t, struct cred *, 109da6c28aaSamw caller_context_t *); 1107c478bd9Sstevel@tonic-gate static int ufs_read(struct vnode *, struct uio *, int, struct cred *, 1117c478bd9Sstevel@tonic-gate struct caller_context *); 1127c478bd9Sstevel@tonic-gate static int ufs_write(struct vnode *, struct uio *, int, struct cred *, 1137c478bd9Sstevel@tonic-gate struct caller_context *); 114da6c28aaSamw static int ufs_ioctl(struct vnode *, int, intptr_t, int, struct cred *, 115da6c28aaSamw int *, caller_context_t *); 116da6c28aaSamw static int ufs_getattr(struct vnode *, struct vattr *, int, struct cred *, 117da6c28aaSamw caller_context_t *); 1187c478bd9Sstevel@tonic-gate static int ufs_setattr(struct vnode *, struct vattr *, int, struct cred *, 1197c478bd9Sstevel@tonic-gate caller_context_t *); 120da6c28aaSamw static int ufs_access(struct vnode *, int, int, struct cred *, 121da6c28aaSamw caller_context_t *); 1227c478bd9Sstevel@tonic-gate static int ufs_lookup(struct vnode *, char *, struct vnode **, 123da6c28aaSamw struct pathname *, int, struct vnode *, struct cred *, 124da6c28aaSamw caller_context_t *, int *, pathname_t *); 1257c478bd9Sstevel@tonic-gate static int ufs_create(struct vnode *, char *, struct vattr *, enum vcexcl, 126da6c28aaSamw int, struct vnode **, struct cred *, int, 127da6c28aaSamw caller_context_t *, vsecattr_t *); 128da6c28aaSamw static int ufs_remove(struct vnode *, char *, struct cred *, 129da6c28aaSamw caller_context_t *, int); 130da6c28aaSamw static int ufs_link(struct vnode *, struct vnode *, char *, struct cred *, 131da6c28aaSamw caller_context_t *, int); 1327c478bd9Sstevel@tonic-gate static int ufs_rename(struct vnode *, char *, struct vnode *, char *, 133da6c28aaSamw struct cred *, caller_context_t *, int); 1347c478bd9Sstevel@tonic-gate static int ufs_mkdir(struct vnode *, char *, struct vattr *, struct vnode **, 135da6c28aaSamw struct cred *, caller_context_t *, int, vsecattr_t *); 136da6c28aaSamw static int ufs_rmdir(struct vnode *, char *, struct vnode *, struct cred *, 137da6c28aaSamw caller_context_t *, int); 138da6c28aaSamw static int ufs_readdir(struct vnode *, struct uio *, struct cred *, int *, 139da6c28aaSamw caller_context_t *, int); 1407c478bd9Sstevel@tonic-gate static int ufs_symlink(struct vnode *, char *, struct vattr *, char *, 141da6c28aaSamw struct cred *, caller_context_t *, int); 142da6c28aaSamw static int ufs_readlink(struct vnode *, struct uio *, struct cred *, 143da6c28aaSamw caller_context_t *); 144da6c28aaSamw static int ufs_fsync(struct vnode *, int, struct cred *, caller_context_t *); 145da6c28aaSamw static void ufs_inactive(struct vnode *, struct cred *, caller_context_t *); 146da6c28aaSamw static int ufs_fid(struct vnode *, struct fid *, caller_context_t *); 1477c478bd9Sstevel@tonic-gate static int ufs_rwlock(struct vnode *, int, caller_context_t *); 1487c478bd9Sstevel@tonic-gate static void ufs_rwunlock(struct vnode *, int, caller_context_t *); 149da6c28aaSamw static int ufs_seek(struct vnode *, offset_t, offset_t *, caller_context_t *); 1507c478bd9Sstevel@tonic-gate static int ufs_frlock(struct vnode *, int, struct flock64 *, int, offset_t, 151da6c28aaSamw struct flk_callback *, struct cred *, 152da6c28aaSamw caller_context_t *); 1537c478bd9Sstevel@tonic-gate static int ufs_space(struct vnode *, int, struct flock64 *, int, offset_t, 1547c478bd9Sstevel@tonic-gate cred_t *, caller_context_t *); 1557c478bd9Sstevel@tonic-gate static int ufs_getpage(struct vnode *, offset_t, size_t, uint_t *, 1567c478bd9Sstevel@tonic-gate struct page **, size_t, struct seg *, caddr_t, 157da6c28aaSamw enum seg_rw, struct cred *, caller_context_t *); 158da6c28aaSamw static int ufs_putpage(struct vnode *, offset_t, size_t, int, struct cred *, 159da6c28aaSamw caller_context_t *); 1607c478bd9Sstevel@tonic-gate static int ufs_putpages(struct vnode *, offset_t, size_t, int, struct cred *); 1617c478bd9Sstevel@tonic-gate static int ufs_map(struct vnode *, offset_t, struct as *, caddr_t *, size_t, 162da6c28aaSamw uchar_t, uchar_t, uint_t, struct cred *, caller_context_t *); 1637c478bd9Sstevel@tonic-gate static int ufs_addmap(struct vnode *, offset_t, struct as *, caddr_t, size_t, 164da6c28aaSamw uchar_t, uchar_t, uint_t, struct cred *, caller_context_t *); 1657c478bd9Sstevel@tonic-gate static int ufs_delmap(struct vnode *, offset_t, struct as *, caddr_t, size_t, 166da6c28aaSamw uint_t, uint_t, uint_t, struct cred *, caller_context_t *); 167da6c28aaSamw static int ufs_poll(vnode_t *, short, int, short *, struct pollhead **, 168da6c28aaSamw caller_context_t *); 169d7334e51Srm15945 static int ufs_dump(vnode_t *, caddr_t, offset_t, offset_t, 170d7334e51Srm15945 caller_context_t *); 171da6c28aaSamw static int ufs_l_pathconf(struct vnode *, int, ulong_t *, struct cred *, 172da6c28aaSamw caller_context_t *); 1737c478bd9Sstevel@tonic-gate static int ufs_pageio(struct vnode *, struct page *, u_offset_t, size_t, int, 174da6c28aaSamw struct cred *, caller_context_t *); 175d7334e51Srm15945 static int ufs_dumpctl(vnode_t *, int, offset_t *, caller_context_t *); 1767c478bd9Sstevel@tonic-gate static daddr32_t *save_dblks(struct inode *, struct ufsvfs *, daddr32_t *, 1777c478bd9Sstevel@tonic-gate daddr32_t *, int, int); 178da6c28aaSamw static int ufs_getsecattr(struct vnode *, vsecattr_t *, int, struct cred *, 179da6c28aaSamw caller_context_t *); 180da6c28aaSamw static int ufs_setsecattr(struct vnode *, vsecattr_t *, int, struct cred *, 181da6c28aaSamw caller_context_t *); 18260c8e821SFrank Batschulat static int ufs_priv_access(void *, int, struct cred *); 1839b5097eeSOwen Roberts static int ufs_eventlookup(struct vnode *, char *, struct cred *, 1849b5097eeSOwen Roberts struct vnode **); 18502ff05a9Svsakar extern int as_map_locked(struct as *, caddr_t, size_t, int ((*)()), void *); 18602ff05a9Svsakar 1877c478bd9Sstevel@tonic-gate /* 1887c478bd9Sstevel@tonic-gate * For lockfs: ulockfs begin/end is now inlined in the ufs_xxx functions. 1897c478bd9Sstevel@tonic-gate * 1907c478bd9Sstevel@tonic-gate * XXX - ULOCKFS in fs_pathconf and ufs_ioctl is not inlined yet. 1917c478bd9Sstevel@tonic-gate */ 1927c478bd9Sstevel@tonic-gate struct vnodeops *ufs_vnodeops; 1937c478bd9Sstevel@tonic-gate 194aa59c4cbSrsb /* NOTE: "not blkd" below means that the operation isn't blocked by lockfs */ 1957c478bd9Sstevel@tonic-gate const fs_operation_def_t ufs_vnodeops_template[] = { 196aa59c4cbSrsb VOPNAME_OPEN, { .vop_open = ufs_open }, /* not blkd */ 197aa59c4cbSrsb VOPNAME_CLOSE, { .vop_close = ufs_close }, /* not blkd */ 198aa59c4cbSrsb VOPNAME_READ, { .vop_read = ufs_read }, 199aa59c4cbSrsb VOPNAME_WRITE, { .vop_write = ufs_write }, 200aa59c4cbSrsb VOPNAME_IOCTL, { .vop_ioctl = ufs_ioctl }, 201aa59c4cbSrsb VOPNAME_GETATTR, { .vop_getattr = ufs_getattr }, 202aa59c4cbSrsb VOPNAME_SETATTR, { .vop_setattr = ufs_setattr }, 203aa59c4cbSrsb VOPNAME_ACCESS, { .vop_access = ufs_access }, 204aa59c4cbSrsb VOPNAME_LOOKUP, { .vop_lookup = ufs_lookup }, 205aa59c4cbSrsb VOPNAME_CREATE, { .vop_create = ufs_create }, 206aa59c4cbSrsb VOPNAME_REMOVE, { .vop_remove = ufs_remove }, 207aa59c4cbSrsb VOPNAME_LINK, { .vop_link = ufs_link }, 208aa59c4cbSrsb VOPNAME_RENAME, { .vop_rename = ufs_rename }, 209aa59c4cbSrsb VOPNAME_MKDIR, { .vop_mkdir = ufs_mkdir }, 210aa59c4cbSrsb VOPNAME_RMDIR, { .vop_rmdir = ufs_rmdir }, 211aa59c4cbSrsb VOPNAME_READDIR, { .vop_readdir = ufs_readdir }, 212aa59c4cbSrsb VOPNAME_SYMLINK, { .vop_symlink = ufs_symlink }, 213aa59c4cbSrsb VOPNAME_READLINK, { .vop_readlink = ufs_readlink }, 214aa59c4cbSrsb VOPNAME_FSYNC, { .vop_fsync = ufs_fsync }, 215aa59c4cbSrsb VOPNAME_INACTIVE, { .vop_inactive = ufs_inactive }, /* not blkd */ 216aa59c4cbSrsb VOPNAME_FID, { .vop_fid = ufs_fid }, 217aa59c4cbSrsb VOPNAME_RWLOCK, { .vop_rwlock = ufs_rwlock }, /* not blkd */ 218aa59c4cbSrsb VOPNAME_RWUNLOCK, { .vop_rwunlock = ufs_rwunlock }, /* not blkd */ 219aa59c4cbSrsb VOPNAME_SEEK, { .vop_seek = ufs_seek }, 220aa59c4cbSrsb VOPNAME_FRLOCK, { .vop_frlock = ufs_frlock }, 221aa59c4cbSrsb VOPNAME_SPACE, { .vop_space = ufs_space }, 222aa59c4cbSrsb VOPNAME_GETPAGE, { .vop_getpage = ufs_getpage }, 223aa59c4cbSrsb VOPNAME_PUTPAGE, { .vop_putpage = ufs_putpage }, 224aa59c4cbSrsb VOPNAME_MAP, { .vop_map = ufs_map }, 225aa59c4cbSrsb VOPNAME_ADDMAP, { .vop_addmap = ufs_addmap }, /* not blkd */ 226aa59c4cbSrsb VOPNAME_DELMAP, { .vop_delmap = ufs_delmap }, /* not blkd */ 227aa59c4cbSrsb VOPNAME_POLL, { .vop_poll = ufs_poll }, /* not blkd */ 228aa59c4cbSrsb VOPNAME_DUMP, { .vop_dump = ufs_dump }, 229aa59c4cbSrsb VOPNAME_PATHCONF, { .vop_pathconf = ufs_l_pathconf }, 230aa59c4cbSrsb VOPNAME_PAGEIO, { .vop_pageio = ufs_pageio }, 231aa59c4cbSrsb VOPNAME_DUMPCTL, { .vop_dumpctl = ufs_dumpctl }, 232aa59c4cbSrsb VOPNAME_GETSECATTR, { .vop_getsecattr = ufs_getsecattr }, 233aa59c4cbSrsb VOPNAME_SETSECATTR, { .vop_setsecattr = ufs_setsecattr }, 234aa59c4cbSrsb VOPNAME_VNEVENT, { .vop_vnevent = fs_vnevent_support }, 2357c478bd9Sstevel@tonic-gate NULL, NULL 2367c478bd9Sstevel@tonic-gate }; 2377c478bd9Sstevel@tonic-gate 2387c478bd9Sstevel@tonic-gate #define MAX_BACKFILE_COUNT 9999 2397c478bd9Sstevel@tonic-gate 2407c478bd9Sstevel@tonic-gate /* 2417c478bd9Sstevel@tonic-gate * Created by ufs_dumpctl() to store a file's disk block info into memory. 2427c478bd9Sstevel@tonic-gate * Used by ufs_dump() to dump data to disk directly. 2437c478bd9Sstevel@tonic-gate */ 2447c478bd9Sstevel@tonic-gate struct dump { 2457c478bd9Sstevel@tonic-gate struct inode *ip; /* the file we contain */ 2467c478bd9Sstevel@tonic-gate daddr_t fsbs; /* number of blocks stored */ 2477c478bd9Sstevel@tonic-gate struct timeval32 time; /* time stamp for the struct */ 2487c478bd9Sstevel@tonic-gate daddr32_t dblk[1]; /* place holder for block info */ 2497c478bd9Sstevel@tonic-gate }; 2507c478bd9Sstevel@tonic-gate 2517c478bd9Sstevel@tonic-gate static struct dump *dump_info = NULL; 2527c478bd9Sstevel@tonic-gate 2537c478bd9Sstevel@tonic-gate /* 2547c478bd9Sstevel@tonic-gate * Previously there was no special action required for ordinary files. 2557c478bd9Sstevel@tonic-gate * (Devices are handled through the device file system.) 2567c478bd9Sstevel@tonic-gate * Now we support Large Files and Large File API requires open to 2577c478bd9Sstevel@tonic-gate * fail if file is large. 2587c478bd9Sstevel@tonic-gate * We could take care to prevent data corruption 2597c478bd9Sstevel@tonic-gate * by doing an atomic check of size and truncate if file is opened with 2607c478bd9Sstevel@tonic-gate * FTRUNC flag set but traditionally this is being done by the vfs/vnode 2617c478bd9Sstevel@tonic-gate * layers. So taking care of truncation here is a change in the existing 2627c478bd9Sstevel@tonic-gate * semantics of VOP_OPEN and therefore we chose not to implement any thing 2637c478bd9Sstevel@tonic-gate * here. The check for the size of the file > 2GB is being done at the 2647c478bd9Sstevel@tonic-gate * vfs layer in routine vn_open(). 2657c478bd9Sstevel@tonic-gate */ 2667c478bd9Sstevel@tonic-gate 2677c478bd9Sstevel@tonic-gate /* ARGSUSED */ 2687c478bd9Sstevel@tonic-gate static int 269da6c28aaSamw ufs_open(struct vnode **vpp, int flag, struct cred *cr, caller_context_t *ct) 2707c478bd9Sstevel@tonic-gate { 2717c478bd9Sstevel@tonic-gate return (0); 2727c478bd9Sstevel@tonic-gate } 2737c478bd9Sstevel@tonic-gate 2747c478bd9Sstevel@tonic-gate /*ARGSUSED*/ 2757c478bd9Sstevel@tonic-gate static int 2767c478bd9Sstevel@tonic-gate ufs_close(struct vnode *vp, int flag, int count, offset_t offset, 277da6c28aaSamw struct cred *cr, caller_context_t *ct) 2787c478bd9Sstevel@tonic-gate { 2797c478bd9Sstevel@tonic-gate cleanlocks(vp, ttoproc(curthread)->p_pid, 0); 2807c478bd9Sstevel@tonic-gate cleanshares(vp, ttoproc(curthread)->p_pid); 2817c478bd9Sstevel@tonic-gate 2827c478bd9Sstevel@tonic-gate /* 2837c478bd9Sstevel@tonic-gate * Push partially filled cluster at last close. 2847c478bd9Sstevel@tonic-gate * ``last close'' is approximated because the dnlc 2857c478bd9Sstevel@tonic-gate * may have a hold on the vnode. 2867c478bd9Sstevel@tonic-gate * Checking for VBAD here will also act as a forced umount check. 2877c478bd9Sstevel@tonic-gate */ 2887c478bd9Sstevel@tonic-gate if (vp->v_count <= 2 && vp->v_type != VBAD) { 2897c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 2907c478bd9Sstevel@tonic-gate if (ip->i_delaylen) { 2917c478bd9Sstevel@tonic-gate ins.in_poc.value.ul++; 2927c478bd9Sstevel@tonic-gate (void) ufs_putpages(vp, ip->i_delayoff, ip->i_delaylen, 2937c478bd9Sstevel@tonic-gate B_ASYNC | B_FREE, cr); 2947c478bd9Sstevel@tonic-gate ip->i_delaylen = 0; 2957c478bd9Sstevel@tonic-gate } 2967c478bd9Sstevel@tonic-gate } 2977c478bd9Sstevel@tonic-gate 2987c478bd9Sstevel@tonic-gate return (0); 2997c478bd9Sstevel@tonic-gate } 3007c478bd9Sstevel@tonic-gate 3017c478bd9Sstevel@tonic-gate /*ARGSUSED*/ 3027c478bd9Sstevel@tonic-gate static int 3037c478bd9Sstevel@tonic-gate ufs_read(struct vnode *vp, struct uio *uiop, int ioflag, struct cred *cr, 3047c478bd9Sstevel@tonic-gate struct caller_context *ct) 3057c478bd9Sstevel@tonic-gate { 3067c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 3077c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp; 3087c478bd9Sstevel@tonic-gate struct ulockfs *ulp = NULL; 3097c478bd9Sstevel@tonic-gate int error = 0; 3107c478bd9Sstevel@tonic-gate int intrans = 0; 3117c478bd9Sstevel@tonic-gate 3127c478bd9Sstevel@tonic-gate ASSERT(RW_READ_HELD(&ip->i_rwlock)); 3137c478bd9Sstevel@tonic-gate 3147c478bd9Sstevel@tonic-gate /* 3157c478bd9Sstevel@tonic-gate * Mandatory locking needs to be done before ufs_lockfs_begin() 3167c478bd9Sstevel@tonic-gate * and TRANS_BEGIN_SYNC() calls since mandatory locks can sleep. 3177c478bd9Sstevel@tonic-gate */ 3187c478bd9Sstevel@tonic-gate if (MANDLOCK(vp, ip->i_mode)) { 3197c478bd9Sstevel@tonic-gate /* 3207c478bd9Sstevel@tonic-gate * ufs_getattr ends up being called by chklock 3217c478bd9Sstevel@tonic-gate */ 3227c478bd9Sstevel@tonic-gate error = chklock(vp, FREAD, uiop->uio_loffset, 3237c478bd9Sstevel@tonic-gate uiop->uio_resid, uiop->uio_fmode, ct); 3247c478bd9Sstevel@tonic-gate if (error) 3257c478bd9Sstevel@tonic-gate goto out; 3267c478bd9Sstevel@tonic-gate } 3277c478bd9Sstevel@tonic-gate 3287c478bd9Sstevel@tonic-gate ufsvfsp = ip->i_ufsvfs; 3297c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_READ_MASK); 3307c478bd9Sstevel@tonic-gate if (error) 3317c478bd9Sstevel@tonic-gate goto out; 3327c478bd9Sstevel@tonic-gate 3337c478bd9Sstevel@tonic-gate /* 3347c478bd9Sstevel@tonic-gate * In the case that a directory is opened for reading as a file 3357c478bd9Sstevel@tonic-gate * (eg "cat .") with the O_RSYNC, O_SYNC and O_DSYNC flags set. 3367c478bd9Sstevel@tonic-gate * The locking order had to be changed to avoid a deadlock with 3377c478bd9Sstevel@tonic-gate * an update taking place on that directory at the same time. 3387c478bd9Sstevel@tonic-gate */ 3397c478bd9Sstevel@tonic-gate if ((ip->i_mode & IFMT) == IFDIR) { 3407c478bd9Sstevel@tonic-gate 3417c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_READER); 3427c478bd9Sstevel@tonic-gate error = rdip(ip, uiop, ioflag, cr); 3437c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 3447c478bd9Sstevel@tonic-gate 3457c478bd9Sstevel@tonic-gate if (error) { 3467c478bd9Sstevel@tonic-gate if (ulp) 3477c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 3487c478bd9Sstevel@tonic-gate goto out; 3497c478bd9Sstevel@tonic-gate } 3507c478bd9Sstevel@tonic-gate 3517c478bd9Sstevel@tonic-gate if (ulp && (ioflag & FRSYNC) && (ioflag & (FSYNC | FDSYNC)) && 3527c478bd9Sstevel@tonic-gate TRANS_ISTRANS(ufsvfsp)) { 3537c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 3547c478bd9Sstevel@tonic-gate TRANS_BEGIN_SYNC(ufsvfsp, TOP_READ_SYNC, TOP_READ_SIZE, 3557c478bd9Sstevel@tonic-gate error); 3567c478bd9Sstevel@tonic-gate ASSERT(!error); 3577c478bd9Sstevel@tonic-gate TRANS_END_SYNC(ufsvfsp, error, TOP_READ_SYNC, 3587c478bd9Sstevel@tonic-gate TOP_READ_SIZE); 3597c478bd9Sstevel@tonic-gate rw_enter(&ip->i_rwlock, RW_READER); 3607c478bd9Sstevel@tonic-gate } 3617c478bd9Sstevel@tonic-gate } else { 3627c478bd9Sstevel@tonic-gate /* 3637c478bd9Sstevel@tonic-gate * Only transact reads to files opened for sync-read and 3647c478bd9Sstevel@tonic-gate * sync-write on a file system that is not write locked. 3657c478bd9Sstevel@tonic-gate * 3667c478bd9Sstevel@tonic-gate * The ``not write locked'' check prevents problems with 3677c478bd9Sstevel@tonic-gate * enabling/disabling logging on a busy file system. E.g., 3687c478bd9Sstevel@tonic-gate * logging exists at the beginning of the read but does not 3697c478bd9Sstevel@tonic-gate * at the end. 3707c478bd9Sstevel@tonic-gate * 3717c478bd9Sstevel@tonic-gate */ 3727c478bd9Sstevel@tonic-gate if (ulp && (ioflag & FRSYNC) && (ioflag & (FSYNC | FDSYNC)) && 3737c478bd9Sstevel@tonic-gate TRANS_ISTRANS(ufsvfsp)) { 3747c478bd9Sstevel@tonic-gate TRANS_BEGIN_SYNC(ufsvfsp, TOP_READ_SYNC, TOP_READ_SIZE, 3757c478bd9Sstevel@tonic-gate error); 3767c478bd9Sstevel@tonic-gate ASSERT(!error); 3777c478bd9Sstevel@tonic-gate intrans = 1; 3787c478bd9Sstevel@tonic-gate } 3797c478bd9Sstevel@tonic-gate 3807c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_READER); 3817c478bd9Sstevel@tonic-gate error = rdip(ip, uiop, ioflag, cr); 3827c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 3837c478bd9Sstevel@tonic-gate 3847c478bd9Sstevel@tonic-gate if (intrans) { 3857c478bd9Sstevel@tonic-gate TRANS_END_SYNC(ufsvfsp, error, TOP_READ_SYNC, 3867c478bd9Sstevel@tonic-gate TOP_READ_SIZE); 3877c478bd9Sstevel@tonic-gate } 3887c478bd9Sstevel@tonic-gate } 3897c478bd9Sstevel@tonic-gate 3907c478bd9Sstevel@tonic-gate if (ulp) { 3917c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 3927c478bd9Sstevel@tonic-gate } 3937c478bd9Sstevel@tonic-gate out: 3947c478bd9Sstevel@tonic-gate 3957c478bd9Sstevel@tonic-gate return (error); 3967c478bd9Sstevel@tonic-gate } 3977c478bd9Sstevel@tonic-gate 3987c478bd9Sstevel@tonic-gate extern int ufs_HW; /* high water mark */ 3997c478bd9Sstevel@tonic-gate extern int ufs_LW; /* low water mark */ 4007c478bd9Sstevel@tonic-gate int ufs_WRITES = 1; /* XXX - enable/disable */ 4017c478bd9Sstevel@tonic-gate int ufs_throttles = 0; /* throttling count */ 4027c478bd9Sstevel@tonic-gate int ufs_allow_shared_writes = 1; /* directio shared writes */ 4037c478bd9Sstevel@tonic-gate 4047c478bd9Sstevel@tonic-gate static int 4057c478bd9Sstevel@tonic-gate ufs_check_rewrite(struct inode *ip, struct uio *uiop, int ioflag) 4067c478bd9Sstevel@tonic-gate { 407f90bab20Sswilcox int shared_write; 408f90bab20Sswilcox 409f90bab20Sswilcox /* 410f90bab20Sswilcox * If the FDSYNC flag is set then ignore the global 411f90bab20Sswilcox * ufs_allow_shared_writes in this case. 412f90bab20Sswilcox */ 413f90bab20Sswilcox shared_write = (ioflag & FDSYNC) | ufs_allow_shared_writes; 4147c478bd9Sstevel@tonic-gate 4157c478bd9Sstevel@tonic-gate /* 4167c478bd9Sstevel@tonic-gate * Filter to determine if this request is suitable as a 4177c478bd9Sstevel@tonic-gate * concurrent rewrite. This write must not allocate blocks 4187c478bd9Sstevel@tonic-gate * by extending the file or filling in holes. No use trying 4197c478bd9Sstevel@tonic-gate * through FSYNC descriptors as the inode will be synchronously 4207c478bd9Sstevel@tonic-gate * updated after the write. The uio structure has not yet been 4217c478bd9Sstevel@tonic-gate * checked for sanity, so assume nothing. 4227c478bd9Sstevel@tonic-gate */ 4237c478bd9Sstevel@tonic-gate return (((ip->i_mode & IFMT) == IFREG) && !(ioflag & FAPPEND) && 4247c478bd9Sstevel@tonic-gate (uiop->uio_loffset >= (offset_t)0) && 4257c478bd9Sstevel@tonic-gate (uiop->uio_loffset < ip->i_size) && (uiop->uio_resid > 0) && 4267c478bd9Sstevel@tonic-gate ((ip->i_size - uiop->uio_loffset) >= uiop->uio_resid) && 4277c478bd9Sstevel@tonic-gate !(ioflag & FSYNC) && !bmap_has_holes(ip) && 428f90bab20Sswilcox shared_write); 4297c478bd9Sstevel@tonic-gate } 4307c478bd9Sstevel@tonic-gate 4317c478bd9Sstevel@tonic-gate /*ARGSUSED*/ 4327c478bd9Sstevel@tonic-gate static int 4337c478bd9Sstevel@tonic-gate ufs_write(struct vnode *vp, struct uio *uiop, int ioflag, cred_t *cr, 4347c478bd9Sstevel@tonic-gate caller_context_t *ct) 4357c478bd9Sstevel@tonic-gate { 4367c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 4377c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp; 4387c478bd9Sstevel@tonic-gate struct ulockfs *ulp; 4397c478bd9Sstevel@tonic-gate int retry = 1; 4407c478bd9Sstevel@tonic-gate int error, resv, resid = 0; 4417c478bd9Sstevel@tonic-gate int directio_status; 4427c478bd9Sstevel@tonic-gate int exclusive; 443f90bab20Sswilcox int rewriteflg; 4447c478bd9Sstevel@tonic-gate long start_resid = uiop->uio_resid; 4457c478bd9Sstevel@tonic-gate 4467c478bd9Sstevel@tonic-gate ASSERT(RW_LOCK_HELD(&ip->i_rwlock)); 4477c478bd9Sstevel@tonic-gate 4487c478bd9Sstevel@tonic-gate retry_mandlock: 4497c478bd9Sstevel@tonic-gate /* 4507c478bd9Sstevel@tonic-gate * Mandatory locking needs to be done before ufs_lockfs_begin() 4517c478bd9Sstevel@tonic-gate * and TRANS_BEGIN_[A]SYNC() calls since mandatory locks can sleep. 4527c478bd9Sstevel@tonic-gate * Check for forced unmounts normally done in ufs_lockfs_begin(). 4537c478bd9Sstevel@tonic-gate */ 4547c478bd9Sstevel@tonic-gate if ((ufsvfsp = ip->i_ufsvfs) == NULL) { 4557c478bd9Sstevel@tonic-gate error = EIO; 4567c478bd9Sstevel@tonic-gate goto out; 4577c478bd9Sstevel@tonic-gate } 4587c478bd9Sstevel@tonic-gate if (MANDLOCK(vp, ip->i_mode)) { 4597c478bd9Sstevel@tonic-gate 4607c478bd9Sstevel@tonic-gate ASSERT(RW_WRITE_HELD(&ip->i_rwlock)); 4617c478bd9Sstevel@tonic-gate 4627c478bd9Sstevel@tonic-gate /* 4637c478bd9Sstevel@tonic-gate * ufs_getattr ends up being called by chklock 4647c478bd9Sstevel@tonic-gate */ 4657c478bd9Sstevel@tonic-gate error = chklock(vp, FWRITE, uiop->uio_loffset, 4667c478bd9Sstevel@tonic-gate uiop->uio_resid, uiop->uio_fmode, ct); 4677c478bd9Sstevel@tonic-gate if (error) 4687c478bd9Sstevel@tonic-gate goto out; 4697c478bd9Sstevel@tonic-gate } 4707c478bd9Sstevel@tonic-gate 4717c478bd9Sstevel@tonic-gate /* i_rwlock can change in chklock */ 4727c478bd9Sstevel@tonic-gate exclusive = rw_write_held(&ip->i_rwlock); 473f90bab20Sswilcox rewriteflg = ufs_check_rewrite(ip, uiop, ioflag); 4747c478bd9Sstevel@tonic-gate 4757c478bd9Sstevel@tonic-gate /* 4767c478bd9Sstevel@tonic-gate * Check for fast-path special case of directio re-writes. 4777c478bd9Sstevel@tonic-gate */ 4787c478bd9Sstevel@tonic-gate if ((ip->i_flag & IDIRECTIO || ufsvfsp->vfs_forcedirectio) && 479f90bab20Sswilcox !exclusive && rewriteflg) { 4807c478bd9Sstevel@tonic-gate 4817c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_WRITE_MASK); 4827c478bd9Sstevel@tonic-gate if (error) 4837c478bd9Sstevel@tonic-gate goto out; 4847c478bd9Sstevel@tonic-gate 4857c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_READER); 4867c478bd9Sstevel@tonic-gate error = ufs_directio_write(ip, uiop, ioflag, 1, cr, 4877c478bd9Sstevel@tonic-gate &directio_status); 4887c478bd9Sstevel@tonic-gate if (directio_status == DIRECTIO_SUCCESS) { 4897c478bd9Sstevel@tonic-gate uint_t i_flag_save; 4907c478bd9Sstevel@tonic-gate 4917c478bd9Sstevel@tonic-gate if (start_resid != uiop->uio_resid) 4927c478bd9Sstevel@tonic-gate error = 0; 4937c478bd9Sstevel@tonic-gate /* 4947c478bd9Sstevel@tonic-gate * Special treatment of access times for re-writes. 4957c478bd9Sstevel@tonic-gate * If IMOD is not already set, then convert it 4967c478bd9Sstevel@tonic-gate * to IMODACC for this operation. This defers 4977c478bd9Sstevel@tonic-gate * entering a delta into the log until the inode 4987c478bd9Sstevel@tonic-gate * is flushed. This mimics what is done for read 4997c478bd9Sstevel@tonic-gate * operations and inode access time. 5007c478bd9Sstevel@tonic-gate */ 5017c478bd9Sstevel@tonic-gate mutex_enter(&ip->i_tlock); 5027c478bd9Sstevel@tonic-gate i_flag_save = ip->i_flag; 5037c478bd9Sstevel@tonic-gate ip->i_flag |= IUPD | ICHG; 5047c478bd9Sstevel@tonic-gate ip->i_seq++; 5057c478bd9Sstevel@tonic-gate ITIMES_NOLOCK(ip); 5067c478bd9Sstevel@tonic-gate if ((i_flag_save & IMOD) == 0) { 5077c478bd9Sstevel@tonic-gate ip->i_flag &= ~IMOD; 5087c478bd9Sstevel@tonic-gate ip->i_flag |= IMODACC; 5097c478bd9Sstevel@tonic-gate } 5107c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 5117c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 5127c478bd9Sstevel@tonic-gate if (ulp) 5137c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 5147c478bd9Sstevel@tonic-gate goto out; 5157c478bd9Sstevel@tonic-gate } 5167c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 5177c478bd9Sstevel@tonic-gate if (ulp) 5187c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 5197c478bd9Sstevel@tonic-gate } 5207c478bd9Sstevel@tonic-gate 5217c478bd9Sstevel@tonic-gate if (!exclusive && !rw_tryupgrade(&ip->i_rwlock)) { 5227c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 5237c478bd9Sstevel@tonic-gate rw_enter(&ip->i_rwlock, RW_WRITER); 5247c478bd9Sstevel@tonic-gate /* 5257c478bd9Sstevel@tonic-gate * Mandatory locking could have been enabled 5267c478bd9Sstevel@tonic-gate * after dropping the i_rwlock. 5277c478bd9Sstevel@tonic-gate */ 5287c478bd9Sstevel@tonic-gate if (MANDLOCK(vp, ip->i_mode)) 5297c478bd9Sstevel@tonic-gate goto retry_mandlock; 5307c478bd9Sstevel@tonic-gate } 5317c478bd9Sstevel@tonic-gate 5327c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_WRITE_MASK); 5337c478bd9Sstevel@tonic-gate if (error) 5347c478bd9Sstevel@tonic-gate goto out; 5357c478bd9Sstevel@tonic-gate 5367c478bd9Sstevel@tonic-gate /* 5377c478bd9Sstevel@tonic-gate * Amount of log space needed for this write 5387c478bd9Sstevel@tonic-gate */ 539f90bab20Sswilcox if (!rewriteflg || !(ioflag & FDSYNC)) 5407c478bd9Sstevel@tonic-gate TRANS_WRITE_RESV(ip, uiop, ulp, &resv, &resid); 5417c478bd9Sstevel@tonic-gate 5427c478bd9Sstevel@tonic-gate /* 5437c478bd9Sstevel@tonic-gate * Throttle writes. 5447c478bd9Sstevel@tonic-gate */ 5457c478bd9Sstevel@tonic-gate if (ufs_WRITES && (ip->i_writes > ufs_HW)) { 5467c478bd9Sstevel@tonic-gate mutex_enter(&ip->i_tlock); 5477c478bd9Sstevel@tonic-gate while (ip->i_writes > ufs_HW) { 5487c478bd9Sstevel@tonic-gate ufs_throttles++; 5497c478bd9Sstevel@tonic-gate cv_wait(&ip->i_wrcv, &ip->i_tlock); 5507c478bd9Sstevel@tonic-gate } 5517c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 5527c478bd9Sstevel@tonic-gate } 5537c478bd9Sstevel@tonic-gate 5547c478bd9Sstevel@tonic-gate /* 5557c478bd9Sstevel@tonic-gate * Enter Transaction 556f90bab20Sswilcox * 557f90bab20Sswilcox * If the write is a rewrite there is no need to open a transaction 558f90bab20Sswilcox * if the FDSYNC flag is set and not the FSYNC. In this case just 559f90bab20Sswilcox * set the IMODACC flag to modify do the update at a later time 560f90bab20Sswilcox * thus avoiding the overhead of the logging transaction that is 561f90bab20Sswilcox * not required. 5627c478bd9Sstevel@tonic-gate */ 5637c478bd9Sstevel@tonic-gate if (ioflag & (FSYNC|FDSYNC)) { 5647c478bd9Sstevel@tonic-gate if (ulp) { 565f90bab20Sswilcox if (rewriteflg) { 566f90bab20Sswilcox uint_t i_flag_save; 567f90bab20Sswilcox 568f90bab20Sswilcox rw_enter(&ip->i_contents, RW_READER); 569f90bab20Sswilcox mutex_enter(&ip->i_tlock); 570f90bab20Sswilcox i_flag_save = ip->i_flag; 571f90bab20Sswilcox ip->i_flag |= IUPD | ICHG; 572f90bab20Sswilcox ip->i_seq++; 573f90bab20Sswilcox ITIMES_NOLOCK(ip); 574f90bab20Sswilcox if ((i_flag_save & IMOD) == 0) { 575f90bab20Sswilcox ip->i_flag &= ~IMOD; 576f90bab20Sswilcox ip->i_flag |= IMODACC; 577f90bab20Sswilcox } 578f90bab20Sswilcox mutex_exit(&ip->i_tlock); 579f90bab20Sswilcox rw_exit(&ip->i_contents); 580f90bab20Sswilcox } else { 5817c478bd9Sstevel@tonic-gate int terr = 0; 582f90bab20Sswilcox TRANS_BEGIN_SYNC(ufsvfsp, TOP_WRITE_SYNC, resv, 583f90bab20Sswilcox terr); 5847c478bd9Sstevel@tonic-gate ASSERT(!terr); 5857c478bd9Sstevel@tonic-gate } 586f90bab20Sswilcox } 5877c478bd9Sstevel@tonic-gate } else { 5887c478bd9Sstevel@tonic-gate if (ulp) 5897c478bd9Sstevel@tonic-gate TRANS_BEGIN_ASYNC(ufsvfsp, TOP_WRITE, resv); 5907c478bd9Sstevel@tonic-gate } 5917c478bd9Sstevel@tonic-gate 5927c478bd9Sstevel@tonic-gate /* 5937c478bd9Sstevel@tonic-gate * Write the file 5947c478bd9Sstevel@tonic-gate */ 5957c478bd9Sstevel@tonic-gate rw_enter(&ufsvfsp->vfs_dqrwlock, RW_READER); 5967c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_WRITER); 5977c478bd9Sstevel@tonic-gate if ((ioflag & FAPPEND) != 0 && (ip->i_mode & IFMT) == IFREG) { 5987c478bd9Sstevel@tonic-gate /* 5997c478bd9Sstevel@tonic-gate * In append mode start at end of file. 6007c478bd9Sstevel@tonic-gate */ 6017c478bd9Sstevel@tonic-gate uiop->uio_loffset = ip->i_size; 6027c478bd9Sstevel@tonic-gate } 6037c478bd9Sstevel@tonic-gate 6047c478bd9Sstevel@tonic-gate /* 6057c478bd9Sstevel@tonic-gate * Mild optimisation, don't call ufs_trans_write() unless we have to 6067c478bd9Sstevel@tonic-gate * Also, suppress file system full messages if we will retry. 6077c478bd9Sstevel@tonic-gate */ 6087c478bd9Sstevel@tonic-gate if (retry) 6097c478bd9Sstevel@tonic-gate ip->i_flag |= IQUIET; 6107c478bd9Sstevel@tonic-gate if (resid) { 6117c478bd9Sstevel@tonic-gate TRANS_WRITE(ip, uiop, ioflag, error, ulp, cr, resv, resid); 6127c478bd9Sstevel@tonic-gate } else { 6137c478bd9Sstevel@tonic-gate error = wrip(ip, uiop, ioflag, cr); 6147c478bd9Sstevel@tonic-gate } 6157c478bd9Sstevel@tonic-gate ip->i_flag &= ~IQUIET; 6167c478bd9Sstevel@tonic-gate 6177c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 6187c478bd9Sstevel@tonic-gate rw_exit(&ufsvfsp->vfs_dqrwlock); 6197c478bd9Sstevel@tonic-gate 6207c478bd9Sstevel@tonic-gate /* 6217c478bd9Sstevel@tonic-gate * Leave Transaction 6227c478bd9Sstevel@tonic-gate */ 6237c478bd9Sstevel@tonic-gate if (ulp) { 6247c478bd9Sstevel@tonic-gate if (ioflag & (FSYNC|FDSYNC)) { 625f90bab20Sswilcox if (!rewriteflg) { 6267c478bd9Sstevel@tonic-gate int terr = 0; 627f90bab20Sswilcox 628f90bab20Sswilcox TRANS_END_SYNC(ufsvfsp, terr, TOP_WRITE_SYNC, 629f90bab20Sswilcox resv); 6307c478bd9Sstevel@tonic-gate if (error == 0) 6317c478bd9Sstevel@tonic-gate error = terr; 632f90bab20Sswilcox } 6337c478bd9Sstevel@tonic-gate } else { 6347c478bd9Sstevel@tonic-gate TRANS_END_ASYNC(ufsvfsp, TOP_WRITE, resv); 6357c478bd9Sstevel@tonic-gate } 6367c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 6377c478bd9Sstevel@tonic-gate } 6387c478bd9Sstevel@tonic-gate out: 6397c478bd9Sstevel@tonic-gate if ((error == ENOSPC) && retry && TRANS_ISTRANS(ufsvfsp)) { 6407c478bd9Sstevel@tonic-gate /* 6417c478bd9Sstevel@tonic-gate * Any blocks tied up in pending deletes? 6427c478bd9Sstevel@tonic-gate */ 6437c478bd9Sstevel@tonic-gate ufs_delete_drain_wait(ufsvfsp, 1); 6447c478bd9Sstevel@tonic-gate retry = 0; 6457c478bd9Sstevel@tonic-gate goto retry_mandlock; 6467c478bd9Sstevel@tonic-gate } 6477c478bd9Sstevel@tonic-gate 6487c478bd9Sstevel@tonic-gate if (error == ENOSPC && (start_resid != uiop->uio_resid)) 6497c478bd9Sstevel@tonic-gate error = 0; 6507c478bd9Sstevel@tonic-gate 6517c478bd9Sstevel@tonic-gate return (error); 6527c478bd9Sstevel@tonic-gate } 6537c478bd9Sstevel@tonic-gate 6547c478bd9Sstevel@tonic-gate /* 6557c478bd9Sstevel@tonic-gate * Don't cache write blocks to files with the sticky bit set. 6567c478bd9Sstevel@tonic-gate * Used to keep swap files from blowing the page cache on a server. 6577c478bd9Sstevel@tonic-gate */ 6587c478bd9Sstevel@tonic-gate int stickyhack = 1; 6597c478bd9Sstevel@tonic-gate 6607c478bd9Sstevel@tonic-gate /* 6617c478bd9Sstevel@tonic-gate * Free behind hacks. The pager is busted. 6627c478bd9Sstevel@tonic-gate * XXX - need to pass the information down to writedone() in a flag like B_SEQ 6637c478bd9Sstevel@tonic-gate * or B_FREE_IF_TIGHT_ON_MEMORY. 6647c478bd9Sstevel@tonic-gate */ 6657c478bd9Sstevel@tonic-gate int freebehind = 1; 666562eee46Srbourbon int smallfile = 0; 667562eee46Srbourbon u_offset_t smallfile64 = 32 * 1024; 6687c478bd9Sstevel@tonic-gate 6697c478bd9Sstevel@tonic-gate /* 6707c478bd9Sstevel@tonic-gate * While we should, in most cases, cache the pages for write, we 6717c478bd9Sstevel@tonic-gate * may also want to cache the pages for read as long as they are 6727c478bd9Sstevel@tonic-gate * frequently re-usable. 6737c478bd9Sstevel@tonic-gate * 6747c478bd9Sstevel@tonic-gate * If cache_read_ahead = 1, the pages for read will go to the tail 6757c478bd9Sstevel@tonic-gate * of the cache list when they are released, otherwise go to the head. 6767c478bd9Sstevel@tonic-gate */ 6777c478bd9Sstevel@tonic-gate int cache_read_ahead = 0; 6787c478bd9Sstevel@tonic-gate 6797c478bd9Sstevel@tonic-gate /* 680562eee46Srbourbon * Freebehind exists so that as we read large files sequentially we 681562eee46Srbourbon * don't consume most of memory with pages from a few files. It takes 682562eee46Srbourbon * longer to re-read from disk multiple small files as it does reading 683562eee46Srbourbon * one large one sequentially. As system memory grows customers need 684562eee46Srbourbon * to retain bigger chunks of files in memory. The advent of the 685562eee46Srbourbon * cachelist opens up of the possibility freeing pages to the head or 686562eee46Srbourbon * tail of the list. 687562eee46Srbourbon * 688562eee46Srbourbon * Not freeing a page is a bet that the page will be read again before 689562eee46Srbourbon * it's segmap slot is needed for something else. If we loose the bet, 690562eee46Srbourbon * it means some other thread is burdened with the page free we did 691562eee46Srbourbon * not do. If we win we save a free and reclaim. 692562eee46Srbourbon * 693562eee46Srbourbon * Freeing it at the tail vs the head of cachelist is a bet that the 694562eee46Srbourbon * page will survive until the next read. It's also saying that this 695562eee46Srbourbon * page is more likely to be re-used than a page freed some time ago 696562eee46Srbourbon * and never reclaimed. 697562eee46Srbourbon * 698562eee46Srbourbon * Freebehind maintains a range of file offset [smallfile1; smallfile2] 699562eee46Srbourbon * 700562eee46Srbourbon * 0 < offset < smallfile1 : pages are not freed. 701562eee46Srbourbon * smallfile1 < offset < smallfile2 : pages freed to tail of cachelist. 702562eee46Srbourbon * smallfile2 < offset : pages freed to head of cachelist. 703562eee46Srbourbon * 704562eee46Srbourbon * The range is computed at most once per second and depends on 705562eee46Srbourbon * freemem and ncpus_online. Both parameters are bounded to be 706562eee46Srbourbon * >= smallfile && >= smallfile64. 707562eee46Srbourbon * 708562eee46Srbourbon * smallfile1 = (free memory / ncpu) / 1000 709562eee46Srbourbon * smallfile2 = (free memory / ncpu) / 10 710562eee46Srbourbon * 711562eee46Srbourbon * A few examples values: 712562eee46Srbourbon * 713562eee46Srbourbon * Free Mem (in Bytes) [smallfile1; smallfile2] [smallfile1; smallfile2] 714562eee46Srbourbon * ncpus_online = 4 ncpus_online = 64 715562eee46Srbourbon * ------------------ ----------------------- ----------------------- 716562eee46Srbourbon * 1G [256K; 25M] [32K; 1.5M] 717562eee46Srbourbon * 10G [2.5M; 250M] [156K; 15M] 718562eee46Srbourbon * 100G [25M; 2.5G] [1.5M; 150M] 719562eee46Srbourbon * 720562eee46Srbourbon */ 721562eee46Srbourbon 722562eee46Srbourbon #define SMALLFILE1_D 1000 723562eee46Srbourbon #define SMALLFILE2_D 10 724562eee46Srbourbon static u_offset_t smallfile1 = 32 * 1024; 725562eee46Srbourbon static u_offset_t smallfile2 = 32 * 1024; 726562eee46Srbourbon static clock_t smallfile_update = 0; /* lbolt value of when to recompute */ 727562eee46Srbourbon uint_t smallfile1_d = SMALLFILE1_D; 728562eee46Srbourbon uint_t smallfile2_d = SMALLFILE2_D; 729562eee46Srbourbon 730562eee46Srbourbon /* 7317c478bd9Sstevel@tonic-gate * wrip does the real work of write requests for ufs. 7327c478bd9Sstevel@tonic-gate */ 7337c478bd9Sstevel@tonic-gate int 7347c478bd9Sstevel@tonic-gate wrip(struct inode *ip, struct uio *uio, int ioflag, struct cred *cr) 7357c478bd9Sstevel@tonic-gate { 7367c478bd9Sstevel@tonic-gate rlim64_t limit = uio->uio_llimit; 7377c478bd9Sstevel@tonic-gate u_offset_t off; 7387c478bd9Sstevel@tonic-gate u_offset_t old_i_size; 7397c478bd9Sstevel@tonic-gate struct fs *fs; 7407c478bd9Sstevel@tonic-gate struct vnode *vp; 7417c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp; 7427c478bd9Sstevel@tonic-gate caddr_t base; 7437c478bd9Sstevel@tonic-gate long start_resid = uio->uio_resid; /* save starting resid */ 7447c478bd9Sstevel@tonic-gate long premove_resid; /* resid before uiomove() */ 7457c478bd9Sstevel@tonic-gate uint_t flags; 7467c478bd9Sstevel@tonic-gate int newpage; 7477c478bd9Sstevel@tonic-gate int iupdat_flag, directio_status; 7487c478bd9Sstevel@tonic-gate int n, on, mapon; 7497c478bd9Sstevel@tonic-gate int error, pagecreate; 7507c478bd9Sstevel@tonic-gate int do_dqrwlock; /* drop/reacquire vfs_dqrwlock */ 7517c478bd9Sstevel@tonic-gate int32_t iblocks; 7527c478bd9Sstevel@tonic-gate int new_iblocks; 7537c478bd9Sstevel@tonic-gate 7547c478bd9Sstevel@tonic-gate /* 7557c478bd9Sstevel@tonic-gate * ip->i_size is incremented before the uiomove 7567c478bd9Sstevel@tonic-gate * is done on a write. If the move fails (bad user 7577c478bd9Sstevel@tonic-gate * address) reset ip->i_size. 7587c478bd9Sstevel@tonic-gate * The better way would be to increment ip->i_size 7597c478bd9Sstevel@tonic-gate * only if the uiomove succeeds. 7607c478bd9Sstevel@tonic-gate */ 7617c478bd9Sstevel@tonic-gate int i_size_changed = 0; 7627c478bd9Sstevel@tonic-gate o_mode_t type; 7637c478bd9Sstevel@tonic-gate int i_seq_needed = 0; 7647c478bd9Sstevel@tonic-gate 7657c478bd9Sstevel@tonic-gate vp = ITOV(ip); 7667c478bd9Sstevel@tonic-gate 7677c478bd9Sstevel@tonic-gate /* 7687c478bd9Sstevel@tonic-gate * check for forced unmount - should not happen as 7697c478bd9Sstevel@tonic-gate * the request passed the lockfs checks. 7707c478bd9Sstevel@tonic-gate */ 7717c478bd9Sstevel@tonic-gate if ((ufsvfsp = ip->i_ufsvfs) == NULL) 7727c478bd9Sstevel@tonic-gate return (EIO); 7737c478bd9Sstevel@tonic-gate 7747c478bd9Sstevel@tonic-gate fs = ip->i_fs; 7757c478bd9Sstevel@tonic-gate 7767c478bd9Sstevel@tonic-gate ASSERT(RW_WRITE_HELD(&ip->i_contents)); 7777c478bd9Sstevel@tonic-gate 7787c478bd9Sstevel@tonic-gate /* check for valid filetype */ 7797c478bd9Sstevel@tonic-gate type = ip->i_mode & IFMT; 7807c478bd9Sstevel@tonic-gate if ((type != IFREG) && (type != IFDIR) && (type != IFATTRDIR) && 7817c478bd9Sstevel@tonic-gate (type != IFLNK) && (type != IFSHAD)) { 7827c478bd9Sstevel@tonic-gate return (EIO); 7837c478bd9Sstevel@tonic-gate } 7847c478bd9Sstevel@tonic-gate 7857c478bd9Sstevel@tonic-gate /* 7867c478bd9Sstevel@tonic-gate * the actual limit of UFS file size 7877c478bd9Sstevel@tonic-gate * is UFS_MAXOFFSET_T 7887c478bd9Sstevel@tonic-gate */ 7897c478bd9Sstevel@tonic-gate if (limit == RLIM64_INFINITY || limit > MAXOFFSET_T) 7907c478bd9Sstevel@tonic-gate limit = MAXOFFSET_T; 7917c478bd9Sstevel@tonic-gate 7927c478bd9Sstevel@tonic-gate if (uio->uio_loffset >= limit) { 7937c478bd9Sstevel@tonic-gate proc_t *p = ttoproc(curthread); 7947c478bd9Sstevel@tonic-gate 7957c478bd9Sstevel@tonic-gate mutex_enter(&p->p_lock); 7967c478bd9Sstevel@tonic-gate (void) rctl_action(rctlproc_legacy[RLIMIT_FSIZE], p->p_rctls, 7977c478bd9Sstevel@tonic-gate p, RCA_UNSAFE_SIGINFO); 7987c478bd9Sstevel@tonic-gate mutex_exit(&p->p_lock); 7997c478bd9Sstevel@tonic-gate return (EFBIG); 8007c478bd9Sstevel@tonic-gate } 8017c478bd9Sstevel@tonic-gate 8027c478bd9Sstevel@tonic-gate /* 8037c478bd9Sstevel@tonic-gate * if largefiles are disallowed, the limit is 8047c478bd9Sstevel@tonic-gate * the pre-largefiles value of 2GB 8057c478bd9Sstevel@tonic-gate */ 8067c478bd9Sstevel@tonic-gate if (ufsvfsp->vfs_lfflags & UFS_LARGEFILES) 8077c478bd9Sstevel@tonic-gate limit = MIN(UFS_MAXOFFSET_T, limit); 8087c478bd9Sstevel@tonic-gate else 8097c478bd9Sstevel@tonic-gate limit = MIN(MAXOFF32_T, limit); 8107c478bd9Sstevel@tonic-gate 8117c478bd9Sstevel@tonic-gate if (uio->uio_loffset < (offset_t)0) { 8127c478bd9Sstevel@tonic-gate return (EINVAL); 8137c478bd9Sstevel@tonic-gate } 8147c478bd9Sstevel@tonic-gate if (uio->uio_resid == 0) { 8157c478bd9Sstevel@tonic-gate return (0); 8167c478bd9Sstevel@tonic-gate } 8177c478bd9Sstevel@tonic-gate 8187c478bd9Sstevel@tonic-gate if (uio->uio_loffset >= limit) 8197c478bd9Sstevel@tonic-gate return (EFBIG); 8207c478bd9Sstevel@tonic-gate 8217c478bd9Sstevel@tonic-gate ip->i_flag |= INOACC; /* don't update ref time in getpage */ 8227c478bd9Sstevel@tonic-gate 8237c478bd9Sstevel@tonic-gate if (ioflag & (FSYNC|FDSYNC)) { 8247c478bd9Sstevel@tonic-gate ip->i_flag |= ISYNC; 8257c478bd9Sstevel@tonic-gate iupdat_flag = 1; 8267c478bd9Sstevel@tonic-gate } 8277c478bd9Sstevel@tonic-gate /* 8287c478bd9Sstevel@tonic-gate * Try to go direct 8297c478bd9Sstevel@tonic-gate */ 8307c478bd9Sstevel@tonic-gate if (ip->i_flag & IDIRECTIO || ufsvfsp->vfs_forcedirectio) { 8317c478bd9Sstevel@tonic-gate uio->uio_llimit = limit; 8327c478bd9Sstevel@tonic-gate error = ufs_directio_write(ip, uio, ioflag, 0, cr, 8337c478bd9Sstevel@tonic-gate &directio_status); 8347c478bd9Sstevel@tonic-gate /* 8357c478bd9Sstevel@tonic-gate * If ufs_directio wrote to the file or set the flags, 8367c478bd9Sstevel@tonic-gate * we need to update i_seq, but it may be deferred. 8377c478bd9Sstevel@tonic-gate */ 8387c478bd9Sstevel@tonic-gate if (start_resid != uio->uio_resid || 8397c478bd9Sstevel@tonic-gate (ip->i_flag & (ICHG|IUPD))) { 8407c478bd9Sstevel@tonic-gate i_seq_needed = 1; 8417c478bd9Sstevel@tonic-gate ip->i_flag |= ISEQ; 8427c478bd9Sstevel@tonic-gate } 8437c478bd9Sstevel@tonic-gate if (directio_status == DIRECTIO_SUCCESS) 8447c478bd9Sstevel@tonic-gate goto out; 8457c478bd9Sstevel@tonic-gate } 8467c478bd9Sstevel@tonic-gate 8477c478bd9Sstevel@tonic-gate /* 8487c478bd9Sstevel@tonic-gate * Behavior with respect to dropping/reacquiring vfs_dqrwlock: 8497c478bd9Sstevel@tonic-gate * 8507c478bd9Sstevel@tonic-gate * o shadow inodes: vfs_dqrwlock is not held at all 8517c478bd9Sstevel@tonic-gate * o quota updates: vfs_dqrwlock is read or write held 8527c478bd9Sstevel@tonic-gate * o other updates: vfs_dqrwlock is read held 8537c478bd9Sstevel@tonic-gate * 8547c478bd9Sstevel@tonic-gate * The first case is the only one where we do not hold 8557c478bd9Sstevel@tonic-gate * vfs_dqrwlock at all while entering wrip(). 8567c478bd9Sstevel@tonic-gate * We must make sure not to downgrade/drop vfs_dqrwlock if we 8577c478bd9Sstevel@tonic-gate * have it as writer, i.e. if we are updating the quota inode. 8587c478bd9Sstevel@tonic-gate * There is no potential deadlock scenario in this case as 8597c478bd9Sstevel@tonic-gate * ufs_getpage() takes care of this and avoids reacquiring 8607c478bd9Sstevel@tonic-gate * vfs_dqrwlock in that case. 8617c478bd9Sstevel@tonic-gate * 8627c478bd9Sstevel@tonic-gate * This check is done here since the above conditions do not change 8637c478bd9Sstevel@tonic-gate * and we possibly loop below, so save a few cycles. 8647c478bd9Sstevel@tonic-gate */ 8657c478bd9Sstevel@tonic-gate if ((type == IFSHAD) || 8667c478bd9Sstevel@tonic-gate (rw_owner(&ufsvfsp->vfs_dqrwlock) == curthread)) { 8677c478bd9Sstevel@tonic-gate do_dqrwlock = 0; 8687c478bd9Sstevel@tonic-gate } else { 8697c478bd9Sstevel@tonic-gate do_dqrwlock = 1; 8707c478bd9Sstevel@tonic-gate } 8717c478bd9Sstevel@tonic-gate 8727c478bd9Sstevel@tonic-gate /* 8737c478bd9Sstevel@tonic-gate * Large Files: We cast MAXBMASK to offset_t 8747c478bd9Sstevel@tonic-gate * inorder to mask out the higher bits. Since offset_t 8757c478bd9Sstevel@tonic-gate * is a signed value, the high order bit set in MAXBMASK 8767c478bd9Sstevel@tonic-gate * value makes it do the right thing by having all bits 1 8777c478bd9Sstevel@tonic-gate * in the higher word. May be removed for _SOLARIS64_. 8787c478bd9Sstevel@tonic-gate */ 8797c478bd9Sstevel@tonic-gate 8807c478bd9Sstevel@tonic-gate fs = ip->i_fs; 8817c478bd9Sstevel@tonic-gate do { 8827c478bd9Sstevel@tonic-gate u_offset_t uoff = uio->uio_loffset; 8837c478bd9Sstevel@tonic-gate off = uoff & (offset_t)MAXBMASK; 8847c478bd9Sstevel@tonic-gate mapon = (int)(uoff & (offset_t)MAXBOFFSET); 8857c478bd9Sstevel@tonic-gate on = (int)blkoff(fs, uoff); 8867c478bd9Sstevel@tonic-gate n = (int)MIN(fs->fs_bsize - on, uio->uio_resid); 8877c478bd9Sstevel@tonic-gate new_iblocks = 1; 8887c478bd9Sstevel@tonic-gate 8897c478bd9Sstevel@tonic-gate if (type == IFREG && uoff + n >= limit) { 8907c478bd9Sstevel@tonic-gate if (uoff >= limit) { 8917c478bd9Sstevel@tonic-gate error = EFBIG; 8927c478bd9Sstevel@tonic-gate goto out; 8937c478bd9Sstevel@tonic-gate } 8947c478bd9Sstevel@tonic-gate /* 8957c478bd9Sstevel@tonic-gate * since uoff + n >= limit, 8967c478bd9Sstevel@tonic-gate * therefore n >= limit - uoff, and n is an int 8977c478bd9Sstevel@tonic-gate * so it is safe to cast it to an int 8987c478bd9Sstevel@tonic-gate */ 8997c478bd9Sstevel@tonic-gate n = (int)(limit - (rlim64_t)uoff); 9007c478bd9Sstevel@tonic-gate } 9017c478bd9Sstevel@tonic-gate if (uoff + n > ip->i_size) { 9027c478bd9Sstevel@tonic-gate /* 9037c478bd9Sstevel@tonic-gate * We are extending the length of the file. 9047c478bd9Sstevel@tonic-gate * bmap is used so that we are sure that 9057c478bd9Sstevel@tonic-gate * if we need to allocate new blocks, that it 9067c478bd9Sstevel@tonic-gate * is done here before we up the file size. 9077c478bd9Sstevel@tonic-gate */ 9087c478bd9Sstevel@tonic-gate error = bmap_write(ip, uoff, (int)(on + n), 909303bf60bSsdebnath mapon == 0, NULL, cr); 9107c478bd9Sstevel@tonic-gate /* 9117c478bd9Sstevel@tonic-gate * bmap_write never drops i_contents so if 9127c478bd9Sstevel@tonic-gate * the flags are set it changed the file. 9137c478bd9Sstevel@tonic-gate */ 9147c478bd9Sstevel@tonic-gate if (ip->i_flag & (ICHG|IUPD)) { 9157c478bd9Sstevel@tonic-gate i_seq_needed = 1; 9167c478bd9Sstevel@tonic-gate ip->i_flag |= ISEQ; 9177c478bd9Sstevel@tonic-gate } 9187c478bd9Sstevel@tonic-gate if (error) 9197c478bd9Sstevel@tonic-gate break; 9207c478bd9Sstevel@tonic-gate /* 9217c478bd9Sstevel@tonic-gate * There is a window of vulnerability here. 9227c478bd9Sstevel@tonic-gate * The sequence of operations: allocate file 9237c478bd9Sstevel@tonic-gate * system blocks, uiomove the data into pages, 9247c478bd9Sstevel@tonic-gate * and then update the size of the file in the 9257c478bd9Sstevel@tonic-gate * inode, must happen atomically. However, due 9267c478bd9Sstevel@tonic-gate * to current locking constraints, this can not 9277c478bd9Sstevel@tonic-gate * be done. 9287c478bd9Sstevel@tonic-gate */ 9297c478bd9Sstevel@tonic-gate ASSERT(ip->i_writer == NULL); 9307c478bd9Sstevel@tonic-gate ip->i_writer = curthread; 9317c478bd9Sstevel@tonic-gate i_size_changed = 1; 9327c478bd9Sstevel@tonic-gate /* 9337c478bd9Sstevel@tonic-gate * If we are writing from the beginning of 9347c478bd9Sstevel@tonic-gate * the mapping, we can just create the 9357c478bd9Sstevel@tonic-gate * pages without having to read them. 9367c478bd9Sstevel@tonic-gate */ 9377c478bd9Sstevel@tonic-gate pagecreate = (mapon == 0); 9387c478bd9Sstevel@tonic-gate } else if (n == MAXBSIZE) { 9397c478bd9Sstevel@tonic-gate /* 9407c478bd9Sstevel@tonic-gate * Going to do a whole mappings worth, 9417c478bd9Sstevel@tonic-gate * so we can just create the pages w/o 9427c478bd9Sstevel@tonic-gate * having to read them in. But before 9437c478bd9Sstevel@tonic-gate * we do that, we need to make sure any 9447c478bd9Sstevel@tonic-gate * needed blocks are allocated first. 9457c478bd9Sstevel@tonic-gate */ 9467c478bd9Sstevel@tonic-gate iblocks = ip->i_blocks; 947303bf60bSsdebnath error = bmap_write(ip, uoff, (int)(on + n), 948303bf60bSsdebnath BI_ALLOC_ONLY, NULL, cr); 9497c478bd9Sstevel@tonic-gate /* 9507c478bd9Sstevel@tonic-gate * bmap_write never drops i_contents so if 9517c478bd9Sstevel@tonic-gate * the flags are set it changed the file. 9527c478bd9Sstevel@tonic-gate */ 9537c478bd9Sstevel@tonic-gate if (ip->i_flag & (ICHG|IUPD)) { 9547c478bd9Sstevel@tonic-gate i_seq_needed = 1; 9557c478bd9Sstevel@tonic-gate ip->i_flag |= ISEQ; 9567c478bd9Sstevel@tonic-gate } 9577c478bd9Sstevel@tonic-gate if (error) 9587c478bd9Sstevel@tonic-gate break; 9597c478bd9Sstevel@tonic-gate pagecreate = 1; 9607c478bd9Sstevel@tonic-gate /* 9617c478bd9Sstevel@tonic-gate * check if the new created page needed the 9627c478bd9Sstevel@tonic-gate * allocation of new disk blocks. 9637c478bd9Sstevel@tonic-gate */ 9647c478bd9Sstevel@tonic-gate if (iblocks == ip->i_blocks) 9657c478bd9Sstevel@tonic-gate new_iblocks = 0; /* no new blocks allocated */ 9667c478bd9Sstevel@tonic-gate } else { 9677c478bd9Sstevel@tonic-gate pagecreate = 0; 9687c478bd9Sstevel@tonic-gate /* 9697c478bd9Sstevel@tonic-gate * In sync mode flush the indirect blocks which 9707c478bd9Sstevel@tonic-gate * may have been allocated and not written on 9717c478bd9Sstevel@tonic-gate * disk. In above cases bmap_write will allocate 9727c478bd9Sstevel@tonic-gate * in sync mode. 9737c478bd9Sstevel@tonic-gate */ 9747c478bd9Sstevel@tonic-gate if (ioflag & (FSYNC|FDSYNC)) { 9757c478bd9Sstevel@tonic-gate error = ufs_indirblk_sync(ip, uoff); 9767c478bd9Sstevel@tonic-gate if (error) 9777c478bd9Sstevel@tonic-gate break; 9787c478bd9Sstevel@tonic-gate } 9797c478bd9Sstevel@tonic-gate } 9807c478bd9Sstevel@tonic-gate 9817c478bd9Sstevel@tonic-gate /* 9827c478bd9Sstevel@tonic-gate * At this point we can enter ufs_getpage() in one 9837c478bd9Sstevel@tonic-gate * of two ways: 9847c478bd9Sstevel@tonic-gate * 1) segmap_getmapflt() calls ufs_getpage() when the 9857c478bd9Sstevel@tonic-gate * forcefault parameter is true (pagecreate == 0) 9867c478bd9Sstevel@tonic-gate * 2) uiomove() causes a page fault. 9877c478bd9Sstevel@tonic-gate * 9887c478bd9Sstevel@tonic-gate * We have to drop the contents lock to prevent the VM 989da6c28aaSamw * system from trying to reacquire it in ufs_getpage() 9907c478bd9Sstevel@tonic-gate * should the uiomove cause a pagefault. 9917c478bd9Sstevel@tonic-gate * 9927c478bd9Sstevel@tonic-gate * We have to drop the reader vfs_dqrwlock here as well. 9937c478bd9Sstevel@tonic-gate */ 9947c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 9957c478bd9Sstevel@tonic-gate if (do_dqrwlock) { 9967c478bd9Sstevel@tonic-gate ASSERT(RW_LOCK_HELD(&ufsvfsp->vfs_dqrwlock)); 9977c478bd9Sstevel@tonic-gate ASSERT(!(RW_WRITE_HELD(&ufsvfsp->vfs_dqrwlock))); 9987c478bd9Sstevel@tonic-gate rw_exit(&ufsvfsp->vfs_dqrwlock); 9997c478bd9Sstevel@tonic-gate } 10007c478bd9Sstevel@tonic-gate 1001a5652762Spraks newpage = 0; 1002a5652762Spraks premove_resid = uio->uio_resid; 10036f5f1c63SDonghai Qiao 10046f5f1c63SDonghai Qiao /* 10056f5f1c63SDonghai Qiao * Touch the page and fault it in if it is not in core 10066f5f1c63SDonghai Qiao * before segmap_getmapflt or vpm_data_copy can lock it. 10076f5f1c63SDonghai Qiao * This is to avoid the deadlock if the buffer is mapped 10086f5f1c63SDonghai Qiao * to the same file through mmap which we want to write. 10096f5f1c63SDonghai Qiao */ 10106f5f1c63SDonghai Qiao uio_prefaultpages((long)n, uio); 10116f5f1c63SDonghai Qiao 1012a5652762Spraks if (vpm_enable) { 1013a5652762Spraks /* 1014a5652762Spraks * Copy data. If new pages are created, part of 1015a5652762Spraks * the page that is not written will be initizliazed 1016a5652762Spraks * with zeros. 1017a5652762Spraks */ 1018a5652762Spraks error = vpm_data_copy(vp, (off + mapon), (uint_t)n, 1019a5652762Spraks uio, !pagecreate, &newpage, 0, S_WRITE); 1020a5652762Spraks } else { 1021a5652762Spraks 10227c478bd9Sstevel@tonic-gate base = segmap_getmapflt(segkmap, vp, (off + mapon), 10237c478bd9Sstevel@tonic-gate (uint_t)n, !pagecreate, S_WRITE); 10247c478bd9Sstevel@tonic-gate 10257c478bd9Sstevel@tonic-gate /* 10267c478bd9Sstevel@tonic-gate * segmap_pagecreate() returns 1 if it calls 10277c478bd9Sstevel@tonic-gate * page_create_va() to allocate any pages. 10287c478bd9Sstevel@tonic-gate */ 10297c478bd9Sstevel@tonic-gate 10307c478bd9Sstevel@tonic-gate if (pagecreate) 10317c478bd9Sstevel@tonic-gate newpage = segmap_pagecreate(segkmap, base, 10327c478bd9Sstevel@tonic-gate (size_t)n, 0); 10337c478bd9Sstevel@tonic-gate 10347c478bd9Sstevel@tonic-gate error = uiomove(base + mapon, (long)n, UIO_WRITE, uio); 1035a5652762Spraks } 10367c478bd9Sstevel@tonic-gate 10377c478bd9Sstevel@tonic-gate /* 10387c478bd9Sstevel@tonic-gate * If "newpage" is set, then a new page was created and it 10397c478bd9Sstevel@tonic-gate * does not contain valid data, so it needs to be initialized 10407c478bd9Sstevel@tonic-gate * at this point. 10417c478bd9Sstevel@tonic-gate * Otherwise the page contains old data, which was overwritten 10427c478bd9Sstevel@tonic-gate * partially or as a whole in uiomove. 10437c478bd9Sstevel@tonic-gate * If there is only one iovec structure within uio, then 10447c478bd9Sstevel@tonic-gate * on error uiomove will not be able to update uio->uio_loffset 10457c478bd9Sstevel@tonic-gate * and we would zero the whole page here! 10467c478bd9Sstevel@tonic-gate * 10477c478bd9Sstevel@tonic-gate * If uiomove fails because of an error, the old valid data 10487c478bd9Sstevel@tonic-gate * is kept instead of filling the rest of the page with zero's. 10497c478bd9Sstevel@tonic-gate */ 1050a5652762Spraks if (!vpm_enable && newpage && 10517c478bd9Sstevel@tonic-gate uio->uio_loffset < roundup(off + mapon + n, PAGESIZE)) { 10527c478bd9Sstevel@tonic-gate /* 10537c478bd9Sstevel@tonic-gate * We created pages w/o initializing them completely, 10547c478bd9Sstevel@tonic-gate * thus we need to zero the part that wasn't set up. 10557c478bd9Sstevel@tonic-gate * This happens on most EOF write cases and if 10567c478bd9Sstevel@tonic-gate * we had some sort of error during the uiomove. 10577c478bd9Sstevel@tonic-gate */ 10587c478bd9Sstevel@tonic-gate int nzero, nmoved; 10597c478bd9Sstevel@tonic-gate 10607c478bd9Sstevel@tonic-gate nmoved = (int)(uio->uio_loffset - (off + mapon)); 10617c478bd9Sstevel@tonic-gate ASSERT(nmoved >= 0 && nmoved <= n); 10627c478bd9Sstevel@tonic-gate nzero = roundup(on + n, PAGESIZE) - nmoved; 10637c478bd9Sstevel@tonic-gate ASSERT(nzero > 0 && mapon + nmoved + nzero <= MAXBSIZE); 10647c478bd9Sstevel@tonic-gate (void) kzero(base + mapon + nmoved, (uint_t)nzero); 10657c478bd9Sstevel@tonic-gate } 10667c478bd9Sstevel@tonic-gate 10677c478bd9Sstevel@tonic-gate /* 10687c478bd9Sstevel@tonic-gate * Unlock the pages allocated by page_create_va() 10697c478bd9Sstevel@tonic-gate * in segmap_pagecreate() 10707c478bd9Sstevel@tonic-gate */ 1071a5652762Spraks if (!vpm_enable && newpage) 10727c478bd9Sstevel@tonic-gate segmap_pageunlock(segkmap, base, (size_t)n, S_WRITE); 10737c478bd9Sstevel@tonic-gate 10747c478bd9Sstevel@tonic-gate /* 10757c478bd9Sstevel@tonic-gate * If the size of the file changed, then update the 10767c478bd9Sstevel@tonic-gate * size field in the inode now. This can't be done 10777c478bd9Sstevel@tonic-gate * before the call to segmap_pageunlock or there is 10787c478bd9Sstevel@tonic-gate * a potential deadlock with callers to ufs_putpage(). 10797c478bd9Sstevel@tonic-gate * They will be holding i_contents and trying to lock 10807c478bd9Sstevel@tonic-gate * a page, while this thread is holding a page locked 10817c478bd9Sstevel@tonic-gate * and trying to acquire i_contents. 10827c478bd9Sstevel@tonic-gate */ 10837c478bd9Sstevel@tonic-gate if (i_size_changed) { 10847c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_WRITER); 10857c478bd9Sstevel@tonic-gate old_i_size = ip->i_size; 10867c478bd9Sstevel@tonic-gate UFS_SET_ISIZE(uoff + n, ip); 10877c478bd9Sstevel@tonic-gate TRANS_INODE(ufsvfsp, ip); 10887c478bd9Sstevel@tonic-gate /* 10897c478bd9Sstevel@tonic-gate * file has grown larger than 2GB. Set flag 10907c478bd9Sstevel@tonic-gate * in superblock to indicate this, if it 10917c478bd9Sstevel@tonic-gate * is not already set. 10927c478bd9Sstevel@tonic-gate */ 10937c478bd9Sstevel@tonic-gate if ((ip->i_size > MAXOFF32_T) && 10947c478bd9Sstevel@tonic-gate !(fs->fs_flags & FSLARGEFILES)) { 10957c478bd9Sstevel@tonic-gate ASSERT(ufsvfsp->vfs_lfflags & UFS_LARGEFILES); 10967c478bd9Sstevel@tonic-gate mutex_enter(&ufsvfsp->vfs_lock); 10977c478bd9Sstevel@tonic-gate fs->fs_flags |= FSLARGEFILES; 10987c478bd9Sstevel@tonic-gate ufs_sbwrite(ufsvfsp); 10997c478bd9Sstevel@tonic-gate mutex_exit(&ufsvfsp->vfs_lock); 11007c478bd9Sstevel@tonic-gate } 11017c478bd9Sstevel@tonic-gate mutex_enter(&ip->i_tlock); 11027c478bd9Sstevel@tonic-gate ip->i_writer = NULL; 11037c478bd9Sstevel@tonic-gate cv_broadcast(&ip->i_wrcv); 11047c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 11057c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 11067c478bd9Sstevel@tonic-gate } 11077c478bd9Sstevel@tonic-gate 11087c478bd9Sstevel@tonic-gate if (error) { 11097c478bd9Sstevel@tonic-gate /* 11107c478bd9Sstevel@tonic-gate * If we failed on a write, we may have already 11117c478bd9Sstevel@tonic-gate * allocated file blocks as well as pages. It's 11127c478bd9Sstevel@tonic-gate * hard to undo the block allocation, but we must 11137c478bd9Sstevel@tonic-gate * be sure to invalidate any pages that may have 11147c478bd9Sstevel@tonic-gate * been allocated. 11157c478bd9Sstevel@tonic-gate * 11167c478bd9Sstevel@tonic-gate * If the page was created without initialization 11177c478bd9Sstevel@tonic-gate * then we must check if it should be possible 11187c478bd9Sstevel@tonic-gate * to destroy the new page and to keep the old data 11197c478bd9Sstevel@tonic-gate * on the disk. 11207c478bd9Sstevel@tonic-gate * 11217c478bd9Sstevel@tonic-gate * It is possible to destroy the page without 11227c478bd9Sstevel@tonic-gate * having to write back its contents only when 11237c478bd9Sstevel@tonic-gate * - the size of the file keeps unchanged 11247c478bd9Sstevel@tonic-gate * - bmap_write() did not allocate new disk blocks 11257c478bd9Sstevel@tonic-gate * it is possible to create big files using "seek" and 11267c478bd9Sstevel@tonic-gate * write to the end of the file. A "write" to a 11277c478bd9Sstevel@tonic-gate * position before the end of the file would not 11287c478bd9Sstevel@tonic-gate * change the size of the file but it would allocate 11297c478bd9Sstevel@tonic-gate * new disk blocks. 11307c478bd9Sstevel@tonic-gate * - uiomove intended to overwrite the whole page. 11317c478bd9Sstevel@tonic-gate * - a new page was created (newpage == 1). 11327c478bd9Sstevel@tonic-gate */ 11337c478bd9Sstevel@tonic-gate 11347c478bd9Sstevel@tonic-gate if (i_size_changed == 0 && new_iblocks == 0 && 11357c478bd9Sstevel@tonic-gate newpage) { 11367c478bd9Sstevel@tonic-gate 11377c478bd9Sstevel@tonic-gate /* unwind what uiomove eventually last did */ 11387c478bd9Sstevel@tonic-gate uio->uio_resid = premove_resid; 11397c478bd9Sstevel@tonic-gate 11407c478bd9Sstevel@tonic-gate /* 11417c478bd9Sstevel@tonic-gate * destroy the page, do not write ambiguous 11427c478bd9Sstevel@tonic-gate * data to the disk. 11437c478bd9Sstevel@tonic-gate */ 11447c478bd9Sstevel@tonic-gate flags = SM_DESTROY; 11457c478bd9Sstevel@tonic-gate } else { 11467c478bd9Sstevel@tonic-gate /* 11477c478bd9Sstevel@tonic-gate * write the page back to the disk, if dirty, 11487c478bd9Sstevel@tonic-gate * and remove the page from the cache. 11497c478bd9Sstevel@tonic-gate */ 11507c478bd9Sstevel@tonic-gate flags = SM_INVAL; 11517c478bd9Sstevel@tonic-gate } 1152a5652762Spraks 1153a5652762Spraks if (vpm_enable) { 1154a5652762Spraks /* 1155a5652762Spraks * Flush pages. 1156a5652762Spraks */ 1157a5652762Spraks (void) vpm_sync_pages(vp, off, n, flags); 1158a5652762Spraks } else { 11597c478bd9Sstevel@tonic-gate (void) segmap_release(segkmap, base, flags); 1160a5652762Spraks } 11617c478bd9Sstevel@tonic-gate } else { 11627c478bd9Sstevel@tonic-gate flags = 0; 11637c478bd9Sstevel@tonic-gate /* 11647c478bd9Sstevel@tonic-gate * Force write back for synchronous write cases. 11657c478bd9Sstevel@tonic-gate */ 11667c478bd9Sstevel@tonic-gate if ((ioflag & (FSYNC|FDSYNC)) || type == IFDIR) { 11677c478bd9Sstevel@tonic-gate /* 11687c478bd9Sstevel@tonic-gate * If the sticky bit is set but the 11697c478bd9Sstevel@tonic-gate * execute bit is not set, we do a 11707c478bd9Sstevel@tonic-gate * synchronous write back and free 11717c478bd9Sstevel@tonic-gate * the page when done. We set up swap 11727c478bd9Sstevel@tonic-gate * files to be handled this way to 11737c478bd9Sstevel@tonic-gate * prevent servers from keeping around 11747c478bd9Sstevel@tonic-gate * the client's swap pages too long. 11757c478bd9Sstevel@tonic-gate * XXX - there ought to be a better way. 11767c478bd9Sstevel@tonic-gate */ 11777c478bd9Sstevel@tonic-gate if (IS_SWAPVP(vp)) { 11787c478bd9Sstevel@tonic-gate flags = SM_WRITE | SM_FREE | 11797c478bd9Sstevel@tonic-gate SM_DONTNEED; 11807c478bd9Sstevel@tonic-gate iupdat_flag = 0; 11817c478bd9Sstevel@tonic-gate } else { 11827c478bd9Sstevel@tonic-gate flags = SM_WRITE; 11837c478bd9Sstevel@tonic-gate } 11847c478bd9Sstevel@tonic-gate } else if (n + on == MAXBSIZE || IS_SWAPVP(vp)) { 11857c478bd9Sstevel@tonic-gate /* 11867c478bd9Sstevel@tonic-gate * Have written a whole block. 11877c478bd9Sstevel@tonic-gate * Start an asynchronous write and 11887c478bd9Sstevel@tonic-gate * mark the buffer to indicate that 11897c478bd9Sstevel@tonic-gate * it won't be needed again soon. 11907c478bd9Sstevel@tonic-gate */ 11917c478bd9Sstevel@tonic-gate flags = SM_WRITE | SM_ASYNC | SM_DONTNEED; 11927c478bd9Sstevel@tonic-gate } 1193a5652762Spraks if (vpm_enable) { 1194a5652762Spraks /* 1195a5652762Spraks * Flush pages. 1196a5652762Spraks */ 11973bd1497bSpraks error = vpm_sync_pages(vp, off, n, flags); 1198a5652762Spraks } else { 11993bd1497bSpraks error = segmap_release(segkmap, base, flags); 1200a5652762Spraks } 12017c478bd9Sstevel@tonic-gate /* 12027c478bd9Sstevel@tonic-gate * If the operation failed and is synchronous, 12037c478bd9Sstevel@tonic-gate * then we need to unwind what uiomove() last 12047c478bd9Sstevel@tonic-gate * did so we can potentially return an error to 12057c478bd9Sstevel@tonic-gate * the caller. If this write operation was 12067c478bd9Sstevel@tonic-gate * done in two pieces and the first succeeded, 12077c478bd9Sstevel@tonic-gate * then we won't return an error for the second 12087c478bd9Sstevel@tonic-gate * piece that failed. However, we only want to 12097c478bd9Sstevel@tonic-gate * return a resid value that reflects what was 12107c478bd9Sstevel@tonic-gate * really done. 12117c478bd9Sstevel@tonic-gate * 12127c478bd9Sstevel@tonic-gate * Failures for non-synchronous operations can 12137c478bd9Sstevel@tonic-gate * be ignored since the page subsystem will 12147c478bd9Sstevel@tonic-gate * retry the operation until it succeeds or the 12157c478bd9Sstevel@tonic-gate * file system is unmounted. 12167c478bd9Sstevel@tonic-gate */ 12177c478bd9Sstevel@tonic-gate if (error) { 12187c478bd9Sstevel@tonic-gate if ((ioflag & (FSYNC | FDSYNC)) || 12197c478bd9Sstevel@tonic-gate type == IFDIR) { 12207c478bd9Sstevel@tonic-gate uio->uio_resid = premove_resid; 12217c478bd9Sstevel@tonic-gate } else { 12227c478bd9Sstevel@tonic-gate error = 0; 12237c478bd9Sstevel@tonic-gate } 12247c478bd9Sstevel@tonic-gate } 12257c478bd9Sstevel@tonic-gate } 12267c478bd9Sstevel@tonic-gate 12277c478bd9Sstevel@tonic-gate /* 12287c478bd9Sstevel@tonic-gate * Re-acquire contents lock. 12297c478bd9Sstevel@tonic-gate * If it was dropped, reacquire reader vfs_dqrwlock as well. 12307c478bd9Sstevel@tonic-gate */ 12317c478bd9Sstevel@tonic-gate if (do_dqrwlock) 12327c478bd9Sstevel@tonic-gate rw_enter(&ufsvfsp->vfs_dqrwlock, RW_READER); 12337c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_WRITER); 12347c478bd9Sstevel@tonic-gate 12357c478bd9Sstevel@tonic-gate /* 12367c478bd9Sstevel@tonic-gate * If the uiomove() failed or if a synchronous 12377c478bd9Sstevel@tonic-gate * page push failed, fix up i_size. 12387c478bd9Sstevel@tonic-gate */ 12397c478bd9Sstevel@tonic-gate if (error) { 12407c478bd9Sstevel@tonic-gate if (i_size_changed) { 12417c478bd9Sstevel@tonic-gate /* 12427c478bd9Sstevel@tonic-gate * The uiomove failed, and we 12437c478bd9Sstevel@tonic-gate * allocated blocks,so get rid 12447c478bd9Sstevel@tonic-gate * of them. 12457c478bd9Sstevel@tonic-gate */ 12467c478bd9Sstevel@tonic-gate (void) ufs_itrunc(ip, old_i_size, 0, cr); 12477c478bd9Sstevel@tonic-gate } 12487c478bd9Sstevel@tonic-gate } else { 12497c478bd9Sstevel@tonic-gate /* 12507c478bd9Sstevel@tonic-gate * XXX - Can this be out of the loop? 12517c478bd9Sstevel@tonic-gate */ 12527c478bd9Sstevel@tonic-gate ip->i_flag |= IUPD | ICHG; 12537c478bd9Sstevel@tonic-gate /* 12547c478bd9Sstevel@tonic-gate * Only do one increase of i_seq for multiple 12557c478bd9Sstevel@tonic-gate * pieces. Because we drop locks, record 12567c478bd9Sstevel@tonic-gate * the fact that we changed the timestamp and 12577c478bd9Sstevel@tonic-gate * are deferring the increase in case another thread 12587c478bd9Sstevel@tonic-gate * pushes our timestamp update. 12597c478bd9Sstevel@tonic-gate */ 12607c478bd9Sstevel@tonic-gate i_seq_needed = 1; 12617c478bd9Sstevel@tonic-gate ip->i_flag |= ISEQ; 12627c478bd9Sstevel@tonic-gate if (i_size_changed) 12637c478bd9Sstevel@tonic-gate ip->i_flag |= IATTCHG; 12647c478bd9Sstevel@tonic-gate if ((ip->i_mode & (IEXEC | (IEXEC >> 3) | 12657c478bd9Sstevel@tonic-gate (IEXEC >> 6))) != 0 && 12667c478bd9Sstevel@tonic-gate (ip->i_mode & (ISUID | ISGID)) != 0 && 12677c478bd9Sstevel@tonic-gate secpolicy_vnode_setid_retain(cr, 12687c478bd9Sstevel@tonic-gate (ip->i_mode & ISUID) != 0 && ip->i_uid == 0) != 0) { 12697c478bd9Sstevel@tonic-gate /* 12707c478bd9Sstevel@tonic-gate * Clear Set-UID & Set-GID bits on 12717c478bd9Sstevel@tonic-gate * successful write if not privileged 12727c478bd9Sstevel@tonic-gate * and at least one of the execute bits 12737c478bd9Sstevel@tonic-gate * is set. If we always clear Set-GID, 12747c478bd9Sstevel@tonic-gate * mandatory file and record locking is 12757c478bd9Sstevel@tonic-gate * unuseable. 12767c478bd9Sstevel@tonic-gate */ 12777c478bd9Sstevel@tonic-gate ip->i_mode &= ~(ISUID | ISGID); 12787c478bd9Sstevel@tonic-gate } 12797c478bd9Sstevel@tonic-gate } 1280f90bab20Sswilcox /* 1281f90bab20Sswilcox * In the case the FDSYNC flag is set and this is a 1282f90bab20Sswilcox * "rewrite" we won't log a delta. 1283f90bab20Sswilcox * The FSYNC flag overrides all cases. 1284f90bab20Sswilcox */ 1285f90bab20Sswilcox if (!ufs_check_rewrite(ip, uio, ioflag) || !(ioflag & FDSYNC)) { 12867c478bd9Sstevel@tonic-gate TRANS_INODE(ufsvfsp, ip); 1287f90bab20Sswilcox } 12887c478bd9Sstevel@tonic-gate } while (error == 0 && uio->uio_resid > 0 && n != 0); 12897c478bd9Sstevel@tonic-gate 12907c478bd9Sstevel@tonic-gate out: 12917c478bd9Sstevel@tonic-gate /* 12927c478bd9Sstevel@tonic-gate * Make sure i_seq is increased at least once per write 12937c478bd9Sstevel@tonic-gate */ 12947c478bd9Sstevel@tonic-gate if (i_seq_needed) { 12957c478bd9Sstevel@tonic-gate ip->i_seq++; 12967c478bd9Sstevel@tonic-gate ip->i_flag &= ~ISEQ; /* no longer deferred */ 12977c478bd9Sstevel@tonic-gate } 12987c478bd9Sstevel@tonic-gate 12997c478bd9Sstevel@tonic-gate /* 13007c478bd9Sstevel@tonic-gate * Inode is updated according to this table - 13017c478bd9Sstevel@tonic-gate * 13027c478bd9Sstevel@tonic-gate * FSYNC FDSYNC(posix.4) 13037c478bd9Sstevel@tonic-gate * -------------------------- 13047c478bd9Sstevel@tonic-gate * always@ IATTCHG|IBDWRITE 13057c478bd9Sstevel@tonic-gate * 13067c478bd9Sstevel@tonic-gate * @ - If we are doing synchronous write the only time we should 13077c478bd9Sstevel@tonic-gate * not be sync'ing the ip here is if we have the stickyhack 13087c478bd9Sstevel@tonic-gate * activated, the file is marked with the sticky bit and 13097c478bd9Sstevel@tonic-gate * no exec bit, the file length has not been changed and 13107c478bd9Sstevel@tonic-gate * no new blocks have been allocated during this write. 13117c478bd9Sstevel@tonic-gate */ 13127c478bd9Sstevel@tonic-gate 13137c478bd9Sstevel@tonic-gate if ((ip->i_flag & ISYNC) != 0) { 13147c478bd9Sstevel@tonic-gate /* 13157c478bd9Sstevel@tonic-gate * we have eliminated nosync 13167c478bd9Sstevel@tonic-gate */ 13177c478bd9Sstevel@tonic-gate if ((ip->i_flag & (IATTCHG|IBDWRITE)) || 13187c478bd9Sstevel@tonic-gate ((ioflag & FSYNC) && iupdat_flag)) { 13197c478bd9Sstevel@tonic-gate ufs_iupdat(ip, 1); 13207c478bd9Sstevel@tonic-gate } 13217c478bd9Sstevel@tonic-gate } 13227c478bd9Sstevel@tonic-gate 13237c478bd9Sstevel@tonic-gate /* 13247c478bd9Sstevel@tonic-gate * If we've already done a partial-write, terminate 13257c478bd9Sstevel@tonic-gate * the write but return no error unless the error is ENOSPC 13267c478bd9Sstevel@tonic-gate * because the caller can detect this and free resources and 13277c478bd9Sstevel@tonic-gate * try again. 13287c478bd9Sstevel@tonic-gate */ 13297c478bd9Sstevel@tonic-gate if ((start_resid != uio->uio_resid) && (error != ENOSPC)) 13307c478bd9Sstevel@tonic-gate error = 0; 13317c478bd9Sstevel@tonic-gate 13327c478bd9Sstevel@tonic-gate ip->i_flag &= ~(INOACC | ISYNC); 13337c478bd9Sstevel@tonic-gate ITIMES_NOLOCK(ip); 13347c478bd9Sstevel@tonic-gate return (error); 13357c478bd9Sstevel@tonic-gate } 13367c478bd9Sstevel@tonic-gate 13377c478bd9Sstevel@tonic-gate /* 13387c478bd9Sstevel@tonic-gate * rdip does the real work of read requests for ufs. 13397c478bd9Sstevel@tonic-gate */ 13407c478bd9Sstevel@tonic-gate int 13417c478bd9Sstevel@tonic-gate rdip(struct inode *ip, struct uio *uio, int ioflag, cred_t *cr) 13427c478bd9Sstevel@tonic-gate { 13437c478bd9Sstevel@tonic-gate u_offset_t off; 13447c478bd9Sstevel@tonic-gate caddr_t base; 13457c478bd9Sstevel@tonic-gate struct fs *fs; 13467c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp; 13477c478bd9Sstevel@tonic-gate struct vnode *vp; 13487c478bd9Sstevel@tonic-gate long oresid = uio->uio_resid; 13497c478bd9Sstevel@tonic-gate u_offset_t n, on, mapon; 13507c478bd9Sstevel@tonic-gate int error = 0; 13517c478bd9Sstevel@tonic-gate int doupdate = 1; 1352562eee46Srbourbon uint_t flags; 13537c478bd9Sstevel@tonic-gate int dofree, directio_status; 13547c478bd9Sstevel@tonic-gate krw_t rwtype; 13557c478bd9Sstevel@tonic-gate o_mode_t type; 1356d3d50737SRafael Vanoni clock_t now; 13577c478bd9Sstevel@tonic-gate 13587c478bd9Sstevel@tonic-gate vp = ITOV(ip); 13597c478bd9Sstevel@tonic-gate 13607c478bd9Sstevel@tonic-gate ASSERT(RW_LOCK_HELD(&ip->i_contents)); 13617c478bd9Sstevel@tonic-gate 13627c478bd9Sstevel@tonic-gate ufsvfsp = ip->i_ufsvfs; 13637c478bd9Sstevel@tonic-gate 13647c478bd9Sstevel@tonic-gate if (ufsvfsp == NULL) 13657c478bd9Sstevel@tonic-gate return (EIO); 13667c478bd9Sstevel@tonic-gate 13677c478bd9Sstevel@tonic-gate fs = ufsvfsp->vfs_fs; 13687c478bd9Sstevel@tonic-gate 13697c478bd9Sstevel@tonic-gate /* check for valid filetype */ 13707c478bd9Sstevel@tonic-gate type = ip->i_mode & IFMT; 13717c478bd9Sstevel@tonic-gate if ((type != IFREG) && (type != IFDIR) && (type != IFATTRDIR) && 13727c478bd9Sstevel@tonic-gate (type != IFLNK) && (type != IFSHAD)) { 13737c478bd9Sstevel@tonic-gate return (EIO); 13747c478bd9Sstevel@tonic-gate } 13757c478bd9Sstevel@tonic-gate 13767c478bd9Sstevel@tonic-gate if (uio->uio_loffset > UFS_MAXOFFSET_T) { 13777c478bd9Sstevel@tonic-gate error = 0; 13787c478bd9Sstevel@tonic-gate goto out; 13797c478bd9Sstevel@tonic-gate } 13807c478bd9Sstevel@tonic-gate if (uio->uio_loffset < (offset_t)0) { 13817c478bd9Sstevel@tonic-gate return (EINVAL); 13827c478bd9Sstevel@tonic-gate } 13837c478bd9Sstevel@tonic-gate if (uio->uio_resid == 0) { 13847c478bd9Sstevel@tonic-gate return (0); 13857c478bd9Sstevel@tonic-gate } 13867c478bd9Sstevel@tonic-gate 13877c478bd9Sstevel@tonic-gate if (!ULOCKFS_IS_NOIACC(ITOUL(ip)) && (fs->fs_ronly == 0) && 13887c478bd9Sstevel@tonic-gate (!ufsvfsp->vfs_noatime)) { 13897c478bd9Sstevel@tonic-gate mutex_enter(&ip->i_tlock); 13907c478bd9Sstevel@tonic-gate ip->i_flag |= IACC; 13917c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 13927c478bd9Sstevel@tonic-gate } 13937c478bd9Sstevel@tonic-gate /* 13947c478bd9Sstevel@tonic-gate * Try to go direct 13957c478bd9Sstevel@tonic-gate */ 13967c478bd9Sstevel@tonic-gate if (ip->i_flag & IDIRECTIO || ufsvfsp->vfs_forcedirectio) { 13977c478bd9Sstevel@tonic-gate error = ufs_directio_read(ip, uio, cr, &directio_status); 13987c478bd9Sstevel@tonic-gate if (directio_status == DIRECTIO_SUCCESS) 13997c478bd9Sstevel@tonic-gate goto out; 14007c478bd9Sstevel@tonic-gate } 14017c478bd9Sstevel@tonic-gate 14027c478bd9Sstevel@tonic-gate rwtype = (rw_write_held(&ip->i_contents)?RW_WRITER:RW_READER); 14037c478bd9Sstevel@tonic-gate 14047c478bd9Sstevel@tonic-gate do { 14057c478bd9Sstevel@tonic-gate offset_t diff; 14067c478bd9Sstevel@tonic-gate u_offset_t uoff = uio->uio_loffset; 14077c478bd9Sstevel@tonic-gate off = uoff & (offset_t)MAXBMASK; 14087c478bd9Sstevel@tonic-gate mapon = (u_offset_t)(uoff & (offset_t)MAXBOFFSET); 14097c478bd9Sstevel@tonic-gate on = (u_offset_t)blkoff(fs, uoff); 14107c478bd9Sstevel@tonic-gate n = MIN((u_offset_t)fs->fs_bsize - on, 14117c478bd9Sstevel@tonic-gate (u_offset_t)uio->uio_resid); 14127c478bd9Sstevel@tonic-gate 14137c478bd9Sstevel@tonic-gate diff = ip->i_size - uoff; 14147c478bd9Sstevel@tonic-gate 14157c478bd9Sstevel@tonic-gate if (diff <= (offset_t)0) { 14167c478bd9Sstevel@tonic-gate error = 0; 14177c478bd9Sstevel@tonic-gate goto out; 14187c478bd9Sstevel@tonic-gate } 14197c478bd9Sstevel@tonic-gate if (diff < (offset_t)n) 14207c478bd9Sstevel@tonic-gate n = (int)diff; 1421562eee46Srbourbon 1422562eee46Srbourbon /* 1423562eee46Srbourbon * We update smallfile2 and smallfile1 at most every second. 1424562eee46Srbourbon */ 1425d3d50737SRafael Vanoni now = ddi_get_lbolt(); 1426d3d50737SRafael Vanoni if (now >= smallfile_update) { 1427562eee46Srbourbon uint64_t percpufreeb; 1428562eee46Srbourbon if (smallfile1_d == 0) smallfile1_d = SMALLFILE1_D; 1429562eee46Srbourbon if (smallfile2_d == 0) smallfile2_d = SMALLFILE2_D; 1430562eee46Srbourbon percpufreeb = ptob((uint64_t)freemem) / ncpus_online; 1431562eee46Srbourbon smallfile1 = percpufreeb / smallfile1_d; 1432562eee46Srbourbon smallfile2 = percpufreeb / smallfile2_d; 1433562eee46Srbourbon smallfile1 = MAX(smallfile1, smallfile); 1434562eee46Srbourbon smallfile1 = MAX(smallfile1, smallfile64); 1435562eee46Srbourbon smallfile2 = MAX(smallfile1, smallfile2); 1436d3d50737SRafael Vanoni smallfile_update = now + hz; 1437562eee46Srbourbon } 1438562eee46Srbourbon 14397c478bd9Sstevel@tonic-gate dofree = freebehind && 1440562eee46Srbourbon ip->i_nextr == (off & PAGEMASK) && off > smallfile1; 14417c478bd9Sstevel@tonic-gate 14427c478bd9Sstevel@tonic-gate /* 14437c478bd9Sstevel@tonic-gate * At this point we can enter ufs_getpage() in one of two 14447c478bd9Sstevel@tonic-gate * ways: 14457c478bd9Sstevel@tonic-gate * 1) segmap_getmapflt() calls ufs_getpage() when the 14467c478bd9Sstevel@tonic-gate * forcefault parameter is true (value of 1 is passed) 14477c478bd9Sstevel@tonic-gate * 2) uiomove() causes a page fault. 14487c478bd9Sstevel@tonic-gate * 14497c478bd9Sstevel@tonic-gate * We cannot hold onto an i_contents reader lock without 14507c478bd9Sstevel@tonic-gate * risking deadlock in ufs_getpage() so drop a reader lock. 14517c478bd9Sstevel@tonic-gate * The ufs_getpage() dolock logic already allows for a 14527c478bd9Sstevel@tonic-gate * thread holding i_contents as writer to work properly 14537c478bd9Sstevel@tonic-gate * so we keep a writer lock. 14547c478bd9Sstevel@tonic-gate */ 14557c478bd9Sstevel@tonic-gate if (rwtype == RW_READER) 14567c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 1457a5652762Spraks 1458a5652762Spraks if (vpm_enable) { 1459a5652762Spraks /* 1460a5652762Spraks * Copy data. 1461a5652762Spraks */ 1462a5652762Spraks error = vpm_data_copy(vp, (off + mapon), (uint_t)n, 1463a5652762Spraks uio, 1, NULL, 0, S_READ); 1464a5652762Spraks } else { 14657c478bd9Sstevel@tonic-gate base = segmap_getmapflt(segkmap, vp, (off + mapon), 14667c478bd9Sstevel@tonic-gate (uint_t)n, 1, S_READ); 14677c478bd9Sstevel@tonic-gate error = uiomove(base + mapon, (long)n, UIO_READ, uio); 1468a5652762Spraks } 14697c478bd9Sstevel@tonic-gate 14707c478bd9Sstevel@tonic-gate flags = 0; 14717c478bd9Sstevel@tonic-gate if (!error) { 14727c478bd9Sstevel@tonic-gate /* 1473562eee46Srbourbon * If reading sequential we won't need this 1474562eee46Srbourbon * buffer again soon. For offsets in range 1475562eee46Srbourbon * [smallfile1, smallfile2] release the pages 1476562eee46Srbourbon * at the tail of the cache list, larger 1477562eee46Srbourbon * offsets are released at the head. 14787c478bd9Sstevel@tonic-gate */ 14797c478bd9Sstevel@tonic-gate if (dofree) { 1480562eee46Srbourbon flags = SM_FREE | SM_ASYNC; 1481562eee46Srbourbon if ((cache_read_ahead == 0) && 1482562eee46Srbourbon (off > smallfile2)) 1483562eee46Srbourbon flags |= SM_DONTNEED; 14847c478bd9Sstevel@tonic-gate } 14857c478bd9Sstevel@tonic-gate /* 14867c478bd9Sstevel@tonic-gate * In POSIX SYNC (FSYNC and FDSYNC) read mode, 14877c478bd9Sstevel@tonic-gate * we want to make sure that the page which has 14887c478bd9Sstevel@tonic-gate * been read, is written on disk if it is dirty. 14897c478bd9Sstevel@tonic-gate * And corresponding indirect blocks should also 14907c478bd9Sstevel@tonic-gate * be flushed out. 14917c478bd9Sstevel@tonic-gate */ 14927c478bd9Sstevel@tonic-gate if ((ioflag & FRSYNC) && (ioflag & (FSYNC|FDSYNC))) { 14937c478bd9Sstevel@tonic-gate flags &= ~SM_ASYNC; 14947c478bd9Sstevel@tonic-gate flags |= SM_WRITE; 14957c478bd9Sstevel@tonic-gate } 1496a5652762Spraks if (vpm_enable) { 1497a5652762Spraks error = vpm_sync_pages(vp, off, n, flags); 1498a5652762Spraks } else { 14997c478bd9Sstevel@tonic-gate error = segmap_release(segkmap, base, flags); 1500a5652762Spraks } 1501a5652762Spraks } else { 1502a5652762Spraks if (vpm_enable) { 1503a5652762Spraks (void) vpm_sync_pages(vp, off, n, flags); 1504a5652762Spraks } else { 15057c478bd9Sstevel@tonic-gate (void) segmap_release(segkmap, base, flags); 1506a5652762Spraks } 1507a5652762Spraks } 15087c478bd9Sstevel@tonic-gate 15097c478bd9Sstevel@tonic-gate if (rwtype == RW_READER) 15107c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, rwtype); 15117c478bd9Sstevel@tonic-gate } while (error == 0 && uio->uio_resid > 0 && n != 0); 15127c478bd9Sstevel@tonic-gate out: 15137c478bd9Sstevel@tonic-gate /* 15147c478bd9Sstevel@tonic-gate * Inode is updated according to this table if FRSYNC is set. 15157c478bd9Sstevel@tonic-gate * 15167c478bd9Sstevel@tonic-gate * FSYNC FDSYNC(posix.4) 15177c478bd9Sstevel@tonic-gate * -------------------------- 15187c478bd9Sstevel@tonic-gate * always IATTCHG|IBDWRITE 15197c478bd9Sstevel@tonic-gate */ 15207c478bd9Sstevel@tonic-gate /* 15217c478bd9Sstevel@tonic-gate * The inode is not updated if we're logging and the inode is a 15227c478bd9Sstevel@tonic-gate * directory with FRSYNC, FSYNC and FDSYNC flags set. 15237c478bd9Sstevel@tonic-gate */ 15247c478bd9Sstevel@tonic-gate if (ioflag & FRSYNC) { 15257c478bd9Sstevel@tonic-gate if (TRANS_ISTRANS(ufsvfsp) && ((ip->i_mode & IFMT) == IFDIR)) { 15267c478bd9Sstevel@tonic-gate doupdate = 0; 15277c478bd9Sstevel@tonic-gate } 15287c478bd9Sstevel@tonic-gate if (doupdate) { 15297c478bd9Sstevel@tonic-gate if ((ioflag & FSYNC) || 15307c478bd9Sstevel@tonic-gate ((ioflag & FDSYNC) && 15317c478bd9Sstevel@tonic-gate (ip->i_flag & (IATTCHG|IBDWRITE)))) { 15327c478bd9Sstevel@tonic-gate ufs_iupdat(ip, 1); 15337c478bd9Sstevel@tonic-gate } 15347c478bd9Sstevel@tonic-gate } 15357c478bd9Sstevel@tonic-gate } 15367c478bd9Sstevel@tonic-gate /* 15377c478bd9Sstevel@tonic-gate * If we've already done a partial read, terminate 15387c478bd9Sstevel@tonic-gate * the read but return no error. 15397c478bd9Sstevel@tonic-gate */ 15407c478bd9Sstevel@tonic-gate if (oresid != uio->uio_resid) 15417c478bd9Sstevel@tonic-gate error = 0; 15427c478bd9Sstevel@tonic-gate ITIMES(ip); 15437c478bd9Sstevel@tonic-gate 15447c478bd9Sstevel@tonic-gate return (error); 15457c478bd9Sstevel@tonic-gate } 15467c478bd9Sstevel@tonic-gate 15477c478bd9Sstevel@tonic-gate /* ARGSUSED */ 15487c478bd9Sstevel@tonic-gate static int 15497c478bd9Sstevel@tonic-gate ufs_ioctl( 15507c478bd9Sstevel@tonic-gate struct vnode *vp, 15517c478bd9Sstevel@tonic-gate int cmd, 15527c478bd9Sstevel@tonic-gate intptr_t arg, 15537c478bd9Sstevel@tonic-gate int flag, 15547c478bd9Sstevel@tonic-gate struct cred *cr, 1555da6c28aaSamw int *rvalp, 1556da6c28aaSamw caller_context_t *ct) 15577c478bd9Sstevel@tonic-gate { 15587c478bd9Sstevel@tonic-gate struct lockfs lockfs, lockfs_out; 15597c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp = VTOI(vp)->i_ufsvfs; 15607c478bd9Sstevel@tonic-gate char *comment, *original_comment; 15617c478bd9Sstevel@tonic-gate struct fs *fs; 15627c478bd9Sstevel@tonic-gate struct ulockfs *ulp; 15637c478bd9Sstevel@tonic-gate offset_t off; 15647c478bd9Sstevel@tonic-gate extern int maxphys; 15657c478bd9Sstevel@tonic-gate int error; 15667c478bd9Sstevel@tonic-gate int issync; 15677c478bd9Sstevel@tonic-gate int trans_size; 15687c478bd9Sstevel@tonic-gate 15697c478bd9Sstevel@tonic-gate 15707c478bd9Sstevel@tonic-gate /* 15717c478bd9Sstevel@tonic-gate * forcibly unmounted 15727c478bd9Sstevel@tonic-gate */ 157302ffed0eSjr26306 if (ufsvfsp == NULL || vp->v_vfsp == NULL || 157402ffed0eSjr26306 vp->v_vfsp->vfs_flag & VFS_UNMOUNTED) 15757c478bd9Sstevel@tonic-gate return (EIO); 15767c478bd9Sstevel@tonic-gate fs = ufsvfsp->vfs_fs; 15777c478bd9Sstevel@tonic-gate 15787c478bd9Sstevel@tonic-gate if (cmd == Q_QUOTACTL) { 15797c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_QUOTA_MASK); 15807c478bd9Sstevel@tonic-gate if (error) 15817c478bd9Sstevel@tonic-gate return (error); 15827c478bd9Sstevel@tonic-gate 15837c478bd9Sstevel@tonic-gate if (ulp) { 15847c478bd9Sstevel@tonic-gate TRANS_BEGIN_ASYNC(ufsvfsp, TOP_QUOTA, 15857c478bd9Sstevel@tonic-gate TOP_SETQUOTA_SIZE(fs)); 15867c478bd9Sstevel@tonic-gate } 15877c478bd9Sstevel@tonic-gate 15887c478bd9Sstevel@tonic-gate error = quotactl(vp, arg, flag, cr); 15897c478bd9Sstevel@tonic-gate 15907c478bd9Sstevel@tonic-gate if (ulp) { 15917c478bd9Sstevel@tonic-gate TRANS_END_ASYNC(ufsvfsp, TOP_QUOTA, 15927c478bd9Sstevel@tonic-gate TOP_SETQUOTA_SIZE(fs)); 15937c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 15947c478bd9Sstevel@tonic-gate } 15957c478bd9Sstevel@tonic-gate return (error); 15967c478bd9Sstevel@tonic-gate } 15977c478bd9Sstevel@tonic-gate 15987c478bd9Sstevel@tonic-gate switch (cmd) { 15997c478bd9Sstevel@tonic-gate case _FIOLFS: 16007c478bd9Sstevel@tonic-gate /* 16017c478bd9Sstevel@tonic-gate * file system locking 16027c478bd9Sstevel@tonic-gate */ 16037c478bd9Sstevel@tonic-gate if (secpolicy_fs_config(cr, ufsvfsp->vfs_vfs) != 0) 16047c478bd9Sstevel@tonic-gate return (EPERM); 16057c478bd9Sstevel@tonic-gate 16067c478bd9Sstevel@tonic-gate if ((flag & DATAMODEL_MASK) == DATAMODEL_NATIVE) { 16077c478bd9Sstevel@tonic-gate if (copyin((caddr_t)arg, &lockfs, 16087c478bd9Sstevel@tonic-gate sizeof (struct lockfs))) 16097c478bd9Sstevel@tonic-gate return (EFAULT); 16107c478bd9Sstevel@tonic-gate } 16117c478bd9Sstevel@tonic-gate #ifdef _SYSCALL32_IMPL 16127c478bd9Sstevel@tonic-gate else { 16137c478bd9Sstevel@tonic-gate struct lockfs32 lockfs32; 16147c478bd9Sstevel@tonic-gate /* Translate ILP32 lockfs to LP64 lockfs */ 16157c478bd9Sstevel@tonic-gate if (copyin((caddr_t)arg, &lockfs32, 16167c478bd9Sstevel@tonic-gate sizeof (struct lockfs32))) 16177c478bd9Sstevel@tonic-gate return (EFAULT); 16187c478bd9Sstevel@tonic-gate lockfs.lf_lock = (ulong_t)lockfs32.lf_lock; 16197c478bd9Sstevel@tonic-gate lockfs.lf_flags = (ulong_t)lockfs32.lf_flags; 16207c478bd9Sstevel@tonic-gate lockfs.lf_key = (ulong_t)lockfs32.lf_key; 16217c478bd9Sstevel@tonic-gate lockfs.lf_comlen = (ulong_t)lockfs32.lf_comlen; 16227c478bd9Sstevel@tonic-gate lockfs.lf_comment = 16237c478bd9Sstevel@tonic-gate (caddr_t)(uintptr_t)lockfs32.lf_comment; 16247c478bd9Sstevel@tonic-gate } 16257c478bd9Sstevel@tonic-gate #endif /* _SYSCALL32_IMPL */ 16267c478bd9Sstevel@tonic-gate 16277c478bd9Sstevel@tonic-gate if (lockfs.lf_comlen) { 16287c478bd9Sstevel@tonic-gate if (lockfs.lf_comlen > LOCKFS_MAXCOMMENTLEN) 16297c478bd9Sstevel@tonic-gate return (ENAMETOOLONG); 163080d34432Sfrankho comment = 163180d34432Sfrankho kmem_alloc(lockfs.lf_comlen, KM_SLEEP); 16327c478bd9Sstevel@tonic-gate if (copyin(lockfs.lf_comment, comment, 16337c478bd9Sstevel@tonic-gate lockfs.lf_comlen)) { 16347c478bd9Sstevel@tonic-gate kmem_free(comment, lockfs.lf_comlen); 16357c478bd9Sstevel@tonic-gate return (EFAULT); 16367c478bd9Sstevel@tonic-gate } 16377c478bd9Sstevel@tonic-gate original_comment = lockfs.lf_comment; 16387c478bd9Sstevel@tonic-gate lockfs.lf_comment = comment; 16397c478bd9Sstevel@tonic-gate } 16407c478bd9Sstevel@tonic-gate if ((error = ufs_fiolfs(vp, &lockfs, 0)) == 0) { 16417c478bd9Sstevel@tonic-gate lockfs.lf_comment = original_comment; 16427c478bd9Sstevel@tonic-gate 16437c478bd9Sstevel@tonic-gate if ((flag & DATAMODEL_MASK) == 16447c478bd9Sstevel@tonic-gate DATAMODEL_NATIVE) { 16457c478bd9Sstevel@tonic-gate (void) copyout(&lockfs, (caddr_t)arg, 16467c478bd9Sstevel@tonic-gate sizeof (struct lockfs)); 16477c478bd9Sstevel@tonic-gate } 16487c478bd9Sstevel@tonic-gate #ifdef _SYSCALL32_IMPL 16497c478bd9Sstevel@tonic-gate else { 16507c478bd9Sstevel@tonic-gate struct lockfs32 lockfs32; 16517c478bd9Sstevel@tonic-gate /* Translate LP64 to ILP32 lockfs */ 16527c478bd9Sstevel@tonic-gate lockfs32.lf_lock = 16537c478bd9Sstevel@tonic-gate (uint32_t)lockfs.lf_lock; 16547c478bd9Sstevel@tonic-gate lockfs32.lf_flags = 16557c478bd9Sstevel@tonic-gate (uint32_t)lockfs.lf_flags; 16567c478bd9Sstevel@tonic-gate lockfs32.lf_key = 16577c478bd9Sstevel@tonic-gate (uint32_t)lockfs.lf_key; 16587c478bd9Sstevel@tonic-gate lockfs32.lf_comlen = 16597c478bd9Sstevel@tonic-gate (uint32_t)lockfs.lf_comlen; 16607c478bd9Sstevel@tonic-gate lockfs32.lf_comment = 166180d34432Sfrankho (uint32_t)(uintptr_t) 166280d34432Sfrankho lockfs.lf_comment; 16637c478bd9Sstevel@tonic-gate (void) copyout(&lockfs32, (caddr_t)arg, 16647c478bd9Sstevel@tonic-gate sizeof (struct lockfs32)); 16657c478bd9Sstevel@tonic-gate } 16667c478bd9Sstevel@tonic-gate #endif /* _SYSCALL32_IMPL */ 16677c478bd9Sstevel@tonic-gate 16687c478bd9Sstevel@tonic-gate } else { 16697c478bd9Sstevel@tonic-gate if (lockfs.lf_comlen) 16707c478bd9Sstevel@tonic-gate kmem_free(comment, lockfs.lf_comlen); 16717c478bd9Sstevel@tonic-gate } 16727c478bd9Sstevel@tonic-gate return (error); 16737c478bd9Sstevel@tonic-gate 16747c478bd9Sstevel@tonic-gate case _FIOLFSS: 16757c478bd9Sstevel@tonic-gate /* 16767c478bd9Sstevel@tonic-gate * get file system locking status 16777c478bd9Sstevel@tonic-gate */ 16787c478bd9Sstevel@tonic-gate 16797c478bd9Sstevel@tonic-gate if ((flag & DATAMODEL_MASK) == DATAMODEL_NATIVE) { 16807c478bd9Sstevel@tonic-gate if (copyin((caddr_t)arg, &lockfs, 16817c478bd9Sstevel@tonic-gate sizeof (struct lockfs))) 16827c478bd9Sstevel@tonic-gate return (EFAULT); 16837c478bd9Sstevel@tonic-gate } 16847c478bd9Sstevel@tonic-gate #ifdef _SYSCALL32_IMPL 16857c478bd9Sstevel@tonic-gate else { 16867c478bd9Sstevel@tonic-gate struct lockfs32 lockfs32; 16877c478bd9Sstevel@tonic-gate /* Translate ILP32 lockfs to LP64 lockfs */ 16887c478bd9Sstevel@tonic-gate if (copyin((caddr_t)arg, &lockfs32, 16897c478bd9Sstevel@tonic-gate sizeof (struct lockfs32))) 16907c478bd9Sstevel@tonic-gate return (EFAULT); 16917c478bd9Sstevel@tonic-gate lockfs.lf_lock = (ulong_t)lockfs32.lf_lock; 16927c478bd9Sstevel@tonic-gate lockfs.lf_flags = (ulong_t)lockfs32.lf_flags; 16937c478bd9Sstevel@tonic-gate lockfs.lf_key = (ulong_t)lockfs32.lf_key; 16947c478bd9Sstevel@tonic-gate lockfs.lf_comlen = (ulong_t)lockfs32.lf_comlen; 16957c478bd9Sstevel@tonic-gate lockfs.lf_comment = 16967c478bd9Sstevel@tonic-gate (caddr_t)(uintptr_t)lockfs32.lf_comment; 16977c478bd9Sstevel@tonic-gate } 16987c478bd9Sstevel@tonic-gate #endif /* _SYSCALL32_IMPL */ 16997c478bd9Sstevel@tonic-gate 17007c478bd9Sstevel@tonic-gate if (error = ufs_fiolfss(vp, &lockfs_out)) 17017c478bd9Sstevel@tonic-gate return (error); 17027c478bd9Sstevel@tonic-gate lockfs.lf_lock = lockfs_out.lf_lock; 17037c478bd9Sstevel@tonic-gate lockfs.lf_key = lockfs_out.lf_key; 17047c478bd9Sstevel@tonic-gate lockfs.lf_flags = lockfs_out.lf_flags; 17057c478bd9Sstevel@tonic-gate lockfs.lf_comlen = MIN(lockfs.lf_comlen, 17067c478bd9Sstevel@tonic-gate lockfs_out.lf_comlen); 17077c478bd9Sstevel@tonic-gate 17087c478bd9Sstevel@tonic-gate if ((flag & DATAMODEL_MASK) == DATAMODEL_NATIVE) { 17097c478bd9Sstevel@tonic-gate if (copyout(&lockfs, (caddr_t)arg, 17107c478bd9Sstevel@tonic-gate sizeof (struct lockfs))) 17117c478bd9Sstevel@tonic-gate return (EFAULT); 17127c478bd9Sstevel@tonic-gate } 17137c478bd9Sstevel@tonic-gate #ifdef _SYSCALL32_IMPL 17147c478bd9Sstevel@tonic-gate else { 17157c478bd9Sstevel@tonic-gate /* Translate LP64 to ILP32 lockfs */ 17167c478bd9Sstevel@tonic-gate struct lockfs32 lockfs32; 17177c478bd9Sstevel@tonic-gate lockfs32.lf_lock = (uint32_t)lockfs.lf_lock; 17187c478bd9Sstevel@tonic-gate lockfs32.lf_flags = (uint32_t)lockfs.lf_flags; 17197c478bd9Sstevel@tonic-gate lockfs32.lf_key = (uint32_t)lockfs.lf_key; 17207c478bd9Sstevel@tonic-gate lockfs32.lf_comlen = (uint32_t)lockfs.lf_comlen; 17217c478bd9Sstevel@tonic-gate lockfs32.lf_comment = 17227c478bd9Sstevel@tonic-gate (uint32_t)(uintptr_t)lockfs.lf_comment; 17237c478bd9Sstevel@tonic-gate if (copyout(&lockfs32, (caddr_t)arg, 17247c478bd9Sstevel@tonic-gate sizeof (struct lockfs32))) 17257c478bd9Sstevel@tonic-gate return (EFAULT); 17267c478bd9Sstevel@tonic-gate } 17277c478bd9Sstevel@tonic-gate #endif /* _SYSCALL32_IMPL */ 17287c478bd9Sstevel@tonic-gate 17297c478bd9Sstevel@tonic-gate if (lockfs.lf_comlen && 17307c478bd9Sstevel@tonic-gate lockfs.lf_comment && lockfs_out.lf_comment) 17317c478bd9Sstevel@tonic-gate if (copyout(lockfs_out.lf_comment, 173280d34432Sfrankho lockfs.lf_comment, lockfs.lf_comlen)) 17337c478bd9Sstevel@tonic-gate return (EFAULT); 17347c478bd9Sstevel@tonic-gate return (0); 17357c478bd9Sstevel@tonic-gate 17367c478bd9Sstevel@tonic-gate case _FIOSATIME: 17377c478bd9Sstevel@tonic-gate /* 17387c478bd9Sstevel@tonic-gate * set access time 17397c478bd9Sstevel@tonic-gate */ 17407c478bd9Sstevel@tonic-gate 17417c478bd9Sstevel@tonic-gate /* 17427c478bd9Sstevel@tonic-gate * if mounted w/o atime, return quietly. 17437c478bd9Sstevel@tonic-gate * I briefly thought about returning ENOSYS, but 17447c478bd9Sstevel@tonic-gate * figured that most apps would consider this fatal 17457c478bd9Sstevel@tonic-gate * but the idea is to make this as seamless as poss. 17467c478bd9Sstevel@tonic-gate */ 17477c478bd9Sstevel@tonic-gate if (ufsvfsp->vfs_noatime) 17487c478bd9Sstevel@tonic-gate return (0); 17497c478bd9Sstevel@tonic-gate 17507c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, 17517c478bd9Sstevel@tonic-gate ULOCKFS_SETATTR_MASK); 17527c478bd9Sstevel@tonic-gate if (error) 17537c478bd9Sstevel@tonic-gate return (error); 17547c478bd9Sstevel@tonic-gate 17557c478bd9Sstevel@tonic-gate if (ulp) { 17567c478bd9Sstevel@tonic-gate trans_size = (int)TOP_SETATTR_SIZE(VTOI(vp)); 17577c478bd9Sstevel@tonic-gate TRANS_BEGIN_CSYNC(ufsvfsp, issync, 17587c478bd9Sstevel@tonic-gate TOP_SETATTR, trans_size); 17597c478bd9Sstevel@tonic-gate } 17607c478bd9Sstevel@tonic-gate 17617c478bd9Sstevel@tonic-gate error = ufs_fiosatime(vp, (struct timeval *)arg, 17627c478bd9Sstevel@tonic-gate flag, cr); 17637c478bd9Sstevel@tonic-gate 17647c478bd9Sstevel@tonic-gate if (ulp) { 17657c478bd9Sstevel@tonic-gate TRANS_END_CSYNC(ufsvfsp, error, issync, 17667c478bd9Sstevel@tonic-gate TOP_SETATTR, trans_size); 17677c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 17687c478bd9Sstevel@tonic-gate } 17697c478bd9Sstevel@tonic-gate return (error); 17707c478bd9Sstevel@tonic-gate 17717c478bd9Sstevel@tonic-gate case _FIOSDIO: 17727c478bd9Sstevel@tonic-gate /* 17737c478bd9Sstevel@tonic-gate * set delayed-io 17747c478bd9Sstevel@tonic-gate */ 17757c478bd9Sstevel@tonic-gate return (ufs_fiosdio(vp, (uint_t *)arg, flag, cr)); 17767c478bd9Sstevel@tonic-gate 17777c478bd9Sstevel@tonic-gate case _FIOGDIO: 17787c478bd9Sstevel@tonic-gate /* 17797c478bd9Sstevel@tonic-gate * get delayed-io 17807c478bd9Sstevel@tonic-gate */ 17817c478bd9Sstevel@tonic-gate return (ufs_fiogdio(vp, (uint_t *)arg, flag, cr)); 17827c478bd9Sstevel@tonic-gate 17837c478bd9Sstevel@tonic-gate case _FIOIO: 17847c478bd9Sstevel@tonic-gate /* 17857c478bd9Sstevel@tonic-gate * inode open 17867c478bd9Sstevel@tonic-gate */ 17877c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, 17887c478bd9Sstevel@tonic-gate ULOCKFS_VGET_MASK); 17897c478bd9Sstevel@tonic-gate if (error) 17907c478bd9Sstevel@tonic-gate return (error); 17917c478bd9Sstevel@tonic-gate 17927c478bd9Sstevel@tonic-gate error = ufs_fioio(vp, (struct fioio *)arg, flag, cr); 17937c478bd9Sstevel@tonic-gate 17947c478bd9Sstevel@tonic-gate if (ulp) { 17957c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 17967c478bd9Sstevel@tonic-gate } 17977c478bd9Sstevel@tonic-gate return (error); 17987c478bd9Sstevel@tonic-gate 17997c478bd9Sstevel@tonic-gate case _FIOFFS: 18007c478bd9Sstevel@tonic-gate /* 18017c478bd9Sstevel@tonic-gate * file system flush (push w/invalidate) 18027c478bd9Sstevel@tonic-gate */ 18037c478bd9Sstevel@tonic-gate if ((caddr_t)arg != NULL) 18047c478bd9Sstevel@tonic-gate return (EINVAL); 18057c478bd9Sstevel@tonic-gate return (ufs_fioffs(vp, NULL, cr)); 18067c478bd9Sstevel@tonic-gate 18077c478bd9Sstevel@tonic-gate case _FIOISBUSY: 18087c478bd9Sstevel@tonic-gate /* 18097c478bd9Sstevel@tonic-gate * Contract-private interface for Legato 18107c478bd9Sstevel@tonic-gate * Purge this vnode from the DNLC and decide 18117c478bd9Sstevel@tonic-gate * if this vnode is busy (*arg == 1) or not 18127c478bd9Sstevel@tonic-gate * (*arg == 0) 18137c478bd9Sstevel@tonic-gate */ 18147c478bd9Sstevel@tonic-gate if (secpolicy_fs_config(cr, ufsvfsp->vfs_vfs) != 0) 18157c478bd9Sstevel@tonic-gate return (EPERM); 18167c478bd9Sstevel@tonic-gate error = ufs_fioisbusy(vp, (int *)arg, cr); 18177c478bd9Sstevel@tonic-gate return (error); 18187c478bd9Sstevel@tonic-gate 18197c478bd9Sstevel@tonic-gate case _FIODIRECTIO: 18207c478bd9Sstevel@tonic-gate return (ufs_fiodirectio(vp, (int)arg, cr)); 18217c478bd9Sstevel@tonic-gate 18227c478bd9Sstevel@tonic-gate case _FIOTUNE: 18237c478bd9Sstevel@tonic-gate /* 18247c478bd9Sstevel@tonic-gate * Tune the file system (aka setting fs attributes) 18257c478bd9Sstevel@tonic-gate */ 18267c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, 18277c478bd9Sstevel@tonic-gate ULOCKFS_SETATTR_MASK); 18287c478bd9Sstevel@tonic-gate if (error) 18297c478bd9Sstevel@tonic-gate return (error); 18307c478bd9Sstevel@tonic-gate 18317c478bd9Sstevel@tonic-gate error = ufs_fiotune(vp, (struct fiotune *)arg, cr); 18327c478bd9Sstevel@tonic-gate 18337c478bd9Sstevel@tonic-gate if (ulp) 18347c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 18357c478bd9Sstevel@tonic-gate return (error); 18367c478bd9Sstevel@tonic-gate 18377c478bd9Sstevel@tonic-gate case _FIOLOGENABLE: 18387c478bd9Sstevel@tonic-gate if (secpolicy_fs_config(cr, ufsvfsp->vfs_vfs) != 0) 18397c478bd9Sstevel@tonic-gate return (EPERM); 18407c478bd9Sstevel@tonic-gate return (ufs_fiologenable(vp, (void *)arg, cr, flag)); 18417c478bd9Sstevel@tonic-gate 18427c478bd9Sstevel@tonic-gate case _FIOLOGDISABLE: 18437c478bd9Sstevel@tonic-gate if (secpolicy_fs_config(cr, ufsvfsp->vfs_vfs) != 0) 18447c478bd9Sstevel@tonic-gate return (EPERM); 18457c478bd9Sstevel@tonic-gate return (ufs_fiologdisable(vp, (void *)arg, cr, flag)); 18467c478bd9Sstevel@tonic-gate 18477c478bd9Sstevel@tonic-gate case _FIOISLOG: 18487c478bd9Sstevel@tonic-gate return (ufs_fioislog(vp, (void *)arg, cr, flag)); 18497c478bd9Sstevel@tonic-gate 18507c478bd9Sstevel@tonic-gate case _FIOSNAPSHOTCREATE_MULTI: 18517c478bd9Sstevel@tonic-gate { 18527c478bd9Sstevel@tonic-gate struct fiosnapcreate_multi fc, *fcp; 18537c478bd9Sstevel@tonic-gate size_t fcm_size; 18547c478bd9Sstevel@tonic-gate 18557c478bd9Sstevel@tonic-gate if (copyin((void *)arg, &fc, sizeof (fc))) 18567c478bd9Sstevel@tonic-gate return (EFAULT); 18577c478bd9Sstevel@tonic-gate if (fc.backfilecount > MAX_BACKFILE_COUNT) 18587c478bd9Sstevel@tonic-gate return (EINVAL); 18597c478bd9Sstevel@tonic-gate fcm_size = sizeof (struct fiosnapcreate_multi) + 18607c478bd9Sstevel@tonic-gate (fc.backfilecount - 1) * sizeof (int); 18617c478bd9Sstevel@tonic-gate fcp = (struct fiosnapcreate_multi *) 18627c478bd9Sstevel@tonic-gate kmem_alloc(fcm_size, KM_SLEEP); 18637c478bd9Sstevel@tonic-gate if (copyin((void *)arg, fcp, fcm_size)) { 18647c478bd9Sstevel@tonic-gate kmem_free(fcp, fcm_size); 18657c478bd9Sstevel@tonic-gate return (EFAULT); 18667c478bd9Sstevel@tonic-gate } 18677c478bd9Sstevel@tonic-gate error = ufs_snap_create(vp, fcp, cr); 186895b7dc1dSmishra /* 186995b7dc1dSmishra * Do copyout even if there is an error because 187095b7dc1dSmishra * the details of error is stored in fcp. 187195b7dc1dSmishra */ 187295b7dc1dSmishra if (copyout(fcp, (void *)arg, fcm_size)) 18737c478bd9Sstevel@tonic-gate error = EFAULT; 18747c478bd9Sstevel@tonic-gate kmem_free(fcp, fcm_size); 18757c478bd9Sstevel@tonic-gate return (error); 18767c478bd9Sstevel@tonic-gate } 18777c478bd9Sstevel@tonic-gate 18787c478bd9Sstevel@tonic-gate case _FIOSNAPSHOTDELETE: 18797c478bd9Sstevel@tonic-gate { 18807c478bd9Sstevel@tonic-gate struct fiosnapdelete fc; 18817c478bd9Sstevel@tonic-gate 18827c478bd9Sstevel@tonic-gate if (copyin((void *)arg, &fc, sizeof (fc))) 18837c478bd9Sstevel@tonic-gate return (EFAULT); 18847c478bd9Sstevel@tonic-gate error = ufs_snap_delete(vp, &fc, cr); 18857c478bd9Sstevel@tonic-gate if (!error && copyout(&fc, (void *)arg, sizeof (fc))) 18867c478bd9Sstevel@tonic-gate error = EFAULT; 18877c478bd9Sstevel@tonic-gate return (error); 18887c478bd9Sstevel@tonic-gate } 18897c478bd9Sstevel@tonic-gate 18907c478bd9Sstevel@tonic-gate case _FIOGETSUPERBLOCK: 18917c478bd9Sstevel@tonic-gate if (copyout(fs, (void *)arg, SBSIZE)) 18927c478bd9Sstevel@tonic-gate return (EFAULT); 18937c478bd9Sstevel@tonic-gate return (0); 18947c478bd9Sstevel@tonic-gate 18957c478bd9Sstevel@tonic-gate case _FIOGETMAXPHYS: 18967c478bd9Sstevel@tonic-gate if (copyout(&maxphys, (void *)arg, sizeof (maxphys))) 18977c478bd9Sstevel@tonic-gate return (EFAULT); 18987c478bd9Sstevel@tonic-gate return (0); 18997c478bd9Sstevel@tonic-gate 19007c478bd9Sstevel@tonic-gate /* 19017c478bd9Sstevel@tonic-gate * The following 3 ioctls are for TSufs support 19027c478bd9Sstevel@tonic-gate * although could potentially be used elsewhere 19037c478bd9Sstevel@tonic-gate */ 19047c478bd9Sstevel@tonic-gate case _FIO_SET_LUFS_DEBUG: 19057c478bd9Sstevel@tonic-gate if (secpolicy_fs_config(cr, ufsvfsp->vfs_vfs) != 0) 19067c478bd9Sstevel@tonic-gate return (EPERM); 19077c478bd9Sstevel@tonic-gate lufs_debug = (uint32_t)arg; 19087c478bd9Sstevel@tonic-gate return (0); 19097c478bd9Sstevel@tonic-gate 19107c478bd9Sstevel@tonic-gate case _FIO_SET_LUFS_ERROR: 19117c478bd9Sstevel@tonic-gate if (secpolicy_fs_config(cr, ufsvfsp->vfs_vfs) != 0) 19127c478bd9Sstevel@tonic-gate return (EPERM); 19137c478bd9Sstevel@tonic-gate TRANS_SETERROR(ufsvfsp); 19147c478bd9Sstevel@tonic-gate return (0); 19157c478bd9Sstevel@tonic-gate 19167c478bd9Sstevel@tonic-gate case _FIO_GET_TOP_STATS: 19177c478bd9Sstevel@tonic-gate { 19187c478bd9Sstevel@tonic-gate fio_lufs_stats_t *ls; 19197c478bd9Sstevel@tonic-gate ml_unit_t *ul = ufsvfsp->vfs_log; 19207c478bd9Sstevel@tonic-gate 19217c478bd9Sstevel@tonic-gate ls = kmem_zalloc(sizeof (*ls), KM_SLEEP); 19227c478bd9Sstevel@tonic-gate ls->ls_debug = ul->un_debug; /* return debug value */ 19237c478bd9Sstevel@tonic-gate /* Copy stucture if statistics are being kept */ 19247c478bd9Sstevel@tonic-gate if (ul->un_logmap->mtm_tops) { 19257c478bd9Sstevel@tonic-gate ls->ls_topstats = *(ul->un_logmap->mtm_tops); 19267c478bd9Sstevel@tonic-gate } 19277c478bd9Sstevel@tonic-gate error = 0; 19287c478bd9Sstevel@tonic-gate if (copyout(ls, (void *)arg, sizeof (*ls))) 19297c478bd9Sstevel@tonic-gate error = EFAULT; 19307c478bd9Sstevel@tonic-gate kmem_free(ls, sizeof (*ls)); 19317c478bd9Sstevel@tonic-gate return (error); 19327c478bd9Sstevel@tonic-gate } 19337c478bd9Sstevel@tonic-gate 19347c478bd9Sstevel@tonic-gate case _FIO_SEEK_DATA: 19357c478bd9Sstevel@tonic-gate case _FIO_SEEK_HOLE: 19367c478bd9Sstevel@tonic-gate if (ddi_copyin((void *)arg, &off, sizeof (off), flag)) 19377c478bd9Sstevel@tonic-gate return (EFAULT); 19387c478bd9Sstevel@tonic-gate /* offset paramater is in/out */ 19397c478bd9Sstevel@tonic-gate error = ufs_fio_holey(vp, cmd, &off); 19407c478bd9Sstevel@tonic-gate if (error) 19417c478bd9Sstevel@tonic-gate return (error); 19427c478bd9Sstevel@tonic-gate if (ddi_copyout(&off, (void *)arg, sizeof (off), flag)) 19437c478bd9Sstevel@tonic-gate return (EFAULT); 19447c478bd9Sstevel@tonic-gate return (0); 19457c478bd9Sstevel@tonic-gate 1946986fd29aSsetje case _FIO_COMPRESSED: 1947986fd29aSsetje { 1948986fd29aSsetje /* 1949986fd29aSsetje * This is a project private ufs ioctl() to mark 1950986fd29aSsetje * the inode as that belonging to a compressed 1951986fd29aSsetje * file. This is used to mark individual 1952342440ecSPrasad Singamsetty * compressed files in a miniroot archive. 1953986fd29aSsetje * The files compressed in this manner are 1954986fd29aSsetje * automatically decompressed by the dcfs filesystem 1955986fd29aSsetje * (via an interception in ufs_lookup - see decompvp()) 1956986fd29aSsetje * which is layered on top of ufs on a system running 1957342440ecSPrasad Singamsetty * from the archive. See uts/common/fs/dcfs for details. 1958986fd29aSsetje * This ioctl only marks the file as compressed - the 1959986fd29aSsetje * actual compression is done by fiocompress (a 1960986fd29aSsetje * userland utility) which invokes this ioctl(). 1961986fd29aSsetje */ 1962986fd29aSsetje struct inode *ip = VTOI(vp); 1963986fd29aSsetje 1964986fd29aSsetje error = ufs_lockfs_begin(ufsvfsp, &ulp, 1965986fd29aSsetje ULOCKFS_SETATTR_MASK); 1966986fd29aSsetje if (error) 1967986fd29aSsetje return (error); 1968986fd29aSsetje 1969986fd29aSsetje if (ulp) { 1970986fd29aSsetje TRANS_BEGIN_ASYNC(ufsvfsp, TOP_IUPDAT, 1971986fd29aSsetje TOP_IUPDAT_SIZE(ip)); 1972986fd29aSsetje } 1973986fd29aSsetje 1974986fd29aSsetje error = ufs_mark_compressed(vp); 1975986fd29aSsetje 1976986fd29aSsetje if (ulp) { 1977986fd29aSsetje TRANS_END_ASYNC(ufsvfsp, TOP_IUPDAT, 1978986fd29aSsetje TOP_IUPDAT_SIZE(ip)); 1979986fd29aSsetje ufs_lockfs_end(ulp); 1980986fd29aSsetje } 1981986fd29aSsetje 1982986fd29aSsetje return (error); 1983986fd29aSsetje 1984986fd29aSsetje } 1985986fd29aSsetje 19867c478bd9Sstevel@tonic-gate default: 19877c478bd9Sstevel@tonic-gate return (ENOTTY); 19887c478bd9Sstevel@tonic-gate } 19897c478bd9Sstevel@tonic-gate } 19907c478bd9Sstevel@tonic-gate 1991986fd29aSsetje 19927c478bd9Sstevel@tonic-gate /* ARGSUSED */ 19937c478bd9Sstevel@tonic-gate static int 19947c478bd9Sstevel@tonic-gate ufs_getattr(struct vnode *vp, struct vattr *vap, int flags, 1995da6c28aaSamw struct cred *cr, caller_context_t *ct) 19967c478bd9Sstevel@tonic-gate { 19977c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 19987c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp; 19997c478bd9Sstevel@tonic-gate int err; 20007c478bd9Sstevel@tonic-gate 20017c478bd9Sstevel@tonic-gate if (vap->va_mask == AT_SIZE) { 20027c478bd9Sstevel@tonic-gate /* 20037c478bd9Sstevel@tonic-gate * for performance, if only the size is requested don't bother 20047c478bd9Sstevel@tonic-gate * with anything else. 20057c478bd9Sstevel@tonic-gate */ 20067c478bd9Sstevel@tonic-gate UFS_GET_ISIZE(&vap->va_size, ip); 20077c478bd9Sstevel@tonic-gate return (0); 20087c478bd9Sstevel@tonic-gate } 20097c478bd9Sstevel@tonic-gate 20107c478bd9Sstevel@tonic-gate /* 20117c478bd9Sstevel@tonic-gate * inlined lockfs checks 20127c478bd9Sstevel@tonic-gate */ 20137c478bd9Sstevel@tonic-gate ufsvfsp = ip->i_ufsvfs; 20147c478bd9Sstevel@tonic-gate if ((ufsvfsp == NULL) || ULOCKFS_IS_HLOCK(&ufsvfsp->vfs_ulockfs)) { 20157c478bd9Sstevel@tonic-gate err = EIO; 20167c478bd9Sstevel@tonic-gate goto out; 20177c478bd9Sstevel@tonic-gate } 20187c478bd9Sstevel@tonic-gate 20197c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_READER); 20207c478bd9Sstevel@tonic-gate /* 20217c478bd9Sstevel@tonic-gate * Return all the attributes. This should be refined so 20227c478bd9Sstevel@tonic-gate * that it only returns what's asked for. 20237c478bd9Sstevel@tonic-gate */ 20247c478bd9Sstevel@tonic-gate 20257c478bd9Sstevel@tonic-gate /* 20267c478bd9Sstevel@tonic-gate * Copy from inode table. 20277c478bd9Sstevel@tonic-gate */ 20287c478bd9Sstevel@tonic-gate vap->va_type = vp->v_type; 20297c478bd9Sstevel@tonic-gate vap->va_mode = ip->i_mode & MODEMASK; 20307c478bd9Sstevel@tonic-gate /* 20317c478bd9Sstevel@tonic-gate * If there is an ACL and there is a mask entry, then do the 20327c478bd9Sstevel@tonic-gate * extra work that completes the equivalent of an acltomode(3) 20337c478bd9Sstevel@tonic-gate * call. According to POSIX P1003.1e, the acl mask should be 20347c478bd9Sstevel@tonic-gate * returned in the group permissions field. 20357c478bd9Sstevel@tonic-gate * 20367c478bd9Sstevel@tonic-gate * - start with the original permission and mode bits (from above) 20377c478bd9Sstevel@tonic-gate * - clear the group owner bits 20387c478bd9Sstevel@tonic-gate * - add in the mask bits. 20397c478bd9Sstevel@tonic-gate */ 20407c478bd9Sstevel@tonic-gate if (ip->i_ufs_acl && ip->i_ufs_acl->aclass.acl_ismask) { 20417c478bd9Sstevel@tonic-gate vap->va_mode &= ~((VREAD | VWRITE | VEXEC) >> 3); 20427c478bd9Sstevel@tonic-gate vap->va_mode |= 20437c478bd9Sstevel@tonic-gate (ip->i_ufs_acl->aclass.acl_maskbits & PERMMASK) << 3; 20447c478bd9Sstevel@tonic-gate } 20457c478bd9Sstevel@tonic-gate vap->va_uid = ip->i_uid; 20467c478bd9Sstevel@tonic-gate vap->va_gid = ip->i_gid; 20477c478bd9Sstevel@tonic-gate vap->va_fsid = ip->i_dev; 20487c478bd9Sstevel@tonic-gate vap->va_nodeid = (ino64_t)ip->i_number; 20497c478bd9Sstevel@tonic-gate vap->va_nlink = ip->i_nlink; 20507c478bd9Sstevel@tonic-gate vap->va_size = ip->i_size; 20517c478bd9Sstevel@tonic-gate if (vp->v_type == VCHR || vp->v_type == VBLK) 20527c478bd9Sstevel@tonic-gate vap->va_rdev = ip->i_rdev; 20537c478bd9Sstevel@tonic-gate else 20547c478bd9Sstevel@tonic-gate vap->va_rdev = 0; /* not a b/c spec. */ 20557c478bd9Sstevel@tonic-gate mutex_enter(&ip->i_tlock); 20567c478bd9Sstevel@tonic-gate ITIMES_NOLOCK(ip); /* mark correct time in inode */ 20577c478bd9Sstevel@tonic-gate vap->va_seq = ip->i_seq; 20587c478bd9Sstevel@tonic-gate vap->va_atime.tv_sec = (time_t)ip->i_atime.tv_sec; 20597c478bd9Sstevel@tonic-gate vap->va_atime.tv_nsec = ip->i_atime.tv_usec*1000; 20607c478bd9Sstevel@tonic-gate vap->va_mtime.tv_sec = (time_t)ip->i_mtime.tv_sec; 20617c478bd9Sstevel@tonic-gate vap->va_mtime.tv_nsec = ip->i_mtime.tv_usec*1000; 20627c478bd9Sstevel@tonic-gate vap->va_ctime.tv_sec = (time_t)ip->i_ctime.tv_sec; 20637c478bd9Sstevel@tonic-gate vap->va_ctime.tv_nsec = ip->i_ctime.tv_usec*1000; 20647c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 20657c478bd9Sstevel@tonic-gate 20667c478bd9Sstevel@tonic-gate switch (ip->i_mode & IFMT) { 20677c478bd9Sstevel@tonic-gate 20687c478bd9Sstevel@tonic-gate case IFBLK: 20697c478bd9Sstevel@tonic-gate vap->va_blksize = MAXBSIZE; /* was BLKDEV_IOSIZE */ 20707c478bd9Sstevel@tonic-gate break; 20717c478bd9Sstevel@tonic-gate 20727c478bd9Sstevel@tonic-gate case IFCHR: 20737c478bd9Sstevel@tonic-gate vap->va_blksize = MAXBSIZE; 20747c478bd9Sstevel@tonic-gate break; 20757c478bd9Sstevel@tonic-gate 20767c478bd9Sstevel@tonic-gate default: 20777c478bd9Sstevel@tonic-gate vap->va_blksize = ip->i_fs->fs_bsize; 20787c478bd9Sstevel@tonic-gate break; 20797c478bd9Sstevel@tonic-gate } 20807c478bd9Sstevel@tonic-gate vap->va_nblocks = (fsblkcnt64_t)ip->i_blocks; 20817c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 20827c478bd9Sstevel@tonic-gate err = 0; 20837c478bd9Sstevel@tonic-gate 20847c478bd9Sstevel@tonic-gate out: 20857c478bd9Sstevel@tonic-gate return (err); 20867c478bd9Sstevel@tonic-gate } 20877c478bd9Sstevel@tonic-gate 208860c8e821SFrank Batschulat /* 208960c8e821SFrank Batschulat * Special wrapper to provide a callback for secpolicy_vnode_setattr(). 209060c8e821SFrank Batschulat * The i_contents lock is already held by the caller and we need to 209160c8e821SFrank Batschulat * declare the inode as 'void *' argument. 209260c8e821SFrank Batschulat */ 209360c8e821SFrank Batschulat static int 209460c8e821SFrank Batschulat ufs_priv_access(void *vip, int mode, struct cred *cr) 209560c8e821SFrank Batschulat { 209660c8e821SFrank Batschulat struct inode *ip = vip; 209760c8e821SFrank Batschulat 209860c8e821SFrank Batschulat return (ufs_iaccess(ip, mode, cr, 0)); 209960c8e821SFrank Batschulat } 210060c8e821SFrank Batschulat 21017c478bd9Sstevel@tonic-gate /*ARGSUSED4*/ 21027c478bd9Sstevel@tonic-gate static int 21037c478bd9Sstevel@tonic-gate ufs_setattr( 21047c478bd9Sstevel@tonic-gate struct vnode *vp, 21057c478bd9Sstevel@tonic-gate struct vattr *vap, 21067c478bd9Sstevel@tonic-gate int flags, 21077c478bd9Sstevel@tonic-gate struct cred *cr, 21087c478bd9Sstevel@tonic-gate caller_context_t *ct) 21097c478bd9Sstevel@tonic-gate { 21107c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 21117c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp = ip->i_ufsvfs; 21127c478bd9Sstevel@tonic-gate struct fs *fs; 21137c478bd9Sstevel@tonic-gate struct ulockfs *ulp; 21147c478bd9Sstevel@tonic-gate char *errmsg1; 21157c478bd9Sstevel@tonic-gate char *errmsg2; 21167c478bd9Sstevel@tonic-gate long blocks; 21177c478bd9Sstevel@tonic-gate long int mask = vap->va_mask; 21187c478bd9Sstevel@tonic-gate size_t len1, len2; 21197c478bd9Sstevel@tonic-gate int issync; 21207c478bd9Sstevel@tonic-gate int trans_size; 21217c478bd9Sstevel@tonic-gate int dotrans; 21227c478bd9Sstevel@tonic-gate int dorwlock; 21237c478bd9Sstevel@tonic-gate int error; 21247c478bd9Sstevel@tonic-gate int owner_change; 21257c478bd9Sstevel@tonic-gate int dodqlock; 21267c478bd9Sstevel@tonic-gate timestruc_t now; 21277c478bd9Sstevel@tonic-gate vattr_t oldva; 21287c478bd9Sstevel@tonic-gate int retry = 1; 212902ff05a9Svsakar int indeadlock; 21307c478bd9Sstevel@tonic-gate 21317c478bd9Sstevel@tonic-gate /* 21327c478bd9Sstevel@tonic-gate * Cannot set these attributes. 21337c478bd9Sstevel@tonic-gate */ 2134da6c28aaSamw if ((mask & AT_NOSET) || (mask & AT_XVATTR)) 2135da6c28aaSamw return (EINVAL); 21367c478bd9Sstevel@tonic-gate 21377c478bd9Sstevel@tonic-gate /* 21387c478bd9Sstevel@tonic-gate * check for forced unmount 21397c478bd9Sstevel@tonic-gate */ 21407c478bd9Sstevel@tonic-gate if (ufsvfsp == NULL) 21417c478bd9Sstevel@tonic-gate return (EIO); 21427c478bd9Sstevel@tonic-gate 21437c478bd9Sstevel@tonic-gate fs = ufsvfsp->vfs_fs; 21447c478bd9Sstevel@tonic-gate if (fs->fs_ronly != 0) 21457c478bd9Sstevel@tonic-gate return (EROFS); 21467c478bd9Sstevel@tonic-gate 21477c478bd9Sstevel@tonic-gate again: 21487c478bd9Sstevel@tonic-gate errmsg1 = NULL; 21497c478bd9Sstevel@tonic-gate errmsg2 = NULL; 21507c478bd9Sstevel@tonic-gate dotrans = 0; 21517c478bd9Sstevel@tonic-gate dorwlock = 0; 21527c478bd9Sstevel@tonic-gate dodqlock = 0; 21537c478bd9Sstevel@tonic-gate 21547c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_SETATTR_MASK); 21557c478bd9Sstevel@tonic-gate if (error) 21567c478bd9Sstevel@tonic-gate goto out; 21577c478bd9Sstevel@tonic-gate 21587c478bd9Sstevel@tonic-gate /* 21597c478bd9Sstevel@tonic-gate * Acquire i_rwlock before TRANS_BEGIN_CSYNC() if this is a file. 21607c478bd9Sstevel@tonic-gate * This follows the protocol for read()/write(). 21617c478bd9Sstevel@tonic-gate */ 21627c478bd9Sstevel@tonic-gate if (vp->v_type != VDIR) { 216302ff05a9Svsakar /* 216402ff05a9Svsakar * ufs_tryirwlock uses rw_tryenter and checks for SLOCK to 216502ff05a9Svsakar * avoid i_rwlock, ufs_lockfs_begin deadlock. If deadlock 216602ff05a9Svsakar * possible, retries the operation. 216702ff05a9Svsakar */ 216802ff05a9Svsakar ufs_tryirwlock(&ip->i_rwlock, RW_WRITER, retry_file); 216902ff05a9Svsakar if (indeadlock) { 217002ff05a9Svsakar if (ulp) 217102ff05a9Svsakar ufs_lockfs_end(ulp); 217202ff05a9Svsakar goto again; 217302ff05a9Svsakar } 21747c478bd9Sstevel@tonic-gate dorwlock = 1; 21757c478bd9Sstevel@tonic-gate } 21767c478bd9Sstevel@tonic-gate 21777c478bd9Sstevel@tonic-gate /* 21787c478bd9Sstevel@tonic-gate * Truncate file. Must have write permission and not be a directory. 21797c478bd9Sstevel@tonic-gate */ 21807c478bd9Sstevel@tonic-gate if (mask & AT_SIZE) { 21817c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_WRITER); 21827c478bd9Sstevel@tonic-gate if (vp->v_type == VDIR) { 21837c478bd9Sstevel@tonic-gate error = EISDIR; 21847c478bd9Sstevel@tonic-gate goto update_inode; 21857c478bd9Sstevel@tonic-gate } 218660c8e821SFrank Batschulat if (error = ufs_iaccess(ip, IWRITE, cr, 0)) 21877c478bd9Sstevel@tonic-gate goto update_inode; 21887c478bd9Sstevel@tonic-gate 21897c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 21907c478bd9Sstevel@tonic-gate error = TRANS_ITRUNC(ip, vap->va_size, 0, cr); 21917c478bd9Sstevel@tonic-gate if (error) { 21927c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_WRITER); 21937c478bd9Sstevel@tonic-gate goto update_inode; 21947c478bd9Sstevel@tonic-gate } 219572102e74SBryan Cantrill 219672102e74SBryan Cantrill if (error == 0 && vap->va_size) 219772102e74SBryan Cantrill vnevent_truncate(vp, ct); 21987c478bd9Sstevel@tonic-gate } 21997c478bd9Sstevel@tonic-gate 22007c478bd9Sstevel@tonic-gate if (ulp) { 22017c478bd9Sstevel@tonic-gate trans_size = (int)TOP_SETATTR_SIZE(ip); 22027c478bd9Sstevel@tonic-gate TRANS_BEGIN_CSYNC(ufsvfsp, issync, TOP_SETATTR, trans_size); 22037c478bd9Sstevel@tonic-gate ++dotrans; 22047c478bd9Sstevel@tonic-gate } 22057c478bd9Sstevel@tonic-gate 22067c478bd9Sstevel@tonic-gate /* 22077c478bd9Sstevel@tonic-gate * Acquire i_rwlock after TRANS_BEGIN_CSYNC() if this is a directory. 22087c478bd9Sstevel@tonic-gate * This follows the protocol established by 22097c478bd9Sstevel@tonic-gate * ufs_link/create/remove/rename/mkdir/rmdir/symlink. 22107c478bd9Sstevel@tonic-gate */ 22117c478bd9Sstevel@tonic-gate if (vp->v_type == VDIR) { 221202ff05a9Svsakar ufs_tryirwlock_trans(&ip->i_rwlock, RW_WRITER, TOP_SETATTR, 221302ff05a9Svsakar retry_dir); 221402ff05a9Svsakar if (indeadlock) 221502ff05a9Svsakar goto again; 22167c478bd9Sstevel@tonic-gate dorwlock = 1; 22177c478bd9Sstevel@tonic-gate } 22187c478bd9Sstevel@tonic-gate 22197c478bd9Sstevel@tonic-gate /* 22207c478bd9Sstevel@tonic-gate * Grab quota lock if we are changing the file's owner. 22217c478bd9Sstevel@tonic-gate */ 22227c478bd9Sstevel@tonic-gate if (mask & AT_UID) { 22237c478bd9Sstevel@tonic-gate rw_enter(&ufsvfsp->vfs_dqrwlock, RW_READER); 22247c478bd9Sstevel@tonic-gate dodqlock = 1; 22257c478bd9Sstevel@tonic-gate } 22267c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_WRITER); 22277c478bd9Sstevel@tonic-gate 22287c478bd9Sstevel@tonic-gate oldva.va_mode = ip->i_mode; 22297c478bd9Sstevel@tonic-gate oldva.va_uid = ip->i_uid; 22307c478bd9Sstevel@tonic-gate oldva.va_gid = ip->i_gid; 22317c478bd9Sstevel@tonic-gate 22327c478bd9Sstevel@tonic-gate vap->va_mask &= ~AT_SIZE; 223360c8e821SFrank Batschulat 22347c478bd9Sstevel@tonic-gate error = secpolicy_vnode_setattr(cr, vp, vap, &oldva, flags, 223560c8e821SFrank Batschulat ufs_priv_access, ip); 22367c478bd9Sstevel@tonic-gate if (error) 22377c478bd9Sstevel@tonic-gate goto update_inode; 22387c478bd9Sstevel@tonic-gate 22397c478bd9Sstevel@tonic-gate mask = vap->va_mask; 22407c478bd9Sstevel@tonic-gate 22417c478bd9Sstevel@tonic-gate /* 22427c478bd9Sstevel@tonic-gate * Change file access modes. 22437c478bd9Sstevel@tonic-gate */ 22447c478bd9Sstevel@tonic-gate if (mask & AT_MODE) { 22457c478bd9Sstevel@tonic-gate ip->i_mode = (ip->i_mode & IFMT) | (vap->va_mode & ~IFMT); 22467c478bd9Sstevel@tonic-gate TRANS_INODE(ufsvfsp, ip); 22477c478bd9Sstevel@tonic-gate ip->i_flag |= ICHG; 22487c478bd9Sstevel@tonic-gate if (stickyhack) { 22497c478bd9Sstevel@tonic-gate mutex_enter(&vp->v_lock); 22507c478bd9Sstevel@tonic-gate if ((ip->i_mode & (ISVTX | IEXEC | IFDIR)) == ISVTX) 22517c478bd9Sstevel@tonic-gate vp->v_flag |= VSWAPLIKE; 22527c478bd9Sstevel@tonic-gate else 22537c478bd9Sstevel@tonic-gate vp->v_flag &= ~VSWAPLIKE; 22547c478bd9Sstevel@tonic-gate mutex_exit(&vp->v_lock); 22557c478bd9Sstevel@tonic-gate } 22567c478bd9Sstevel@tonic-gate } 22577c478bd9Sstevel@tonic-gate if (mask & (AT_UID|AT_GID)) { 22587c478bd9Sstevel@tonic-gate if (mask & AT_UID) { 22597c478bd9Sstevel@tonic-gate /* 22607c478bd9Sstevel@tonic-gate * Don't change ownership of the quota inode. 22617c478bd9Sstevel@tonic-gate */ 22627c478bd9Sstevel@tonic-gate if (ufsvfsp->vfs_qinod == ip) { 22637c478bd9Sstevel@tonic-gate ASSERT(ufsvfsp->vfs_qflags & MQ_ENABLED); 22647c478bd9Sstevel@tonic-gate error = EINVAL; 22657c478bd9Sstevel@tonic-gate goto update_inode; 22667c478bd9Sstevel@tonic-gate } 22677c478bd9Sstevel@tonic-gate 22687c478bd9Sstevel@tonic-gate /* 22697c478bd9Sstevel@tonic-gate * No real ownership change. 22707c478bd9Sstevel@tonic-gate */ 22717c478bd9Sstevel@tonic-gate if (ip->i_uid == vap->va_uid) { 22727c478bd9Sstevel@tonic-gate blocks = 0; 22737c478bd9Sstevel@tonic-gate owner_change = 0; 22747c478bd9Sstevel@tonic-gate } 22757c478bd9Sstevel@tonic-gate /* 22767c478bd9Sstevel@tonic-gate * Remove the blocks and the file, from the old user's 22777c478bd9Sstevel@tonic-gate * quota. 22787c478bd9Sstevel@tonic-gate */ 22797c478bd9Sstevel@tonic-gate else { 22807c478bd9Sstevel@tonic-gate blocks = ip->i_blocks; 22817c478bd9Sstevel@tonic-gate owner_change = 1; 22827c478bd9Sstevel@tonic-gate 22837c478bd9Sstevel@tonic-gate (void) chkdq(ip, -blocks, /* force */ 1, cr, 22847c478bd9Sstevel@tonic-gate (char **)NULL, (size_t *)NULL); 22857c478bd9Sstevel@tonic-gate (void) chkiq(ufsvfsp, /* change */ -1, ip, 228680d34432Sfrankho (uid_t)ip->i_uid, /* force */ 1, cr, 22877c478bd9Sstevel@tonic-gate (char **)NULL, (size_t *)NULL); 22887c478bd9Sstevel@tonic-gate dqrele(ip->i_dquot); 22897c478bd9Sstevel@tonic-gate } 22907c478bd9Sstevel@tonic-gate 22917c478bd9Sstevel@tonic-gate ip->i_uid = vap->va_uid; 22927c478bd9Sstevel@tonic-gate 22937c478bd9Sstevel@tonic-gate /* 22947c478bd9Sstevel@tonic-gate * There is a real ownership change. 22957c478bd9Sstevel@tonic-gate */ 22967c478bd9Sstevel@tonic-gate if (owner_change) { 22977c478bd9Sstevel@tonic-gate /* 22987c478bd9Sstevel@tonic-gate * Add the blocks and the file to the new 22997c478bd9Sstevel@tonic-gate * user's quota. 23007c478bd9Sstevel@tonic-gate */ 23017c478bd9Sstevel@tonic-gate ip->i_dquot = getinoquota(ip); 23027c478bd9Sstevel@tonic-gate (void) chkdq(ip, blocks, /* force */ 1, cr, 23037c478bd9Sstevel@tonic-gate &errmsg1, &len1); 23047c478bd9Sstevel@tonic-gate (void) chkiq(ufsvfsp, /* change */ 1, 230580d34432Sfrankho (struct inode *)NULL, (uid_t)ip->i_uid, 230680d34432Sfrankho /* force */ 1, cr, &errmsg2, &len2); 23077c478bd9Sstevel@tonic-gate } 23087c478bd9Sstevel@tonic-gate } 23097c478bd9Sstevel@tonic-gate if (mask & AT_GID) { 23107c478bd9Sstevel@tonic-gate ip->i_gid = vap->va_gid; 23117c478bd9Sstevel@tonic-gate } 23127c478bd9Sstevel@tonic-gate TRANS_INODE(ufsvfsp, ip); 23137c478bd9Sstevel@tonic-gate ip->i_flag |= ICHG; 23147c478bd9Sstevel@tonic-gate } 23157c478bd9Sstevel@tonic-gate /* 23167c478bd9Sstevel@tonic-gate * Change file access or modified times. 23177c478bd9Sstevel@tonic-gate */ 23187c478bd9Sstevel@tonic-gate if (mask & (AT_ATIME|AT_MTIME)) { 23197c478bd9Sstevel@tonic-gate /* Check that the time value is within ufs range */ 23207c478bd9Sstevel@tonic-gate if (((mask & AT_ATIME) && TIMESPEC_OVERFLOW(&vap->va_atime)) || 23217c478bd9Sstevel@tonic-gate ((mask & AT_MTIME) && TIMESPEC_OVERFLOW(&vap->va_mtime))) { 23227c478bd9Sstevel@tonic-gate error = EOVERFLOW; 23237c478bd9Sstevel@tonic-gate goto update_inode; 23247c478bd9Sstevel@tonic-gate } 23257c478bd9Sstevel@tonic-gate 23267c478bd9Sstevel@tonic-gate /* 23277c478bd9Sstevel@tonic-gate * if the "noaccess" mount option is set and only atime 23287c478bd9Sstevel@tonic-gate * update is requested, do nothing. No error is returned. 23297c478bd9Sstevel@tonic-gate */ 23307c478bd9Sstevel@tonic-gate if ((ufsvfsp->vfs_noatime) && 23317c478bd9Sstevel@tonic-gate ((mask & (AT_ATIME|AT_MTIME)) == AT_ATIME)) 23327c478bd9Sstevel@tonic-gate goto skip_atime; 23337c478bd9Sstevel@tonic-gate 23347c478bd9Sstevel@tonic-gate if (mask & AT_ATIME) { 23357c478bd9Sstevel@tonic-gate ip->i_atime.tv_sec = vap->va_atime.tv_sec; 23367c478bd9Sstevel@tonic-gate ip->i_atime.tv_usec = vap->va_atime.tv_nsec / 1000; 23377c478bd9Sstevel@tonic-gate ip->i_flag &= ~IACC; 23387c478bd9Sstevel@tonic-gate } 23397c478bd9Sstevel@tonic-gate if (mask & AT_MTIME) { 23407c478bd9Sstevel@tonic-gate ip->i_mtime.tv_sec = vap->va_mtime.tv_sec; 23417c478bd9Sstevel@tonic-gate ip->i_mtime.tv_usec = vap->va_mtime.tv_nsec / 1000; 23427c478bd9Sstevel@tonic-gate gethrestime(&now); 23437c478bd9Sstevel@tonic-gate if (now.tv_sec > TIME32_MAX) { 23447c478bd9Sstevel@tonic-gate /* 23457c478bd9Sstevel@tonic-gate * In 2038, ctime sticks forever.. 23467c478bd9Sstevel@tonic-gate */ 23477c478bd9Sstevel@tonic-gate ip->i_ctime.tv_sec = TIME32_MAX; 23487c478bd9Sstevel@tonic-gate ip->i_ctime.tv_usec = 0; 23497c478bd9Sstevel@tonic-gate } else { 23507c478bd9Sstevel@tonic-gate ip->i_ctime.tv_sec = now.tv_sec; 23517c478bd9Sstevel@tonic-gate ip->i_ctime.tv_usec = now.tv_nsec / 1000; 23527c478bd9Sstevel@tonic-gate } 23537c478bd9Sstevel@tonic-gate ip->i_flag &= ~(IUPD|ICHG); 23547c478bd9Sstevel@tonic-gate ip->i_flag |= IMODTIME; 23557c478bd9Sstevel@tonic-gate } 23567c478bd9Sstevel@tonic-gate TRANS_INODE(ufsvfsp, ip); 23577c478bd9Sstevel@tonic-gate ip->i_flag |= IMOD; 23587c478bd9Sstevel@tonic-gate } 23597c478bd9Sstevel@tonic-gate 23607c478bd9Sstevel@tonic-gate skip_atime: 23617c478bd9Sstevel@tonic-gate /* 23627c478bd9Sstevel@tonic-gate * The presence of a shadow inode may indicate an ACL, but does 23637c478bd9Sstevel@tonic-gate * not imply an ACL. Future FSD types should be handled here too 23647c478bd9Sstevel@tonic-gate * and check for the presence of the attribute-specific data 23657c478bd9Sstevel@tonic-gate * before referencing it. 23667c478bd9Sstevel@tonic-gate */ 23677c478bd9Sstevel@tonic-gate if (ip->i_shadow) { 23687c478bd9Sstevel@tonic-gate /* 23697c478bd9Sstevel@tonic-gate * XXX if ufs_iupdat is changed to sandbagged write fix 23707c478bd9Sstevel@tonic-gate * ufs_acl_setattr to push ip to keep acls consistent 23717c478bd9Sstevel@tonic-gate * 23727c478bd9Sstevel@tonic-gate * Suppress out of inodes messages if we will retry. 23737c478bd9Sstevel@tonic-gate */ 23747c478bd9Sstevel@tonic-gate if (retry) 23757c478bd9Sstevel@tonic-gate ip->i_flag |= IQUIET; 23767c478bd9Sstevel@tonic-gate error = ufs_acl_setattr(ip, vap, cr); 23777c478bd9Sstevel@tonic-gate ip->i_flag &= ~IQUIET; 23787c478bd9Sstevel@tonic-gate } 23797c478bd9Sstevel@tonic-gate 23807c478bd9Sstevel@tonic-gate update_inode: 23817c478bd9Sstevel@tonic-gate /* 23827c478bd9Sstevel@tonic-gate * Setattr always increases the sequence number 23837c478bd9Sstevel@tonic-gate */ 23847c478bd9Sstevel@tonic-gate ip->i_seq++; 23857c478bd9Sstevel@tonic-gate 23867c478bd9Sstevel@tonic-gate /* 23877c478bd9Sstevel@tonic-gate * if nfsd and not logging; push synchronously 23887c478bd9Sstevel@tonic-gate */ 23897c478bd9Sstevel@tonic-gate if ((curthread->t_flag & T_DONTPEND) && !TRANS_ISTRANS(ufsvfsp)) { 23907c478bd9Sstevel@tonic-gate ufs_iupdat(ip, 1); 23917c478bd9Sstevel@tonic-gate } else { 23927c478bd9Sstevel@tonic-gate ITIMES_NOLOCK(ip); 23937c478bd9Sstevel@tonic-gate } 23947c478bd9Sstevel@tonic-gate 23957c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 23967c478bd9Sstevel@tonic-gate if (dodqlock) { 23977c478bd9Sstevel@tonic-gate rw_exit(&ufsvfsp->vfs_dqrwlock); 23987c478bd9Sstevel@tonic-gate } 23997c478bd9Sstevel@tonic-gate if (dorwlock) 24007c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 24017c478bd9Sstevel@tonic-gate 24027c478bd9Sstevel@tonic-gate if (ulp) { 24037c478bd9Sstevel@tonic-gate if (dotrans) { 24047c478bd9Sstevel@tonic-gate int terr = 0; 24057c478bd9Sstevel@tonic-gate TRANS_END_CSYNC(ufsvfsp, terr, issync, TOP_SETATTR, 24067c478bd9Sstevel@tonic-gate trans_size); 24077c478bd9Sstevel@tonic-gate if (error == 0) 24087c478bd9Sstevel@tonic-gate error = terr; 24097c478bd9Sstevel@tonic-gate } 24107c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 24117c478bd9Sstevel@tonic-gate } 24127c478bd9Sstevel@tonic-gate out: 24137c478bd9Sstevel@tonic-gate /* 24147c478bd9Sstevel@tonic-gate * If out of inodes or blocks, see if we can free something 24157c478bd9Sstevel@tonic-gate * up from the delete queue. 24167c478bd9Sstevel@tonic-gate */ 24177c478bd9Sstevel@tonic-gate if ((error == ENOSPC) && retry && TRANS_ISTRANS(ufsvfsp)) { 24187c478bd9Sstevel@tonic-gate ufs_delete_drain_wait(ufsvfsp, 1); 24197c478bd9Sstevel@tonic-gate retry = 0; 24207c478bd9Sstevel@tonic-gate if (errmsg1 != NULL) 24217c478bd9Sstevel@tonic-gate kmem_free(errmsg1, len1); 24227c478bd9Sstevel@tonic-gate if (errmsg2 != NULL) 24237c478bd9Sstevel@tonic-gate kmem_free(errmsg2, len2); 24247c478bd9Sstevel@tonic-gate goto again; 24257c478bd9Sstevel@tonic-gate } 24267c478bd9Sstevel@tonic-gate if (errmsg1 != NULL) { 24277c478bd9Sstevel@tonic-gate uprintf(errmsg1); 24287c478bd9Sstevel@tonic-gate kmem_free(errmsg1, len1); 24297c478bd9Sstevel@tonic-gate } 24307c478bd9Sstevel@tonic-gate if (errmsg2 != NULL) { 24317c478bd9Sstevel@tonic-gate uprintf(errmsg2); 24327c478bd9Sstevel@tonic-gate kmem_free(errmsg2, len2); 24337c478bd9Sstevel@tonic-gate } 24347c478bd9Sstevel@tonic-gate return (error); 24357c478bd9Sstevel@tonic-gate } 24367c478bd9Sstevel@tonic-gate 24377c478bd9Sstevel@tonic-gate /*ARGSUSED*/ 24387c478bd9Sstevel@tonic-gate static int 2439da6c28aaSamw ufs_access(struct vnode *vp, int mode, int flags, struct cred *cr, 2440da6c28aaSamw caller_context_t *ct) 24417c478bd9Sstevel@tonic-gate { 24427c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 24437c478bd9Sstevel@tonic-gate 24447c478bd9Sstevel@tonic-gate if (ip->i_ufsvfs == NULL) 24457c478bd9Sstevel@tonic-gate return (EIO); 24467c478bd9Sstevel@tonic-gate 24477c478bd9Sstevel@tonic-gate /* 24487c478bd9Sstevel@tonic-gate * The ufs_iaccess function wants to be called with 24497c478bd9Sstevel@tonic-gate * mode bits expressed as "ufs specific" bits. 24507c478bd9Sstevel@tonic-gate * I.e., VWRITE|VREAD|VEXEC do not make sense to 24517c478bd9Sstevel@tonic-gate * ufs_iaccess() but IWRITE|IREAD|IEXEC do. 24527c478bd9Sstevel@tonic-gate * But since they're the same we just pass the vnode mode 24537c478bd9Sstevel@tonic-gate * bit but just verify that assumption at compile time. 24547c478bd9Sstevel@tonic-gate */ 24557c478bd9Sstevel@tonic-gate #if IWRITE != VWRITE || IREAD != VREAD || IEXEC != VEXEC 24567c478bd9Sstevel@tonic-gate #error "ufs_access needs to map Vmodes to Imodes" 24577c478bd9Sstevel@tonic-gate #endif 245860c8e821SFrank Batschulat return (ufs_iaccess(ip, mode, cr, 1)); 24597c478bd9Sstevel@tonic-gate } 24607c478bd9Sstevel@tonic-gate 24617c478bd9Sstevel@tonic-gate /* ARGSUSED */ 24627c478bd9Sstevel@tonic-gate static int 2463da6c28aaSamw ufs_readlink(struct vnode *vp, struct uio *uiop, struct cred *cr, 2464da6c28aaSamw caller_context_t *ct) 24657c478bd9Sstevel@tonic-gate { 24667c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 24677c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp; 24687c478bd9Sstevel@tonic-gate struct ulockfs *ulp; 24697c478bd9Sstevel@tonic-gate int error; 24707c478bd9Sstevel@tonic-gate int fastsymlink; 24717c478bd9Sstevel@tonic-gate 24727c478bd9Sstevel@tonic-gate if (vp->v_type != VLNK) { 24737c478bd9Sstevel@tonic-gate error = EINVAL; 24747c478bd9Sstevel@tonic-gate goto nolockout; 24757c478bd9Sstevel@tonic-gate } 24767c478bd9Sstevel@tonic-gate 24777c478bd9Sstevel@tonic-gate /* 24787c478bd9Sstevel@tonic-gate * If the symbolic link is empty there is nothing to read. 24797c478bd9Sstevel@tonic-gate * Fast-track these empty symbolic links 24807c478bd9Sstevel@tonic-gate */ 24817c478bd9Sstevel@tonic-gate if (ip->i_size == 0) { 24827c478bd9Sstevel@tonic-gate error = 0; 24837c478bd9Sstevel@tonic-gate goto nolockout; 24847c478bd9Sstevel@tonic-gate } 24857c478bd9Sstevel@tonic-gate 24867c478bd9Sstevel@tonic-gate ufsvfsp = ip->i_ufsvfs; 24877c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_READLINK_MASK); 24887c478bd9Sstevel@tonic-gate if (error) 24897c478bd9Sstevel@tonic-gate goto nolockout; 24907c478bd9Sstevel@tonic-gate /* 24917c478bd9Sstevel@tonic-gate * The ip->i_rwlock protects the data blocks used for FASTSYMLINK 24927c478bd9Sstevel@tonic-gate */ 24937c478bd9Sstevel@tonic-gate again: 24947c478bd9Sstevel@tonic-gate fastsymlink = 0; 24957c478bd9Sstevel@tonic-gate if (ip->i_flag & IFASTSYMLNK) { 24967c478bd9Sstevel@tonic-gate rw_enter(&ip->i_rwlock, RW_READER); 24977c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_READER); 24987c478bd9Sstevel@tonic-gate if (ip->i_flag & IFASTSYMLNK) { 24997c478bd9Sstevel@tonic-gate if (!ULOCKFS_IS_NOIACC(ITOUL(ip)) && 25007c478bd9Sstevel@tonic-gate (ip->i_fs->fs_ronly == 0) && 25017c478bd9Sstevel@tonic-gate (!ufsvfsp->vfs_noatime)) { 25027c478bd9Sstevel@tonic-gate mutex_enter(&ip->i_tlock); 25037c478bd9Sstevel@tonic-gate ip->i_flag |= IACC; 25047c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 25057c478bd9Sstevel@tonic-gate } 25067c478bd9Sstevel@tonic-gate error = uiomove((caddr_t)&ip->i_db[1], 25077c478bd9Sstevel@tonic-gate MIN(ip->i_size, uiop->uio_resid), 25087c478bd9Sstevel@tonic-gate UIO_READ, uiop); 25097c478bd9Sstevel@tonic-gate ITIMES(ip); 25107c478bd9Sstevel@tonic-gate ++fastsymlink; 25117c478bd9Sstevel@tonic-gate } 25127c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 25137c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 25147c478bd9Sstevel@tonic-gate } 25157c478bd9Sstevel@tonic-gate if (!fastsymlink) { 25167c478bd9Sstevel@tonic-gate ssize_t size; /* number of bytes read */ 25177c478bd9Sstevel@tonic-gate caddr_t basep; /* pointer to input data */ 25187c478bd9Sstevel@tonic-gate ino_t ino; 25197c478bd9Sstevel@tonic-gate long igen; 25207c478bd9Sstevel@tonic-gate struct uio tuio; /* temp uio struct */ 25217c478bd9Sstevel@tonic-gate struct uio *tuiop; 25227c478bd9Sstevel@tonic-gate iovec_t tiov; /* temp iovec struct */ 25237c478bd9Sstevel@tonic-gate char kbuf[FSL_SIZE]; /* buffer to hold fast symlink */ 25247c478bd9Sstevel@tonic-gate int tflag = 0; /* flag to indicate temp vars used */ 25257c478bd9Sstevel@tonic-gate 25267c478bd9Sstevel@tonic-gate ino = ip->i_number; 25277c478bd9Sstevel@tonic-gate igen = ip->i_gen; 25287c478bd9Sstevel@tonic-gate size = uiop->uio_resid; 25297c478bd9Sstevel@tonic-gate basep = uiop->uio_iov->iov_base; 25307c478bd9Sstevel@tonic-gate tuiop = uiop; 25317c478bd9Sstevel@tonic-gate 25327c478bd9Sstevel@tonic-gate rw_enter(&ip->i_rwlock, RW_WRITER); 25337c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_WRITER); 25347c478bd9Sstevel@tonic-gate if (ip->i_flag & IFASTSYMLNK) { 25357c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 25367c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 25377c478bd9Sstevel@tonic-gate goto again; 25387c478bd9Sstevel@tonic-gate } 25397c478bd9Sstevel@tonic-gate 25407c478bd9Sstevel@tonic-gate /* can this be a fast symlink and is it a user buffer? */ 25417c478bd9Sstevel@tonic-gate if (ip->i_size <= FSL_SIZE && 25427c478bd9Sstevel@tonic-gate (uiop->uio_segflg == UIO_USERSPACE || 25437c478bd9Sstevel@tonic-gate uiop->uio_segflg == UIO_USERISPACE)) { 25447c478bd9Sstevel@tonic-gate 25457c478bd9Sstevel@tonic-gate bzero(&tuio, sizeof (struct uio)); 25467c478bd9Sstevel@tonic-gate /* 25477c478bd9Sstevel@tonic-gate * setup a kernel buffer to read link into. this 25487c478bd9Sstevel@tonic-gate * is to fix a race condition where the user buffer 25497c478bd9Sstevel@tonic-gate * got corrupted before copying it into the inode. 25507c478bd9Sstevel@tonic-gate */ 25517c478bd9Sstevel@tonic-gate size = ip->i_size; 25527c478bd9Sstevel@tonic-gate tiov.iov_len = size; 25537c478bd9Sstevel@tonic-gate tiov.iov_base = kbuf; 25547c478bd9Sstevel@tonic-gate tuio.uio_iov = &tiov; 25557c478bd9Sstevel@tonic-gate tuio.uio_iovcnt = 1; 25567c478bd9Sstevel@tonic-gate tuio.uio_offset = uiop->uio_offset; 25577c478bd9Sstevel@tonic-gate tuio.uio_segflg = UIO_SYSSPACE; 25587c478bd9Sstevel@tonic-gate tuio.uio_fmode = uiop->uio_fmode; 25597c478bd9Sstevel@tonic-gate tuio.uio_extflg = uiop->uio_extflg; 25607c478bd9Sstevel@tonic-gate tuio.uio_limit = uiop->uio_limit; 25617c478bd9Sstevel@tonic-gate tuio.uio_resid = size; 25627c478bd9Sstevel@tonic-gate 25637c478bd9Sstevel@tonic-gate basep = tuio.uio_iov->iov_base; 25647c478bd9Sstevel@tonic-gate tuiop = &tuio; 25657c478bd9Sstevel@tonic-gate tflag = 1; 25667c478bd9Sstevel@tonic-gate } 25677c478bd9Sstevel@tonic-gate 25687c478bd9Sstevel@tonic-gate error = rdip(ip, tuiop, 0, cr); 25697c478bd9Sstevel@tonic-gate if (!(error == 0 && ip->i_number == ino && ip->i_gen == igen)) { 25707c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 25717c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 25727c478bd9Sstevel@tonic-gate goto out; 25737c478bd9Sstevel@tonic-gate } 25747c478bd9Sstevel@tonic-gate 25757c478bd9Sstevel@tonic-gate if (tflag == 0) 25767c478bd9Sstevel@tonic-gate size -= uiop->uio_resid; 25777c478bd9Sstevel@tonic-gate 25787c478bd9Sstevel@tonic-gate if ((tflag == 0 && ip->i_size <= FSL_SIZE && 25797c478bd9Sstevel@tonic-gate ip->i_size == size) || (tflag == 1 && 25807c478bd9Sstevel@tonic-gate tuio.uio_resid == 0)) { 25817c478bd9Sstevel@tonic-gate error = kcopy(basep, &ip->i_db[1], ip->i_size); 25827c478bd9Sstevel@tonic-gate if (error == 0) { 25837c478bd9Sstevel@tonic-gate ip->i_flag |= IFASTSYMLNK; 25847c478bd9Sstevel@tonic-gate /* 25857c478bd9Sstevel@tonic-gate * free page 25867c478bd9Sstevel@tonic-gate */ 25877c478bd9Sstevel@tonic-gate (void) VOP_PUTPAGE(ITOV(ip), 25887c478bd9Sstevel@tonic-gate (offset_t)0, PAGESIZE, 25897c478bd9Sstevel@tonic-gate (B_DONTNEED | B_FREE | B_FORCE | B_ASYNC), 2590da6c28aaSamw cr, ct); 25917c478bd9Sstevel@tonic-gate } else { 25927c478bd9Sstevel@tonic-gate int i; 25937c478bd9Sstevel@tonic-gate /* error, clear garbage left behind */ 25947c478bd9Sstevel@tonic-gate for (i = 1; i < NDADDR; i++) 25957c478bd9Sstevel@tonic-gate ip->i_db[i] = 0; 25967c478bd9Sstevel@tonic-gate for (i = 0; i < NIADDR; i++) 25977c478bd9Sstevel@tonic-gate ip->i_ib[i] = 0; 25987c478bd9Sstevel@tonic-gate } 25997c478bd9Sstevel@tonic-gate } 26007c478bd9Sstevel@tonic-gate if (tflag == 1) { 26017c478bd9Sstevel@tonic-gate /* now, copy it into the user buffer */ 26027c478bd9Sstevel@tonic-gate error = uiomove((caddr_t)kbuf, 26037c478bd9Sstevel@tonic-gate MIN(size, uiop->uio_resid), 26047c478bd9Sstevel@tonic-gate UIO_READ, uiop); 26057c478bd9Sstevel@tonic-gate } 26067c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 26077c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 26087c478bd9Sstevel@tonic-gate } 26097c478bd9Sstevel@tonic-gate out: 26107c478bd9Sstevel@tonic-gate if (ulp) { 26117c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 26127c478bd9Sstevel@tonic-gate } 26137c478bd9Sstevel@tonic-gate nolockout: 26147c478bd9Sstevel@tonic-gate return (error); 26157c478bd9Sstevel@tonic-gate } 26167c478bd9Sstevel@tonic-gate 26177c478bd9Sstevel@tonic-gate /* ARGSUSED */ 26187c478bd9Sstevel@tonic-gate static int 2619da6c28aaSamw ufs_fsync(struct vnode *vp, int syncflag, struct cred *cr, 2620da6c28aaSamw caller_context_t *ct) 26217c478bd9Sstevel@tonic-gate { 26227c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 26237c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp = ip->i_ufsvfs; 26247c478bd9Sstevel@tonic-gate struct ulockfs *ulp; 26257c478bd9Sstevel@tonic-gate int error; 26267c478bd9Sstevel@tonic-gate 26277c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_FSYNC_MASK); 26287c478bd9Sstevel@tonic-gate if (error) 26297c478bd9Sstevel@tonic-gate return (error); 26307c478bd9Sstevel@tonic-gate 26317c478bd9Sstevel@tonic-gate if (TRANS_ISTRANS(ufsvfsp)) { 26327c478bd9Sstevel@tonic-gate /* 26337c478bd9Sstevel@tonic-gate * First push out any data pages 26347c478bd9Sstevel@tonic-gate */ 26357c478bd9Sstevel@tonic-gate if (vn_has_cached_data(vp) && !(syncflag & FNODSYNC) && 26367c478bd9Sstevel@tonic-gate (vp->v_type != VCHR) && !(IS_SWAPVP(vp))) { 26377c478bd9Sstevel@tonic-gate error = VOP_PUTPAGE(vp, (offset_t)0, (size_t)0, 2638da6c28aaSamw 0, CRED(), ct); 26397c478bd9Sstevel@tonic-gate if (error) 26407c478bd9Sstevel@tonic-gate goto out; 26417c478bd9Sstevel@tonic-gate } 26427c478bd9Sstevel@tonic-gate 26437c478bd9Sstevel@tonic-gate /* 26447c478bd9Sstevel@tonic-gate * Delta any delayed inode times updates 26457c478bd9Sstevel@tonic-gate * and push inode to log. 26467c478bd9Sstevel@tonic-gate * All other inode deltas will have already been delta'd 26477c478bd9Sstevel@tonic-gate * and will be pushed during the commit. 26487c478bd9Sstevel@tonic-gate */ 26497c478bd9Sstevel@tonic-gate if (!(syncflag & FDSYNC) && 26507c478bd9Sstevel@tonic-gate ((ip->i_flag & (IMOD|IMODACC)) == IMODACC)) { 26517c478bd9Sstevel@tonic-gate if (ulp) { 26527c478bd9Sstevel@tonic-gate TRANS_BEGIN_ASYNC(ufsvfsp, TOP_FSYNC, 26537c478bd9Sstevel@tonic-gate TOP_SYNCIP_SIZE); 26547c478bd9Sstevel@tonic-gate } 26557c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_READER); 26567c478bd9Sstevel@tonic-gate mutex_enter(&ip->i_tlock); 26577c478bd9Sstevel@tonic-gate ip->i_flag &= ~IMODTIME; 26587c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 26597c478bd9Sstevel@tonic-gate ufs_iupdat(ip, I_SYNC); 26607c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 26617c478bd9Sstevel@tonic-gate if (ulp) { 26627c478bd9Sstevel@tonic-gate TRANS_END_ASYNC(ufsvfsp, TOP_FSYNC, 26637c478bd9Sstevel@tonic-gate TOP_SYNCIP_SIZE); 26647c478bd9Sstevel@tonic-gate } 26657c478bd9Sstevel@tonic-gate } 26667c478bd9Sstevel@tonic-gate 26677c478bd9Sstevel@tonic-gate /* 26687c478bd9Sstevel@tonic-gate * Commit the Moby transaction 26697c478bd9Sstevel@tonic-gate * 26707c478bd9Sstevel@tonic-gate * Deltas have already been made so we just need to 26717c478bd9Sstevel@tonic-gate * commit them with a synchronous transaction. 26727c478bd9Sstevel@tonic-gate * TRANS_BEGIN_SYNC() will return an error 26737c478bd9Sstevel@tonic-gate * if there are no deltas to commit, for an 26747c478bd9Sstevel@tonic-gate * empty transaction. 26757c478bd9Sstevel@tonic-gate */ 26767c478bd9Sstevel@tonic-gate if (ulp) { 26777c478bd9Sstevel@tonic-gate TRANS_BEGIN_SYNC(ufsvfsp, TOP_FSYNC, TOP_COMMIT_SIZE, 26787c478bd9Sstevel@tonic-gate error); 26797c478bd9Sstevel@tonic-gate if (error) { 26807c478bd9Sstevel@tonic-gate error = 0; /* commit wasn't needed */ 26817c478bd9Sstevel@tonic-gate goto out; 26827c478bd9Sstevel@tonic-gate } 26837c478bd9Sstevel@tonic-gate TRANS_END_SYNC(ufsvfsp, error, TOP_FSYNC, 26847c478bd9Sstevel@tonic-gate TOP_COMMIT_SIZE); 26857c478bd9Sstevel@tonic-gate } 26867c478bd9Sstevel@tonic-gate } else { /* not logging */ 26877c478bd9Sstevel@tonic-gate if (!(IS_SWAPVP(vp))) 26887c478bd9Sstevel@tonic-gate if (syncflag & FNODSYNC) { 26897c478bd9Sstevel@tonic-gate /* Just update the inode only */ 26907c478bd9Sstevel@tonic-gate TRANS_IUPDAT(ip, 1); 26917c478bd9Sstevel@tonic-gate error = 0; 26927c478bd9Sstevel@tonic-gate } else if (syncflag & FDSYNC) 26937c478bd9Sstevel@tonic-gate /* Do data-synchronous writes */ 26947c478bd9Sstevel@tonic-gate error = TRANS_SYNCIP(ip, 0, I_DSYNC, TOP_FSYNC); 26957c478bd9Sstevel@tonic-gate else 26967c478bd9Sstevel@tonic-gate /* Do synchronous writes */ 26977c478bd9Sstevel@tonic-gate error = TRANS_SYNCIP(ip, 0, I_SYNC, TOP_FSYNC); 26987c478bd9Sstevel@tonic-gate 26997c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_WRITER); 27007c478bd9Sstevel@tonic-gate if (!error) 27017c478bd9Sstevel@tonic-gate error = ufs_sync_indir(ip); 27027c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 27037c478bd9Sstevel@tonic-gate } 27047c478bd9Sstevel@tonic-gate out: 27057c478bd9Sstevel@tonic-gate if (ulp) { 27067c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 27077c478bd9Sstevel@tonic-gate } 27087c478bd9Sstevel@tonic-gate return (error); 27097c478bd9Sstevel@tonic-gate } 27107c478bd9Sstevel@tonic-gate 27117c478bd9Sstevel@tonic-gate /*ARGSUSED*/ 27127c478bd9Sstevel@tonic-gate static void 2713da6c28aaSamw ufs_inactive(struct vnode *vp, struct cred *cr, caller_context_t *ct) 27147c478bd9Sstevel@tonic-gate { 27157c478bd9Sstevel@tonic-gate ufs_iinactive(VTOI(vp)); 27167c478bd9Sstevel@tonic-gate } 27177c478bd9Sstevel@tonic-gate 27187c478bd9Sstevel@tonic-gate /* 27197c478bd9Sstevel@tonic-gate * Unix file system operations having to do with directory manipulation. 27207c478bd9Sstevel@tonic-gate */ 27217c478bd9Sstevel@tonic-gate int ufs_lookup_idle_count = 2; /* Number of inodes to idle each time */ 27227c478bd9Sstevel@tonic-gate /* ARGSUSED */ 27237c478bd9Sstevel@tonic-gate static int 27247c478bd9Sstevel@tonic-gate ufs_lookup(struct vnode *dvp, char *nm, struct vnode **vpp, 2725da6c28aaSamw struct pathname *pnp, int flags, struct vnode *rdir, struct cred *cr, 2726da6c28aaSamw caller_context_t *ct, int *direntflags, pathname_t *realpnp) 27277c478bd9Sstevel@tonic-gate { 27287c478bd9Sstevel@tonic-gate struct inode *ip; 27297c478bd9Sstevel@tonic-gate struct inode *sip; 27307c478bd9Sstevel@tonic-gate struct inode *xip; 27317c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp; 27327c478bd9Sstevel@tonic-gate struct ulockfs *ulp; 27337c478bd9Sstevel@tonic-gate struct vnode *vp; 27347c478bd9Sstevel@tonic-gate int error; 27357c478bd9Sstevel@tonic-gate 27367c478bd9Sstevel@tonic-gate /* 27377c478bd9Sstevel@tonic-gate * Check flags for type of lookup (regular file or attribute file) 27387c478bd9Sstevel@tonic-gate */ 27397c478bd9Sstevel@tonic-gate 27407c478bd9Sstevel@tonic-gate ip = VTOI(dvp); 27417c478bd9Sstevel@tonic-gate 27427c478bd9Sstevel@tonic-gate if (flags & LOOKUP_XATTR) { 27437c478bd9Sstevel@tonic-gate 27447c478bd9Sstevel@tonic-gate /* 2745da6c28aaSamw * If not mounted with XATTR support then return EINVAL 2746da6c28aaSamw */ 2747da6c28aaSamw 2748da6c28aaSamw if (!(ip->i_ufsvfs->vfs_vfs->vfs_flag & VFS_XATTR)) 2749da6c28aaSamw return (EINVAL); 2750da6c28aaSamw /* 27517c478bd9Sstevel@tonic-gate * We don't allow recursive attributes... 27527c478bd9Sstevel@tonic-gate * Maybe someday we will. 27537c478bd9Sstevel@tonic-gate */ 27547c478bd9Sstevel@tonic-gate if ((ip->i_cflags & IXATTR)) { 27557c478bd9Sstevel@tonic-gate return (EINVAL); 27567c478bd9Sstevel@tonic-gate } 27577c478bd9Sstevel@tonic-gate 27587c478bd9Sstevel@tonic-gate if ((vp = dnlc_lookup(dvp, XATTR_DIR_NAME)) == NULL) { 27597c478bd9Sstevel@tonic-gate error = ufs_xattr_getattrdir(dvp, &sip, flags, cr); 27607c478bd9Sstevel@tonic-gate if (error) { 27617c478bd9Sstevel@tonic-gate *vpp = NULL; 27627c478bd9Sstevel@tonic-gate goto out; 27637c478bd9Sstevel@tonic-gate } 27647c478bd9Sstevel@tonic-gate 27657c478bd9Sstevel@tonic-gate vp = ITOV(sip); 27667c478bd9Sstevel@tonic-gate dnlc_update(dvp, XATTR_DIR_NAME, vp); 27677c478bd9Sstevel@tonic-gate } 27687c478bd9Sstevel@tonic-gate 27697c478bd9Sstevel@tonic-gate /* 27707c478bd9Sstevel@tonic-gate * Check accessibility of directory. 27717c478bd9Sstevel@tonic-gate */ 27727c478bd9Sstevel@tonic-gate if (vp == DNLC_NO_VNODE) { 27737c478bd9Sstevel@tonic-gate VN_RELE(vp); 27747c478bd9Sstevel@tonic-gate error = ENOENT; 27757c478bd9Sstevel@tonic-gate goto out; 27767c478bd9Sstevel@tonic-gate } 277760c8e821SFrank Batschulat if ((error = ufs_iaccess(VTOI(vp), IEXEC, cr, 1)) != 0) { 27787c478bd9Sstevel@tonic-gate VN_RELE(vp); 27797c478bd9Sstevel@tonic-gate goto out; 27807c478bd9Sstevel@tonic-gate } 27817c478bd9Sstevel@tonic-gate 27827c478bd9Sstevel@tonic-gate *vpp = vp; 27837c478bd9Sstevel@tonic-gate return (0); 27847c478bd9Sstevel@tonic-gate } 27857c478bd9Sstevel@tonic-gate 27867c478bd9Sstevel@tonic-gate /* 27877c478bd9Sstevel@tonic-gate * Check for a null component, which we should treat as 27887c478bd9Sstevel@tonic-gate * looking at dvp from within it's parent, so we don't 27897c478bd9Sstevel@tonic-gate * need a call to ufs_iaccess(), as it has already been 27907c478bd9Sstevel@tonic-gate * done. 27917c478bd9Sstevel@tonic-gate */ 27927c478bd9Sstevel@tonic-gate if (nm[0] == 0) { 27937c478bd9Sstevel@tonic-gate VN_HOLD(dvp); 27947c478bd9Sstevel@tonic-gate error = 0; 27957c478bd9Sstevel@tonic-gate *vpp = dvp; 27967c478bd9Sstevel@tonic-gate goto out; 27977c478bd9Sstevel@tonic-gate } 27987c478bd9Sstevel@tonic-gate 27997c478bd9Sstevel@tonic-gate /* 28007c478bd9Sstevel@tonic-gate * Check for "." ie itself. this is a quick check and 28017c478bd9Sstevel@tonic-gate * avoids adding "." into the dnlc (which have been seen 28027c478bd9Sstevel@tonic-gate * to occupy >10% of the cache). 28037c478bd9Sstevel@tonic-gate */ 28047c478bd9Sstevel@tonic-gate if ((nm[0] == '.') && (nm[1] == 0)) { 28057c478bd9Sstevel@tonic-gate /* 28067c478bd9Sstevel@tonic-gate * Don't return without checking accessibility 28077c478bd9Sstevel@tonic-gate * of the directory. We only need the lock if 28087c478bd9Sstevel@tonic-gate * we are going to return it. 28097c478bd9Sstevel@tonic-gate */ 281060c8e821SFrank Batschulat if ((error = ufs_iaccess(ip, IEXEC, cr, 1)) == 0) { 28117c478bd9Sstevel@tonic-gate VN_HOLD(dvp); 28127c478bd9Sstevel@tonic-gate *vpp = dvp; 28137c478bd9Sstevel@tonic-gate } 28147c478bd9Sstevel@tonic-gate goto out; 28157c478bd9Sstevel@tonic-gate } 28167c478bd9Sstevel@tonic-gate 28177c478bd9Sstevel@tonic-gate /* 28187c478bd9Sstevel@tonic-gate * Fast path: Check the directory name lookup cache. 28197c478bd9Sstevel@tonic-gate */ 28207c478bd9Sstevel@tonic-gate if (vp = dnlc_lookup(dvp, nm)) { 28217c478bd9Sstevel@tonic-gate /* 28227c478bd9Sstevel@tonic-gate * Check accessibility of directory. 28237c478bd9Sstevel@tonic-gate */ 282460c8e821SFrank Batschulat if ((error = ufs_iaccess(ip, IEXEC, cr, 1)) != 0) { 28257c478bd9Sstevel@tonic-gate VN_RELE(vp); 28267c478bd9Sstevel@tonic-gate goto out; 28277c478bd9Sstevel@tonic-gate } 28287c478bd9Sstevel@tonic-gate if (vp == DNLC_NO_VNODE) { 28297c478bd9Sstevel@tonic-gate VN_RELE(vp); 28307c478bd9Sstevel@tonic-gate error = ENOENT; 28317c478bd9Sstevel@tonic-gate goto out; 28327c478bd9Sstevel@tonic-gate } 28337c478bd9Sstevel@tonic-gate xip = VTOI(vp); 28347c478bd9Sstevel@tonic-gate ulp = NULL; 28357c478bd9Sstevel@tonic-gate goto fastpath; 28367c478bd9Sstevel@tonic-gate } 28377c478bd9Sstevel@tonic-gate 28387c478bd9Sstevel@tonic-gate /* 28397c478bd9Sstevel@tonic-gate * Keep the idle queue from getting too long by 28407c478bd9Sstevel@tonic-gate * idling two inodes before attempting to allocate another. 28417c478bd9Sstevel@tonic-gate * This operation must be performed before entering 28427c478bd9Sstevel@tonic-gate * lockfs or a transaction. 28437c478bd9Sstevel@tonic-gate */ 28447c478bd9Sstevel@tonic-gate if (ufs_idle_q.uq_ne > ufs_idle_q.uq_hiwat) 28457c478bd9Sstevel@tonic-gate if ((curthread->t_flag & T_DONTBLOCK) == 0) { 28467c478bd9Sstevel@tonic-gate ins.in_lidles.value.ul += ufs_lookup_idle_count; 28477c478bd9Sstevel@tonic-gate ufs_idle_some(ufs_lookup_idle_count); 28487c478bd9Sstevel@tonic-gate } 28497c478bd9Sstevel@tonic-gate 285002ff05a9Svsakar retry_lookup: 285160c8e821SFrank Batschulat /* 285260c8e821SFrank Batschulat * Check accessibility of directory. 285360c8e821SFrank Batschulat */ 285460c8e821SFrank Batschulat if (error = ufs_diraccess(ip, IEXEC, cr)) 285560c8e821SFrank Batschulat goto out; 285660c8e821SFrank Batschulat 28577c478bd9Sstevel@tonic-gate ufsvfsp = ip->i_ufsvfs; 28587c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_LOOKUP_MASK); 28597c478bd9Sstevel@tonic-gate if (error) 28607c478bd9Sstevel@tonic-gate goto out; 28617c478bd9Sstevel@tonic-gate 28629b5097eeSOwen Roberts error = ufs_dirlook(ip, nm, &xip, cr, 1, 0); 28637c478bd9Sstevel@tonic-gate 28647c478bd9Sstevel@tonic-gate fastpath: 28657c478bd9Sstevel@tonic-gate if (error == 0) { 28667c478bd9Sstevel@tonic-gate ip = xip; 28677c478bd9Sstevel@tonic-gate *vpp = ITOV(ip); 28687c478bd9Sstevel@tonic-gate 28697c478bd9Sstevel@tonic-gate /* 28707c478bd9Sstevel@tonic-gate * If vnode is a device return special vnode instead. 28717c478bd9Sstevel@tonic-gate */ 28727c478bd9Sstevel@tonic-gate if (IS_DEVVP(*vpp)) { 28737c478bd9Sstevel@tonic-gate struct vnode *newvp; 28747c478bd9Sstevel@tonic-gate 28757c478bd9Sstevel@tonic-gate newvp = specvp(*vpp, (*vpp)->v_rdev, (*vpp)->v_type, 28767c478bd9Sstevel@tonic-gate cr); 28777c478bd9Sstevel@tonic-gate VN_RELE(*vpp); 28787c478bd9Sstevel@tonic-gate if (newvp == NULL) 28797c478bd9Sstevel@tonic-gate error = ENOSYS; 28807c478bd9Sstevel@tonic-gate else 28817c478bd9Sstevel@tonic-gate *vpp = newvp; 2882986fd29aSsetje } else if (ip->i_cflags & ICOMPRESS) { 2883986fd29aSsetje struct vnode *newvp; 2884986fd29aSsetje 2885986fd29aSsetje /* 2886986fd29aSsetje * Compressed file, substitute dcfs vnode 2887986fd29aSsetje */ 2888986fd29aSsetje newvp = decompvp(*vpp, cr, ct); 2889986fd29aSsetje VN_RELE(*vpp); 2890986fd29aSsetje if (newvp == NULL) 2891986fd29aSsetje error = ENOSYS; 2892986fd29aSsetje else 2893986fd29aSsetje *vpp = newvp; 28947c478bd9Sstevel@tonic-gate } 28957c478bd9Sstevel@tonic-gate } 28967c478bd9Sstevel@tonic-gate if (ulp) { 28977c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 28987c478bd9Sstevel@tonic-gate } 28997c478bd9Sstevel@tonic-gate 290002ff05a9Svsakar if (error == EAGAIN) 290102ff05a9Svsakar goto retry_lookup; 290202ff05a9Svsakar 29037c478bd9Sstevel@tonic-gate out: 29047c478bd9Sstevel@tonic-gate return (error); 29057c478bd9Sstevel@tonic-gate } 29067c478bd9Sstevel@tonic-gate 2907da6c28aaSamw /*ARGSUSED*/ 29087c478bd9Sstevel@tonic-gate static int 29097c478bd9Sstevel@tonic-gate ufs_create(struct vnode *dvp, char *name, struct vattr *vap, enum vcexcl excl, 2910da6c28aaSamw int mode, struct vnode **vpp, struct cred *cr, int flag, 2911da6c28aaSamw caller_context_t *ct, vsecattr_t *vsecp) 29127c478bd9Sstevel@tonic-gate { 29137c478bd9Sstevel@tonic-gate struct inode *ip; 29147c478bd9Sstevel@tonic-gate struct inode *xip; 29157c478bd9Sstevel@tonic-gate struct inode *dip; 29167c478bd9Sstevel@tonic-gate struct vnode *xvp; 29177c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp; 29187c478bd9Sstevel@tonic-gate struct ulockfs *ulp; 29197c478bd9Sstevel@tonic-gate int error; 29207c478bd9Sstevel@tonic-gate int issync; 29217c478bd9Sstevel@tonic-gate int truncflag; 29227c478bd9Sstevel@tonic-gate int trans_size; 29237c478bd9Sstevel@tonic-gate int noentry; 29247c478bd9Sstevel@tonic-gate int defer_dip_seq_update = 0; /* need to defer update of dip->i_seq */ 29257c478bd9Sstevel@tonic-gate int retry = 1; 292602ff05a9Svsakar int indeadlock; 29277c478bd9Sstevel@tonic-gate 29287c478bd9Sstevel@tonic-gate again: 29297c478bd9Sstevel@tonic-gate ip = VTOI(dvp); 29307c478bd9Sstevel@tonic-gate ufsvfsp = ip->i_ufsvfs; 29317c478bd9Sstevel@tonic-gate truncflag = 0; 29327c478bd9Sstevel@tonic-gate 29337c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_CREATE_MASK); 29347c478bd9Sstevel@tonic-gate if (error) 29357c478bd9Sstevel@tonic-gate goto out; 29367c478bd9Sstevel@tonic-gate 29377c478bd9Sstevel@tonic-gate if (ulp) { 29387c478bd9Sstevel@tonic-gate trans_size = (int)TOP_CREATE_SIZE(ip); 29397c478bd9Sstevel@tonic-gate TRANS_BEGIN_CSYNC(ufsvfsp, issync, TOP_CREATE, trans_size); 29407c478bd9Sstevel@tonic-gate } 29417c478bd9Sstevel@tonic-gate 29427c478bd9Sstevel@tonic-gate if ((vap->va_mode & VSVTX) && secpolicy_vnode_stky_modify(cr) != 0) 29437c478bd9Sstevel@tonic-gate vap->va_mode &= ~VSVTX; 29447c478bd9Sstevel@tonic-gate 29457c478bd9Sstevel@tonic-gate if (*name == '\0') { 29467c478bd9Sstevel@tonic-gate /* 29477c478bd9Sstevel@tonic-gate * Null component name refers to the directory itself. 29487c478bd9Sstevel@tonic-gate */ 29497c478bd9Sstevel@tonic-gate VN_HOLD(dvp); 29507c478bd9Sstevel@tonic-gate /* 29517c478bd9Sstevel@tonic-gate * Even though this is an error case, we need to grab the 29527c478bd9Sstevel@tonic-gate * quota lock since the error handling code below is common. 29537c478bd9Sstevel@tonic-gate */ 29547c478bd9Sstevel@tonic-gate rw_enter(&ufsvfsp->vfs_dqrwlock, RW_READER); 29557c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_WRITER); 29567c478bd9Sstevel@tonic-gate error = EEXIST; 29577c478bd9Sstevel@tonic-gate } else { 29587c478bd9Sstevel@tonic-gate xip = NULL; 29597c478bd9Sstevel@tonic-gate noentry = 0; 296002ff05a9Svsakar /* 296102ff05a9Svsakar * ufs_tryirwlock_trans uses rw_tryenter and checks for SLOCK 296202ff05a9Svsakar * to avoid i_rwlock, ufs_lockfs_begin deadlock. If deadlock 296302ff05a9Svsakar * possible, retries the operation. 296402ff05a9Svsakar */ 296502ff05a9Svsakar ufs_tryirwlock_trans(&ip->i_rwlock, RW_WRITER, TOP_CREATE, 296602ff05a9Svsakar retry_dir); 296702ff05a9Svsakar if (indeadlock) 296802ff05a9Svsakar goto again; 296902ff05a9Svsakar 29707c478bd9Sstevel@tonic-gate xvp = dnlc_lookup(dvp, name); 29717c478bd9Sstevel@tonic-gate if (xvp == DNLC_NO_VNODE) { 29727c478bd9Sstevel@tonic-gate noentry = 1; 29737c478bd9Sstevel@tonic-gate VN_RELE(xvp); 29747c478bd9Sstevel@tonic-gate xvp = NULL; 29757c478bd9Sstevel@tonic-gate } 29767c478bd9Sstevel@tonic-gate if (xvp) { 29777c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 297860c8e821SFrank Batschulat if (error = ufs_iaccess(ip, IEXEC, cr, 1)) { 29797c478bd9Sstevel@tonic-gate VN_RELE(xvp); 29807c478bd9Sstevel@tonic-gate } else { 29817c478bd9Sstevel@tonic-gate error = EEXIST; 29827c478bd9Sstevel@tonic-gate xip = VTOI(xvp); 29837c478bd9Sstevel@tonic-gate } 29847c478bd9Sstevel@tonic-gate } else { 29857c478bd9Sstevel@tonic-gate /* 29867c478bd9Sstevel@tonic-gate * Suppress file system full message if we will retry 29877c478bd9Sstevel@tonic-gate */ 29887c478bd9Sstevel@tonic-gate error = ufs_direnter_cm(ip, name, DE_CREATE, 298980d34432Sfrankho vap, &xip, cr, (noentry | (retry ? IQUIET : 0))); 299002ff05a9Svsakar if (error == EAGAIN) { 299102ff05a9Svsakar if (ulp) { 299202ff05a9Svsakar TRANS_END_CSYNC(ufsvfsp, error, issync, 299302ff05a9Svsakar TOP_CREATE, trans_size); 299402ff05a9Svsakar ufs_lockfs_end(ulp); 299502ff05a9Svsakar } 299602ff05a9Svsakar goto again; 299702ff05a9Svsakar } 29987c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 29997c478bd9Sstevel@tonic-gate } 30007c478bd9Sstevel@tonic-gate ip = xip; 30017c478bd9Sstevel@tonic-gate if (ip != NULL) { 30027c478bd9Sstevel@tonic-gate rw_enter(&ufsvfsp->vfs_dqrwlock, RW_READER); 30037c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_WRITER); 30047c478bd9Sstevel@tonic-gate } 30057c478bd9Sstevel@tonic-gate } 30067c478bd9Sstevel@tonic-gate 30077c478bd9Sstevel@tonic-gate /* 30087c478bd9Sstevel@tonic-gate * If the file already exists and this is a non-exclusive create, 30097c478bd9Sstevel@tonic-gate * check permissions and allow access for non-directories. 30107c478bd9Sstevel@tonic-gate * Read-only create of an existing directory is also allowed. 30117c478bd9Sstevel@tonic-gate * We fail an exclusive create of anything which already exists. 30127c478bd9Sstevel@tonic-gate */ 30137c478bd9Sstevel@tonic-gate if (error == EEXIST) { 30147c478bd9Sstevel@tonic-gate dip = VTOI(dvp); 30157c478bd9Sstevel@tonic-gate if (excl == NONEXCL) { 30167c478bd9Sstevel@tonic-gate if ((((ip->i_mode & IFMT) == IFDIR) || 30177c478bd9Sstevel@tonic-gate ((ip->i_mode & IFMT) == IFATTRDIR)) && 30187c478bd9Sstevel@tonic-gate (mode & IWRITE)) 30197c478bd9Sstevel@tonic-gate error = EISDIR; 30207c478bd9Sstevel@tonic-gate else if (mode) 302160c8e821SFrank Batschulat error = ufs_iaccess(ip, mode, cr, 0); 30227c478bd9Sstevel@tonic-gate else 30237c478bd9Sstevel@tonic-gate error = 0; 30247c478bd9Sstevel@tonic-gate } 30257c478bd9Sstevel@tonic-gate if (error) { 30267c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 30277c478bd9Sstevel@tonic-gate rw_exit(&ufsvfsp->vfs_dqrwlock); 30287c478bd9Sstevel@tonic-gate VN_RELE(ITOV(ip)); 30297c478bd9Sstevel@tonic-gate goto unlock; 30307c478bd9Sstevel@tonic-gate } 30317c478bd9Sstevel@tonic-gate /* 30327c478bd9Sstevel@tonic-gate * If the error EEXIST was set, then i_seq can not 30337c478bd9Sstevel@tonic-gate * have been updated. The sequence number interface 30347c478bd9Sstevel@tonic-gate * is defined such that a non-error VOP_CREATE must 30357c478bd9Sstevel@tonic-gate * increase the dir va_seq it by at least one. If we 30367c478bd9Sstevel@tonic-gate * have cleared the error, increase i_seq. Note that 30377c478bd9Sstevel@tonic-gate * we are increasing the dir i_seq and in rare cases 30387c478bd9Sstevel@tonic-gate * ip may actually be from the dvp, so we already have 30397c478bd9Sstevel@tonic-gate * the locks and it will not be subject to truncation. 30407c478bd9Sstevel@tonic-gate * In case we have to update i_seq of the parent 30417c478bd9Sstevel@tonic-gate * directory dip, we have to defer it till we have 30427c478bd9Sstevel@tonic-gate * released our locks on ip due to lock ordering requirements. 30437c478bd9Sstevel@tonic-gate */ 30447c478bd9Sstevel@tonic-gate if (ip != dip) 30457c478bd9Sstevel@tonic-gate defer_dip_seq_update = 1; 30467c478bd9Sstevel@tonic-gate else 30477c478bd9Sstevel@tonic-gate ip->i_seq++; 30487c478bd9Sstevel@tonic-gate 30497c478bd9Sstevel@tonic-gate if (((ip->i_mode & IFMT) == IFREG) && 30507c478bd9Sstevel@tonic-gate (vap->va_mask & AT_SIZE) && vap->va_size == 0) { 30517c478bd9Sstevel@tonic-gate /* 30527c478bd9Sstevel@tonic-gate * Truncate regular files, if requested by caller. 30537c478bd9Sstevel@tonic-gate * Grab i_rwlock to make sure no one else is 30547c478bd9Sstevel@tonic-gate * currently writing to the file (we promised 30557c478bd9Sstevel@tonic-gate * bmap we would do this). 30567c478bd9Sstevel@tonic-gate * Must get the locks in the correct order. 30577c478bd9Sstevel@tonic-gate */ 30587c478bd9Sstevel@tonic-gate if (ip->i_size == 0) { 30597c478bd9Sstevel@tonic-gate ip->i_flag |= ICHG | IUPD; 30607c478bd9Sstevel@tonic-gate ip->i_seq++; 30617c478bd9Sstevel@tonic-gate TRANS_INODE(ufsvfsp, ip); 30627c478bd9Sstevel@tonic-gate } else { 30637c478bd9Sstevel@tonic-gate /* 30647c478bd9Sstevel@tonic-gate * Large Files: Why this check here? 30657c478bd9Sstevel@tonic-gate * Though we do it in vn_create() we really 30667c478bd9Sstevel@tonic-gate * want to guarantee that we do not destroy 30677c478bd9Sstevel@tonic-gate * Large file data by atomically checking 30687c478bd9Sstevel@tonic-gate * the size while holding the contents 30697c478bd9Sstevel@tonic-gate * lock. 30707c478bd9Sstevel@tonic-gate */ 30717c478bd9Sstevel@tonic-gate if (flag && !(flag & FOFFMAX) && 30727c478bd9Sstevel@tonic-gate ((ip->i_mode & IFMT) == IFREG) && 30737c478bd9Sstevel@tonic-gate (ip->i_size > (offset_t)MAXOFF32_T)) { 30747c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 30757c478bd9Sstevel@tonic-gate rw_exit(&ufsvfsp->vfs_dqrwlock); 30767c478bd9Sstevel@tonic-gate error = EOVERFLOW; 30777c478bd9Sstevel@tonic-gate goto unlock; 30787c478bd9Sstevel@tonic-gate } 30797c478bd9Sstevel@tonic-gate if (TRANS_ISTRANS(ufsvfsp)) 30807c478bd9Sstevel@tonic-gate truncflag++; 30817c478bd9Sstevel@tonic-gate else { 30827c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 30837c478bd9Sstevel@tonic-gate rw_exit(&ufsvfsp->vfs_dqrwlock); 308402ff05a9Svsakar ufs_tryirwlock_trans(&ip->i_rwlock, 308502ff05a9Svsakar RW_WRITER, TOP_CREATE, 308602ff05a9Svsakar retry_file); 308702ff05a9Svsakar if (indeadlock) { 308802ff05a9Svsakar VN_RELE(ITOV(ip)); 308902ff05a9Svsakar goto again; 309002ff05a9Svsakar } 30917c478bd9Sstevel@tonic-gate rw_enter(&ufsvfsp->vfs_dqrwlock, 30927c478bd9Sstevel@tonic-gate RW_READER); 30937c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_WRITER); 30947c478bd9Sstevel@tonic-gate (void) ufs_itrunc(ip, (u_offset_t)0, 0, 30957c478bd9Sstevel@tonic-gate cr); 30967c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 30977c478bd9Sstevel@tonic-gate } 3098df2381bfSpraks 3099df2381bfSpraks } 3100df2381bfSpraks if (error == 0) { 3101da6c28aaSamw vnevent_create(ITOV(ip), ct); 31027c478bd9Sstevel@tonic-gate } 31037c478bd9Sstevel@tonic-gate } 31047c478bd9Sstevel@tonic-gate } 31057c478bd9Sstevel@tonic-gate 31067c478bd9Sstevel@tonic-gate if (error) { 31077c478bd9Sstevel@tonic-gate if (ip != NULL) { 31087c478bd9Sstevel@tonic-gate rw_exit(&ufsvfsp->vfs_dqrwlock); 31097c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 31107c478bd9Sstevel@tonic-gate } 31117c478bd9Sstevel@tonic-gate goto unlock; 31127c478bd9Sstevel@tonic-gate } 31137c478bd9Sstevel@tonic-gate 31147c478bd9Sstevel@tonic-gate *vpp = ITOV(ip); 31157c478bd9Sstevel@tonic-gate ITIMES(ip); 31167c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 31177c478bd9Sstevel@tonic-gate rw_exit(&ufsvfsp->vfs_dqrwlock); 31187c478bd9Sstevel@tonic-gate 31197c478bd9Sstevel@tonic-gate /* 31207c478bd9Sstevel@tonic-gate * If vnode is a device return special vnode instead. 31217c478bd9Sstevel@tonic-gate */ 31227c478bd9Sstevel@tonic-gate if (!error && IS_DEVVP(*vpp)) { 31237c478bd9Sstevel@tonic-gate struct vnode *newvp; 31247c478bd9Sstevel@tonic-gate 31257c478bd9Sstevel@tonic-gate newvp = specvp(*vpp, (*vpp)->v_rdev, (*vpp)->v_type, cr); 31267c478bd9Sstevel@tonic-gate VN_RELE(*vpp); 31277c478bd9Sstevel@tonic-gate if (newvp == NULL) { 31287c478bd9Sstevel@tonic-gate error = ENOSYS; 31297c478bd9Sstevel@tonic-gate goto unlock; 31307c478bd9Sstevel@tonic-gate } 31317c478bd9Sstevel@tonic-gate truncflag = 0; 31327c478bd9Sstevel@tonic-gate *vpp = newvp; 31337c478bd9Sstevel@tonic-gate } 31347c478bd9Sstevel@tonic-gate unlock: 31357c478bd9Sstevel@tonic-gate 31367c478bd9Sstevel@tonic-gate /* 31377c478bd9Sstevel@tonic-gate * Do the deferred update of the parent directory's sequence 31387c478bd9Sstevel@tonic-gate * number now. 31397c478bd9Sstevel@tonic-gate */ 31407c478bd9Sstevel@tonic-gate if (defer_dip_seq_update == 1) { 31417c478bd9Sstevel@tonic-gate rw_enter(&dip->i_contents, RW_READER); 31427c478bd9Sstevel@tonic-gate mutex_enter(&dip->i_tlock); 31437c478bd9Sstevel@tonic-gate dip->i_seq++; 31447c478bd9Sstevel@tonic-gate mutex_exit(&dip->i_tlock); 31457c478bd9Sstevel@tonic-gate rw_exit(&dip->i_contents); 31467c478bd9Sstevel@tonic-gate } 31477c478bd9Sstevel@tonic-gate 31487c478bd9Sstevel@tonic-gate if (ulp) { 31497c478bd9Sstevel@tonic-gate int terr = 0; 31507c478bd9Sstevel@tonic-gate 31517c478bd9Sstevel@tonic-gate TRANS_END_CSYNC(ufsvfsp, terr, issync, TOP_CREATE, 31527c478bd9Sstevel@tonic-gate trans_size); 31537c478bd9Sstevel@tonic-gate 31547c478bd9Sstevel@tonic-gate /* 31557c478bd9Sstevel@tonic-gate * If we haven't had a more interesting failure 31567c478bd9Sstevel@tonic-gate * already, then anything that might've happened 31577c478bd9Sstevel@tonic-gate * here should be reported. 31587c478bd9Sstevel@tonic-gate */ 31597c478bd9Sstevel@tonic-gate if (error == 0) 31607c478bd9Sstevel@tonic-gate error = terr; 31617c478bd9Sstevel@tonic-gate } 31627c478bd9Sstevel@tonic-gate 31637c478bd9Sstevel@tonic-gate if (!error && truncflag) { 316402ff05a9Svsakar ufs_tryirwlock(&ip->i_rwlock, RW_WRITER, retry_trunc); 316502ff05a9Svsakar if (indeadlock) { 316602ff05a9Svsakar if (ulp) 316702ff05a9Svsakar ufs_lockfs_end(ulp); 316802ff05a9Svsakar VN_RELE(ITOV(ip)); 316902ff05a9Svsakar goto again; 317002ff05a9Svsakar } 31717c478bd9Sstevel@tonic-gate (void) TRANS_ITRUNC(ip, (u_offset_t)0, 0, cr); 31727c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 31737c478bd9Sstevel@tonic-gate } 31747c478bd9Sstevel@tonic-gate 31757c478bd9Sstevel@tonic-gate if (ulp) 31767c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 31777c478bd9Sstevel@tonic-gate 31787c478bd9Sstevel@tonic-gate /* 31797c478bd9Sstevel@tonic-gate * If no inodes available, try to free one up out of the 31807c478bd9Sstevel@tonic-gate * pending delete queue. 31817c478bd9Sstevel@tonic-gate */ 31827c478bd9Sstevel@tonic-gate if ((error == ENOSPC) && retry && TRANS_ISTRANS(ufsvfsp)) { 31837c478bd9Sstevel@tonic-gate ufs_delete_drain_wait(ufsvfsp, 1); 31847c478bd9Sstevel@tonic-gate retry = 0; 31857c478bd9Sstevel@tonic-gate goto again; 31867c478bd9Sstevel@tonic-gate } 31877c478bd9Sstevel@tonic-gate 31887c478bd9Sstevel@tonic-gate out: 31897c478bd9Sstevel@tonic-gate return (error); 31907c478bd9Sstevel@tonic-gate } 31917c478bd9Sstevel@tonic-gate 31927c478bd9Sstevel@tonic-gate extern int ufs_idle_max; 31937c478bd9Sstevel@tonic-gate /*ARGSUSED*/ 31947c478bd9Sstevel@tonic-gate static int 3195da6c28aaSamw ufs_remove(struct vnode *vp, char *nm, struct cred *cr, 3196da6c28aaSamw caller_context_t *ct, int flags) 31977c478bd9Sstevel@tonic-gate { 31987c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 31997c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp = ip->i_ufsvfs; 32007c478bd9Sstevel@tonic-gate struct ulockfs *ulp; 32017c478bd9Sstevel@tonic-gate vnode_t *rmvp = NULL; /* Vnode corresponding to name being removed */ 320202ff05a9Svsakar int indeadlock; 32037c478bd9Sstevel@tonic-gate int error; 32047c478bd9Sstevel@tonic-gate int issync; 32057c478bd9Sstevel@tonic-gate int trans_size; 32067c478bd9Sstevel@tonic-gate 32077c478bd9Sstevel@tonic-gate /* 32087c478bd9Sstevel@tonic-gate * don't let the delete queue get too long 32097c478bd9Sstevel@tonic-gate */ 32107c478bd9Sstevel@tonic-gate if (ufsvfsp == NULL) { 32117c478bd9Sstevel@tonic-gate error = EIO; 32127c478bd9Sstevel@tonic-gate goto out; 32137c478bd9Sstevel@tonic-gate } 32147c478bd9Sstevel@tonic-gate if (ufsvfsp->vfs_delete.uq_ne > ufs_idle_max) 32157c478bd9Sstevel@tonic-gate ufs_delete_drain(vp->v_vfsp, 1, 1); 32167c478bd9Sstevel@tonic-gate 32179b5097eeSOwen Roberts error = ufs_eventlookup(vp, nm, cr, &rmvp); 32189b5097eeSOwen Roberts if (rmvp != NULL) { 32199b5097eeSOwen Roberts /* Only send the event if there were no errors */ 32209b5097eeSOwen Roberts if (error == 0) 32219b5097eeSOwen Roberts vnevent_remove(rmvp, vp, nm, ct); 32229b5097eeSOwen Roberts VN_RELE(rmvp); 32239b5097eeSOwen Roberts } 32249b5097eeSOwen Roberts 322502ff05a9Svsakar retry_remove: 32267c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_REMOVE_MASK); 32277c478bd9Sstevel@tonic-gate if (error) 32287c478bd9Sstevel@tonic-gate goto out; 32297c478bd9Sstevel@tonic-gate 32307c478bd9Sstevel@tonic-gate if (ulp) 32317c478bd9Sstevel@tonic-gate TRANS_BEGIN_CSYNC(ufsvfsp, issync, TOP_REMOVE, 32327c478bd9Sstevel@tonic-gate trans_size = (int)TOP_REMOVE_SIZE(VTOI(vp))); 32337c478bd9Sstevel@tonic-gate 323402ff05a9Svsakar /* 323502ff05a9Svsakar * ufs_tryirwlock_trans uses rw_tryenter and checks for SLOCK 323602ff05a9Svsakar * to avoid i_rwlock, ufs_lockfs_begin deadlock. If deadlock 323702ff05a9Svsakar * possible, retries the operation. 323802ff05a9Svsakar */ 323902ff05a9Svsakar ufs_tryirwlock_trans(&ip->i_rwlock, RW_WRITER, TOP_REMOVE, retry); 324002ff05a9Svsakar if (indeadlock) 324102ff05a9Svsakar goto retry_remove; 32427c478bd9Sstevel@tonic-gate error = ufs_dirremove(ip, nm, (struct inode *)0, (struct vnode *)0, 32439b5097eeSOwen Roberts DR_REMOVE, cr); 32447c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 32457c478bd9Sstevel@tonic-gate 32467c478bd9Sstevel@tonic-gate if (ulp) { 32477c478bd9Sstevel@tonic-gate TRANS_END_CSYNC(ufsvfsp, error, issync, TOP_REMOVE, trans_size); 32487c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 32497c478bd9Sstevel@tonic-gate } 32507c478bd9Sstevel@tonic-gate 32517c478bd9Sstevel@tonic-gate out: 32527c478bd9Sstevel@tonic-gate return (error); 32537c478bd9Sstevel@tonic-gate } 32547c478bd9Sstevel@tonic-gate 32557c478bd9Sstevel@tonic-gate /* 32567c478bd9Sstevel@tonic-gate * Link a file or a directory. Only privileged processes are allowed to 32577c478bd9Sstevel@tonic-gate * make links to directories. 32587c478bd9Sstevel@tonic-gate */ 3259da6c28aaSamw /*ARGSUSED*/ 32607c478bd9Sstevel@tonic-gate static int 3261da6c28aaSamw ufs_link(struct vnode *tdvp, struct vnode *svp, char *tnm, struct cred *cr, 3262da6c28aaSamw caller_context_t *ct, int flags) 32637c478bd9Sstevel@tonic-gate { 32647c478bd9Sstevel@tonic-gate struct inode *sip; 32657c478bd9Sstevel@tonic-gate struct inode *tdp = VTOI(tdvp); 32667c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp = tdp->i_ufsvfs; 32677c478bd9Sstevel@tonic-gate struct ulockfs *ulp; 32687c478bd9Sstevel@tonic-gate struct vnode *realvp; 32697c478bd9Sstevel@tonic-gate int error; 32707c478bd9Sstevel@tonic-gate int issync; 32717c478bd9Sstevel@tonic-gate int trans_size; 32727c478bd9Sstevel@tonic-gate int isdev; 327302ff05a9Svsakar int indeadlock; 32747c478bd9Sstevel@tonic-gate 327502ff05a9Svsakar retry_link: 32767c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_LINK_MASK); 32777c478bd9Sstevel@tonic-gate if (error) 32787c478bd9Sstevel@tonic-gate goto out; 32797c478bd9Sstevel@tonic-gate 32807c478bd9Sstevel@tonic-gate if (ulp) 32817c478bd9Sstevel@tonic-gate TRANS_BEGIN_CSYNC(ufsvfsp, issync, TOP_LINK, 32827c478bd9Sstevel@tonic-gate trans_size = (int)TOP_LINK_SIZE(VTOI(tdvp))); 32837c478bd9Sstevel@tonic-gate 3284da6c28aaSamw if (VOP_REALVP(svp, &realvp, ct) == 0) 32857c478bd9Sstevel@tonic-gate svp = realvp; 32867c478bd9Sstevel@tonic-gate 32877c478bd9Sstevel@tonic-gate /* 32887c478bd9Sstevel@tonic-gate * Make sure link for extended attributes is valid 32897c478bd9Sstevel@tonic-gate * We only support hard linking of attr in ATTRDIR to ATTRDIR 32907c478bd9Sstevel@tonic-gate * 32917c478bd9Sstevel@tonic-gate * Make certain we don't attempt to look at a device node as 32927c478bd9Sstevel@tonic-gate * a ufs inode. 32937c478bd9Sstevel@tonic-gate */ 32947c478bd9Sstevel@tonic-gate 32957c478bd9Sstevel@tonic-gate isdev = IS_DEVVP(svp); 32967c478bd9Sstevel@tonic-gate if (((isdev == 0) && ((VTOI(svp)->i_cflags & IXATTR) == 0) && 32977c478bd9Sstevel@tonic-gate ((tdp->i_mode & IFMT) == IFATTRDIR)) || 32987c478bd9Sstevel@tonic-gate ((isdev == 0) && (VTOI(svp)->i_cflags & IXATTR) && 32997c478bd9Sstevel@tonic-gate ((tdp->i_mode & IFMT) == IFDIR))) { 33007c478bd9Sstevel@tonic-gate error = EINVAL; 33017c478bd9Sstevel@tonic-gate goto unlock; 33027c478bd9Sstevel@tonic-gate } 33037c478bd9Sstevel@tonic-gate 33047c478bd9Sstevel@tonic-gate sip = VTOI(svp); 33057c478bd9Sstevel@tonic-gate if ((svp->v_type == VDIR && 33067c478bd9Sstevel@tonic-gate secpolicy_fs_linkdir(cr, ufsvfsp->vfs_vfs) != 0) || 33077c478bd9Sstevel@tonic-gate (sip->i_uid != crgetuid(cr) && secpolicy_basic_link(cr) != 0)) { 33087c478bd9Sstevel@tonic-gate error = EPERM; 33097c478bd9Sstevel@tonic-gate goto unlock; 33107c478bd9Sstevel@tonic-gate } 331102ff05a9Svsakar 331202ff05a9Svsakar /* 331302ff05a9Svsakar * ufs_tryirwlock_trans uses rw_tryenter and checks for SLOCK 331402ff05a9Svsakar * to avoid i_rwlock, ufs_lockfs_begin deadlock. If deadlock 331502ff05a9Svsakar * possible, retries the operation. 331602ff05a9Svsakar */ 331702ff05a9Svsakar ufs_tryirwlock_trans(&tdp->i_rwlock, RW_WRITER, TOP_LINK, retry); 331802ff05a9Svsakar if (indeadlock) 331902ff05a9Svsakar goto retry_link; 33207c478bd9Sstevel@tonic-gate error = ufs_direnter_lr(tdp, tnm, DE_LINK, (struct inode *)0, 33219b5097eeSOwen Roberts sip, cr); 33227c478bd9Sstevel@tonic-gate rw_exit(&tdp->i_rwlock); 33237c478bd9Sstevel@tonic-gate 33247c478bd9Sstevel@tonic-gate unlock: 33257c478bd9Sstevel@tonic-gate if (ulp) { 33267c478bd9Sstevel@tonic-gate TRANS_END_CSYNC(ufsvfsp, error, issync, TOP_LINK, trans_size); 33277c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 33287c478bd9Sstevel@tonic-gate } 3329df2381bfSpraks 3330df2381bfSpraks if (!error) { 3331da6c28aaSamw vnevent_link(svp, ct); 3332df2381bfSpraks } 33337c478bd9Sstevel@tonic-gate out: 33347c478bd9Sstevel@tonic-gate return (error); 33357c478bd9Sstevel@tonic-gate } 33367c478bd9Sstevel@tonic-gate 33377c478bd9Sstevel@tonic-gate uint64_t ufs_rename_retry_cnt; 33387c478bd9Sstevel@tonic-gate uint64_t ufs_rename_upgrade_retry_cnt; 33397c478bd9Sstevel@tonic-gate uint64_t ufs_rename_dircheck_retry_cnt; 33407c478bd9Sstevel@tonic-gate clock_t ufs_rename_backoff_delay = 1; 33417c478bd9Sstevel@tonic-gate 33427c478bd9Sstevel@tonic-gate /* 33437c478bd9Sstevel@tonic-gate * Rename a file or directory. 33447c478bd9Sstevel@tonic-gate * We are given the vnode and entry string of the source and the 33457c478bd9Sstevel@tonic-gate * vnode and entry string of the place we want to move the source 33467c478bd9Sstevel@tonic-gate * to (the target). The essential operation is: 33477c478bd9Sstevel@tonic-gate * unlink(target); 33487c478bd9Sstevel@tonic-gate * link(source, target); 33497c478bd9Sstevel@tonic-gate * unlink(source); 33507c478bd9Sstevel@tonic-gate * but "atomically". Can't do full commit without saving state in 33517c478bd9Sstevel@tonic-gate * the inode on disk, which isn't feasible at this time. Best we 33527c478bd9Sstevel@tonic-gate * can do is always guarantee that the TARGET exists. 33537c478bd9Sstevel@tonic-gate */ 3354baa4d099Sswilcox 33557c478bd9Sstevel@tonic-gate /*ARGSUSED*/ 33567c478bd9Sstevel@tonic-gate static int 33577c478bd9Sstevel@tonic-gate ufs_rename( 33587c478bd9Sstevel@tonic-gate struct vnode *sdvp, /* old (source) parent vnode */ 33597c478bd9Sstevel@tonic-gate char *snm, /* old (source) entry name */ 33607c478bd9Sstevel@tonic-gate struct vnode *tdvp, /* new (target) parent vnode */ 33617c478bd9Sstevel@tonic-gate char *tnm, /* new (target) entry name */ 3362da6c28aaSamw struct cred *cr, 3363da6c28aaSamw caller_context_t *ct, 3364da6c28aaSamw int flags) 33657c478bd9Sstevel@tonic-gate { 33667c478bd9Sstevel@tonic-gate struct inode *sip = NULL; /* source inode */ 3367baa4d099Sswilcox struct inode *ip = NULL; /* check inode */ 33687c478bd9Sstevel@tonic-gate struct inode *sdp; /* old (source) parent inode */ 33697c478bd9Sstevel@tonic-gate struct inode *tdp; /* new (target) parent inode */ 33709b5097eeSOwen Roberts struct vnode *svp = NULL; /* source vnode */ 33717c478bd9Sstevel@tonic-gate struct vnode *tvp = NULL; /* target vnode, if it exists */ 33727c478bd9Sstevel@tonic-gate struct vnode *realvp; 33737c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp; 33747c478bd9Sstevel@tonic-gate struct ulockfs *ulp; 33755b024a5bSbatschul struct ufs_slot slot; 3376baa4d099Sswilcox timestruc_t now; 33777c478bd9Sstevel@tonic-gate int error; 33787c478bd9Sstevel@tonic-gate int issync; 33797c478bd9Sstevel@tonic-gate int trans_size; 338002ff05a9Svsakar krwlock_t *first_lock; 338102ff05a9Svsakar krwlock_t *second_lock; 338202ff05a9Svsakar krwlock_t *reverse_lock; 33839b5097eeSOwen Roberts int serr, terr; 33847c478bd9Sstevel@tonic-gate 33857c478bd9Sstevel@tonic-gate sdp = VTOI(sdvp); 3386baa4d099Sswilcox slot.fbp = NULL; 33877c478bd9Sstevel@tonic-gate ufsvfsp = sdp->i_ufsvfs; 33889b5097eeSOwen Roberts 33899b5097eeSOwen Roberts if (VOP_REALVP(tdvp, &realvp, ct) == 0) 33909b5097eeSOwen Roberts tdvp = realvp; 33919b5097eeSOwen Roberts 33929b5097eeSOwen Roberts terr = ufs_eventlookup(tdvp, tnm, cr, &tvp); 33939b5097eeSOwen Roberts serr = ufs_eventlookup(sdvp, snm, cr, &svp); 33949b5097eeSOwen Roberts 33959b5097eeSOwen Roberts if ((serr == 0) && ((terr == 0) || (terr == ENOENT))) { 33969b5097eeSOwen Roberts if (tvp != NULL) 33979b5097eeSOwen Roberts vnevent_rename_dest(tvp, tdvp, tnm, ct); 33989b5097eeSOwen Roberts 33999b5097eeSOwen Roberts /* 34009b5097eeSOwen Roberts * Notify the target directory of the rename event 34019b5097eeSOwen Roberts * if source and target directories are not the same. 34029b5097eeSOwen Roberts */ 34039b5097eeSOwen Roberts if (sdvp != tdvp) 34049b5097eeSOwen Roberts vnevent_rename_dest_dir(tdvp, ct); 34059b5097eeSOwen Roberts 34069b5097eeSOwen Roberts if (svp != NULL) 34079b5097eeSOwen Roberts vnevent_rename_src(svp, sdvp, snm, ct); 34089b5097eeSOwen Roberts } 34099b5097eeSOwen Roberts 34109b5097eeSOwen Roberts if (tvp != NULL) 34119b5097eeSOwen Roberts VN_RELE(tvp); 34129b5097eeSOwen Roberts 34139b5097eeSOwen Roberts if (svp != NULL) 34149b5097eeSOwen Roberts VN_RELE(svp); 34159b5097eeSOwen Roberts 341602ff05a9Svsakar retry_rename: 34177c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_RENAME_MASK); 34187c478bd9Sstevel@tonic-gate if (error) 34197c478bd9Sstevel@tonic-gate goto out; 34207c478bd9Sstevel@tonic-gate 34217c478bd9Sstevel@tonic-gate if (ulp) 34227c478bd9Sstevel@tonic-gate TRANS_BEGIN_CSYNC(ufsvfsp, issync, TOP_RENAME, 34237c478bd9Sstevel@tonic-gate trans_size = (int)TOP_RENAME_SIZE(sdp)); 34247c478bd9Sstevel@tonic-gate 3425da6c28aaSamw if (VOP_REALVP(tdvp, &realvp, ct) == 0) 34267c478bd9Sstevel@tonic-gate tdvp = realvp; 34277c478bd9Sstevel@tonic-gate 34287c478bd9Sstevel@tonic-gate tdp = VTOI(tdvp); 34297c478bd9Sstevel@tonic-gate 34307c478bd9Sstevel@tonic-gate /* 34317c478bd9Sstevel@tonic-gate * We only allow renaming of attributes from ATTRDIR to ATTRDIR. 34327c478bd9Sstevel@tonic-gate */ 34337c478bd9Sstevel@tonic-gate if ((tdp->i_mode & IFMT) != (sdp->i_mode & IFMT)) { 34347c478bd9Sstevel@tonic-gate error = EINVAL; 34357c478bd9Sstevel@tonic-gate goto unlock; 34367c478bd9Sstevel@tonic-gate } 34377c478bd9Sstevel@tonic-gate 34387c478bd9Sstevel@tonic-gate /* 343960c8e821SFrank Batschulat * Check accessibility of directory. 344060c8e821SFrank Batschulat */ 344160c8e821SFrank Batschulat if (error = ufs_diraccess(sdp, IEXEC, cr)) 344260c8e821SFrank Batschulat goto unlock; 344360c8e821SFrank Batschulat 344460c8e821SFrank Batschulat /* 34457c478bd9Sstevel@tonic-gate * Look up inode of file we're supposed to rename. 34467c478bd9Sstevel@tonic-gate */ 3447baa4d099Sswilcox gethrestime(&now); 34489b5097eeSOwen Roberts if (error = ufs_dirlook(sdp, snm, &sip, cr, 0, 0)) { 344902ff05a9Svsakar if (error == EAGAIN) { 345002ff05a9Svsakar if (ulp) { 345102ff05a9Svsakar TRANS_END_CSYNC(ufsvfsp, error, issync, 345202ff05a9Svsakar TOP_RENAME, trans_size); 345302ff05a9Svsakar ufs_lockfs_end(ulp); 345402ff05a9Svsakar } 345502ff05a9Svsakar goto retry_rename; 345602ff05a9Svsakar } 345702ff05a9Svsakar 34587c478bd9Sstevel@tonic-gate goto unlock; 34597c478bd9Sstevel@tonic-gate } 34607c478bd9Sstevel@tonic-gate 34617c478bd9Sstevel@tonic-gate /* 34627c478bd9Sstevel@tonic-gate * Lock both the source and target directories (they may be 34637c478bd9Sstevel@tonic-gate * the same) to provide the atomicity semantics that was 34647c478bd9Sstevel@tonic-gate * previously provided by the per file system vfs_rename_lock 34657c478bd9Sstevel@tonic-gate * 34667c478bd9Sstevel@tonic-gate * with vfs_rename_lock removed to allow simultaneous renames 34677c478bd9Sstevel@tonic-gate * within a file system, ufs_dircheckpath can deadlock while 34687c478bd9Sstevel@tonic-gate * traversing back to ensure that source is not a parent directory 34697c478bd9Sstevel@tonic-gate * of target parent directory. This is because we get into 34707c478bd9Sstevel@tonic-gate * ufs_dircheckpath with the sdp and tdp locks held as RW_WRITER. 34717c478bd9Sstevel@tonic-gate * If the tdp and sdp of the simultaneous renames happen to be 34727c478bd9Sstevel@tonic-gate * in the path of each other, it can lead to a deadlock. This 34737c478bd9Sstevel@tonic-gate * can be avoided by getting the locks as RW_READER here and then 34747c478bd9Sstevel@tonic-gate * upgrading to RW_WRITER after completing the ufs_dircheckpath. 347502ff05a9Svsakar * 347602ff05a9Svsakar * We hold the target directory's i_rwlock after calling 347702ff05a9Svsakar * ufs_lockfs_begin but in many other operations (like ufs_readdir) 347802ff05a9Svsakar * VOP_RWLOCK is explicitly called by the filesystem independent code 347902ff05a9Svsakar * before calling the file system operation. In these cases the order 348002ff05a9Svsakar * is reversed (i.e i_rwlock is taken first and then ufs_lockfs_begin 348102ff05a9Svsakar * is called). This is fine as long as ufs_lockfs_begin acts as a VOP 348202ff05a9Svsakar * counter but with ufs_quiesce setting the SLOCK bit this becomes a 348302ff05a9Svsakar * synchronizing object which might lead to a deadlock. So we use 348402ff05a9Svsakar * rw_tryenter instead of rw_enter. If we fail to get this lock and 348502ff05a9Svsakar * find that SLOCK bit is set, we call ufs_lockfs_end and restart the 348602ff05a9Svsakar * operation. 34877c478bd9Sstevel@tonic-gate */ 34887c478bd9Sstevel@tonic-gate retry: 348902ff05a9Svsakar first_lock = &tdp->i_rwlock; 349002ff05a9Svsakar second_lock = &sdp->i_rwlock; 349102ff05a9Svsakar retry_firstlock: 349202ff05a9Svsakar if (!rw_tryenter(first_lock, RW_READER)) { 34937c478bd9Sstevel@tonic-gate /* 349402ff05a9Svsakar * We didn't get the lock. Check if the SLOCK is set in the 349502ff05a9Svsakar * ufsvfs. If yes, we might be in a deadlock. Safer to give up 349602ff05a9Svsakar * and wait for SLOCK to be cleared. 34977c478bd9Sstevel@tonic-gate */ 349802ff05a9Svsakar 349902ff05a9Svsakar if (ulp && ULOCKFS_IS_SLOCK(ulp)) { 350002ff05a9Svsakar TRANS_END_CSYNC(ufsvfsp, error, issync, TOP_RENAME, 350102ff05a9Svsakar trans_size); 350202ff05a9Svsakar ufs_lockfs_end(ulp); 350302ff05a9Svsakar goto retry_rename; 350402ff05a9Svsakar 350502ff05a9Svsakar } else { 35067c478bd9Sstevel@tonic-gate /* 350702ff05a9Svsakar * SLOCK isn't set so this is a genuine synchronization 350802ff05a9Svsakar * case. Let's try again after giving them a breather. 35097c478bd9Sstevel@tonic-gate */ 351002ff05a9Svsakar delay(RETRY_LOCK_DELAY); 351102ff05a9Svsakar goto retry_firstlock; 35127c478bd9Sstevel@tonic-gate } 35137c478bd9Sstevel@tonic-gate } 351402ff05a9Svsakar /* 351502ff05a9Svsakar * Need to check if the tdp and sdp are same !!! 351602ff05a9Svsakar */ 351702ff05a9Svsakar if ((tdp != sdp) && (!rw_tryenter(second_lock, RW_READER))) { 351802ff05a9Svsakar /* 351902ff05a9Svsakar * We didn't get the lock. Check if the SLOCK is set in the 352002ff05a9Svsakar * ufsvfs. If yes, we might be in a deadlock. Safer to give up 352102ff05a9Svsakar * and wait for SLOCK to be cleared. 352202ff05a9Svsakar */ 352302ff05a9Svsakar 352402ff05a9Svsakar rw_exit(first_lock); 352502ff05a9Svsakar if (ulp && ULOCKFS_IS_SLOCK(ulp)) { 352602ff05a9Svsakar TRANS_END_CSYNC(ufsvfsp, error, issync, TOP_RENAME, 352702ff05a9Svsakar trans_size); 352802ff05a9Svsakar ufs_lockfs_end(ulp); 352902ff05a9Svsakar goto retry_rename; 353002ff05a9Svsakar 353102ff05a9Svsakar } else { 353202ff05a9Svsakar /* 353302ff05a9Svsakar * So we couldn't get the second level peer lock *and* 353402ff05a9Svsakar * the SLOCK bit isn't set. Too bad we can be 353502ff05a9Svsakar * contentding with someone wanting these locks otherway 353602ff05a9Svsakar * round. Reverse the locks in case there is a heavy 353702ff05a9Svsakar * contention for the second level lock. 353802ff05a9Svsakar */ 353902ff05a9Svsakar reverse_lock = first_lock; 354002ff05a9Svsakar first_lock = second_lock; 354102ff05a9Svsakar second_lock = reverse_lock; 354202ff05a9Svsakar ufs_rename_retry_cnt++; 354302ff05a9Svsakar goto retry_firstlock; 354402ff05a9Svsakar } 35457c478bd9Sstevel@tonic-gate } 35467c478bd9Sstevel@tonic-gate 35477c478bd9Sstevel@tonic-gate if (sip == tdp) { 35487c478bd9Sstevel@tonic-gate error = EINVAL; 35497c478bd9Sstevel@tonic-gate goto errout; 35507c478bd9Sstevel@tonic-gate } 35517c478bd9Sstevel@tonic-gate /* 35527c478bd9Sstevel@tonic-gate * Make sure we can delete the source entry. This requires 35537c478bd9Sstevel@tonic-gate * write permission on the containing directory. 35547c478bd9Sstevel@tonic-gate * Check for sticky directories. 35557c478bd9Sstevel@tonic-gate */ 35567c478bd9Sstevel@tonic-gate rw_enter(&sdp->i_contents, RW_READER); 35577c478bd9Sstevel@tonic-gate rw_enter(&sip->i_contents, RW_READER); 355860c8e821SFrank Batschulat if ((error = ufs_iaccess(sdp, IWRITE, cr, 0)) != 0 || 35597c478bd9Sstevel@tonic-gate (error = ufs_sticky_remove_access(sdp, sip, cr)) != 0) { 35607c478bd9Sstevel@tonic-gate rw_exit(&sip->i_contents); 35617c478bd9Sstevel@tonic-gate rw_exit(&sdp->i_contents); 35627c478bd9Sstevel@tonic-gate goto errout; 35637c478bd9Sstevel@tonic-gate } 35647c478bd9Sstevel@tonic-gate 35657c478bd9Sstevel@tonic-gate /* 35667c478bd9Sstevel@tonic-gate * If this is a rename of a directory and the parent is 35677c478bd9Sstevel@tonic-gate * different (".." must be changed), then the source 35687c478bd9Sstevel@tonic-gate * directory must not be in the directory hierarchy 35697c478bd9Sstevel@tonic-gate * above the target, as this would orphan everything 35707c478bd9Sstevel@tonic-gate * below the source directory. Also the user must have 35717c478bd9Sstevel@tonic-gate * write permission in the source so as to be able to 35727c478bd9Sstevel@tonic-gate * change "..". 35737c478bd9Sstevel@tonic-gate */ 35747c478bd9Sstevel@tonic-gate if ((((sip->i_mode & IFMT) == IFDIR) || 35757c478bd9Sstevel@tonic-gate ((sip->i_mode & IFMT) == IFATTRDIR)) && sdp != tdp) { 35767c478bd9Sstevel@tonic-gate ino_t inum; 35777c478bd9Sstevel@tonic-gate 357860c8e821SFrank Batschulat if (error = ufs_iaccess(sip, IWRITE, cr, 0)) { 35797c478bd9Sstevel@tonic-gate rw_exit(&sip->i_contents); 35807c478bd9Sstevel@tonic-gate rw_exit(&sdp->i_contents); 35817c478bd9Sstevel@tonic-gate goto errout; 35827c478bd9Sstevel@tonic-gate } 35837c478bd9Sstevel@tonic-gate inum = sip->i_number; 35847c478bd9Sstevel@tonic-gate rw_exit(&sip->i_contents); 35857c478bd9Sstevel@tonic-gate rw_exit(&sdp->i_contents); 35867c478bd9Sstevel@tonic-gate if ((error = ufs_dircheckpath(inum, tdp, sdp, cr))) { 35877c478bd9Sstevel@tonic-gate /* 35887c478bd9Sstevel@tonic-gate * If we got EAGAIN ufs_dircheckpath detected a 35897c478bd9Sstevel@tonic-gate * potential deadlock and backed out. We need 35907c478bd9Sstevel@tonic-gate * to retry the operation since sdp and tdp have 35917c478bd9Sstevel@tonic-gate * to be released to avoid the deadlock. 35927c478bd9Sstevel@tonic-gate */ 35937c478bd9Sstevel@tonic-gate if (error == EAGAIN) { 35947c478bd9Sstevel@tonic-gate rw_exit(&tdp->i_rwlock); 35957c478bd9Sstevel@tonic-gate if (tdp != sdp) 35967c478bd9Sstevel@tonic-gate rw_exit(&sdp->i_rwlock); 35977c478bd9Sstevel@tonic-gate delay(ufs_rename_backoff_delay); 35987c478bd9Sstevel@tonic-gate ufs_rename_dircheck_retry_cnt++; 35997c478bd9Sstevel@tonic-gate goto retry; 36007c478bd9Sstevel@tonic-gate } 36017c478bd9Sstevel@tonic-gate goto errout; 36027c478bd9Sstevel@tonic-gate } 36037c478bd9Sstevel@tonic-gate } else { 36047c478bd9Sstevel@tonic-gate rw_exit(&sip->i_contents); 36057c478bd9Sstevel@tonic-gate rw_exit(&sdp->i_contents); 36067c478bd9Sstevel@tonic-gate } 36077c478bd9Sstevel@tonic-gate 36087c478bd9Sstevel@tonic-gate 36097c478bd9Sstevel@tonic-gate /* 36107c478bd9Sstevel@tonic-gate * Check for renaming '.' or '..' or alias of '.' 36117c478bd9Sstevel@tonic-gate */ 36127c478bd9Sstevel@tonic-gate if (strcmp(snm, ".") == 0 || strcmp(snm, "..") == 0 || sdp == sip) { 36137c478bd9Sstevel@tonic-gate error = EINVAL; 36147c478bd9Sstevel@tonic-gate goto errout; 36157c478bd9Sstevel@tonic-gate } 36167c478bd9Sstevel@tonic-gate 36177c478bd9Sstevel@tonic-gate /* 36187c478bd9Sstevel@tonic-gate * Simultaneous renames can deadlock in ufs_dircheckpath since it 36197c478bd9Sstevel@tonic-gate * tries to traverse back the file tree with both tdp and sdp held 36207c478bd9Sstevel@tonic-gate * as RW_WRITER. To avoid that we have to hold the tdp and sdp locks 36217c478bd9Sstevel@tonic-gate * as RW_READERS till ufs_dircheckpath is done. 36227c478bd9Sstevel@tonic-gate * Now that ufs_dircheckpath is done with, we can upgrade the locks 36237c478bd9Sstevel@tonic-gate * to RW_WRITER. 36247c478bd9Sstevel@tonic-gate */ 36257c478bd9Sstevel@tonic-gate if (!rw_tryupgrade(&tdp->i_rwlock)) { 36267c478bd9Sstevel@tonic-gate /* 36277c478bd9Sstevel@tonic-gate * The upgrade failed. We got to give away the lock 36287c478bd9Sstevel@tonic-gate * as to avoid deadlocking with someone else who is 36297c478bd9Sstevel@tonic-gate * waiting for writer lock. With the lock gone, we 36307c478bd9Sstevel@tonic-gate * cannot be sure the checks done above will hold 36317c478bd9Sstevel@tonic-gate * good when we eventually get them back as writer. 36327c478bd9Sstevel@tonic-gate * So if we can't upgrade we drop the locks and retry 36337c478bd9Sstevel@tonic-gate * everything again. 36347c478bd9Sstevel@tonic-gate */ 36357c478bd9Sstevel@tonic-gate rw_exit(&tdp->i_rwlock); 36367c478bd9Sstevel@tonic-gate if (tdp != sdp) 36377c478bd9Sstevel@tonic-gate rw_exit(&sdp->i_rwlock); 36387c478bd9Sstevel@tonic-gate delay(ufs_rename_backoff_delay); 36397c478bd9Sstevel@tonic-gate ufs_rename_upgrade_retry_cnt++; 36407c478bd9Sstevel@tonic-gate goto retry; 36417c478bd9Sstevel@tonic-gate } 36427c478bd9Sstevel@tonic-gate if (tdp != sdp) { 36437c478bd9Sstevel@tonic-gate if (!rw_tryupgrade(&sdp->i_rwlock)) { 36447c478bd9Sstevel@tonic-gate /* 36457c478bd9Sstevel@tonic-gate * The upgrade failed. We got to give away the lock 36467c478bd9Sstevel@tonic-gate * as to avoid deadlocking with someone else who is 36477c478bd9Sstevel@tonic-gate * waiting for writer lock. With the lock gone, we 36487c478bd9Sstevel@tonic-gate * cannot be sure the checks done above will hold 36497c478bd9Sstevel@tonic-gate * good when we eventually get them back as writer. 36507c478bd9Sstevel@tonic-gate * So if we can't upgrade we drop the locks and retry 36517c478bd9Sstevel@tonic-gate * everything again. 36527c478bd9Sstevel@tonic-gate */ 36537c478bd9Sstevel@tonic-gate rw_exit(&tdp->i_rwlock); 36547c478bd9Sstevel@tonic-gate rw_exit(&sdp->i_rwlock); 36557c478bd9Sstevel@tonic-gate delay(ufs_rename_backoff_delay); 36567c478bd9Sstevel@tonic-gate ufs_rename_upgrade_retry_cnt++; 36577c478bd9Sstevel@tonic-gate goto retry; 36587c478bd9Sstevel@tonic-gate } 36597c478bd9Sstevel@tonic-gate } 3660baa4d099Sswilcox 3661baa4d099Sswilcox /* 3662baa4d099Sswilcox * Now that all the locks are held check to make sure another thread 3663baa4d099Sswilcox * didn't slip in and take out the sip. 3664baa4d099Sswilcox */ 3665baa4d099Sswilcox slot.status = NONE; 3666baa4d099Sswilcox if ((sip->i_ctime.tv_usec * 1000) > now.tv_nsec || 3667baa4d099Sswilcox sip->i_ctime.tv_sec > now.tv_sec) { 3668baa4d099Sswilcox rw_enter(&sdp->i_ufsvfs->vfs_dqrwlock, RW_READER); 3669baa4d099Sswilcox rw_enter(&sdp->i_contents, RW_WRITER); 3670baa4d099Sswilcox error = ufs_dircheckforname(sdp, snm, strlen(snm), &slot, 3671baa4d099Sswilcox &ip, cr, 0); 3672baa4d099Sswilcox rw_exit(&sdp->i_contents); 3673baa4d099Sswilcox rw_exit(&sdp->i_ufsvfs->vfs_dqrwlock); 3674baa4d099Sswilcox if (error) { 3675baa4d099Sswilcox goto errout; 3676baa4d099Sswilcox } 3677baa4d099Sswilcox if (ip == NULL) { 3678baa4d099Sswilcox error = ENOENT; 3679baa4d099Sswilcox goto errout; 3680baa4d099Sswilcox } else { 3681baa4d099Sswilcox /* 3682baa4d099Sswilcox * If the inode was found need to drop the v_count 3683baa4d099Sswilcox * so as not to keep the filesystem from being 3684baa4d099Sswilcox * unmounted at a later time. 3685baa4d099Sswilcox */ 3686baa4d099Sswilcox VN_RELE(ITOV(ip)); 3687baa4d099Sswilcox } 3688baa4d099Sswilcox 3689baa4d099Sswilcox /* 3690baa4d099Sswilcox * Release the slot.fbp that has the page mapped and 3691baa4d099Sswilcox * locked SE_SHARED, and could be used in in 3692baa4d099Sswilcox * ufs_direnter_lr() which needs to get the SE_EXCL lock 3693baa4d099Sswilcox * on said page. 3694baa4d099Sswilcox */ 3695baa4d099Sswilcox if (slot.fbp) { 3696baa4d099Sswilcox fbrelse(slot.fbp, S_OTHER); 3697baa4d099Sswilcox slot.fbp = NULL; 3698baa4d099Sswilcox } 3699baa4d099Sswilcox } 3700baa4d099Sswilcox 37017c478bd9Sstevel@tonic-gate /* 37029b5097eeSOwen Roberts * Link source to the target. 37037c478bd9Sstevel@tonic-gate */ 37049b5097eeSOwen Roberts if (error = ufs_direnter_lr(tdp, tnm, DE_RENAME, sdp, sip, cr)) { 37057c478bd9Sstevel@tonic-gate /* 37067c478bd9Sstevel@tonic-gate * ESAME isn't really an error; it indicates that the 37077c478bd9Sstevel@tonic-gate * operation should not be done because the source and target 37087c478bd9Sstevel@tonic-gate * are the same file, but that no error should be reported. 37097c478bd9Sstevel@tonic-gate */ 37107c478bd9Sstevel@tonic-gate if (error == ESAME) 37117c478bd9Sstevel@tonic-gate error = 0; 37127c478bd9Sstevel@tonic-gate goto errout; 37137c478bd9Sstevel@tonic-gate } 37147c478bd9Sstevel@tonic-gate 37157c478bd9Sstevel@tonic-gate /* 37167c478bd9Sstevel@tonic-gate * Unlink the source. 37177c478bd9Sstevel@tonic-gate * Remove the source entry. ufs_dirremove() checks that the entry 37187c478bd9Sstevel@tonic-gate * still reflects sip, and returns an error if it doesn't. 37197c478bd9Sstevel@tonic-gate * If the entry has changed just forget about it. Release 37207c478bd9Sstevel@tonic-gate * the source inode. 37217c478bd9Sstevel@tonic-gate */ 37227c478bd9Sstevel@tonic-gate if ((error = ufs_dirremove(sdp, snm, sip, (struct vnode *)0, 37239b5097eeSOwen Roberts DR_RENAME, cr)) == ENOENT) 37247c478bd9Sstevel@tonic-gate error = 0; 37257c478bd9Sstevel@tonic-gate 37267c478bd9Sstevel@tonic-gate errout: 3727baa4d099Sswilcox if (slot.fbp) 3728baa4d099Sswilcox fbrelse(slot.fbp, S_OTHER); 3729baa4d099Sswilcox 37307c478bd9Sstevel@tonic-gate rw_exit(&tdp->i_rwlock); 37317c478bd9Sstevel@tonic-gate if (sdp != tdp) { 37327c478bd9Sstevel@tonic-gate rw_exit(&sdp->i_rwlock); 37337c478bd9Sstevel@tonic-gate } 37347c478bd9Sstevel@tonic-gate 37359b5097eeSOwen Roberts VN_RELE(ITOV(sip)); 37369b5097eeSOwen Roberts 37377c478bd9Sstevel@tonic-gate unlock: 37387c478bd9Sstevel@tonic-gate if (ulp) { 37397c478bd9Sstevel@tonic-gate TRANS_END_CSYNC(ufsvfsp, error, issync, TOP_RENAME, trans_size); 37407c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 37417c478bd9Sstevel@tonic-gate } 37427c478bd9Sstevel@tonic-gate 37437c478bd9Sstevel@tonic-gate out: 37447c478bd9Sstevel@tonic-gate return (error); 37457c478bd9Sstevel@tonic-gate } 37467c478bd9Sstevel@tonic-gate 37477c478bd9Sstevel@tonic-gate /*ARGSUSED*/ 37487c478bd9Sstevel@tonic-gate static int 37497c478bd9Sstevel@tonic-gate ufs_mkdir(struct vnode *dvp, char *dirname, struct vattr *vap, 3750da6c28aaSamw struct vnode **vpp, struct cred *cr, caller_context_t *ct, int flags, 3751da6c28aaSamw vsecattr_t *vsecp) 37527c478bd9Sstevel@tonic-gate { 37537c478bd9Sstevel@tonic-gate struct inode *ip; 37547c478bd9Sstevel@tonic-gate struct inode *xip; 37557c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp; 37567c478bd9Sstevel@tonic-gate struct ulockfs *ulp; 37577c478bd9Sstevel@tonic-gate int error; 37587c478bd9Sstevel@tonic-gate int issync; 37597c478bd9Sstevel@tonic-gate int trans_size; 376002ff05a9Svsakar int indeadlock; 37617c478bd9Sstevel@tonic-gate int retry = 1; 37627c478bd9Sstevel@tonic-gate 37637c478bd9Sstevel@tonic-gate ASSERT((vap->va_mask & (AT_TYPE|AT_MODE)) == (AT_TYPE|AT_MODE)); 37647c478bd9Sstevel@tonic-gate 37657c478bd9Sstevel@tonic-gate /* 37667c478bd9Sstevel@tonic-gate * Can't make directory in attr hidden dir 37677c478bd9Sstevel@tonic-gate */ 37687c478bd9Sstevel@tonic-gate if ((VTOI(dvp)->i_mode & IFMT) == IFATTRDIR) 37697c478bd9Sstevel@tonic-gate return (EINVAL); 37707c478bd9Sstevel@tonic-gate 37717c478bd9Sstevel@tonic-gate again: 37727c478bd9Sstevel@tonic-gate ip = VTOI(dvp); 37737c478bd9Sstevel@tonic-gate ufsvfsp = ip->i_ufsvfs; 37747c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_MKDIR_MASK); 37757c478bd9Sstevel@tonic-gate if (error) 37767c478bd9Sstevel@tonic-gate goto out; 37777c478bd9Sstevel@tonic-gate if (ulp) 37787c478bd9Sstevel@tonic-gate TRANS_BEGIN_CSYNC(ufsvfsp, issync, TOP_MKDIR, 37797c478bd9Sstevel@tonic-gate trans_size = (int)TOP_MKDIR_SIZE(ip)); 37807c478bd9Sstevel@tonic-gate 378102ff05a9Svsakar /* 378202ff05a9Svsakar * ufs_tryirwlock_trans uses rw_tryenter and checks for SLOCK 378302ff05a9Svsakar * to avoid i_rwlock, ufs_lockfs_begin deadlock. If deadlock 378402ff05a9Svsakar * possible, retries the operation. 378502ff05a9Svsakar */ 378602ff05a9Svsakar ufs_tryirwlock_trans(&ip->i_rwlock, RW_WRITER, TOP_MKDIR, retry); 378702ff05a9Svsakar if (indeadlock) 378802ff05a9Svsakar goto again; 37897c478bd9Sstevel@tonic-gate 37907c478bd9Sstevel@tonic-gate error = ufs_direnter_cm(ip, dirname, DE_MKDIR, vap, &xip, cr, 37917c478bd9Sstevel@tonic-gate (retry ? IQUIET : 0)); 379202ff05a9Svsakar if (error == EAGAIN) { 379302ff05a9Svsakar if (ulp) { 379402ff05a9Svsakar TRANS_END_CSYNC(ufsvfsp, error, issync, TOP_MKDIR, 379502ff05a9Svsakar trans_size); 379602ff05a9Svsakar ufs_lockfs_end(ulp); 379702ff05a9Svsakar } 379802ff05a9Svsakar goto again; 379902ff05a9Svsakar } 38007c478bd9Sstevel@tonic-gate 38017c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 38027c478bd9Sstevel@tonic-gate if (error == 0) { 38037c478bd9Sstevel@tonic-gate ip = xip; 38047c478bd9Sstevel@tonic-gate *vpp = ITOV(ip); 38057c478bd9Sstevel@tonic-gate } else if (error == EEXIST) 38067c478bd9Sstevel@tonic-gate VN_RELE(ITOV(xip)); 38077c478bd9Sstevel@tonic-gate 38087c478bd9Sstevel@tonic-gate if (ulp) { 38097c478bd9Sstevel@tonic-gate int terr = 0; 38107c478bd9Sstevel@tonic-gate TRANS_END_CSYNC(ufsvfsp, terr, issync, TOP_MKDIR, trans_size); 38117c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 38127c478bd9Sstevel@tonic-gate if (error == 0) 38137c478bd9Sstevel@tonic-gate error = terr; 38147c478bd9Sstevel@tonic-gate } 38157c478bd9Sstevel@tonic-gate out: 38167c478bd9Sstevel@tonic-gate if ((error == ENOSPC) && retry && TRANS_ISTRANS(ufsvfsp)) { 38177c478bd9Sstevel@tonic-gate ufs_delete_drain_wait(ufsvfsp, 1); 38187c478bd9Sstevel@tonic-gate retry = 0; 38197c478bd9Sstevel@tonic-gate goto again; 38207c478bd9Sstevel@tonic-gate } 38217c478bd9Sstevel@tonic-gate 38227c478bd9Sstevel@tonic-gate return (error); 38237c478bd9Sstevel@tonic-gate } 38247c478bd9Sstevel@tonic-gate 38257c478bd9Sstevel@tonic-gate /*ARGSUSED*/ 38267c478bd9Sstevel@tonic-gate static int 3827da6c28aaSamw ufs_rmdir(struct vnode *vp, char *nm, struct vnode *cdir, struct cred *cr, 3828da6c28aaSamw caller_context_t *ct, int flags) 38297c478bd9Sstevel@tonic-gate { 38307c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 38317c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp = ip->i_ufsvfs; 38327c478bd9Sstevel@tonic-gate struct ulockfs *ulp; 38337c478bd9Sstevel@tonic-gate vnode_t *rmvp = NULL; /* Vnode of removed directory */ 38347c478bd9Sstevel@tonic-gate int error; 38357c478bd9Sstevel@tonic-gate int issync; 383602ff05a9Svsakar int trans_size; 383702ff05a9Svsakar int indeadlock; 38387c478bd9Sstevel@tonic-gate 38397c478bd9Sstevel@tonic-gate /* 38407c478bd9Sstevel@tonic-gate * don't let the delete queue get too long 38417c478bd9Sstevel@tonic-gate */ 38427c478bd9Sstevel@tonic-gate if (ufsvfsp == NULL) { 38437c478bd9Sstevel@tonic-gate error = EIO; 38447c478bd9Sstevel@tonic-gate goto out; 38457c478bd9Sstevel@tonic-gate } 38467c478bd9Sstevel@tonic-gate if (ufsvfsp->vfs_delete.uq_ne > ufs_idle_max) 38477c478bd9Sstevel@tonic-gate ufs_delete_drain(vp->v_vfsp, 1, 1); 38487c478bd9Sstevel@tonic-gate 38499b5097eeSOwen Roberts error = ufs_eventlookup(vp, nm, cr, &rmvp); 38509b5097eeSOwen Roberts if (rmvp != NULL) { 38519b5097eeSOwen Roberts /* Only send the event if there were no errors */ 38529b5097eeSOwen Roberts if (error == 0) 38539b5097eeSOwen Roberts vnevent_rmdir(rmvp, vp, nm, ct); 38549b5097eeSOwen Roberts VN_RELE(rmvp); 38559b5097eeSOwen Roberts } 38569b5097eeSOwen Roberts 385702ff05a9Svsakar retry_rmdir: 38587c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_RMDIR_MASK); 38597c478bd9Sstevel@tonic-gate if (error) 38607c478bd9Sstevel@tonic-gate goto out; 38617c478bd9Sstevel@tonic-gate 38627c478bd9Sstevel@tonic-gate if (ulp) 386302ff05a9Svsakar TRANS_BEGIN_CSYNC(ufsvfsp, issync, TOP_RMDIR, 386402ff05a9Svsakar trans_size = TOP_RMDIR_SIZE); 38657c478bd9Sstevel@tonic-gate 386602ff05a9Svsakar /* 386702ff05a9Svsakar * ufs_tryirwlock_trans uses rw_tryenter and checks for SLOCK 386802ff05a9Svsakar * to avoid i_rwlock, ufs_lockfs_begin deadlock. If deadlock 386902ff05a9Svsakar * possible, retries the operation. 387002ff05a9Svsakar */ 387102ff05a9Svsakar ufs_tryirwlock_trans(&ip->i_rwlock, RW_WRITER, TOP_RMDIR, retry); 387202ff05a9Svsakar if (indeadlock) 387302ff05a9Svsakar goto retry_rmdir; 38749b5097eeSOwen Roberts error = ufs_dirremove(ip, nm, (struct inode *)0, cdir, DR_RMDIR, cr); 38759b5097eeSOwen Roberts 38767c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 38777c478bd9Sstevel@tonic-gate 38787c478bd9Sstevel@tonic-gate if (ulp) { 38797c478bd9Sstevel@tonic-gate TRANS_END_CSYNC(ufsvfsp, error, issync, TOP_RMDIR, 388002ff05a9Svsakar trans_size); 38817c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 38827c478bd9Sstevel@tonic-gate } 38837c478bd9Sstevel@tonic-gate 38847c478bd9Sstevel@tonic-gate out: 38857c478bd9Sstevel@tonic-gate return (error); 38867c478bd9Sstevel@tonic-gate } 38877c478bd9Sstevel@tonic-gate 38887c478bd9Sstevel@tonic-gate /* ARGSUSED */ 38897c478bd9Sstevel@tonic-gate static int 38907c478bd9Sstevel@tonic-gate ufs_readdir( 38917c478bd9Sstevel@tonic-gate struct vnode *vp, 38927c478bd9Sstevel@tonic-gate struct uio *uiop, 38937c478bd9Sstevel@tonic-gate struct cred *cr, 3894da6c28aaSamw int *eofp, 3895da6c28aaSamw caller_context_t *ct, 3896da6c28aaSamw int flags) 38977c478bd9Sstevel@tonic-gate { 38987c478bd9Sstevel@tonic-gate struct iovec *iovp; 38997c478bd9Sstevel@tonic-gate struct inode *ip; 39007c478bd9Sstevel@tonic-gate struct direct *idp; 39017c478bd9Sstevel@tonic-gate struct dirent64 *odp; 39027c478bd9Sstevel@tonic-gate struct fbuf *fbp; 39037c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp; 39047c478bd9Sstevel@tonic-gate struct ulockfs *ulp; 39057c478bd9Sstevel@tonic-gate caddr_t outbuf; 39067c478bd9Sstevel@tonic-gate size_t bufsize; 39077c478bd9Sstevel@tonic-gate uint_t offset; 39087c478bd9Sstevel@tonic-gate uint_t bytes_wanted, total_bytes_wanted; 39097c478bd9Sstevel@tonic-gate int incount = 0; 39107c478bd9Sstevel@tonic-gate int outcount = 0; 39117c478bd9Sstevel@tonic-gate int error; 39127c478bd9Sstevel@tonic-gate 39137c478bd9Sstevel@tonic-gate ip = VTOI(vp); 39147c478bd9Sstevel@tonic-gate ASSERT(RW_READ_HELD(&ip->i_rwlock)); 39157c478bd9Sstevel@tonic-gate 39167c478bd9Sstevel@tonic-gate if (uiop->uio_loffset >= MAXOFF32_T) { 39177c478bd9Sstevel@tonic-gate if (eofp) 39187c478bd9Sstevel@tonic-gate *eofp = 1; 39197c478bd9Sstevel@tonic-gate return (0); 39207c478bd9Sstevel@tonic-gate } 39217c478bd9Sstevel@tonic-gate 39227c478bd9Sstevel@tonic-gate /* 39237c478bd9Sstevel@tonic-gate * Check if we have been called with a valid iov_len 39247c478bd9Sstevel@tonic-gate * and bail out if not, otherwise we may potentially loop 39257c478bd9Sstevel@tonic-gate * forever further down. 39267c478bd9Sstevel@tonic-gate */ 39277c478bd9Sstevel@tonic-gate if (uiop->uio_iov->iov_len <= 0) { 39287c478bd9Sstevel@tonic-gate error = EINVAL; 39297c478bd9Sstevel@tonic-gate goto out; 39307c478bd9Sstevel@tonic-gate } 39317c478bd9Sstevel@tonic-gate 39327c478bd9Sstevel@tonic-gate /* 39337c478bd9Sstevel@tonic-gate * Large Files: When we come here we are guaranteed that 39347c478bd9Sstevel@tonic-gate * uio_offset can be used safely. The high word is zero. 39357c478bd9Sstevel@tonic-gate */ 39367c478bd9Sstevel@tonic-gate 39377c478bd9Sstevel@tonic-gate ufsvfsp = ip->i_ufsvfs; 39387c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_READDIR_MASK); 39397c478bd9Sstevel@tonic-gate if (error) 39407c478bd9Sstevel@tonic-gate goto out; 39417c478bd9Sstevel@tonic-gate 39427c478bd9Sstevel@tonic-gate iovp = uiop->uio_iov; 39437c478bd9Sstevel@tonic-gate total_bytes_wanted = iovp->iov_len; 39447c478bd9Sstevel@tonic-gate 39457c478bd9Sstevel@tonic-gate /* Large Files: directory files should not be "large" */ 39467c478bd9Sstevel@tonic-gate 39477c478bd9Sstevel@tonic-gate ASSERT(ip->i_size <= MAXOFF32_T); 39487c478bd9Sstevel@tonic-gate 39497c478bd9Sstevel@tonic-gate /* Force offset to be valid (to guard against bogus lseek() values) */ 39507c478bd9Sstevel@tonic-gate offset = (uint_t)uiop->uio_offset & ~(DIRBLKSIZ - 1); 39517c478bd9Sstevel@tonic-gate 39527c478bd9Sstevel@tonic-gate /* Quit if at end of file or link count of zero (posix) */ 39537c478bd9Sstevel@tonic-gate if (offset >= (uint_t)ip->i_size || ip->i_nlink <= 0) { 39547c478bd9Sstevel@tonic-gate if (eofp) 39557c478bd9Sstevel@tonic-gate *eofp = 1; 39567c478bd9Sstevel@tonic-gate error = 0; 39577c478bd9Sstevel@tonic-gate goto unlock; 39587c478bd9Sstevel@tonic-gate } 39597c478bd9Sstevel@tonic-gate 39607c478bd9Sstevel@tonic-gate /* 39617c478bd9Sstevel@tonic-gate * Get space to change directory entries into fs independent format. 39627c478bd9Sstevel@tonic-gate * Do fast alloc for the most commonly used-request size (filesystem 39637c478bd9Sstevel@tonic-gate * block size). 39647c478bd9Sstevel@tonic-gate */ 39657c478bd9Sstevel@tonic-gate if (uiop->uio_segflg != UIO_SYSSPACE || uiop->uio_iovcnt != 1) { 39667c478bd9Sstevel@tonic-gate bufsize = total_bytes_wanted; 39677c478bd9Sstevel@tonic-gate outbuf = kmem_alloc(bufsize, KM_SLEEP); 39687c478bd9Sstevel@tonic-gate odp = (struct dirent64 *)outbuf; 39697c478bd9Sstevel@tonic-gate } else { 39707c478bd9Sstevel@tonic-gate bufsize = total_bytes_wanted; 39717c478bd9Sstevel@tonic-gate odp = (struct dirent64 *)iovp->iov_base; 39727c478bd9Sstevel@tonic-gate } 39737c478bd9Sstevel@tonic-gate 39747c478bd9Sstevel@tonic-gate nextblk: 39757c478bd9Sstevel@tonic-gate bytes_wanted = total_bytes_wanted; 39767c478bd9Sstevel@tonic-gate 39777c478bd9Sstevel@tonic-gate /* Truncate request to file size */ 39787c478bd9Sstevel@tonic-gate if (offset + bytes_wanted > (int)ip->i_size) 39797c478bd9Sstevel@tonic-gate bytes_wanted = (int)(ip->i_size - offset); 39807c478bd9Sstevel@tonic-gate 39817c478bd9Sstevel@tonic-gate /* Comply with MAXBSIZE boundary restrictions of fbread() */ 39827c478bd9Sstevel@tonic-gate if ((offset & MAXBOFFSET) + bytes_wanted > MAXBSIZE) 39837c478bd9Sstevel@tonic-gate bytes_wanted = MAXBSIZE - (offset & MAXBOFFSET); 39847c478bd9Sstevel@tonic-gate 39857c478bd9Sstevel@tonic-gate /* 39867c478bd9Sstevel@tonic-gate * Read in the next chunk. 39877c478bd9Sstevel@tonic-gate * We are still holding the i_rwlock. 39887c478bd9Sstevel@tonic-gate */ 39897c478bd9Sstevel@tonic-gate error = fbread(vp, (offset_t)offset, bytes_wanted, S_OTHER, &fbp); 39907c478bd9Sstevel@tonic-gate 39917c478bd9Sstevel@tonic-gate if (error) 39927c478bd9Sstevel@tonic-gate goto update_inode; 39937c478bd9Sstevel@tonic-gate if (!ULOCKFS_IS_NOIACC(ITOUL(ip)) && (ip->i_fs->fs_ronly == 0) && 39947c478bd9Sstevel@tonic-gate (!ufsvfsp->vfs_noatime)) { 39957c478bd9Sstevel@tonic-gate ip->i_flag |= IACC; 39967c478bd9Sstevel@tonic-gate } 39977c478bd9Sstevel@tonic-gate incount = 0; 39987c478bd9Sstevel@tonic-gate idp = (struct direct *)fbp->fb_addr; 399980d34432Sfrankho if (idp->d_ino == 0 && idp->d_reclen == 0 && idp->d_namlen == 0) { 40007c478bd9Sstevel@tonic-gate cmn_err(CE_WARN, "ufs_readdir: bad dir, inumber = %llu, " 40017c478bd9Sstevel@tonic-gate "fs = %s\n", 40027c478bd9Sstevel@tonic-gate (u_longlong_t)ip->i_number, ufsvfsp->vfs_fs->fs_fsmnt); 40037c478bd9Sstevel@tonic-gate fbrelse(fbp, S_OTHER); 40047c478bd9Sstevel@tonic-gate error = ENXIO; 40057c478bd9Sstevel@tonic-gate goto update_inode; 40067c478bd9Sstevel@tonic-gate } 40077c478bd9Sstevel@tonic-gate /* Transform to file-system independent format */ 40087c478bd9Sstevel@tonic-gate while (incount < bytes_wanted) { 40097c478bd9Sstevel@tonic-gate /* 40107c478bd9Sstevel@tonic-gate * If the current directory entry is mangled, then skip 40117c478bd9Sstevel@tonic-gate * to the next block. It would be nice to set the FSBAD 40127c478bd9Sstevel@tonic-gate * flag in the super-block so that a fsck is forced on 40137c478bd9Sstevel@tonic-gate * next reboot, but locking is a problem. 40147c478bd9Sstevel@tonic-gate */ 40157c478bd9Sstevel@tonic-gate if (idp->d_reclen & 0x3) { 40167c478bd9Sstevel@tonic-gate offset = (offset + DIRBLKSIZ) & ~(DIRBLKSIZ-1); 40177c478bd9Sstevel@tonic-gate break; 40187c478bd9Sstevel@tonic-gate } 40197c478bd9Sstevel@tonic-gate 40207c478bd9Sstevel@tonic-gate /* Skip to requested offset and skip empty entries */ 40217c478bd9Sstevel@tonic-gate if (idp->d_ino != 0 && offset >= (uint_t)uiop->uio_offset) { 40227c478bd9Sstevel@tonic-gate ushort_t this_reclen = 40237c478bd9Sstevel@tonic-gate DIRENT64_RECLEN(idp->d_namlen); 40247c478bd9Sstevel@tonic-gate /* Buffer too small for any entries */ 40257c478bd9Sstevel@tonic-gate if (!outcount && this_reclen > bufsize) { 40267c478bd9Sstevel@tonic-gate fbrelse(fbp, S_OTHER); 40277c478bd9Sstevel@tonic-gate error = EINVAL; 40287c478bd9Sstevel@tonic-gate goto update_inode; 40297c478bd9Sstevel@tonic-gate } 40307c478bd9Sstevel@tonic-gate /* If would overrun the buffer, quit */ 40317c478bd9Sstevel@tonic-gate if (outcount + this_reclen > bufsize) { 40327c478bd9Sstevel@tonic-gate break; 40337c478bd9Sstevel@tonic-gate } 40347c478bd9Sstevel@tonic-gate /* Take this entry */ 40357c478bd9Sstevel@tonic-gate odp->d_ino = (ino64_t)idp->d_ino; 40367c478bd9Sstevel@tonic-gate odp->d_reclen = (ushort_t)this_reclen; 40377c478bd9Sstevel@tonic-gate odp->d_off = (offset_t)(offset + idp->d_reclen); 40387c478bd9Sstevel@tonic-gate 40397c478bd9Sstevel@tonic-gate /* use strncpy(9f) to zero out uninitialized bytes */ 40407c478bd9Sstevel@tonic-gate 40417c478bd9Sstevel@tonic-gate ASSERT(strlen(idp->d_name) + 1 <= 40427c478bd9Sstevel@tonic-gate DIRENT64_NAMELEN(this_reclen)); 40437c478bd9Sstevel@tonic-gate (void) strncpy(odp->d_name, idp->d_name, 40447c478bd9Sstevel@tonic-gate DIRENT64_NAMELEN(this_reclen)); 40457c478bd9Sstevel@tonic-gate outcount += odp->d_reclen; 404680d34432Sfrankho odp = (struct dirent64 *) 404780d34432Sfrankho ((intptr_t)odp + odp->d_reclen); 40487c478bd9Sstevel@tonic-gate ASSERT(outcount <= bufsize); 40497c478bd9Sstevel@tonic-gate } 40507c478bd9Sstevel@tonic-gate if (idp->d_reclen) { 40517c478bd9Sstevel@tonic-gate incount += idp->d_reclen; 40527c478bd9Sstevel@tonic-gate offset += idp->d_reclen; 40537c478bd9Sstevel@tonic-gate idp = (struct direct *)((intptr_t)idp + idp->d_reclen); 40547c478bd9Sstevel@tonic-gate } else { 40557c478bd9Sstevel@tonic-gate offset = (offset + DIRBLKSIZ) & ~(DIRBLKSIZ-1); 40567c478bd9Sstevel@tonic-gate break; 40577c478bd9Sstevel@tonic-gate } 40587c478bd9Sstevel@tonic-gate } 40597c478bd9Sstevel@tonic-gate /* Release the chunk */ 40607c478bd9Sstevel@tonic-gate fbrelse(fbp, S_OTHER); 40617c478bd9Sstevel@tonic-gate 40627c478bd9Sstevel@tonic-gate /* Read whole block, but got no entries, read another if not eof */ 40637c478bd9Sstevel@tonic-gate 40647c478bd9Sstevel@tonic-gate /* 40657c478bd9Sstevel@tonic-gate * Large Files: casting i_size to int here is not a problem 40667c478bd9Sstevel@tonic-gate * because directory sizes are always less than MAXOFF32_T. 40677c478bd9Sstevel@tonic-gate * See assertion above. 40687c478bd9Sstevel@tonic-gate */ 40697c478bd9Sstevel@tonic-gate 40707c478bd9Sstevel@tonic-gate if (offset < (int)ip->i_size && !outcount) 40717c478bd9Sstevel@tonic-gate goto nextblk; 40727c478bd9Sstevel@tonic-gate 40737c478bd9Sstevel@tonic-gate /* Copy out the entry data */ 40747c478bd9Sstevel@tonic-gate if (uiop->uio_segflg == UIO_SYSSPACE && uiop->uio_iovcnt == 1) { 40757c478bd9Sstevel@tonic-gate iovp->iov_base += outcount; 40767c478bd9Sstevel@tonic-gate iovp->iov_len -= outcount; 40777c478bd9Sstevel@tonic-gate uiop->uio_resid -= outcount; 40787c478bd9Sstevel@tonic-gate uiop->uio_offset = offset; 40797c478bd9Sstevel@tonic-gate } else if ((error = uiomove(outbuf, (long)outcount, UIO_READ, 40807c478bd9Sstevel@tonic-gate uiop)) == 0) 40817c478bd9Sstevel@tonic-gate uiop->uio_offset = offset; 40827c478bd9Sstevel@tonic-gate update_inode: 40837c478bd9Sstevel@tonic-gate ITIMES(ip); 40847c478bd9Sstevel@tonic-gate if (uiop->uio_segflg != UIO_SYSSPACE || uiop->uio_iovcnt != 1) 40857c478bd9Sstevel@tonic-gate kmem_free(outbuf, bufsize); 40867c478bd9Sstevel@tonic-gate 40877c478bd9Sstevel@tonic-gate if (eofp && error == 0) 40887c478bd9Sstevel@tonic-gate *eofp = (uiop->uio_offset >= (int)ip->i_size); 40897c478bd9Sstevel@tonic-gate unlock: 40907c478bd9Sstevel@tonic-gate if (ulp) { 40917c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 40927c478bd9Sstevel@tonic-gate } 40937c478bd9Sstevel@tonic-gate out: 40947c478bd9Sstevel@tonic-gate return (error); 40957c478bd9Sstevel@tonic-gate } 40967c478bd9Sstevel@tonic-gate 40977c478bd9Sstevel@tonic-gate /*ARGSUSED*/ 40987c478bd9Sstevel@tonic-gate static int 40997c478bd9Sstevel@tonic-gate ufs_symlink( 41007c478bd9Sstevel@tonic-gate struct vnode *dvp, /* ptr to parent dir vnode */ 41017c478bd9Sstevel@tonic-gate char *linkname, /* name of symbolic link */ 41027c478bd9Sstevel@tonic-gate struct vattr *vap, /* attributes */ 41037c478bd9Sstevel@tonic-gate char *target, /* target path */ 4104da6c28aaSamw struct cred *cr, /* user credentials */ 4105da6c28aaSamw caller_context_t *ct, 4106da6c28aaSamw int flags) 41077c478bd9Sstevel@tonic-gate { 41087c478bd9Sstevel@tonic-gate struct inode *ip, *dip = VTOI(dvp); 41097c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp = dip->i_ufsvfs; 41107c478bd9Sstevel@tonic-gate struct ulockfs *ulp; 41117c478bd9Sstevel@tonic-gate int error; 41127c478bd9Sstevel@tonic-gate int issync; 41137c478bd9Sstevel@tonic-gate int trans_size; 41147c478bd9Sstevel@tonic-gate int residual; 41157c478bd9Sstevel@tonic-gate int ioflag; 41167c478bd9Sstevel@tonic-gate int retry = 1; 41177c478bd9Sstevel@tonic-gate 41187c478bd9Sstevel@tonic-gate /* 41197c478bd9Sstevel@tonic-gate * No symlinks in attrdirs at this time 41207c478bd9Sstevel@tonic-gate */ 41217c478bd9Sstevel@tonic-gate if ((VTOI(dvp)->i_mode & IFMT) == IFATTRDIR) 41227c478bd9Sstevel@tonic-gate return (EINVAL); 41237c478bd9Sstevel@tonic-gate 41247c478bd9Sstevel@tonic-gate again: 41257c478bd9Sstevel@tonic-gate ip = (struct inode *)NULL; 41267c478bd9Sstevel@tonic-gate vap->va_type = VLNK; 41277c478bd9Sstevel@tonic-gate vap->va_rdev = 0; 41287c478bd9Sstevel@tonic-gate 41297c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_SYMLINK_MASK); 41307c478bd9Sstevel@tonic-gate if (error) 41317c478bd9Sstevel@tonic-gate goto out; 41327c478bd9Sstevel@tonic-gate 41337c478bd9Sstevel@tonic-gate if (ulp) 41347c478bd9Sstevel@tonic-gate TRANS_BEGIN_CSYNC(ufsvfsp, issync, TOP_SYMLINK, 41357c478bd9Sstevel@tonic-gate trans_size = (int)TOP_SYMLINK_SIZE(dip)); 41367c478bd9Sstevel@tonic-gate 41377c478bd9Sstevel@tonic-gate /* 41387c478bd9Sstevel@tonic-gate * We must create the inode before the directory entry, to avoid 41397c478bd9Sstevel@tonic-gate * racing with readlink(). ufs_dirmakeinode requires that we 41407c478bd9Sstevel@tonic-gate * hold the quota lock as reader, and directory locks as writer. 41417c478bd9Sstevel@tonic-gate */ 41427c478bd9Sstevel@tonic-gate 41437c478bd9Sstevel@tonic-gate rw_enter(&dip->i_rwlock, RW_WRITER); 41447c478bd9Sstevel@tonic-gate rw_enter(&ufsvfsp->vfs_dqrwlock, RW_READER); 41457c478bd9Sstevel@tonic-gate rw_enter(&dip->i_contents, RW_WRITER); 41467c478bd9Sstevel@tonic-gate 41477c478bd9Sstevel@tonic-gate /* 41487c478bd9Sstevel@tonic-gate * Suppress any out of inodes messages if we will retry on 41497c478bd9Sstevel@tonic-gate * ENOSP 41507c478bd9Sstevel@tonic-gate */ 41517c478bd9Sstevel@tonic-gate if (retry) 41527c478bd9Sstevel@tonic-gate dip->i_flag |= IQUIET; 41537c478bd9Sstevel@tonic-gate 41547c478bd9Sstevel@tonic-gate error = ufs_dirmakeinode(dip, &ip, vap, DE_SYMLINK, cr); 41557c478bd9Sstevel@tonic-gate 41567c478bd9Sstevel@tonic-gate dip->i_flag &= ~IQUIET; 41577c478bd9Sstevel@tonic-gate 41587c478bd9Sstevel@tonic-gate rw_exit(&dip->i_contents); 41597c478bd9Sstevel@tonic-gate rw_exit(&ufsvfsp->vfs_dqrwlock); 41607c478bd9Sstevel@tonic-gate rw_exit(&dip->i_rwlock); 41617c478bd9Sstevel@tonic-gate 41627c478bd9Sstevel@tonic-gate if (error) 41637c478bd9Sstevel@tonic-gate goto unlock; 41647c478bd9Sstevel@tonic-gate 41657c478bd9Sstevel@tonic-gate /* 41667c478bd9Sstevel@tonic-gate * OK. The inode has been created. Write out the data of the 41677c478bd9Sstevel@tonic-gate * symbolic link. Since symbolic links are metadata, and should 41687c478bd9Sstevel@tonic-gate * remain consistent across a system crash, we need to force the 41697c478bd9Sstevel@tonic-gate * data out synchronously. 41707c478bd9Sstevel@tonic-gate * 41717c478bd9Sstevel@tonic-gate * (This is a change from the semantics in earlier releases, which 41727c478bd9Sstevel@tonic-gate * only created symbolic links synchronously if the semi-documented 41737c478bd9Sstevel@tonic-gate * 'syncdir' option was set, or if we were being invoked by the NFS 41747c478bd9Sstevel@tonic-gate * server, which requires symbolic links to be created synchronously.) 41757c478bd9Sstevel@tonic-gate * 41767c478bd9Sstevel@tonic-gate * We need to pass in a pointer for the residual length; otherwise 41777c478bd9Sstevel@tonic-gate * ufs_rdwri() will always return EIO if it can't write the data, 41787c478bd9Sstevel@tonic-gate * even if the error was really ENOSPC or EDQUOT. 41797c478bd9Sstevel@tonic-gate */ 41807c478bd9Sstevel@tonic-gate 41817c478bd9Sstevel@tonic-gate ioflag = FWRITE | FDSYNC; 41827c478bd9Sstevel@tonic-gate residual = 0; 41837c478bd9Sstevel@tonic-gate 41847c478bd9Sstevel@tonic-gate rw_enter(&ufsvfsp->vfs_dqrwlock, RW_READER); 41857c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_WRITER); 41867c478bd9Sstevel@tonic-gate 41877c478bd9Sstevel@tonic-gate /* 41887c478bd9Sstevel@tonic-gate * Suppress file system full messages if we will retry 41897c478bd9Sstevel@tonic-gate */ 41907c478bd9Sstevel@tonic-gate if (retry) 41917c478bd9Sstevel@tonic-gate ip->i_flag |= IQUIET; 41927c478bd9Sstevel@tonic-gate 41937c478bd9Sstevel@tonic-gate error = ufs_rdwri(UIO_WRITE, ioflag, ip, target, strlen(target), 41947c478bd9Sstevel@tonic-gate (offset_t)0, UIO_SYSSPACE, &residual, cr); 41957c478bd9Sstevel@tonic-gate 41967c478bd9Sstevel@tonic-gate ip->i_flag &= ~IQUIET; 41977c478bd9Sstevel@tonic-gate 41987c478bd9Sstevel@tonic-gate if (error) { 41997c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 42007c478bd9Sstevel@tonic-gate rw_exit(&ufsvfsp->vfs_dqrwlock); 42017c478bd9Sstevel@tonic-gate goto remove; 42027c478bd9Sstevel@tonic-gate } 42037c478bd9Sstevel@tonic-gate 42047c478bd9Sstevel@tonic-gate /* 42057c478bd9Sstevel@tonic-gate * If the link's data is small enough, we can cache it in the inode. 42067c478bd9Sstevel@tonic-gate * This is a "fast symbolic link". We don't use the first direct 42077c478bd9Sstevel@tonic-gate * block because that's actually used to point at the symbolic link's 42087c478bd9Sstevel@tonic-gate * contents on disk; but we know that none of the other direct or 42097c478bd9Sstevel@tonic-gate * indirect blocks can be used because symbolic links are restricted 42107c478bd9Sstevel@tonic-gate * to be smaller than a file system block. 42117c478bd9Sstevel@tonic-gate */ 42127c478bd9Sstevel@tonic-gate 42137c478bd9Sstevel@tonic-gate ASSERT(MAXPATHLEN <= VBSIZE(ITOV(ip))); 42147c478bd9Sstevel@tonic-gate 42157c478bd9Sstevel@tonic-gate if (ip->i_size > 0 && ip->i_size <= FSL_SIZE) { 42167c478bd9Sstevel@tonic-gate if (kcopy(target, &ip->i_db[1], ip->i_size) == 0) { 42177c478bd9Sstevel@tonic-gate ip->i_flag |= IFASTSYMLNK; 42187c478bd9Sstevel@tonic-gate } else { 42197c478bd9Sstevel@tonic-gate int i; 42207c478bd9Sstevel@tonic-gate /* error, clear garbage left behind */ 42217c478bd9Sstevel@tonic-gate for (i = 1; i < NDADDR; i++) 42227c478bd9Sstevel@tonic-gate ip->i_db[i] = 0; 42237c478bd9Sstevel@tonic-gate for (i = 0; i < NIADDR; i++) 42247c478bd9Sstevel@tonic-gate ip->i_ib[i] = 0; 42257c478bd9Sstevel@tonic-gate } 42267c478bd9Sstevel@tonic-gate } 42277c478bd9Sstevel@tonic-gate 42287c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 42297c478bd9Sstevel@tonic-gate rw_exit(&ufsvfsp->vfs_dqrwlock); 42307c478bd9Sstevel@tonic-gate 42317c478bd9Sstevel@tonic-gate /* 42327c478bd9Sstevel@tonic-gate * OK. We've successfully created the symbolic link. All that 42337c478bd9Sstevel@tonic-gate * remains is to insert it into the appropriate directory. 42347c478bd9Sstevel@tonic-gate */ 42357c478bd9Sstevel@tonic-gate 42367c478bd9Sstevel@tonic-gate rw_enter(&dip->i_rwlock, RW_WRITER); 42379b5097eeSOwen Roberts error = ufs_direnter_lr(dip, linkname, DE_SYMLINK, NULL, ip, cr); 42387c478bd9Sstevel@tonic-gate rw_exit(&dip->i_rwlock); 42397c478bd9Sstevel@tonic-gate 42407c478bd9Sstevel@tonic-gate /* 42417c478bd9Sstevel@tonic-gate * Fall through into remove-on-error code. We're either done, or we 42427c478bd9Sstevel@tonic-gate * need to remove the inode (if we couldn't insert it). 42437c478bd9Sstevel@tonic-gate */ 42447c478bd9Sstevel@tonic-gate 42457c478bd9Sstevel@tonic-gate remove: 42467c478bd9Sstevel@tonic-gate if (error && (ip != NULL)) { 42477c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_WRITER); 42487c478bd9Sstevel@tonic-gate ip->i_nlink--; 42497c478bd9Sstevel@tonic-gate ip->i_flag |= ICHG; 42507c478bd9Sstevel@tonic-gate ip->i_seq++; 42517c478bd9Sstevel@tonic-gate ufs_setreclaim(ip); 42527c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 42537c478bd9Sstevel@tonic-gate } 42547c478bd9Sstevel@tonic-gate 42557c478bd9Sstevel@tonic-gate unlock: 42567c478bd9Sstevel@tonic-gate if (ip != NULL) 42577c478bd9Sstevel@tonic-gate VN_RELE(ITOV(ip)); 42587c478bd9Sstevel@tonic-gate 42597c478bd9Sstevel@tonic-gate if (ulp) { 42607c478bd9Sstevel@tonic-gate int terr = 0; 42617c478bd9Sstevel@tonic-gate 42627c478bd9Sstevel@tonic-gate TRANS_END_CSYNC(ufsvfsp, terr, issync, TOP_SYMLINK, 42637c478bd9Sstevel@tonic-gate trans_size); 42647c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 42657c478bd9Sstevel@tonic-gate if (error == 0) 42667c478bd9Sstevel@tonic-gate error = terr; 42677c478bd9Sstevel@tonic-gate } 42687c478bd9Sstevel@tonic-gate 42697c478bd9Sstevel@tonic-gate /* 42707c478bd9Sstevel@tonic-gate * We may have failed due to lack of an inode or of a block to 42717c478bd9Sstevel@tonic-gate * store the target in. Try flushing the delete queue to free 42727c478bd9Sstevel@tonic-gate * logically-available things up and try again. 42737c478bd9Sstevel@tonic-gate */ 42747c478bd9Sstevel@tonic-gate if ((error == ENOSPC) && retry && TRANS_ISTRANS(ufsvfsp)) { 42757c478bd9Sstevel@tonic-gate ufs_delete_drain_wait(ufsvfsp, 1); 42767c478bd9Sstevel@tonic-gate retry = 0; 42777c478bd9Sstevel@tonic-gate goto again; 42787c478bd9Sstevel@tonic-gate } 42797c478bd9Sstevel@tonic-gate 42807c478bd9Sstevel@tonic-gate out: 42817c478bd9Sstevel@tonic-gate return (error); 42827c478bd9Sstevel@tonic-gate } 42837c478bd9Sstevel@tonic-gate 42847c478bd9Sstevel@tonic-gate /* 42857c478bd9Sstevel@tonic-gate * Ufs specific routine used to do ufs io. 42867c478bd9Sstevel@tonic-gate */ 42877c478bd9Sstevel@tonic-gate int 42887c478bd9Sstevel@tonic-gate ufs_rdwri(enum uio_rw rw, int ioflag, struct inode *ip, caddr_t base, 42897c478bd9Sstevel@tonic-gate ssize_t len, offset_t offset, enum uio_seg seg, int *aresid, 42907c478bd9Sstevel@tonic-gate struct cred *cr) 42917c478bd9Sstevel@tonic-gate { 42927c478bd9Sstevel@tonic-gate struct uio auio; 42937c478bd9Sstevel@tonic-gate struct iovec aiov; 42947c478bd9Sstevel@tonic-gate int error; 42957c478bd9Sstevel@tonic-gate 42967c478bd9Sstevel@tonic-gate ASSERT(RW_LOCK_HELD(&ip->i_contents)); 42977c478bd9Sstevel@tonic-gate 42987c478bd9Sstevel@tonic-gate bzero((caddr_t)&auio, sizeof (uio_t)); 42997c478bd9Sstevel@tonic-gate bzero((caddr_t)&aiov, sizeof (iovec_t)); 43007c478bd9Sstevel@tonic-gate 43017c478bd9Sstevel@tonic-gate aiov.iov_base = base; 43027c478bd9Sstevel@tonic-gate aiov.iov_len = len; 43037c478bd9Sstevel@tonic-gate auio.uio_iov = &aiov; 43047c478bd9Sstevel@tonic-gate auio.uio_iovcnt = 1; 43057c478bd9Sstevel@tonic-gate auio.uio_loffset = offset; 43067c478bd9Sstevel@tonic-gate auio.uio_segflg = (short)seg; 43077c478bd9Sstevel@tonic-gate auio.uio_resid = len; 43087c478bd9Sstevel@tonic-gate 43097c478bd9Sstevel@tonic-gate if (rw == UIO_WRITE) { 43107c478bd9Sstevel@tonic-gate auio.uio_fmode = FWRITE; 43117c478bd9Sstevel@tonic-gate auio.uio_extflg = UIO_COPY_DEFAULT; 43127c478bd9Sstevel@tonic-gate auio.uio_llimit = curproc->p_fsz_ctl; 43137c478bd9Sstevel@tonic-gate error = wrip(ip, &auio, ioflag, cr); 43147c478bd9Sstevel@tonic-gate } else { 43157c478bd9Sstevel@tonic-gate auio.uio_fmode = FREAD; 43167c478bd9Sstevel@tonic-gate auio.uio_extflg = UIO_COPY_CACHED; 43177c478bd9Sstevel@tonic-gate auio.uio_llimit = MAXOFFSET_T; 43187c478bd9Sstevel@tonic-gate error = rdip(ip, &auio, ioflag, cr); 43197c478bd9Sstevel@tonic-gate } 43207c478bd9Sstevel@tonic-gate 43217c478bd9Sstevel@tonic-gate if (aresid) { 43227c478bd9Sstevel@tonic-gate *aresid = auio.uio_resid; 43237c478bd9Sstevel@tonic-gate } else if (auio.uio_resid) { 43247c478bd9Sstevel@tonic-gate error = EIO; 43257c478bd9Sstevel@tonic-gate } 43267c478bd9Sstevel@tonic-gate return (error); 43277c478bd9Sstevel@tonic-gate } 43287c478bd9Sstevel@tonic-gate 4329da6c28aaSamw /*ARGSUSED*/ 43307c478bd9Sstevel@tonic-gate static int 4331da6c28aaSamw ufs_fid(struct vnode *vp, struct fid *fidp, caller_context_t *ct) 43327c478bd9Sstevel@tonic-gate { 43337c478bd9Sstevel@tonic-gate struct ufid *ufid; 43347c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 43357c478bd9Sstevel@tonic-gate 43367c478bd9Sstevel@tonic-gate if (ip->i_ufsvfs == NULL) 43377c478bd9Sstevel@tonic-gate return (EIO); 43387c478bd9Sstevel@tonic-gate 43397c478bd9Sstevel@tonic-gate if (fidp->fid_len < (sizeof (struct ufid) - sizeof (ushort_t))) { 43407c478bd9Sstevel@tonic-gate fidp->fid_len = sizeof (struct ufid) - sizeof (ushort_t); 43417c478bd9Sstevel@tonic-gate return (ENOSPC); 43427c478bd9Sstevel@tonic-gate } 43437c478bd9Sstevel@tonic-gate 43447c478bd9Sstevel@tonic-gate ufid = (struct ufid *)fidp; 43457c478bd9Sstevel@tonic-gate bzero((char *)ufid, sizeof (struct ufid)); 43467c478bd9Sstevel@tonic-gate ufid->ufid_len = sizeof (struct ufid) - sizeof (ushort_t); 43477c478bd9Sstevel@tonic-gate ufid->ufid_ino = ip->i_number; 43487c478bd9Sstevel@tonic-gate ufid->ufid_gen = ip->i_gen; 43497c478bd9Sstevel@tonic-gate 43507c478bd9Sstevel@tonic-gate return (0); 43517c478bd9Sstevel@tonic-gate } 43527c478bd9Sstevel@tonic-gate 43537c478bd9Sstevel@tonic-gate /* ARGSUSED2 */ 43547c478bd9Sstevel@tonic-gate static int 43557c478bd9Sstevel@tonic-gate ufs_rwlock(struct vnode *vp, int write_lock, caller_context_t *ctp) 43567c478bd9Sstevel@tonic-gate { 43577c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 43587c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp; 43597c478bd9Sstevel@tonic-gate int forcedirectio; 43607c478bd9Sstevel@tonic-gate 43617c478bd9Sstevel@tonic-gate /* 43627c478bd9Sstevel@tonic-gate * Read case is easy. 43637c478bd9Sstevel@tonic-gate */ 43647c478bd9Sstevel@tonic-gate if (!write_lock) { 43657c478bd9Sstevel@tonic-gate rw_enter(&ip->i_rwlock, RW_READER); 43667c478bd9Sstevel@tonic-gate return (V_WRITELOCK_FALSE); 43677c478bd9Sstevel@tonic-gate } 43687c478bd9Sstevel@tonic-gate 43697c478bd9Sstevel@tonic-gate /* 43707c478bd9Sstevel@tonic-gate * Caller has requested a writer lock, but that inhibits any 43717c478bd9Sstevel@tonic-gate * concurrency in the VOPs that follow. Acquire the lock shared 43727c478bd9Sstevel@tonic-gate * and defer exclusive access until it is known to be needed in 43737c478bd9Sstevel@tonic-gate * other VOP handlers. Some cases can be determined here. 43747c478bd9Sstevel@tonic-gate */ 43757c478bd9Sstevel@tonic-gate 43767c478bd9Sstevel@tonic-gate /* 43777c478bd9Sstevel@tonic-gate * If directio is not set, there is no chance of concurrency, 43787c478bd9Sstevel@tonic-gate * so just acquire the lock exclusive. Beware of a forced 43797c478bd9Sstevel@tonic-gate * unmount before looking at the mount option. 43807c478bd9Sstevel@tonic-gate */ 43817c478bd9Sstevel@tonic-gate ufsvfsp = ip->i_ufsvfs; 43827c478bd9Sstevel@tonic-gate forcedirectio = ufsvfsp ? ufsvfsp->vfs_forcedirectio : 0; 43837c478bd9Sstevel@tonic-gate if (!(ip->i_flag & IDIRECTIO || forcedirectio) || 43847c478bd9Sstevel@tonic-gate !ufs_allow_shared_writes) { 43857c478bd9Sstevel@tonic-gate rw_enter(&ip->i_rwlock, RW_WRITER); 43867c478bd9Sstevel@tonic-gate return (V_WRITELOCK_TRUE); 43877c478bd9Sstevel@tonic-gate } 43887c478bd9Sstevel@tonic-gate 43897c478bd9Sstevel@tonic-gate /* 43907c478bd9Sstevel@tonic-gate * Mandatory locking forces acquiring i_rwlock exclusive. 43917c478bd9Sstevel@tonic-gate */ 43927c478bd9Sstevel@tonic-gate if (MANDLOCK(vp, ip->i_mode)) { 43937c478bd9Sstevel@tonic-gate rw_enter(&ip->i_rwlock, RW_WRITER); 43947c478bd9Sstevel@tonic-gate return (V_WRITELOCK_TRUE); 43957c478bd9Sstevel@tonic-gate } 43967c478bd9Sstevel@tonic-gate 43977c478bd9Sstevel@tonic-gate /* 43987c478bd9Sstevel@tonic-gate * Acquire the lock shared in case a concurrent write follows. 43997c478bd9Sstevel@tonic-gate * Mandatory locking could have become enabled before the lock 44007c478bd9Sstevel@tonic-gate * was acquired. Re-check and upgrade if needed. 44017c478bd9Sstevel@tonic-gate */ 44027c478bd9Sstevel@tonic-gate rw_enter(&ip->i_rwlock, RW_READER); 44037c478bd9Sstevel@tonic-gate if (MANDLOCK(vp, ip->i_mode)) { 44047c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 44057c478bd9Sstevel@tonic-gate rw_enter(&ip->i_rwlock, RW_WRITER); 44067c478bd9Sstevel@tonic-gate return (V_WRITELOCK_TRUE); 44077c478bd9Sstevel@tonic-gate } 44087c478bd9Sstevel@tonic-gate return (V_WRITELOCK_FALSE); 44097c478bd9Sstevel@tonic-gate } 44107c478bd9Sstevel@tonic-gate 44117c478bd9Sstevel@tonic-gate /*ARGSUSED*/ 44127c478bd9Sstevel@tonic-gate static void 44137c478bd9Sstevel@tonic-gate ufs_rwunlock(struct vnode *vp, int write_lock, caller_context_t *ctp) 44147c478bd9Sstevel@tonic-gate { 44157c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 44167c478bd9Sstevel@tonic-gate 44177c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 44187c478bd9Sstevel@tonic-gate } 44197c478bd9Sstevel@tonic-gate 44207c478bd9Sstevel@tonic-gate /* ARGSUSED */ 44217c478bd9Sstevel@tonic-gate static int 4422da6c28aaSamw ufs_seek(struct vnode *vp, offset_t ooff, offset_t *noffp, 4423da6c28aaSamw caller_context_t *ct) 44247c478bd9Sstevel@tonic-gate { 44257c478bd9Sstevel@tonic-gate return ((*noffp < 0 || *noffp > MAXOFFSET_T) ? EINVAL : 0); 44267c478bd9Sstevel@tonic-gate } 44277c478bd9Sstevel@tonic-gate 44287c478bd9Sstevel@tonic-gate /* ARGSUSED */ 44297c478bd9Sstevel@tonic-gate static int 44307c478bd9Sstevel@tonic-gate ufs_frlock(struct vnode *vp, int cmd, struct flock64 *bfp, int flag, 4431da6c28aaSamw offset_t offset, struct flk_callback *flk_cbp, struct cred *cr, 4432da6c28aaSamw caller_context_t *ct) 44337c478bd9Sstevel@tonic-gate { 44347c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 44357c478bd9Sstevel@tonic-gate 44367c478bd9Sstevel@tonic-gate if (ip->i_ufsvfs == NULL) 44377c478bd9Sstevel@tonic-gate return (EIO); 44387c478bd9Sstevel@tonic-gate 44397c478bd9Sstevel@tonic-gate /* 44407c478bd9Sstevel@tonic-gate * If file is being mapped, disallow frlock. 44417c478bd9Sstevel@tonic-gate * XXX I am not holding tlock while checking i_mapcnt because the 44427c478bd9Sstevel@tonic-gate * current locking strategy drops all locks before calling fs_frlock. 44437c478bd9Sstevel@tonic-gate * So, mapcnt could change before we enter fs_frlock making is 44447c478bd9Sstevel@tonic-gate * meaningless to have held tlock in the first place. 44457c478bd9Sstevel@tonic-gate */ 44467c478bd9Sstevel@tonic-gate if (ip->i_mapcnt > 0 && MANDLOCK(vp, ip->i_mode)) 44477c478bd9Sstevel@tonic-gate return (EAGAIN); 4448da6c28aaSamw return (fs_frlock(vp, cmd, bfp, flag, offset, flk_cbp, cr, ct)); 44497c478bd9Sstevel@tonic-gate } 44507c478bd9Sstevel@tonic-gate 44517c478bd9Sstevel@tonic-gate /* ARGSUSED */ 44527c478bd9Sstevel@tonic-gate static int 4453303bf60bSsdebnath ufs_space(struct vnode *vp, int cmd, struct flock64 *bfp, int flag, 4454303bf60bSsdebnath offset_t offset, cred_t *cr, caller_context_t *ct) 44557c478bd9Sstevel@tonic-gate { 44567c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp = VTOI(vp)->i_ufsvfs; 44577c478bd9Sstevel@tonic-gate struct ulockfs *ulp; 44587c478bd9Sstevel@tonic-gate int error; 44597c478bd9Sstevel@tonic-gate 4460303bf60bSsdebnath if ((error = convoff(vp, bfp, 0, offset)) == 0) { 4461303bf60bSsdebnath if (cmd == F_FREESP) { 4462303bf60bSsdebnath error = ufs_lockfs_begin(ufsvfsp, &ulp, 4463303bf60bSsdebnath ULOCKFS_SPACE_MASK); 44647c478bd9Sstevel@tonic-gate if (error) 44657c478bd9Sstevel@tonic-gate return (error); 44667c478bd9Sstevel@tonic-gate error = ufs_freesp(vp, bfp, flag, cr); 446772102e74SBryan Cantrill 446872102e74SBryan Cantrill if (error == 0 && bfp->l_start == 0) 446972102e74SBryan Cantrill vnevent_truncate(vp, ct); 4470303bf60bSsdebnath } else if (cmd == F_ALLOCSP) { 4471303bf60bSsdebnath error = ufs_lockfs_begin(ufsvfsp, &ulp, 4472303bf60bSsdebnath ULOCKFS_FALLOCATE_MASK); 4473303bf60bSsdebnath if (error) 4474303bf60bSsdebnath return (error); 4475303bf60bSsdebnath error = ufs_allocsp(vp, bfp, cr); 4476303bf60bSsdebnath } else 4477303bf60bSsdebnath return (EINVAL); /* Command not handled here */ 44787c478bd9Sstevel@tonic-gate 44797c478bd9Sstevel@tonic-gate if (ulp) 44807c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 4481303bf60bSsdebnath 4482303bf60bSsdebnath } 44837c478bd9Sstevel@tonic-gate return (error); 44847c478bd9Sstevel@tonic-gate } 44857c478bd9Sstevel@tonic-gate 44867c478bd9Sstevel@tonic-gate /* 44877c478bd9Sstevel@tonic-gate * Used to determine if read ahead should be done. Also used to 44887c478bd9Sstevel@tonic-gate * to determine when write back occurs. 44897c478bd9Sstevel@tonic-gate */ 44907c478bd9Sstevel@tonic-gate #define CLUSTSZ(ip) ((ip)->i_ufsvfs->vfs_ioclustsz) 44917c478bd9Sstevel@tonic-gate 44927c478bd9Sstevel@tonic-gate /* 44937c478bd9Sstevel@tonic-gate * A faster version of ufs_getpage. 44947c478bd9Sstevel@tonic-gate * 44957c478bd9Sstevel@tonic-gate * We optimize by inlining the pvn_getpages iterator, eliminating 44967c478bd9Sstevel@tonic-gate * calls to bmap_read if file doesn't have UFS holes, and avoiding 44977c478bd9Sstevel@tonic-gate * the overhead of page_exists(). 44987c478bd9Sstevel@tonic-gate * 44997c478bd9Sstevel@tonic-gate * When files has UFS_HOLES and ufs_getpage is called with S_READ, 45007c478bd9Sstevel@tonic-gate * we set *protp to PROT_READ to avoid calling bmap_read. This approach 45017c478bd9Sstevel@tonic-gate * victimizes performance when a file with UFS holes is faulted 45027c478bd9Sstevel@tonic-gate * first in the S_READ mode, and then in the S_WRITE mode. We will get 45037c478bd9Sstevel@tonic-gate * two MMU faults in this case. 45047c478bd9Sstevel@tonic-gate * 45057c478bd9Sstevel@tonic-gate * XXX - the inode fields which control the sequential mode are not 45067c478bd9Sstevel@tonic-gate * protected by any mutex. The read ahead will act wild if 45077c478bd9Sstevel@tonic-gate * multiple processes will access the file concurrently and 45087c478bd9Sstevel@tonic-gate * some of them in sequential mode. One particulary bad case 45097c478bd9Sstevel@tonic-gate * is if another thread will change the value of i_nextrio between 45107c478bd9Sstevel@tonic-gate * the time this thread tests the i_nextrio value and then reads it 45117c478bd9Sstevel@tonic-gate * again to use it as the offset for the read ahead. 45127c478bd9Sstevel@tonic-gate */ 4513da6c28aaSamw /*ARGSUSED*/ 45147c478bd9Sstevel@tonic-gate static int 45157c478bd9Sstevel@tonic-gate ufs_getpage(struct vnode *vp, offset_t off, size_t len, uint_t *protp, 45167c478bd9Sstevel@tonic-gate page_t *plarr[], size_t plsz, struct seg *seg, caddr_t addr, 4517da6c28aaSamw enum seg_rw rw, struct cred *cr, caller_context_t *ct) 45187c478bd9Sstevel@tonic-gate { 45197c478bd9Sstevel@tonic-gate u_offset_t uoff = (u_offset_t)off; /* type conversion */ 45207c478bd9Sstevel@tonic-gate u_offset_t pgoff; 45217c478bd9Sstevel@tonic-gate u_offset_t eoff; 45227c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 45237c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp = ip->i_ufsvfs; 45247c478bd9Sstevel@tonic-gate struct fs *fs; 45257c478bd9Sstevel@tonic-gate struct ulockfs *ulp; 45267c478bd9Sstevel@tonic-gate page_t **pl; 45277c478bd9Sstevel@tonic-gate caddr_t pgaddr; 45287c478bd9Sstevel@tonic-gate krw_t rwtype; 45297c478bd9Sstevel@tonic-gate int err; 45307c478bd9Sstevel@tonic-gate int has_holes; 45317c478bd9Sstevel@tonic-gate int beyond_eof; 45327c478bd9Sstevel@tonic-gate int seqmode; 45337c478bd9Sstevel@tonic-gate int pgsize = PAGESIZE; 45347c478bd9Sstevel@tonic-gate int dolock; 45357c478bd9Sstevel@tonic-gate int do_qlock; 45367c478bd9Sstevel@tonic-gate int trans_size; 45377c478bd9Sstevel@tonic-gate 45387c478bd9Sstevel@tonic-gate ASSERT((uoff & PAGEOFFSET) == 0); 45397c478bd9Sstevel@tonic-gate 45407c478bd9Sstevel@tonic-gate if (protp) 45417c478bd9Sstevel@tonic-gate *protp = PROT_ALL; 45427c478bd9Sstevel@tonic-gate 45437c478bd9Sstevel@tonic-gate /* 45447c478bd9Sstevel@tonic-gate * Obey the lockfs protocol 45457c478bd9Sstevel@tonic-gate */ 45467c478bd9Sstevel@tonic-gate err = ufs_lockfs_begin_getpage(ufsvfsp, &ulp, seg, 45477c478bd9Sstevel@tonic-gate rw == S_READ || rw == S_EXEC, protp); 45487c478bd9Sstevel@tonic-gate if (err) 45497c478bd9Sstevel@tonic-gate goto out; 45507c478bd9Sstevel@tonic-gate 45517c478bd9Sstevel@tonic-gate fs = ufsvfsp->vfs_fs; 45527c478bd9Sstevel@tonic-gate 45537c478bd9Sstevel@tonic-gate if (ulp && (rw == S_CREATE || rw == S_WRITE) && 45547c478bd9Sstevel@tonic-gate !(vp->v_flag & VISSWAP)) { 45557c478bd9Sstevel@tonic-gate /* 45567c478bd9Sstevel@tonic-gate * Try to start a transaction, will return if blocking is 45577c478bd9Sstevel@tonic-gate * expected to occur and the address space is not the 45587c478bd9Sstevel@tonic-gate * kernel address space. 45597c478bd9Sstevel@tonic-gate */ 45607c478bd9Sstevel@tonic-gate trans_size = TOP_GETPAGE_SIZE(ip); 45617c478bd9Sstevel@tonic-gate if (seg->s_as != &kas) { 45627c478bd9Sstevel@tonic-gate TRANS_TRY_BEGIN_ASYNC(ufsvfsp, TOP_GETPAGE, 45637c478bd9Sstevel@tonic-gate trans_size, err) 45647c478bd9Sstevel@tonic-gate if (err == EWOULDBLOCK) { 45657c478bd9Sstevel@tonic-gate /* 45667c478bd9Sstevel@tonic-gate * Use EDEADLK here because the VM code 45677c478bd9Sstevel@tonic-gate * can normally never see this error. 45687c478bd9Sstevel@tonic-gate */ 45697c478bd9Sstevel@tonic-gate err = EDEADLK; 45707c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 45717c478bd9Sstevel@tonic-gate goto out; 45727c478bd9Sstevel@tonic-gate } 45737c478bd9Sstevel@tonic-gate } else { 45747c478bd9Sstevel@tonic-gate TRANS_BEGIN_ASYNC(ufsvfsp, TOP_GETPAGE, trans_size); 45757c478bd9Sstevel@tonic-gate } 45767c478bd9Sstevel@tonic-gate } 45777c478bd9Sstevel@tonic-gate 45787c478bd9Sstevel@tonic-gate if (vp->v_flag & VNOMAP) { 45797c478bd9Sstevel@tonic-gate err = ENOSYS; 45807c478bd9Sstevel@tonic-gate goto unlock; 45817c478bd9Sstevel@tonic-gate } 45827c478bd9Sstevel@tonic-gate 45837c478bd9Sstevel@tonic-gate seqmode = ip->i_nextr == uoff && rw != S_CREATE; 45847c478bd9Sstevel@tonic-gate 45857c478bd9Sstevel@tonic-gate rwtype = RW_READER; /* start as a reader */ 45867c478bd9Sstevel@tonic-gate dolock = (rw_owner(&ip->i_contents) != curthread); 45877c478bd9Sstevel@tonic-gate /* 45887c478bd9Sstevel@tonic-gate * If this thread owns the lock, i.e., this thread grabbed it 45897c478bd9Sstevel@tonic-gate * as writer somewhere above, then we don't need to grab the 45907c478bd9Sstevel@tonic-gate * lock as reader in this routine. 45917c478bd9Sstevel@tonic-gate */ 45927c478bd9Sstevel@tonic-gate do_qlock = (rw_owner(&ufsvfsp->vfs_dqrwlock) != curthread); 45937c478bd9Sstevel@tonic-gate 45947c478bd9Sstevel@tonic-gate retrylock: 45957c478bd9Sstevel@tonic-gate if (dolock) { 45967c478bd9Sstevel@tonic-gate /* 45977c478bd9Sstevel@tonic-gate * Grab the quota lock if we need to call 45987c478bd9Sstevel@tonic-gate * bmap_write() below (with i_contents as writer). 45997c478bd9Sstevel@tonic-gate */ 46007c478bd9Sstevel@tonic-gate if (do_qlock && rwtype == RW_WRITER) 46017c478bd9Sstevel@tonic-gate rw_enter(&ufsvfsp->vfs_dqrwlock, RW_READER); 46027c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, rwtype); 46037c478bd9Sstevel@tonic-gate } 46047c478bd9Sstevel@tonic-gate 46057c478bd9Sstevel@tonic-gate /* 46067c478bd9Sstevel@tonic-gate * We may be getting called as a side effect of a bmap using 46077c478bd9Sstevel@tonic-gate * fbread() when the blocks might be being allocated and the 46087c478bd9Sstevel@tonic-gate * size has not yet been up'ed. In this case we want to be 46097c478bd9Sstevel@tonic-gate * able to return zero pages if we get back UFS_HOLE from 46107c478bd9Sstevel@tonic-gate * calling bmap for a non write case here. We also might have 46117c478bd9Sstevel@tonic-gate * to read some frags from the disk into a page if we are 46127c478bd9Sstevel@tonic-gate * extending the number of frags for a given lbn in bmap(). 46137c478bd9Sstevel@tonic-gate * Large Files: The read of i_size here is atomic because 46147c478bd9Sstevel@tonic-gate * i_contents is held here. If dolock is zero, the lock 46157c478bd9Sstevel@tonic-gate * is held in bmap routines. 46167c478bd9Sstevel@tonic-gate */ 4617120107ddSmc208700 beyond_eof = uoff + len > 4618bcfec3c7Smc208700 P2ROUNDUP_TYPED(ip->i_size, PAGESIZE, u_offset_t); 46197c478bd9Sstevel@tonic-gate if (beyond_eof && seg != segkmap) { 46207c478bd9Sstevel@tonic-gate if (dolock) { 46217c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 46227c478bd9Sstevel@tonic-gate if (do_qlock && rwtype == RW_WRITER) 46237c478bd9Sstevel@tonic-gate rw_exit(&ufsvfsp->vfs_dqrwlock); 46247c478bd9Sstevel@tonic-gate } 46257c478bd9Sstevel@tonic-gate err = EFAULT; 46267c478bd9Sstevel@tonic-gate goto unlock; 46277c478bd9Sstevel@tonic-gate } 46287c478bd9Sstevel@tonic-gate 46297c478bd9Sstevel@tonic-gate /* 46307c478bd9Sstevel@tonic-gate * Must hold i_contents lock throughout the call to pvn_getpages 46317c478bd9Sstevel@tonic-gate * since locked pages are returned from each call to ufs_getapage. 46327c478bd9Sstevel@tonic-gate * Must *not* return locked pages and then try for contents lock 46337c478bd9Sstevel@tonic-gate * due to lock ordering requirements (inode > page) 46347c478bd9Sstevel@tonic-gate */ 46357c478bd9Sstevel@tonic-gate 46367c478bd9Sstevel@tonic-gate has_holes = bmap_has_holes(ip); 46377c478bd9Sstevel@tonic-gate 46387c478bd9Sstevel@tonic-gate if ((rw == S_WRITE || rw == S_CREATE) && has_holes && !beyond_eof) { 46397c478bd9Sstevel@tonic-gate int blk_size; 46407c478bd9Sstevel@tonic-gate u_offset_t offset; 46417c478bd9Sstevel@tonic-gate 46427c478bd9Sstevel@tonic-gate /* 46437c478bd9Sstevel@tonic-gate * We must acquire the RW_WRITER lock in order to 46447c478bd9Sstevel@tonic-gate * call bmap_write(). 46457c478bd9Sstevel@tonic-gate */ 46467c478bd9Sstevel@tonic-gate if (dolock && rwtype == RW_READER) { 46477c478bd9Sstevel@tonic-gate rwtype = RW_WRITER; 46487c478bd9Sstevel@tonic-gate 46497c478bd9Sstevel@tonic-gate /* 46507c478bd9Sstevel@tonic-gate * Grab the quota lock before 46517c478bd9Sstevel@tonic-gate * upgrading i_contents, but if we can't grab it 46527c478bd9Sstevel@tonic-gate * don't wait here due to lock order: 46537c478bd9Sstevel@tonic-gate * vfs_dqrwlock > i_contents. 46547c478bd9Sstevel@tonic-gate */ 465580d34432Sfrankho if (do_qlock && 465680d34432Sfrankho rw_tryenter(&ufsvfsp->vfs_dqrwlock, RW_READER) 465780d34432Sfrankho == 0) { 46587c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 46597c478bd9Sstevel@tonic-gate goto retrylock; 46607c478bd9Sstevel@tonic-gate } 46617c478bd9Sstevel@tonic-gate if (!rw_tryupgrade(&ip->i_contents)) { 46627c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 46637c478bd9Sstevel@tonic-gate if (do_qlock) 46647c478bd9Sstevel@tonic-gate rw_exit(&ufsvfsp->vfs_dqrwlock); 46657c478bd9Sstevel@tonic-gate goto retrylock; 46667c478bd9Sstevel@tonic-gate } 46677c478bd9Sstevel@tonic-gate } 46687c478bd9Sstevel@tonic-gate 46697c478bd9Sstevel@tonic-gate /* 46707c478bd9Sstevel@tonic-gate * May be allocating disk blocks for holes here as 46717c478bd9Sstevel@tonic-gate * a result of mmap faults. write(2) does the bmap_write 46727c478bd9Sstevel@tonic-gate * in rdip/wrip, not here. We are not dealing with frags 46737c478bd9Sstevel@tonic-gate * in this case. 46747c478bd9Sstevel@tonic-gate */ 46757c478bd9Sstevel@tonic-gate /* 46767c478bd9Sstevel@tonic-gate * Large Files: We cast fs_bmask field to offset_t 46777c478bd9Sstevel@tonic-gate * just as we do for MAXBMASK because uoff is a 64-bit 46787c478bd9Sstevel@tonic-gate * data type. fs_bmask will still be a 32-bit type 46797c478bd9Sstevel@tonic-gate * as we cannot change any ondisk data structures. 46807c478bd9Sstevel@tonic-gate */ 46817c478bd9Sstevel@tonic-gate 46827c478bd9Sstevel@tonic-gate offset = uoff & (offset_t)fs->fs_bmask; 46837c478bd9Sstevel@tonic-gate while (offset < uoff + len) { 46847c478bd9Sstevel@tonic-gate blk_size = (int)blksize(fs, ip, lblkno(fs, offset)); 4685303bf60bSsdebnath err = bmap_write(ip, offset, blk_size, 4686303bf60bSsdebnath BI_NORMAL, NULL, cr); 46877c478bd9Sstevel@tonic-gate if (ip->i_flag & (ICHG|IUPD)) 46887c478bd9Sstevel@tonic-gate ip->i_seq++; 46897c478bd9Sstevel@tonic-gate if (err) 46907c478bd9Sstevel@tonic-gate goto update_inode; 46917c478bd9Sstevel@tonic-gate offset += blk_size; /* XXX - make this contig */ 46927c478bd9Sstevel@tonic-gate } 46937c478bd9Sstevel@tonic-gate } 46947c478bd9Sstevel@tonic-gate 46957c478bd9Sstevel@tonic-gate /* 46967c478bd9Sstevel@tonic-gate * Can be a reader from now on. 46977c478bd9Sstevel@tonic-gate */ 46987c478bd9Sstevel@tonic-gate if (dolock && rwtype == RW_WRITER) { 46997c478bd9Sstevel@tonic-gate rw_downgrade(&ip->i_contents); 47007c478bd9Sstevel@tonic-gate /* 47017c478bd9Sstevel@tonic-gate * We can release vfs_dqrwlock early so do it, but make 47027c478bd9Sstevel@tonic-gate * sure we don't try to release it again at the bottom. 47037c478bd9Sstevel@tonic-gate */ 47047c478bd9Sstevel@tonic-gate if (do_qlock) { 47057c478bd9Sstevel@tonic-gate rw_exit(&ufsvfsp->vfs_dqrwlock); 47067c478bd9Sstevel@tonic-gate do_qlock = 0; 47077c478bd9Sstevel@tonic-gate } 47087c478bd9Sstevel@tonic-gate } 47097c478bd9Sstevel@tonic-gate 47107c478bd9Sstevel@tonic-gate /* 47117c478bd9Sstevel@tonic-gate * We remove PROT_WRITE in cases when the file has UFS holes 47127c478bd9Sstevel@tonic-gate * because we don't want to call bmap_read() to check each 47137c478bd9Sstevel@tonic-gate * page if it is backed with a disk block. 47147c478bd9Sstevel@tonic-gate */ 47157c478bd9Sstevel@tonic-gate if (protp && has_holes && rw != S_WRITE && rw != S_CREATE) 47167c478bd9Sstevel@tonic-gate *protp &= ~PROT_WRITE; 47177c478bd9Sstevel@tonic-gate 47187c478bd9Sstevel@tonic-gate err = 0; 47197c478bd9Sstevel@tonic-gate 47207c478bd9Sstevel@tonic-gate /* 47217c478bd9Sstevel@tonic-gate * The loop looks up pages in the range [off, off + len). 47227c478bd9Sstevel@tonic-gate * For each page, we first check if we should initiate an asynchronous 47237c478bd9Sstevel@tonic-gate * read ahead before we call page_lookup (we may sleep in page_lookup 47247c478bd9Sstevel@tonic-gate * for a previously initiated disk read). 47257c478bd9Sstevel@tonic-gate */ 47267c478bd9Sstevel@tonic-gate eoff = (uoff + len); 47277c478bd9Sstevel@tonic-gate for (pgoff = uoff, pgaddr = addr, pl = plarr; 47287c478bd9Sstevel@tonic-gate pgoff < eoff; /* empty */) { 47297c478bd9Sstevel@tonic-gate page_t *pp; 47307c478bd9Sstevel@tonic-gate u_offset_t nextrio; 47317c478bd9Sstevel@tonic-gate se_t se; 47327c478bd9Sstevel@tonic-gate int retval; 47337c478bd9Sstevel@tonic-gate 47347c478bd9Sstevel@tonic-gate se = ((rw == S_CREATE || rw == S_OTHER) ? SE_EXCL : SE_SHARED); 47357c478bd9Sstevel@tonic-gate 47367c478bd9Sstevel@tonic-gate /* Handle async getpage (faultahead) */ 47377c478bd9Sstevel@tonic-gate if (plarr == NULL) { 47387c478bd9Sstevel@tonic-gate ip->i_nextrio = pgoff; 47397c478bd9Sstevel@tonic-gate (void) ufs_getpage_ra(vp, pgoff, seg, pgaddr); 47407c478bd9Sstevel@tonic-gate pgoff += pgsize; 47417c478bd9Sstevel@tonic-gate pgaddr += pgsize; 47427c478bd9Sstevel@tonic-gate continue; 47437c478bd9Sstevel@tonic-gate } 47447c478bd9Sstevel@tonic-gate /* 47457c478bd9Sstevel@tonic-gate * Check if we should initiate read ahead of next cluster. 47467c478bd9Sstevel@tonic-gate * We call page_exists only when we need to confirm that 47477c478bd9Sstevel@tonic-gate * we have the current page before we initiate the read ahead. 47487c478bd9Sstevel@tonic-gate */ 47497c478bd9Sstevel@tonic-gate nextrio = ip->i_nextrio; 47507c478bd9Sstevel@tonic-gate if (seqmode && 47517c478bd9Sstevel@tonic-gate pgoff + CLUSTSZ(ip) >= nextrio && pgoff <= nextrio && 47527c478bd9Sstevel@tonic-gate nextrio < ip->i_size && page_exists(vp, pgoff)) { 47537c478bd9Sstevel@tonic-gate retval = ufs_getpage_ra(vp, pgoff, seg, pgaddr); 47547c478bd9Sstevel@tonic-gate /* 47557c478bd9Sstevel@tonic-gate * We always read ahead the next cluster of data 47567c478bd9Sstevel@tonic-gate * starting from i_nextrio. If the page (vp,nextrio) 47577c478bd9Sstevel@tonic-gate * is actually in core at this point, the routine 47587c478bd9Sstevel@tonic-gate * ufs_getpage_ra() will stop pre-fetching data 47597c478bd9Sstevel@tonic-gate * until we read that page in a synchronized manner 47607c478bd9Sstevel@tonic-gate * through ufs_getpage_miss(). So, we should increase 47617c478bd9Sstevel@tonic-gate * i_nextrio if the page (vp, nextrio) exists. 47627c478bd9Sstevel@tonic-gate */ 47637c478bd9Sstevel@tonic-gate if ((retval == 0) && page_exists(vp, nextrio)) { 47647c478bd9Sstevel@tonic-gate ip->i_nextrio = nextrio + pgsize; 47657c478bd9Sstevel@tonic-gate } 47667c478bd9Sstevel@tonic-gate } 47677c478bd9Sstevel@tonic-gate 47687c478bd9Sstevel@tonic-gate if ((pp = page_lookup(vp, pgoff, se)) != NULL) { 47697c478bd9Sstevel@tonic-gate /* 47707c478bd9Sstevel@tonic-gate * We found the page in the page cache. 47717c478bd9Sstevel@tonic-gate */ 47727c478bd9Sstevel@tonic-gate *pl++ = pp; 47737c478bd9Sstevel@tonic-gate pgoff += pgsize; 47747c478bd9Sstevel@tonic-gate pgaddr += pgsize; 47757c478bd9Sstevel@tonic-gate len -= pgsize; 47767c478bd9Sstevel@tonic-gate plsz -= pgsize; 47777c478bd9Sstevel@tonic-gate } else { 47787c478bd9Sstevel@tonic-gate /* 47797c478bd9Sstevel@tonic-gate * We have to create the page, or read it from disk. 47807c478bd9Sstevel@tonic-gate */ 47817c478bd9Sstevel@tonic-gate if (err = ufs_getpage_miss(vp, pgoff, len, seg, pgaddr, 47827c478bd9Sstevel@tonic-gate pl, plsz, rw, seqmode)) 47837c478bd9Sstevel@tonic-gate goto error; 47847c478bd9Sstevel@tonic-gate 47857c478bd9Sstevel@tonic-gate while (*pl != NULL) { 47867c478bd9Sstevel@tonic-gate pl++; 47877c478bd9Sstevel@tonic-gate pgoff += pgsize; 47887c478bd9Sstevel@tonic-gate pgaddr += pgsize; 47897c478bd9Sstevel@tonic-gate len -= pgsize; 47907c478bd9Sstevel@tonic-gate plsz -= pgsize; 47917c478bd9Sstevel@tonic-gate } 47927c478bd9Sstevel@tonic-gate } 47937c478bd9Sstevel@tonic-gate } 47947c478bd9Sstevel@tonic-gate 47957c478bd9Sstevel@tonic-gate /* 47967c478bd9Sstevel@tonic-gate * Return pages up to plsz if they are in the page cache. 47977c478bd9Sstevel@tonic-gate * We cannot return pages if there is a chance that they are 47987c478bd9Sstevel@tonic-gate * backed with a UFS hole and rw is S_WRITE or S_CREATE. 47997c478bd9Sstevel@tonic-gate */ 48007c478bd9Sstevel@tonic-gate if (plarr && !(has_holes && (rw == S_WRITE || rw == S_CREATE))) { 48017c478bd9Sstevel@tonic-gate 48027c478bd9Sstevel@tonic-gate ASSERT((protp == NULL) || 48037c478bd9Sstevel@tonic-gate !(has_holes && (*protp & PROT_WRITE))); 48047c478bd9Sstevel@tonic-gate 48057c478bd9Sstevel@tonic-gate eoff = pgoff + plsz; 48067c478bd9Sstevel@tonic-gate while (pgoff < eoff) { 48077c478bd9Sstevel@tonic-gate page_t *pp; 48087c478bd9Sstevel@tonic-gate 48097c478bd9Sstevel@tonic-gate if ((pp = page_lookup_nowait(vp, pgoff, 48107c478bd9Sstevel@tonic-gate SE_SHARED)) == NULL) 48117c478bd9Sstevel@tonic-gate break; 48127c478bd9Sstevel@tonic-gate 48137c478bd9Sstevel@tonic-gate *pl++ = pp; 48147c478bd9Sstevel@tonic-gate pgoff += pgsize; 48157c478bd9Sstevel@tonic-gate plsz -= pgsize; 48167c478bd9Sstevel@tonic-gate } 48177c478bd9Sstevel@tonic-gate } 48187c478bd9Sstevel@tonic-gate 48197c478bd9Sstevel@tonic-gate if (plarr) 48207c478bd9Sstevel@tonic-gate *pl = NULL; /* Terminate page list */ 48217c478bd9Sstevel@tonic-gate ip->i_nextr = pgoff; 48227c478bd9Sstevel@tonic-gate 48237c478bd9Sstevel@tonic-gate error: 48247c478bd9Sstevel@tonic-gate if (err && plarr) { 48257c478bd9Sstevel@tonic-gate /* 48267c478bd9Sstevel@tonic-gate * Release any pages we have locked. 48277c478bd9Sstevel@tonic-gate */ 48287c478bd9Sstevel@tonic-gate while (pl > &plarr[0]) 48297c478bd9Sstevel@tonic-gate page_unlock(*--pl); 48307c478bd9Sstevel@tonic-gate 48317c478bd9Sstevel@tonic-gate plarr[0] = NULL; 48327c478bd9Sstevel@tonic-gate } 48337c478bd9Sstevel@tonic-gate 48347c478bd9Sstevel@tonic-gate update_inode: 48357c478bd9Sstevel@tonic-gate /* 48367c478bd9Sstevel@tonic-gate * If the inode is not already marked for IACC (in rdip() for read) 48377c478bd9Sstevel@tonic-gate * and the inode is not marked for no access time update (in wrip() 48387c478bd9Sstevel@tonic-gate * for write) then update the inode access time and mod time now. 48397c478bd9Sstevel@tonic-gate */ 48407c478bd9Sstevel@tonic-gate if ((ip->i_flag & (IACC | INOACC)) == 0) { 48417c478bd9Sstevel@tonic-gate if ((rw != S_OTHER) && (ip->i_mode & IFMT) != IFDIR) { 48427c478bd9Sstevel@tonic-gate if (!ULOCKFS_IS_NOIACC(ITOUL(ip)) && 48437c478bd9Sstevel@tonic-gate (fs->fs_ronly == 0) && 48447c478bd9Sstevel@tonic-gate (!ufsvfsp->vfs_noatime)) { 48457c478bd9Sstevel@tonic-gate mutex_enter(&ip->i_tlock); 48467c478bd9Sstevel@tonic-gate ip->i_flag |= IACC; 48477c478bd9Sstevel@tonic-gate ITIMES_NOLOCK(ip); 48487c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 48497c478bd9Sstevel@tonic-gate } 48507c478bd9Sstevel@tonic-gate } 48517c478bd9Sstevel@tonic-gate } 48527c478bd9Sstevel@tonic-gate 48537c478bd9Sstevel@tonic-gate if (dolock) { 48547c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 48557c478bd9Sstevel@tonic-gate if (do_qlock && rwtype == RW_WRITER) 48567c478bd9Sstevel@tonic-gate rw_exit(&ufsvfsp->vfs_dqrwlock); 48577c478bd9Sstevel@tonic-gate } 48587c478bd9Sstevel@tonic-gate 48597c478bd9Sstevel@tonic-gate unlock: 48607c478bd9Sstevel@tonic-gate if (ulp) { 48617c478bd9Sstevel@tonic-gate if ((rw == S_CREATE || rw == S_WRITE) && 48627c478bd9Sstevel@tonic-gate !(vp->v_flag & VISSWAP)) { 48637c478bd9Sstevel@tonic-gate TRANS_END_ASYNC(ufsvfsp, TOP_GETPAGE, trans_size); 48647c478bd9Sstevel@tonic-gate } 48657c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 48667c478bd9Sstevel@tonic-gate } 48677c478bd9Sstevel@tonic-gate out: 48687c478bd9Sstevel@tonic-gate return (err); 48697c478bd9Sstevel@tonic-gate } 48707c478bd9Sstevel@tonic-gate 48717c478bd9Sstevel@tonic-gate /* 48727c478bd9Sstevel@tonic-gate * ufs_getpage_miss is called when ufs_getpage missed the page in the page 48737c478bd9Sstevel@tonic-gate * cache. The page is either read from the disk, or it's created. 48747c478bd9Sstevel@tonic-gate * A page is created (without disk read) if rw == S_CREATE, or if 48757c478bd9Sstevel@tonic-gate * the page is not backed with a real disk block (UFS hole). 48767c478bd9Sstevel@tonic-gate */ 48777c478bd9Sstevel@tonic-gate /* ARGSUSED */ 48787c478bd9Sstevel@tonic-gate static int 48797c478bd9Sstevel@tonic-gate ufs_getpage_miss(struct vnode *vp, u_offset_t off, size_t len, struct seg *seg, 48807c478bd9Sstevel@tonic-gate caddr_t addr, page_t *pl[], size_t plsz, enum seg_rw rw, int seq) 48817c478bd9Sstevel@tonic-gate { 48827c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 48837c478bd9Sstevel@tonic-gate page_t *pp; 48847c478bd9Sstevel@tonic-gate daddr_t bn; 48857c478bd9Sstevel@tonic-gate size_t io_len; 4886303bf60bSsdebnath int crpage = 0; 48877c478bd9Sstevel@tonic-gate int err; 48887c478bd9Sstevel@tonic-gate int contig; 48897c478bd9Sstevel@tonic-gate int bsize = ip->i_fs->fs_bsize; 48907c478bd9Sstevel@tonic-gate 48917c478bd9Sstevel@tonic-gate /* 48927c478bd9Sstevel@tonic-gate * Figure out whether the page can be created, or must be 48937c478bd9Sstevel@tonic-gate * must be read from the disk. 48947c478bd9Sstevel@tonic-gate */ 48957c478bd9Sstevel@tonic-gate if (rw == S_CREATE) 48967c478bd9Sstevel@tonic-gate crpage = 1; 48977c478bd9Sstevel@tonic-gate else { 48987c478bd9Sstevel@tonic-gate contig = 0; 48997c478bd9Sstevel@tonic-gate if (err = bmap_read(ip, off, &bn, &contig)) 49007c478bd9Sstevel@tonic-gate return (err); 4901303bf60bSsdebnath 49027c478bd9Sstevel@tonic-gate crpage = (bn == UFS_HOLE); 4903303bf60bSsdebnath 4904303bf60bSsdebnath /* 4905303bf60bSsdebnath * If its also a fallocated block that hasn't been written to 4906303bf60bSsdebnath * yet, we will treat it just like a UFS_HOLE and create 4907303bf60bSsdebnath * a zero page for it 4908303bf60bSsdebnath */ 4909303bf60bSsdebnath if (ISFALLOCBLK(ip, bn)) 4910303bf60bSsdebnath crpage = 1; 49117c478bd9Sstevel@tonic-gate } 49127c478bd9Sstevel@tonic-gate 49137c478bd9Sstevel@tonic-gate if (crpage) { 49147c478bd9Sstevel@tonic-gate if ((pp = page_create_va(vp, off, PAGESIZE, PG_WAIT, seg, 49157c478bd9Sstevel@tonic-gate addr)) == NULL) { 49167c478bd9Sstevel@tonic-gate return (ufs_fault(vp, 49177c478bd9Sstevel@tonic-gate "ufs_getpage_miss: page_create == NULL")); 49187c478bd9Sstevel@tonic-gate } 49197c478bd9Sstevel@tonic-gate 49207c478bd9Sstevel@tonic-gate if (rw != S_CREATE) 49217c478bd9Sstevel@tonic-gate pagezero(pp, 0, PAGESIZE); 4922303bf60bSsdebnath 49237c478bd9Sstevel@tonic-gate io_len = PAGESIZE; 49247c478bd9Sstevel@tonic-gate } else { 49257c478bd9Sstevel@tonic-gate u_offset_t io_off; 49267c478bd9Sstevel@tonic-gate uint_t xlen; 49277c478bd9Sstevel@tonic-gate struct buf *bp; 49287c478bd9Sstevel@tonic-gate ufsvfs_t *ufsvfsp = ip->i_ufsvfs; 49297c478bd9Sstevel@tonic-gate 49307c478bd9Sstevel@tonic-gate /* 49317c478bd9Sstevel@tonic-gate * If access is not in sequential order, we read from disk 49327c478bd9Sstevel@tonic-gate * in bsize units. 49337c478bd9Sstevel@tonic-gate * 49347c478bd9Sstevel@tonic-gate * We limit the size of the transfer to bsize if we are reading 49357c478bd9Sstevel@tonic-gate * from the beginning of the file. Note in this situation we 49367c478bd9Sstevel@tonic-gate * will hedge our bets and initiate an async read ahead of 49377c478bd9Sstevel@tonic-gate * the second block. 49387c478bd9Sstevel@tonic-gate */ 49397c478bd9Sstevel@tonic-gate if (!seq || off == 0) 49407c478bd9Sstevel@tonic-gate contig = MIN(contig, bsize); 49417c478bd9Sstevel@tonic-gate 49427c478bd9Sstevel@tonic-gate pp = pvn_read_kluster(vp, off, seg, addr, &io_off, 49437c478bd9Sstevel@tonic-gate &io_len, off, contig, 0); 49447c478bd9Sstevel@tonic-gate 49457c478bd9Sstevel@tonic-gate /* 49467c478bd9Sstevel@tonic-gate * Some other thread has entered the page. 49477c478bd9Sstevel@tonic-gate * ufs_getpage will retry page_lookup. 49487c478bd9Sstevel@tonic-gate */ 49497c478bd9Sstevel@tonic-gate if (pp == NULL) { 49507c478bd9Sstevel@tonic-gate pl[0] = NULL; 49517c478bd9Sstevel@tonic-gate return (0); 49527c478bd9Sstevel@tonic-gate } 49537c478bd9Sstevel@tonic-gate 49547c478bd9Sstevel@tonic-gate /* 49557c478bd9Sstevel@tonic-gate * Zero part of the page which we are not 49567c478bd9Sstevel@tonic-gate * going to read from the disk. 49577c478bd9Sstevel@tonic-gate */ 49587c478bd9Sstevel@tonic-gate xlen = io_len & PAGEOFFSET; 49597c478bd9Sstevel@tonic-gate if (xlen != 0) 49607c478bd9Sstevel@tonic-gate pagezero(pp->p_prev, xlen, PAGESIZE - xlen); 49617c478bd9Sstevel@tonic-gate 49627c478bd9Sstevel@tonic-gate bp = pageio_setup(pp, io_len, ip->i_devvp, B_READ); 49637c478bd9Sstevel@tonic-gate bp->b_edev = ip->i_dev; 49647c478bd9Sstevel@tonic-gate bp->b_dev = cmpdev(ip->i_dev); 49657c478bd9Sstevel@tonic-gate bp->b_blkno = bn; 49667c478bd9Sstevel@tonic-gate bp->b_un.b_addr = (caddr_t)0; 49677c478bd9Sstevel@tonic-gate bp->b_file = ip->i_vnode; 49687c478bd9Sstevel@tonic-gate bp->b_offset = off; 49697c478bd9Sstevel@tonic-gate 49707c478bd9Sstevel@tonic-gate if (ufsvfsp->vfs_log) { 49717c478bd9Sstevel@tonic-gate lufs_read_strategy(ufsvfsp->vfs_log, bp); 49727c478bd9Sstevel@tonic-gate } else if (ufsvfsp->vfs_snapshot) { 49737c478bd9Sstevel@tonic-gate fssnap_strategy(&ufsvfsp->vfs_snapshot, bp); 49747c478bd9Sstevel@tonic-gate } else { 4975d3d50737SRafael Vanoni ufsvfsp->vfs_iotstamp = ddi_get_lbolt(); 49767c478bd9Sstevel@tonic-gate ub.ub_getpages.value.ul++; 49777c478bd9Sstevel@tonic-gate (void) bdev_strategy(bp); 49787c478bd9Sstevel@tonic-gate lwp_stat_update(LWP_STAT_INBLK, 1); 49797c478bd9Sstevel@tonic-gate } 49807c478bd9Sstevel@tonic-gate 49817c478bd9Sstevel@tonic-gate ip->i_nextrio = off + ((io_len + PAGESIZE - 1) & PAGEMASK); 49827c478bd9Sstevel@tonic-gate 49837c478bd9Sstevel@tonic-gate /* 49847c478bd9Sstevel@tonic-gate * If the file access is sequential, initiate read ahead 49857c478bd9Sstevel@tonic-gate * of the next cluster. 49867c478bd9Sstevel@tonic-gate */ 49877c478bd9Sstevel@tonic-gate if (seq && ip->i_nextrio < ip->i_size) 49887c478bd9Sstevel@tonic-gate (void) ufs_getpage_ra(vp, off, seg, addr); 49897c478bd9Sstevel@tonic-gate err = biowait(bp); 49907c478bd9Sstevel@tonic-gate pageio_done(bp); 49917c478bd9Sstevel@tonic-gate 49927c478bd9Sstevel@tonic-gate if (err) { 49937c478bd9Sstevel@tonic-gate pvn_read_done(pp, B_ERROR); 49947c478bd9Sstevel@tonic-gate return (err); 49957c478bd9Sstevel@tonic-gate } 49967c478bd9Sstevel@tonic-gate } 49977c478bd9Sstevel@tonic-gate 49987c478bd9Sstevel@tonic-gate pvn_plist_init(pp, pl, plsz, off, io_len, rw); 49997c478bd9Sstevel@tonic-gate return (0); 50007c478bd9Sstevel@tonic-gate } 50017c478bd9Sstevel@tonic-gate 50027c478bd9Sstevel@tonic-gate /* 50037c478bd9Sstevel@tonic-gate * Read ahead a cluster from the disk. Returns the length in bytes. 50047c478bd9Sstevel@tonic-gate */ 50057c478bd9Sstevel@tonic-gate static int 50067c478bd9Sstevel@tonic-gate ufs_getpage_ra(struct vnode *vp, u_offset_t off, struct seg *seg, caddr_t addr) 50077c478bd9Sstevel@tonic-gate { 50087c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 50097c478bd9Sstevel@tonic-gate page_t *pp; 50107c478bd9Sstevel@tonic-gate u_offset_t io_off = ip->i_nextrio; 50117c478bd9Sstevel@tonic-gate ufsvfs_t *ufsvfsp; 50127c478bd9Sstevel@tonic-gate caddr_t addr2 = addr + (io_off - off); 50137c478bd9Sstevel@tonic-gate struct buf *bp; 50147c478bd9Sstevel@tonic-gate daddr_t bn; 50157c478bd9Sstevel@tonic-gate size_t io_len; 5016303bf60bSsdebnath int err; 50177c478bd9Sstevel@tonic-gate int contig; 50187c478bd9Sstevel@tonic-gate int xlen; 50197c478bd9Sstevel@tonic-gate int bsize = ip->i_fs->fs_bsize; 50207c478bd9Sstevel@tonic-gate 50217c478bd9Sstevel@tonic-gate /* 50227c478bd9Sstevel@tonic-gate * If the directio advisory is in effect on this file, 50237c478bd9Sstevel@tonic-gate * then do not do buffered read ahead. Read ahead makes 50247c478bd9Sstevel@tonic-gate * it more difficult on threads using directio as they 50257c478bd9Sstevel@tonic-gate * will be forced to flush the pages from this vnode. 50267c478bd9Sstevel@tonic-gate */ 50277c478bd9Sstevel@tonic-gate if ((ufsvfsp = ip->i_ufsvfs) == NULL) 50287c478bd9Sstevel@tonic-gate return (0); 50297c478bd9Sstevel@tonic-gate if (ip->i_flag & IDIRECTIO || ufsvfsp->vfs_forcedirectio) 50307c478bd9Sstevel@tonic-gate return (0); 50317c478bd9Sstevel@tonic-gate 50327c478bd9Sstevel@tonic-gate /* 50337c478bd9Sstevel@tonic-gate * Is this test needed? 50347c478bd9Sstevel@tonic-gate */ 50357c478bd9Sstevel@tonic-gate if (addr2 >= seg->s_base + seg->s_size) 50367c478bd9Sstevel@tonic-gate return (0); 50377c478bd9Sstevel@tonic-gate 50387c478bd9Sstevel@tonic-gate contig = 0; 5039303bf60bSsdebnath err = bmap_read(ip, io_off, &bn, &contig); 5040303bf60bSsdebnath /* 5041303bf60bSsdebnath * If its a UFS_HOLE or a fallocated block, do not perform 5042303bf60bSsdebnath * any read ahead's since there probably is nothing to read ahead 5043303bf60bSsdebnath */ 5044303bf60bSsdebnath if (err || bn == UFS_HOLE || ISFALLOCBLK(ip, bn)) 50457c478bd9Sstevel@tonic-gate return (0); 50467c478bd9Sstevel@tonic-gate 50477c478bd9Sstevel@tonic-gate /* 50487c478bd9Sstevel@tonic-gate * Limit the transfer size to bsize if this is the 2nd block. 50497c478bd9Sstevel@tonic-gate */ 50507c478bd9Sstevel@tonic-gate if (io_off == (u_offset_t)bsize) 50517c478bd9Sstevel@tonic-gate contig = MIN(contig, bsize); 50527c478bd9Sstevel@tonic-gate 50537c478bd9Sstevel@tonic-gate if ((pp = pvn_read_kluster(vp, io_off, seg, addr2, &io_off, 50547c478bd9Sstevel@tonic-gate &io_len, io_off, contig, 1)) == NULL) 50557c478bd9Sstevel@tonic-gate return (0); 50567c478bd9Sstevel@tonic-gate 50577c478bd9Sstevel@tonic-gate /* 50587c478bd9Sstevel@tonic-gate * Zero part of page which we are not going to read from disk 50597c478bd9Sstevel@tonic-gate */ 50607c478bd9Sstevel@tonic-gate if ((xlen = (io_len & PAGEOFFSET)) > 0) 50617c478bd9Sstevel@tonic-gate pagezero(pp->p_prev, xlen, PAGESIZE - xlen); 50627c478bd9Sstevel@tonic-gate 50637c478bd9Sstevel@tonic-gate ip->i_nextrio = (io_off + io_len + PAGESIZE - 1) & PAGEMASK; 50647c478bd9Sstevel@tonic-gate 50657c478bd9Sstevel@tonic-gate bp = pageio_setup(pp, io_len, ip->i_devvp, B_READ | B_ASYNC); 50667c478bd9Sstevel@tonic-gate bp->b_edev = ip->i_dev; 50677c478bd9Sstevel@tonic-gate bp->b_dev = cmpdev(ip->i_dev); 50687c478bd9Sstevel@tonic-gate bp->b_blkno = bn; 50697c478bd9Sstevel@tonic-gate bp->b_un.b_addr = (caddr_t)0; 50707c478bd9Sstevel@tonic-gate bp->b_file = ip->i_vnode; 50717c478bd9Sstevel@tonic-gate bp->b_offset = off; 50727c478bd9Sstevel@tonic-gate 50737c478bd9Sstevel@tonic-gate if (ufsvfsp->vfs_log) { 50747c478bd9Sstevel@tonic-gate lufs_read_strategy(ufsvfsp->vfs_log, bp); 50757c478bd9Sstevel@tonic-gate } else if (ufsvfsp->vfs_snapshot) { 50767c478bd9Sstevel@tonic-gate fssnap_strategy(&ufsvfsp->vfs_snapshot, bp); 50777c478bd9Sstevel@tonic-gate } else { 5078d3d50737SRafael Vanoni ufsvfsp->vfs_iotstamp = ddi_get_lbolt(); 50797c478bd9Sstevel@tonic-gate ub.ub_getras.value.ul++; 50807c478bd9Sstevel@tonic-gate (void) bdev_strategy(bp); 50817c478bd9Sstevel@tonic-gate lwp_stat_update(LWP_STAT_INBLK, 1); 50827c478bd9Sstevel@tonic-gate } 50837c478bd9Sstevel@tonic-gate 50847c478bd9Sstevel@tonic-gate return (io_len); 50857c478bd9Sstevel@tonic-gate } 50867c478bd9Sstevel@tonic-gate 50877c478bd9Sstevel@tonic-gate int ufs_delay = 1; 50887c478bd9Sstevel@tonic-gate /* 50897c478bd9Sstevel@tonic-gate * Flags are composed of {B_INVAL, B_FREE, B_DONTNEED, B_FORCE, B_ASYNC} 50907c478bd9Sstevel@tonic-gate * 50917c478bd9Sstevel@tonic-gate * LMXXX - the inode really ought to contain a pointer to one of these 50927c478bd9Sstevel@tonic-gate * async args. Stuff gunk in there and just hand the whole mess off. 50937c478bd9Sstevel@tonic-gate * This would replace i_delaylen, i_delayoff. 50947c478bd9Sstevel@tonic-gate */ 50957c478bd9Sstevel@tonic-gate /*ARGSUSED*/ 50967c478bd9Sstevel@tonic-gate static int 50977c478bd9Sstevel@tonic-gate ufs_putpage(struct vnode *vp, offset_t off, size_t len, int flags, 5098da6c28aaSamw struct cred *cr, caller_context_t *ct) 50997c478bd9Sstevel@tonic-gate { 51007c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 51017c478bd9Sstevel@tonic-gate int err = 0; 51027c478bd9Sstevel@tonic-gate 51037c478bd9Sstevel@tonic-gate if (vp->v_count == 0) { 51047c478bd9Sstevel@tonic-gate return (ufs_fault(vp, "ufs_putpage: bad v_count == 0")); 51057c478bd9Sstevel@tonic-gate } 51067c478bd9Sstevel@tonic-gate 51077c478bd9Sstevel@tonic-gate /* 51087c478bd9Sstevel@tonic-gate * XXX - Why should this check be made here? 51097c478bd9Sstevel@tonic-gate */ 51107c478bd9Sstevel@tonic-gate if (vp->v_flag & VNOMAP) { 51117c478bd9Sstevel@tonic-gate err = ENOSYS; 51127c478bd9Sstevel@tonic-gate goto errout; 51137c478bd9Sstevel@tonic-gate } 51147c478bd9Sstevel@tonic-gate 51157c478bd9Sstevel@tonic-gate if (ip->i_ufsvfs == NULL) { 51167c478bd9Sstevel@tonic-gate err = EIO; 51177c478bd9Sstevel@tonic-gate goto errout; 51187c478bd9Sstevel@tonic-gate } 51197c478bd9Sstevel@tonic-gate 51207c478bd9Sstevel@tonic-gate if (flags & B_ASYNC) { 51217c478bd9Sstevel@tonic-gate if (ufs_delay && len && 51227c478bd9Sstevel@tonic-gate (flags & ~(B_ASYNC|B_DONTNEED|B_FREE)) == 0) { 51237c478bd9Sstevel@tonic-gate mutex_enter(&ip->i_tlock); 51247c478bd9Sstevel@tonic-gate /* 51257c478bd9Sstevel@tonic-gate * If nobody stalled, start a new cluster. 51267c478bd9Sstevel@tonic-gate */ 51277c478bd9Sstevel@tonic-gate if (ip->i_delaylen == 0) { 51287c478bd9Sstevel@tonic-gate ip->i_delayoff = off; 51297c478bd9Sstevel@tonic-gate ip->i_delaylen = len; 51307c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 51317c478bd9Sstevel@tonic-gate goto errout; 51327c478bd9Sstevel@tonic-gate } 51337c478bd9Sstevel@tonic-gate /* 51347c478bd9Sstevel@tonic-gate * If we have a full cluster or they are not contig, 51357c478bd9Sstevel@tonic-gate * then push last cluster and start over. 51367c478bd9Sstevel@tonic-gate */ 51377c478bd9Sstevel@tonic-gate if (ip->i_delaylen >= CLUSTSZ(ip) || 51387c478bd9Sstevel@tonic-gate ip->i_delayoff + ip->i_delaylen != off) { 51397c478bd9Sstevel@tonic-gate u_offset_t doff; 51407c478bd9Sstevel@tonic-gate size_t dlen; 51417c478bd9Sstevel@tonic-gate 51427c478bd9Sstevel@tonic-gate doff = ip->i_delayoff; 51437c478bd9Sstevel@tonic-gate dlen = ip->i_delaylen; 51447c478bd9Sstevel@tonic-gate ip->i_delayoff = off; 51457c478bd9Sstevel@tonic-gate ip->i_delaylen = len; 51467c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 51477c478bd9Sstevel@tonic-gate err = ufs_putpages(vp, doff, dlen, 51487c478bd9Sstevel@tonic-gate flags, cr); 51497c478bd9Sstevel@tonic-gate /* LMXXX - flags are new val, not old */ 51507c478bd9Sstevel@tonic-gate goto errout; 51517c478bd9Sstevel@tonic-gate } 51527c478bd9Sstevel@tonic-gate /* 51537c478bd9Sstevel@tonic-gate * There is something there, it's not full, and 51547c478bd9Sstevel@tonic-gate * it is contig. 51557c478bd9Sstevel@tonic-gate */ 51567c478bd9Sstevel@tonic-gate ip->i_delaylen += len; 51577c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 51587c478bd9Sstevel@tonic-gate goto errout; 51597c478bd9Sstevel@tonic-gate } 51607c478bd9Sstevel@tonic-gate /* 51617c478bd9Sstevel@tonic-gate * Must have weird flags or we are not clustering. 51627c478bd9Sstevel@tonic-gate */ 51637c478bd9Sstevel@tonic-gate } 51647c478bd9Sstevel@tonic-gate 51657c478bd9Sstevel@tonic-gate err = ufs_putpages(vp, off, len, flags, cr); 51667c478bd9Sstevel@tonic-gate 51677c478bd9Sstevel@tonic-gate errout: 51687c478bd9Sstevel@tonic-gate return (err); 51697c478bd9Sstevel@tonic-gate } 51707c478bd9Sstevel@tonic-gate 51717c478bd9Sstevel@tonic-gate /* 51727c478bd9Sstevel@tonic-gate * If len == 0, do from off to EOF. 51737c478bd9Sstevel@tonic-gate * 51747c478bd9Sstevel@tonic-gate * The normal cases should be len == 0 & off == 0 (entire vp list), 51757c478bd9Sstevel@tonic-gate * len == MAXBSIZE (from segmap_release actions), and len == PAGESIZE 51767c478bd9Sstevel@tonic-gate * (from pageout). 51777c478bd9Sstevel@tonic-gate */ 51787c478bd9Sstevel@tonic-gate /*ARGSUSED*/ 51797c478bd9Sstevel@tonic-gate static int 51807c478bd9Sstevel@tonic-gate ufs_putpages( 51817c478bd9Sstevel@tonic-gate struct vnode *vp, 51827c478bd9Sstevel@tonic-gate offset_t off, 51837c478bd9Sstevel@tonic-gate size_t len, 51847c478bd9Sstevel@tonic-gate int flags, 51857c478bd9Sstevel@tonic-gate struct cred *cr) 51867c478bd9Sstevel@tonic-gate { 51877c478bd9Sstevel@tonic-gate u_offset_t io_off; 51887c478bd9Sstevel@tonic-gate u_offset_t eoff; 51897c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 51907c478bd9Sstevel@tonic-gate page_t *pp; 51917c478bd9Sstevel@tonic-gate size_t io_len; 51927c478bd9Sstevel@tonic-gate int err = 0; 51937c478bd9Sstevel@tonic-gate int dolock; 51947c478bd9Sstevel@tonic-gate 51957c478bd9Sstevel@tonic-gate if (vp->v_count == 0) 51967c478bd9Sstevel@tonic-gate return (ufs_fault(vp, "ufs_putpages: v_count == 0")); 51977c478bd9Sstevel@tonic-gate /* 51987c478bd9Sstevel@tonic-gate * Acquire the readers/write inode lock before locking 51997c478bd9Sstevel@tonic-gate * any pages in this inode. 52007c478bd9Sstevel@tonic-gate * The inode lock is held during i/o. 52017c478bd9Sstevel@tonic-gate */ 52027c478bd9Sstevel@tonic-gate if (len == 0) { 52037c478bd9Sstevel@tonic-gate mutex_enter(&ip->i_tlock); 52047c478bd9Sstevel@tonic-gate ip->i_delayoff = ip->i_delaylen = 0; 52057c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 52067c478bd9Sstevel@tonic-gate } 52077c478bd9Sstevel@tonic-gate dolock = (rw_owner(&ip->i_contents) != curthread); 52087c478bd9Sstevel@tonic-gate if (dolock) { 52097c478bd9Sstevel@tonic-gate /* 52107c478bd9Sstevel@tonic-gate * Must synchronize this thread and any possible thread 52117c478bd9Sstevel@tonic-gate * operating in the window of vulnerability in wrip(). 52127c478bd9Sstevel@tonic-gate * It is dangerous to allow both a thread doing a putpage 52137c478bd9Sstevel@tonic-gate * and a thread writing, so serialize them. The exception 52147c478bd9Sstevel@tonic-gate * is when the thread in wrip() does something which causes 52157c478bd9Sstevel@tonic-gate * a putpage operation. Then, the thread must be allowed 52167c478bd9Sstevel@tonic-gate * to continue. It may encounter a bmap_read problem in 52177c478bd9Sstevel@tonic-gate * ufs_putapage, but that is handled in ufs_putapage. 52187c478bd9Sstevel@tonic-gate * Allow async writers to proceed, we don't want to block 52197c478bd9Sstevel@tonic-gate * the pageout daemon. 52207c478bd9Sstevel@tonic-gate */ 52217c478bd9Sstevel@tonic-gate if (ip->i_writer == curthread) 52227c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_READER); 52237c478bd9Sstevel@tonic-gate else { 52247c478bd9Sstevel@tonic-gate for (;;) { 52257c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_READER); 52267c478bd9Sstevel@tonic-gate mutex_enter(&ip->i_tlock); 52277c478bd9Sstevel@tonic-gate /* 52287c478bd9Sstevel@tonic-gate * If there is no thread in the critical 52297c478bd9Sstevel@tonic-gate * section of wrip(), then proceed. 52307c478bd9Sstevel@tonic-gate * Otherwise, wait until there isn't one. 52317c478bd9Sstevel@tonic-gate */ 52327c478bd9Sstevel@tonic-gate if (ip->i_writer == NULL) { 52337c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 52347c478bd9Sstevel@tonic-gate break; 52357c478bd9Sstevel@tonic-gate } 52367c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 52377c478bd9Sstevel@tonic-gate /* 52387c478bd9Sstevel@tonic-gate * Bounce async writers when we have a writer 52397c478bd9Sstevel@tonic-gate * working on this file so we don't deadlock 52407c478bd9Sstevel@tonic-gate * the pageout daemon. 52417c478bd9Sstevel@tonic-gate */ 52427c478bd9Sstevel@tonic-gate if (flags & B_ASYNC) { 52437c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 52447c478bd9Sstevel@tonic-gate return (0); 52457c478bd9Sstevel@tonic-gate } 52467c478bd9Sstevel@tonic-gate cv_wait(&ip->i_wrcv, &ip->i_tlock); 52477c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 52487c478bd9Sstevel@tonic-gate } 52497c478bd9Sstevel@tonic-gate } 52507c478bd9Sstevel@tonic-gate } 52517c478bd9Sstevel@tonic-gate 52527c478bd9Sstevel@tonic-gate if (!vn_has_cached_data(vp)) { 52537c478bd9Sstevel@tonic-gate if (dolock) 52547c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 52557c478bd9Sstevel@tonic-gate return (0); 52567c478bd9Sstevel@tonic-gate } 52577c478bd9Sstevel@tonic-gate 52587c478bd9Sstevel@tonic-gate if (len == 0) { 52597c478bd9Sstevel@tonic-gate /* 52607c478bd9Sstevel@tonic-gate * Search the entire vp list for pages >= off. 52617c478bd9Sstevel@tonic-gate */ 52627c478bd9Sstevel@tonic-gate err = pvn_vplist_dirty(vp, (u_offset_t)off, ufs_putapage, 52637c478bd9Sstevel@tonic-gate flags, cr); 52647c478bd9Sstevel@tonic-gate } else { 52657c478bd9Sstevel@tonic-gate /* 52667c478bd9Sstevel@tonic-gate * Loop over all offsets in the range looking for 52677c478bd9Sstevel@tonic-gate * pages to deal with. 52687c478bd9Sstevel@tonic-gate */ 52697c478bd9Sstevel@tonic-gate if ((eoff = blkroundup(ip->i_fs, ip->i_size)) != 0) 52707c478bd9Sstevel@tonic-gate eoff = MIN(off + len, eoff); 52717c478bd9Sstevel@tonic-gate else 52727c478bd9Sstevel@tonic-gate eoff = off + len; 52737c478bd9Sstevel@tonic-gate 52747c478bd9Sstevel@tonic-gate for (io_off = off; io_off < eoff; io_off += io_len) { 52757c478bd9Sstevel@tonic-gate /* 52767c478bd9Sstevel@tonic-gate * If we are not invalidating, synchronously 52777c478bd9Sstevel@tonic-gate * freeing or writing pages, use the routine 52787c478bd9Sstevel@tonic-gate * page_lookup_nowait() to prevent reclaiming 52797c478bd9Sstevel@tonic-gate * them from the free list. 52807c478bd9Sstevel@tonic-gate */ 52817c478bd9Sstevel@tonic-gate if ((flags & B_INVAL) || ((flags & B_ASYNC) == 0)) { 52827c478bd9Sstevel@tonic-gate pp = page_lookup(vp, io_off, 52837c478bd9Sstevel@tonic-gate (flags & (B_INVAL | B_FREE)) ? 52847c478bd9Sstevel@tonic-gate SE_EXCL : SE_SHARED); 52857c478bd9Sstevel@tonic-gate } else { 52867c478bd9Sstevel@tonic-gate pp = page_lookup_nowait(vp, io_off, 52877c478bd9Sstevel@tonic-gate (flags & B_FREE) ? SE_EXCL : SE_SHARED); 52887c478bd9Sstevel@tonic-gate } 52897c478bd9Sstevel@tonic-gate 52907c478bd9Sstevel@tonic-gate if (pp == NULL || pvn_getdirty(pp, flags) == 0) 52917c478bd9Sstevel@tonic-gate io_len = PAGESIZE; 52927c478bd9Sstevel@tonic-gate else { 52937c478bd9Sstevel@tonic-gate u_offset_t *io_offp = &io_off; 52947c478bd9Sstevel@tonic-gate 52957c478bd9Sstevel@tonic-gate err = ufs_putapage(vp, pp, io_offp, &io_len, 52967c478bd9Sstevel@tonic-gate flags, cr); 52977c478bd9Sstevel@tonic-gate if (err != 0) 52987c478bd9Sstevel@tonic-gate break; 52997c478bd9Sstevel@tonic-gate /* 53007c478bd9Sstevel@tonic-gate * "io_off" and "io_len" are returned as 53017c478bd9Sstevel@tonic-gate * the range of pages we actually wrote. 53027c478bd9Sstevel@tonic-gate * This allows us to skip ahead more quickly 53037c478bd9Sstevel@tonic-gate * since several pages may've been dealt 53047c478bd9Sstevel@tonic-gate * with by this iteration of the loop. 53057c478bd9Sstevel@tonic-gate */ 53067c478bd9Sstevel@tonic-gate } 53077c478bd9Sstevel@tonic-gate } 53087c478bd9Sstevel@tonic-gate } 53097c478bd9Sstevel@tonic-gate if (err == 0 && off == 0 && (len == 0 || len >= ip->i_size)) { 53107c478bd9Sstevel@tonic-gate /* 53117c478bd9Sstevel@tonic-gate * We have just sync'ed back all the pages on 53127c478bd9Sstevel@tonic-gate * the inode, turn off the IMODTIME flag. 53137c478bd9Sstevel@tonic-gate */ 53147c478bd9Sstevel@tonic-gate mutex_enter(&ip->i_tlock); 53157c478bd9Sstevel@tonic-gate ip->i_flag &= ~IMODTIME; 53167c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 53177c478bd9Sstevel@tonic-gate } 53187c478bd9Sstevel@tonic-gate if (dolock) 53197c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 53207c478bd9Sstevel@tonic-gate return (err); 53217c478bd9Sstevel@tonic-gate } 53227c478bd9Sstevel@tonic-gate 53237c478bd9Sstevel@tonic-gate static void 53247c478bd9Sstevel@tonic-gate ufs_iodone(buf_t *bp) 53257c478bd9Sstevel@tonic-gate { 53267c478bd9Sstevel@tonic-gate struct inode *ip; 53277c478bd9Sstevel@tonic-gate 53287c478bd9Sstevel@tonic-gate ASSERT((bp->b_pages->p_vnode != NULL) && !(bp->b_flags & B_READ)); 53297c478bd9Sstevel@tonic-gate 53307c478bd9Sstevel@tonic-gate bp->b_iodone = NULL; 53317c478bd9Sstevel@tonic-gate 53327c478bd9Sstevel@tonic-gate ip = VTOI(bp->b_pages->p_vnode); 53337c478bd9Sstevel@tonic-gate 53347c478bd9Sstevel@tonic-gate mutex_enter(&ip->i_tlock); 53357c478bd9Sstevel@tonic-gate if (ip->i_writes >= ufs_LW) { 53367c478bd9Sstevel@tonic-gate if ((ip->i_writes -= bp->b_bcount) <= ufs_LW) 53377c478bd9Sstevel@tonic-gate if (ufs_WRITES) 53387c478bd9Sstevel@tonic-gate cv_broadcast(&ip->i_wrcv); /* wake all up */ 53397c478bd9Sstevel@tonic-gate } else { 53407c478bd9Sstevel@tonic-gate ip->i_writes -= bp->b_bcount; 53417c478bd9Sstevel@tonic-gate } 53427c478bd9Sstevel@tonic-gate 53437c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 53447c478bd9Sstevel@tonic-gate iodone(bp); 53457c478bd9Sstevel@tonic-gate } 53467c478bd9Sstevel@tonic-gate 53477c478bd9Sstevel@tonic-gate /* 53487c478bd9Sstevel@tonic-gate * Write out a single page, possibly klustering adjacent 53497c478bd9Sstevel@tonic-gate * dirty pages. The inode lock must be held. 53507c478bd9Sstevel@tonic-gate * 53517c478bd9Sstevel@tonic-gate * LMXXX - bsize < pagesize not done. 53527c478bd9Sstevel@tonic-gate */ 53537c478bd9Sstevel@tonic-gate /*ARGSUSED*/ 53547c478bd9Sstevel@tonic-gate int 53557c478bd9Sstevel@tonic-gate ufs_putapage( 53567c478bd9Sstevel@tonic-gate struct vnode *vp, 53577c478bd9Sstevel@tonic-gate page_t *pp, 53587c478bd9Sstevel@tonic-gate u_offset_t *offp, 53597c478bd9Sstevel@tonic-gate size_t *lenp, /* return values */ 53607c478bd9Sstevel@tonic-gate int flags, 53617c478bd9Sstevel@tonic-gate struct cred *cr) 53627c478bd9Sstevel@tonic-gate { 53637c478bd9Sstevel@tonic-gate u_offset_t io_off; 53647c478bd9Sstevel@tonic-gate u_offset_t off; 53657c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 53667c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp = ip->i_ufsvfs; 53677c478bd9Sstevel@tonic-gate struct fs *fs; 53687c478bd9Sstevel@tonic-gate struct buf *bp; 53697c478bd9Sstevel@tonic-gate size_t io_len; 53707c478bd9Sstevel@tonic-gate daddr_t bn; 53717c478bd9Sstevel@tonic-gate int err; 53727c478bd9Sstevel@tonic-gate int contig; 5373b9dc641bSfrankho int dotrans; 53747c478bd9Sstevel@tonic-gate 53757c478bd9Sstevel@tonic-gate ASSERT(RW_LOCK_HELD(&ip->i_contents)); 53767c478bd9Sstevel@tonic-gate 53777c478bd9Sstevel@tonic-gate if (ufsvfsp == NULL) { 53787c478bd9Sstevel@tonic-gate err = EIO; 53797c478bd9Sstevel@tonic-gate goto out_trace; 53807c478bd9Sstevel@tonic-gate } 53817c478bd9Sstevel@tonic-gate 53827c478bd9Sstevel@tonic-gate fs = ip->i_fs; 53837c478bd9Sstevel@tonic-gate ASSERT(fs->fs_ronly == 0); 53847c478bd9Sstevel@tonic-gate 53857c478bd9Sstevel@tonic-gate /* 53867c478bd9Sstevel@tonic-gate * If the modified time on the inode has not already been 5387956bc340Sdm120769 * set elsewhere (e.g. for write/setattr) we set the time now. 53887c478bd9Sstevel@tonic-gate * This gives us approximate modified times for mmap'ed files 53897c478bd9Sstevel@tonic-gate * which are modified via stores in the user address space. 53907c478bd9Sstevel@tonic-gate */ 5391956bc340Sdm120769 if ((ip->i_flag & IMODTIME) == 0) { 53927c478bd9Sstevel@tonic-gate mutex_enter(&ip->i_tlock); 53937c478bd9Sstevel@tonic-gate ip->i_flag |= IUPD; 53947c478bd9Sstevel@tonic-gate ip->i_seq++; 53957c478bd9Sstevel@tonic-gate ITIMES_NOLOCK(ip); 53967c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 53977c478bd9Sstevel@tonic-gate } 53987c478bd9Sstevel@tonic-gate 53997c478bd9Sstevel@tonic-gate /* 54007c478bd9Sstevel@tonic-gate * Align the request to a block boundry (for old file systems), 54017c478bd9Sstevel@tonic-gate * and go ask bmap() how contiguous things are for this file. 54027c478bd9Sstevel@tonic-gate */ 54037c478bd9Sstevel@tonic-gate off = pp->p_offset & (offset_t)fs->fs_bmask; /* block align it */ 54047c478bd9Sstevel@tonic-gate contig = 0; 54057c478bd9Sstevel@tonic-gate err = bmap_read(ip, off, &bn, &contig); 54067c478bd9Sstevel@tonic-gate if (err) 54077c478bd9Sstevel@tonic-gate goto out; 54087c478bd9Sstevel@tonic-gate if (bn == UFS_HOLE) { /* putpage never allocates */ 54097c478bd9Sstevel@tonic-gate /* 54107c478bd9Sstevel@tonic-gate * logging device is in error mode; simply return EIO 54117c478bd9Sstevel@tonic-gate */ 54127c478bd9Sstevel@tonic-gate if (TRANS_ISERROR(ufsvfsp)) { 54137c478bd9Sstevel@tonic-gate err = EIO; 54147c478bd9Sstevel@tonic-gate goto out; 54157c478bd9Sstevel@tonic-gate } 54167c478bd9Sstevel@tonic-gate /* 54177c478bd9Sstevel@tonic-gate * Oops, the thread in the window in wrip() did some 54187c478bd9Sstevel@tonic-gate * sort of operation which caused a putpage in the bad 54197c478bd9Sstevel@tonic-gate * range. In this case, just return an error which will 54207c478bd9Sstevel@tonic-gate * cause the software modified bit on the page to set 54217c478bd9Sstevel@tonic-gate * and the page will get written out again later. 54227c478bd9Sstevel@tonic-gate */ 54237c478bd9Sstevel@tonic-gate if (ip->i_writer == curthread) { 54247c478bd9Sstevel@tonic-gate err = EIO; 54257c478bd9Sstevel@tonic-gate goto out; 54267c478bd9Sstevel@tonic-gate } 54277c478bd9Sstevel@tonic-gate /* 54287c478bd9Sstevel@tonic-gate * If the pager is trying to push a page in the bad range 54297c478bd9Sstevel@tonic-gate * just tell him to try again later when things are better. 54307c478bd9Sstevel@tonic-gate */ 54317c478bd9Sstevel@tonic-gate if (flags & B_ASYNC) { 54327c478bd9Sstevel@tonic-gate err = EAGAIN; 54337c478bd9Sstevel@tonic-gate goto out; 54347c478bd9Sstevel@tonic-gate } 54357c478bd9Sstevel@tonic-gate err = ufs_fault(ITOV(ip), "ufs_putapage: bn == UFS_HOLE"); 54367c478bd9Sstevel@tonic-gate goto out; 54377c478bd9Sstevel@tonic-gate } 54387c478bd9Sstevel@tonic-gate 54397c478bd9Sstevel@tonic-gate /* 5440303bf60bSsdebnath * If it is an fallocate'd block, reverse the negativity since 5441303bf60bSsdebnath * we are now writing to it 5442303bf60bSsdebnath */ 5443303bf60bSsdebnath if (ISFALLOCBLK(ip, bn)) { 5444303bf60bSsdebnath err = bmap_set_bn(vp, off, dbtofsb(fs, -bn)); 5445303bf60bSsdebnath if (err) 5446303bf60bSsdebnath goto out; 5447303bf60bSsdebnath 5448303bf60bSsdebnath bn = -bn; 5449303bf60bSsdebnath } 5450303bf60bSsdebnath 5451303bf60bSsdebnath /* 54527c478bd9Sstevel@tonic-gate * Take the length (of contiguous bytes) passed back from bmap() 54537c478bd9Sstevel@tonic-gate * and _try_ and get a set of pages covering that extent. 54547c478bd9Sstevel@tonic-gate */ 54557c478bd9Sstevel@tonic-gate pp = pvn_write_kluster(vp, pp, &io_off, &io_len, off, contig, flags); 54567c478bd9Sstevel@tonic-gate 54577c478bd9Sstevel@tonic-gate /* 54587c478bd9Sstevel@tonic-gate * May have run out of memory and not clustered backwards. 54597c478bd9Sstevel@tonic-gate * off p_offset 54607c478bd9Sstevel@tonic-gate * [ pp - 1 ][ pp ] 54617c478bd9Sstevel@tonic-gate * [ block ] 54627c478bd9Sstevel@tonic-gate * We told bmap off, so we have to adjust the bn accordingly. 54637c478bd9Sstevel@tonic-gate */ 54647c478bd9Sstevel@tonic-gate if (io_off > off) { 54657c478bd9Sstevel@tonic-gate bn += btod(io_off - off); 54667c478bd9Sstevel@tonic-gate contig -= (io_off - off); 54677c478bd9Sstevel@tonic-gate } 54687c478bd9Sstevel@tonic-gate 54697c478bd9Sstevel@tonic-gate /* 54707c478bd9Sstevel@tonic-gate * bmap was carefull to tell us the right size so use that. 54717c478bd9Sstevel@tonic-gate * There might be unallocated frags at the end. 54727c478bd9Sstevel@tonic-gate * LMXXX - bzero the end of the page? We must be writing after EOF. 54737c478bd9Sstevel@tonic-gate */ 54747c478bd9Sstevel@tonic-gate if (io_len > contig) { 54757c478bd9Sstevel@tonic-gate ASSERT(io_len - contig < fs->fs_bsize); 54767c478bd9Sstevel@tonic-gate io_len -= (io_len - contig); 54777c478bd9Sstevel@tonic-gate } 54787c478bd9Sstevel@tonic-gate 54797c478bd9Sstevel@tonic-gate /* 54807c478bd9Sstevel@tonic-gate * Handle the case where we are writing the last page after EOF. 54817c478bd9Sstevel@tonic-gate * 54827c478bd9Sstevel@tonic-gate * XXX - just a patch for i-mt3. 54837c478bd9Sstevel@tonic-gate */ 54847c478bd9Sstevel@tonic-gate if (io_len == 0) { 548580d34432Sfrankho ASSERT(pp->p_offset >= 548680d34432Sfrankho (u_offset_t)(roundup(ip->i_size, PAGESIZE))); 54877c478bd9Sstevel@tonic-gate io_len = PAGESIZE; 54887c478bd9Sstevel@tonic-gate } 54897c478bd9Sstevel@tonic-gate 54907c478bd9Sstevel@tonic-gate bp = pageio_setup(pp, io_len, ip->i_devvp, B_WRITE | flags); 54917c478bd9Sstevel@tonic-gate 54927c478bd9Sstevel@tonic-gate ULOCKFS_SET_MOD(ITOUL(ip)); 54937c478bd9Sstevel@tonic-gate 54947c478bd9Sstevel@tonic-gate bp->b_edev = ip->i_dev; 54957c478bd9Sstevel@tonic-gate bp->b_dev = cmpdev(ip->i_dev); 54967c478bd9Sstevel@tonic-gate bp->b_blkno = bn; 54977c478bd9Sstevel@tonic-gate bp->b_un.b_addr = (caddr_t)0; 54987c478bd9Sstevel@tonic-gate bp->b_file = ip->i_vnode; 54997c478bd9Sstevel@tonic-gate 5500b9dc641bSfrankho /* 5501b9dc641bSfrankho * File contents of shadow or quota inodes are metadata, and updates 5502b9dc641bSfrankho * to these need to be put into a logging transaction. All direct 5503b9dc641bSfrankho * callers in UFS do that, but fsflush can come here _before_ the 5504b9dc641bSfrankho * normal codepath. An example would be updating ACL information, for 5505b9dc641bSfrankho * which the normal codepath would be: 5506b9dc641bSfrankho * ufs_si_store() 5507b9dc641bSfrankho * ufs_rdwri() 5508b9dc641bSfrankho * wrip() 5509b9dc641bSfrankho * segmap_release() 5510b9dc641bSfrankho * VOP_PUTPAGE() 5511b9dc641bSfrankho * Here, fsflush can pick up the dirty page before segmap_release() 5512b9dc641bSfrankho * forces it out. If that happens, there's no transaction. 5513b9dc641bSfrankho * We therefore need to test whether a transaction exists, and if not 5514b9dc641bSfrankho * create one - for fsflush. 5515b9dc641bSfrankho */ 5516b9dc641bSfrankho dotrans = 5517b9dc641bSfrankho (((ip->i_mode & IFMT) == IFSHAD || ufsvfsp->vfs_qinod == ip) && 5518b9dc641bSfrankho ((curthread->t_flag & T_DONTBLOCK) == 0) && 5519b9dc641bSfrankho (TRANS_ISTRANS(ufsvfsp))); 5520b9dc641bSfrankho 5521b9dc641bSfrankho if (dotrans) { 5522b9dc641bSfrankho curthread->t_flag |= T_DONTBLOCK; 5523b9dc641bSfrankho TRANS_BEGIN_ASYNC(ufsvfsp, TOP_PUTPAGE, TOP_PUTPAGE_SIZE(ip)); 5524b9dc641bSfrankho } 55257c478bd9Sstevel@tonic-gate if (TRANS_ISTRANS(ufsvfsp)) { 55267c478bd9Sstevel@tonic-gate if ((ip->i_mode & IFMT) == IFSHAD) { 55277c478bd9Sstevel@tonic-gate TRANS_BUF(ufsvfsp, 0, io_len, bp, DT_SHAD); 55287c478bd9Sstevel@tonic-gate } else if (ufsvfsp->vfs_qinod == ip) { 55297c478bd9Sstevel@tonic-gate TRANS_DELTA(ufsvfsp, ldbtob(bn), bp->b_bcount, DT_QR, 55307c478bd9Sstevel@tonic-gate 0, 0); 55317c478bd9Sstevel@tonic-gate } 55327c478bd9Sstevel@tonic-gate } 5533b9dc641bSfrankho if (dotrans) { 5534b9dc641bSfrankho TRANS_END_ASYNC(ufsvfsp, TOP_PUTPAGE, TOP_PUTPAGE_SIZE(ip)); 5535b9dc641bSfrankho curthread->t_flag &= ~T_DONTBLOCK; 5536b9dc641bSfrankho } 55377c478bd9Sstevel@tonic-gate 55387c478bd9Sstevel@tonic-gate /* write throttle */ 55397c478bd9Sstevel@tonic-gate 55407c478bd9Sstevel@tonic-gate ASSERT(bp->b_iodone == NULL); 55417c478bd9Sstevel@tonic-gate bp->b_iodone = (int (*)())ufs_iodone; 55427c478bd9Sstevel@tonic-gate mutex_enter(&ip->i_tlock); 55437c478bd9Sstevel@tonic-gate ip->i_writes += bp->b_bcount; 55447c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 55457c478bd9Sstevel@tonic-gate 55467c478bd9Sstevel@tonic-gate if (bp->b_flags & B_ASYNC) { 55477c478bd9Sstevel@tonic-gate if (ufsvfsp->vfs_log) { 55487c478bd9Sstevel@tonic-gate lufs_write_strategy(ufsvfsp->vfs_log, bp); 55497c478bd9Sstevel@tonic-gate } else if (ufsvfsp->vfs_snapshot) { 55507c478bd9Sstevel@tonic-gate fssnap_strategy(&ufsvfsp->vfs_snapshot, bp); 55517c478bd9Sstevel@tonic-gate } else { 5552d3d50737SRafael Vanoni ufsvfsp->vfs_iotstamp = ddi_get_lbolt(); 55537c478bd9Sstevel@tonic-gate ub.ub_putasyncs.value.ul++; 55547c478bd9Sstevel@tonic-gate (void) bdev_strategy(bp); 55557c478bd9Sstevel@tonic-gate lwp_stat_update(LWP_STAT_OUBLK, 1); 55567c478bd9Sstevel@tonic-gate } 55577c478bd9Sstevel@tonic-gate } else { 55587c478bd9Sstevel@tonic-gate if (ufsvfsp->vfs_log) { 55597c478bd9Sstevel@tonic-gate lufs_write_strategy(ufsvfsp->vfs_log, bp); 55607c478bd9Sstevel@tonic-gate } else if (ufsvfsp->vfs_snapshot) { 55617c478bd9Sstevel@tonic-gate fssnap_strategy(&ufsvfsp->vfs_snapshot, bp); 55627c478bd9Sstevel@tonic-gate } else { 5563d3d50737SRafael Vanoni ufsvfsp->vfs_iotstamp = ddi_get_lbolt(); 55647c478bd9Sstevel@tonic-gate ub.ub_putsyncs.value.ul++; 55657c478bd9Sstevel@tonic-gate (void) bdev_strategy(bp); 55667c478bd9Sstevel@tonic-gate lwp_stat_update(LWP_STAT_OUBLK, 1); 55677c478bd9Sstevel@tonic-gate } 55687c478bd9Sstevel@tonic-gate err = biowait(bp); 55697c478bd9Sstevel@tonic-gate pageio_done(bp); 55707c478bd9Sstevel@tonic-gate pvn_write_done(pp, ((err) ? B_ERROR : 0) | B_WRITE | flags); 55717c478bd9Sstevel@tonic-gate } 55727c478bd9Sstevel@tonic-gate 55737c478bd9Sstevel@tonic-gate pp = NULL; 55747c478bd9Sstevel@tonic-gate 55757c478bd9Sstevel@tonic-gate out: 55767c478bd9Sstevel@tonic-gate if (err != 0 && pp != NULL) 55777c478bd9Sstevel@tonic-gate pvn_write_done(pp, B_ERROR | B_WRITE | flags); 55787c478bd9Sstevel@tonic-gate 55797c478bd9Sstevel@tonic-gate if (offp) 55807c478bd9Sstevel@tonic-gate *offp = io_off; 55817c478bd9Sstevel@tonic-gate if (lenp) 55827c478bd9Sstevel@tonic-gate *lenp = io_len; 55837c478bd9Sstevel@tonic-gate out_trace: 55847c478bd9Sstevel@tonic-gate return (err); 55857c478bd9Sstevel@tonic-gate } 55867c478bd9Sstevel@tonic-gate 55876ac3b8a8Svsakar uint64_t ufs_map_alock_retry_cnt; 55886ac3b8a8Svsakar uint64_t ufs_map_lockfs_retry_cnt; 55896ac3b8a8Svsakar 55907c478bd9Sstevel@tonic-gate /* ARGSUSED */ 55917c478bd9Sstevel@tonic-gate static int 55927c478bd9Sstevel@tonic-gate ufs_map(struct vnode *vp, 55937c478bd9Sstevel@tonic-gate offset_t off, 55947c478bd9Sstevel@tonic-gate struct as *as, 55957c478bd9Sstevel@tonic-gate caddr_t *addrp, 55967c478bd9Sstevel@tonic-gate size_t len, 55977c478bd9Sstevel@tonic-gate uchar_t prot, 55987c478bd9Sstevel@tonic-gate uchar_t maxprot, 55997c478bd9Sstevel@tonic-gate uint_t flags, 5600da6c28aaSamw struct cred *cr, 5601da6c28aaSamw caller_context_t *ct) 56027c478bd9Sstevel@tonic-gate { 56037c478bd9Sstevel@tonic-gate struct segvn_crargs vn_a; 56047c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp = VTOI(vp)->i_ufsvfs; 56057c478bd9Sstevel@tonic-gate struct ulockfs *ulp; 56066ac3b8a8Svsakar int error, sig; 56076ac3b8a8Svsakar k_sigset_t smask; 560857846e82Svsakar caddr_t hint = *addrp; 56097c478bd9Sstevel@tonic-gate 56107c478bd9Sstevel@tonic-gate if (vp->v_flag & VNOMAP) { 56117c478bd9Sstevel@tonic-gate error = ENOSYS; 56126ac3b8a8Svsakar goto out; 56137c478bd9Sstevel@tonic-gate } 56147c478bd9Sstevel@tonic-gate 56157c478bd9Sstevel@tonic-gate if (off < (offset_t)0 || (offset_t)(off + len) < (offset_t)0) { 56167c478bd9Sstevel@tonic-gate error = ENXIO; 56176ac3b8a8Svsakar goto out; 56187c478bd9Sstevel@tonic-gate } 56197c478bd9Sstevel@tonic-gate 56207c478bd9Sstevel@tonic-gate if (vp->v_type != VREG) { 56217c478bd9Sstevel@tonic-gate error = ENODEV; 56226ac3b8a8Svsakar goto out; 56237c478bd9Sstevel@tonic-gate } 56247c478bd9Sstevel@tonic-gate 56256ac3b8a8Svsakar retry_map: 562657846e82Svsakar *addrp = hint; 56277c478bd9Sstevel@tonic-gate /* 56287c478bd9Sstevel@tonic-gate * If file is being locked, disallow mapping. 56297c478bd9Sstevel@tonic-gate */ 56307c478bd9Sstevel@tonic-gate if (vn_has_mandatory_locks(vp, VTOI(vp)->i_mode)) { 56317c478bd9Sstevel@tonic-gate error = EAGAIN; 56326ac3b8a8Svsakar goto out; 56337c478bd9Sstevel@tonic-gate } 56347c478bd9Sstevel@tonic-gate 56357c478bd9Sstevel@tonic-gate as_rangelock(as); 56367c478bd9Sstevel@tonic-gate /* 563760946fe0Smec * Note that if we are retrying (because ufs_lockfs_trybegin failed in 56386ac3b8a8Svsakar * the previous attempt), some other thread could have grabbed 563960946fe0Smec * the same VA range if MAP_FIXED is set. In that case, choose_addr 564060946fe0Smec * would unmap the valid VA range, that is ok. 56417c478bd9Sstevel@tonic-gate */ 564260946fe0Smec error = choose_addr(as, addrp, len, off, ADDR_VACALIGN, flags); 564360946fe0Smec if (error != 0) { 564460946fe0Smec as_rangeunlock(as); 564560946fe0Smec goto out; 56467c478bd9Sstevel@tonic-gate } 56477c478bd9Sstevel@tonic-gate 56486ac3b8a8Svsakar /* 56496ac3b8a8Svsakar * a_lock has to be acquired before entering the lockfs protocol 56506ac3b8a8Svsakar * because that is the order in which pagefault works. Also we cannot 56516ac3b8a8Svsakar * block on a_lock here because this waiting writer will prevent 56526ac3b8a8Svsakar * further readers like ufs_read from progressing and could cause 56536ac3b8a8Svsakar * deadlock between ufs_read/ufs_map/pagefault when a quiesce is 56546ac3b8a8Svsakar * pending. 56556ac3b8a8Svsakar */ 5656*dc32d872SJosef 'Jeff' Sipek while (!AS_LOCK_TRYENTER(as, RW_WRITER)) { 56576ac3b8a8Svsakar ufs_map_alock_retry_cnt++; 56586ac3b8a8Svsakar delay(RETRY_LOCK_DELAY); 56596ac3b8a8Svsakar } 56606ac3b8a8Svsakar 56616ac3b8a8Svsakar /* 56626ac3b8a8Svsakar * We can't hold as->a_lock and wait for lockfs to succeed because 56636ac3b8a8Svsakar * the proc tools might hang on a_lock, so call ufs_lockfs_trybegin() 56646ac3b8a8Svsakar * instead. 56656ac3b8a8Svsakar */ 56666ac3b8a8Svsakar if (error = ufs_lockfs_trybegin(ufsvfsp, &ulp, ULOCKFS_MAP_MASK)) { 56676ac3b8a8Svsakar /* 56686ac3b8a8Svsakar * ufs_lockfs_trybegin() did not succeed. It is safer to give up 56696ac3b8a8Svsakar * as->a_lock and wait for ulp->ul_fs_lock status to change. 56706ac3b8a8Svsakar */ 56716ac3b8a8Svsakar ufs_map_lockfs_retry_cnt++; 5672*dc32d872SJosef 'Jeff' Sipek AS_LOCK_EXIT(as); 56736ac3b8a8Svsakar as_rangeunlock(as); 56746ac3b8a8Svsakar if (error == EIO) 56756ac3b8a8Svsakar goto out; 56766ac3b8a8Svsakar 56776ac3b8a8Svsakar mutex_enter(&ulp->ul_lock); 56786ac3b8a8Svsakar while (ulp->ul_fs_lock & ULOCKFS_MAP_MASK) { 56796ac3b8a8Svsakar if (ULOCKFS_IS_SLOCK(ulp) || ufsvfsp->vfs_nointr) { 56806ac3b8a8Svsakar cv_wait(&ulp->ul_cv, &ulp->ul_lock); 56816ac3b8a8Svsakar } else { 56826ac3b8a8Svsakar sigintr(&smask, 1); 56836ac3b8a8Svsakar sig = cv_wait_sig(&ulp->ul_cv, &ulp->ul_lock); 56846ac3b8a8Svsakar sigunintr(&smask); 56856ac3b8a8Svsakar if (((ulp->ul_fs_lock & ULOCKFS_MAP_MASK) && 56866ac3b8a8Svsakar !sig) || ufsvfsp->vfs_dontblock) { 56876ac3b8a8Svsakar mutex_exit(&ulp->ul_lock); 56886ac3b8a8Svsakar return (EINTR); 56896ac3b8a8Svsakar } 56906ac3b8a8Svsakar } 56916ac3b8a8Svsakar } 56926ac3b8a8Svsakar mutex_exit(&ulp->ul_lock); 56936ac3b8a8Svsakar goto retry_map; 56946ac3b8a8Svsakar } 56956ac3b8a8Svsakar 56967c478bd9Sstevel@tonic-gate vn_a.vp = vp; 56977c478bd9Sstevel@tonic-gate vn_a.offset = (u_offset_t)off; 56987c478bd9Sstevel@tonic-gate vn_a.type = flags & MAP_TYPE; 56997c478bd9Sstevel@tonic-gate vn_a.prot = prot; 57007c478bd9Sstevel@tonic-gate vn_a.maxprot = maxprot; 57017c478bd9Sstevel@tonic-gate vn_a.cred = cr; 57027c478bd9Sstevel@tonic-gate vn_a.amp = NULL; 57037c478bd9Sstevel@tonic-gate vn_a.flags = flags & ~MAP_TYPE; 57047c478bd9Sstevel@tonic-gate vn_a.szc = 0; 57057c478bd9Sstevel@tonic-gate vn_a.lgrp_mem_policy_flags = 0; 57067c478bd9Sstevel@tonic-gate 570702ff05a9Svsakar error = as_map_locked(as, *addrp, len, segvn_create, &vn_a); 57086ac3b8a8Svsakar if (ulp) 57097c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 57106ac3b8a8Svsakar as_rangeunlock(as); 57117c478bd9Sstevel@tonic-gate out: 57127c478bd9Sstevel@tonic-gate return (error); 57137c478bd9Sstevel@tonic-gate } 57147c478bd9Sstevel@tonic-gate 57157c478bd9Sstevel@tonic-gate /* ARGSUSED */ 57167c478bd9Sstevel@tonic-gate static int 57177c478bd9Sstevel@tonic-gate ufs_addmap(struct vnode *vp, 57187c478bd9Sstevel@tonic-gate offset_t off, 57197c478bd9Sstevel@tonic-gate struct as *as, 57207c478bd9Sstevel@tonic-gate caddr_t addr, 57217c478bd9Sstevel@tonic-gate size_t len, 57227c478bd9Sstevel@tonic-gate uchar_t prot, 57237c478bd9Sstevel@tonic-gate uchar_t maxprot, 57247c478bd9Sstevel@tonic-gate uint_t flags, 5725da6c28aaSamw struct cred *cr, 5726da6c28aaSamw caller_context_t *ct) 57277c478bd9Sstevel@tonic-gate { 57287c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 57297c478bd9Sstevel@tonic-gate 57307c478bd9Sstevel@tonic-gate if (vp->v_flag & VNOMAP) { 57317c478bd9Sstevel@tonic-gate return (ENOSYS); 57327c478bd9Sstevel@tonic-gate } 57337c478bd9Sstevel@tonic-gate 57347c478bd9Sstevel@tonic-gate mutex_enter(&ip->i_tlock); 57357c478bd9Sstevel@tonic-gate ip->i_mapcnt += btopr(len); 57367c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 57377c478bd9Sstevel@tonic-gate return (0); 57387c478bd9Sstevel@tonic-gate } 57397c478bd9Sstevel@tonic-gate 57407c478bd9Sstevel@tonic-gate /*ARGSUSED*/ 57417c478bd9Sstevel@tonic-gate static int 57427c478bd9Sstevel@tonic-gate ufs_delmap(struct vnode *vp, offset_t off, struct as *as, caddr_t addr, 57437c478bd9Sstevel@tonic-gate size_t len, uint_t prot, uint_t maxprot, uint_t flags, 5744da6c28aaSamw struct cred *cr, caller_context_t *ct) 57457c478bd9Sstevel@tonic-gate { 57467c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 57477c478bd9Sstevel@tonic-gate 57487c478bd9Sstevel@tonic-gate if (vp->v_flag & VNOMAP) { 57497c478bd9Sstevel@tonic-gate return (ENOSYS); 57507c478bd9Sstevel@tonic-gate } 57517c478bd9Sstevel@tonic-gate 57527c478bd9Sstevel@tonic-gate mutex_enter(&ip->i_tlock); 57537c478bd9Sstevel@tonic-gate ip->i_mapcnt -= btopr(len); /* Count released mappings */ 57547c478bd9Sstevel@tonic-gate ASSERT(ip->i_mapcnt >= 0); 57557c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 57567c478bd9Sstevel@tonic-gate return (0); 57577c478bd9Sstevel@tonic-gate } 57587c478bd9Sstevel@tonic-gate /* 57597c478bd9Sstevel@tonic-gate * Return the answer requested to poll() for non-device files 57607c478bd9Sstevel@tonic-gate */ 57617c478bd9Sstevel@tonic-gate struct pollhead ufs_pollhd; 57627c478bd9Sstevel@tonic-gate 57637c478bd9Sstevel@tonic-gate /* ARGSUSED */ 57647c478bd9Sstevel@tonic-gate int 5765da6c28aaSamw ufs_poll(vnode_t *vp, short ev, int any, short *revp, struct pollhead **phpp, 5766da6c28aaSamw caller_context_t *ct) 57677c478bd9Sstevel@tonic-gate { 57687c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp; 57697c478bd9Sstevel@tonic-gate 57707c478bd9Sstevel@tonic-gate *revp = 0; 57717c478bd9Sstevel@tonic-gate ufsvfsp = VTOI(vp)->i_ufsvfs; 57727c478bd9Sstevel@tonic-gate 57737c478bd9Sstevel@tonic-gate if (!ufsvfsp) { 57747c478bd9Sstevel@tonic-gate *revp = POLLHUP; 57757c478bd9Sstevel@tonic-gate goto out; 57767c478bd9Sstevel@tonic-gate } 57777c478bd9Sstevel@tonic-gate 57787c478bd9Sstevel@tonic-gate if (ULOCKFS_IS_HLOCK(&ufsvfsp->vfs_ulockfs) || 57797c478bd9Sstevel@tonic-gate ULOCKFS_IS_ELOCK(&ufsvfsp->vfs_ulockfs)) { 57807c478bd9Sstevel@tonic-gate *revp |= POLLERR; 57817c478bd9Sstevel@tonic-gate 57827c478bd9Sstevel@tonic-gate } else { 57837c478bd9Sstevel@tonic-gate if ((ev & POLLOUT) && !ufsvfsp->vfs_fs->fs_ronly && 57847c478bd9Sstevel@tonic-gate !ULOCKFS_IS_WLOCK(&ufsvfsp->vfs_ulockfs)) 57857c478bd9Sstevel@tonic-gate *revp |= POLLOUT; 57867c478bd9Sstevel@tonic-gate 57877c478bd9Sstevel@tonic-gate if ((ev & POLLWRBAND) && !ufsvfsp->vfs_fs->fs_ronly && 57887c478bd9Sstevel@tonic-gate !ULOCKFS_IS_WLOCK(&ufsvfsp->vfs_ulockfs)) 57897c478bd9Sstevel@tonic-gate *revp |= POLLWRBAND; 57907c478bd9Sstevel@tonic-gate 57917c478bd9Sstevel@tonic-gate if (ev & POLLIN) 57927c478bd9Sstevel@tonic-gate *revp |= POLLIN; 57937c478bd9Sstevel@tonic-gate 57947c478bd9Sstevel@tonic-gate if (ev & POLLRDNORM) 57957c478bd9Sstevel@tonic-gate *revp |= POLLRDNORM; 57967c478bd9Sstevel@tonic-gate 57977c478bd9Sstevel@tonic-gate if (ev & POLLRDBAND) 57987c478bd9Sstevel@tonic-gate *revp |= POLLRDBAND; 57997c478bd9Sstevel@tonic-gate } 58007c478bd9Sstevel@tonic-gate 58017c478bd9Sstevel@tonic-gate if ((ev & POLLPRI) && (*revp & (POLLERR|POLLHUP))) 58027c478bd9Sstevel@tonic-gate *revp |= POLLPRI; 58037c478bd9Sstevel@tonic-gate out: 58047c478bd9Sstevel@tonic-gate *phpp = !any && !*revp ? &ufs_pollhd : (struct pollhead *)NULL; 58057c478bd9Sstevel@tonic-gate 58067c478bd9Sstevel@tonic-gate return (0); 58077c478bd9Sstevel@tonic-gate } 58087c478bd9Sstevel@tonic-gate 58097c478bd9Sstevel@tonic-gate /* ARGSUSED */ 58107c478bd9Sstevel@tonic-gate static int 5811da6c28aaSamw ufs_l_pathconf(struct vnode *vp, int cmd, ulong_t *valp, struct cred *cr, 5812da6c28aaSamw caller_context_t *ct) 58137c478bd9Sstevel@tonic-gate { 58147c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp = VTOI(vp)->i_ufsvfs; 58157c478bd9Sstevel@tonic-gate struct ulockfs *ulp = NULL; 58167c478bd9Sstevel@tonic-gate struct inode *sip = NULL; 58177c478bd9Sstevel@tonic-gate int error; 58187c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 58197c478bd9Sstevel@tonic-gate int issync; 58207c478bd9Sstevel@tonic-gate 58217c478bd9Sstevel@tonic-gate error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_PATHCONF_MASK); 58227c478bd9Sstevel@tonic-gate if (error) 58237c478bd9Sstevel@tonic-gate return (error); 58247c478bd9Sstevel@tonic-gate 58257c478bd9Sstevel@tonic-gate switch (cmd) { 58267c478bd9Sstevel@tonic-gate /* 58277c478bd9Sstevel@tonic-gate * Have to handle _PC_NAME_MAX here, because the normal way 58287c478bd9Sstevel@tonic-gate * [fs_pathconf() -> VOP_STATVFS() -> ufs_statvfs()] 58297c478bd9Sstevel@tonic-gate * results in a lock ordering reversal between 58307c478bd9Sstevel@tonic-gate * ufs_lockfs_{begin,end}() and 58317c478bd9Sstevel@tonic-gate * ufs_thread_{suspend,continue}(). 58327c478bd9Sstevel@tonic-gate * 58337c478bd9Sstevel@tonic-gate * Keep in sync with ufs_statvfs(). 58347c478bd9Sstevel@tonic-gate */ 58357c478bd9Sstevel@tonic-gate case _PC_NAME_MAX: 58367c478bd9Sstevel@tonic-gate *valp = MAXNAMLEN; 58377c478bd9Sstevel@tonic-gate break; 58387c478bd9Sstevel@tonic-gate 58397c478bd9Sstevel@tonic-gate case _PC_FILESIZEBITS: 58407c478bd9Sstevel@tonic-gate if (ufsvfsp->vfs_lfflags & UFS_LARGEFILES) 58417c478bd9Sstevel@tonic-gate *valp = UFS_FILESIZE_BITS; 58427c478bd9Sstevel@tonic-gate else 58437c478bd9Sstevel@tonic-gate *valp = 32; 58447c478bd9Sstevel@tonic-gate break; 58457c478bd9Sstevel@tonic-gate 58467c478bd9Sstevel@tonic-gate case _PC_XATTR_EXISTS: 58477c478bd9Sstevel@tonic-gate if (vp->v_vfsp->vfs_flag & VFS_XATTR) { 58487c478bd9Sstevel@tonic-gate 584980d34432Sfrankho error = 585080d34432Sfrankho ufs_xattr_getattrdir(vp, &sip, LOOKUP_XATTR, cr); 58517c478bd9Sstevel@tonic-gate if (error == 0 && sip != NULL) { 58527c478bd9Sstevel@tonic-gate /* Start transaction */ 58537c478bd9Sstevel@tonic-gate if (ulp) { 58547c478bd9Sstevel@tonic-gate TRANS_BEGIN_CSYNC(ufsvfsp, issync, 58557c478bd9Sstevel@tonic-gate TOP_RMDIR, TOP_RMDIR_SIZE); 58567c478bd9Sstevel@tonic-gate } 58577c478bd9Sstevel@tonic-gate /* 58587c478bd9Sstevel@tonic-gate * Is directory empty 58597c478bd9Sstevel@tonic-gate */ 58607c478bd9Sstevel@tonic-gate rw_enter(&sip->i_rwlock, RW_WRITER); 58617c478bd9Sstevel@tonic-gate rw_enter(&sip->i_contents, RW_WRITER); 58627c478bd9Sstevel@tonic-gate if (ufs_xattrdirempty(sip, 58637c478bd9Sstevel@tonic-gate sip->i_number, CRED())) { 58647c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_WRITER); 58657c478bd9Sstevel@tonic-gate ufs_unhook_shadow(ip, sip); 58667c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 58677c478bd9Sstevel@tonic-gate 58687c478bd9Sstevel@tonic-gate *valp = 0; 58697c478bd9Sstevel@tonic-gate 58707c478bd9Sstevel@tonic-gate } else 58717c478bd9Sstevel@tonic-gate *valp = 1; 58727c478bd9Sstevel@tonic-gate rw_exit(&sip->i_contents); 58737c478bd9Sstevel@tonic-gate rw_exit(&sip->i_rwlock); 58747c478bd9Sstevel@tonic-gate if (ulp) { 58757c478bd9Sstevel@tonic-gate TRANS_END_CSYNC(ufsvfsp, error, issync, 58767c478bd9Sstevel@tonic-gate TOP_RMDIR, TOP_RMDIR_SIZE); 58777c478bd9Sstevel@tonic-gate } 58787c478bd9Sstevel@tonic-gate VN_RELE(ITOV(sip)); 58797c478bd9Sstevel@tonic-gate } else if (error == ENOENT) { 58807c478bd9Sstevel@tonic-gate *valp = 0; 58817c478bd9Sstevel@tonic-gate error = 0; 58827c478bd9Sstevel@tonic-gate } 58837c478bd9Sstevel@tonic-gate } else { 5884da6c28aaSamw error = fs_pathconf(vp, cmd, valp, cr, ct); 58857c478bd9Sstevel@tonic-gate } 58867c478bd9Sstevel@tonic-gate break; 58877c478bd9Sstevel@tonic-gate 58887c478bd9Sstevel@tonic-gate case _PC_ACL_ENABLED: 58897c478bd9Sstevel@tonic-gate *valp = _ACL_ACLENT_ENABLED; 58907c478bd9Sstevel@tonic-gate break; 58917c478bd9Sstevel@tonic-gate 58927c478bd9Sstevel@tonic-gate case _PC_MIN_HOLE_SIZE: 58937c478bd9Sstevel@tonic-gate *valp = (ulong_t)ip->i_fs->fs_bsize; 58947c478bd9Sstevel@tonic-gate break; 58957c478bd9Sstevel@tonic-gate 5896da6c28aaSamw case _PC_SATTR_ENABLED: 5897da6c28aaSamw case _PC_SATTR_EXISTS: 58989660e5cbSJanice Chang *valp = vfs_has_feature(vp->v_vfsp, VFSFT_SYSATTR_VIEWS) && 5899da6c28aaSamw (vp->v_type == VREG || vp->v_type == VDIR); 5900da6c28aaSamw break; 5901da6c28aaSamw 59023b862e9aSRoger A. Faulkner case _PC_TIMESTAMP_RESOLUTION: 59033b862e9aSRoger A. Faulkner /* 59043b862e9aSRoger A. Faulkner * UFS keeps only microsecond timestamp resolution. 59053b862e9aSRoger A. Faulkner * This is historical and will probably never change. 59063b862e9aSRoger A. Faulkner */ 59073b862e9aSRoger A. Faulkner *valp = 1000L; 59083b862e9aSRoger A. Faulkner break; 59093b862e9aSRoger A. Faulkner 59107c478bd9Sstevel@tonic-gate default: 5911da6c28aaSamw error = fs_pathconf(vp, cmd, valp, cr, ct); 59123b862e9aSRoger A. Faulkner break; 59137c478bd9Sstevel@tonic-gate } 59147c478bd9Sstevel@tonic-gate 59157c478bd9Sstevel@tonic-gate if (ulp != NULL) { 59167c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 59177c478bd9Sstevel@tonic-gate } 59187c478bd9Sstevel@tonic-gate return (error); 59197c478bd9Sstevel@tonic-gate } 59207c478bd9Sstevel@tonic-gate 59217c478bd9Sstevel@tonic-gate int ufs_pageio_writes, ufs_pageio_reads; 59227c478bd9Sstevel@tonic-gate 59237c478bd9Sstevel@tonic-gate /*ARGSUSED*/ 59247c478bd9Sstevel@tonic-gate static int 59257c478bd9Sstevel@tonic-gate ufs_pageio(struct vnode *vp, page_t *pp, u_offset_t io_off, size_t io_len, 5926da6c28aaSamw int flags, struct cred *cr, caller_context_t *ct) 59277c478bd9Sstevel@tonic-gate { 59287c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 59297c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp; 59307c478bd9Sstevel@tonic-gate page_t *npp = NULL, *opp = NULL, *cpp = pp; 59317c478bd9Sstevel@tonic-gate struct buf *bp; 59327c478bd9Sstevel@tonic-gate daddr_t bn; 59337c478bd9Sstevel@tonic-gate size_t done_len = 0, cur_len = 0; 59347c478bd9Sstevel@tonic-gate int err = 0; 59357c478bd9Sstevel@tonic-gate int contig = 0; 59367c478bd9Sstevel@tonic-gate int dolock; 59377c478bd9Sstevel@tonic-gate int vmpss = 0; 5938bc69f433Saguzovsk struct ulockfs *ulp; 59397c478bd9Sstevel@tonic-gate 59407c478bd9Sstevel@tonic-gate if ((flags & B_READ) && pp != NULL && pp->p_vnode == vp && 59417c478bd9Sstevel@tonic-gate vp->v_mpssdata != NULL) { 59427c478bd9Sstevel@tonic-gate vmpss = 1; 59437c478bd9Sstevel@tonic-gate } 59447c478bd9Sstevel@tonic-gate 59457c478bd9Sstevel@tonic-gate dolock = (rw_owner(&ip->i_contents) != curthread); 59467c478bd9Sstevel@tonic-gate /* 59477c478bd9Sstevel@tonic-gate * We need a better check. Ideally, we would use another 59487c478bd9Sstevel@tonic-gate * vnodeops so that hlocked and forcibly unmounted file 59497c478bd9Sstevel@tonic-gate * systems would return EIO where appropriate and w/o the 59507c478bd9Sstevel@tonic-gate * need for these checks. 59517c478bd9Sstevel@tonic-gate */ 59527c478bd9Sstevel@tonic-gate if ((ufsvfsp = ip->i_ufsvfs) == NULL) 59537c478bd9Sstevel@tonic-gate return (EIO); 59547c478bd9Sstevel@tonic-gate 5955bc69f433Saguzovsk /* 5956bc69f433Saguzovsk * For vmpss (pp can be NULL) case respect the quiesce protocol. 5957bc69f433Saguzovsk * ul_lock must be taken before locking pages so we can't use it here 5958bc69f433Saguzovsk * if pp is non NULL because segvn already locked pages 5959bc69f433Saguzovsk * SE_EXCL. Instead we rely on the fact that a forced umount or 5960bc69f433Saguzovsk * applying a filesystem lock via ufs_fiolfs() will block in the 5961bc69f433Saguzovsk * implicit call to ufs_flush() until we unlock the pages after the 5962bc69f433Saguzovsk * return to segvn. Other ufs_quiesce() callers keep ufs_quiesce_pend 5963bc69f433Saguzovsk * above 0 until they are done. We have to be careful not to increment 5964bc69f433Saguzovsk * ul_vnops_cnt here after forceful unmount hlocks the file system. 5965bc69f433Saguzovsk * 5966bc69f433Saguzovsk * If pp is NULL use ul_lock to make sure we don't increment 5967bc69f433Saguzovsk * ul_vnops_cnt after forceful unmount hlocks the file system. 5968bc69f433Saguzovsk */ 5969bc69f433Saguzovsk if (vmpss || pp == NULL) { 5970bc69f433Saguzovsk ulp = &ufsvfsp->vfs_ulockfs; 5971bc69f433Saguzovsk if (pp == NULL) 5972bc69f433Saguzovsk mutex_enter(&ulp->ul_lock); 5973bc69f433Saguzovsk if (ulp->ul_fs_lock & ULOCKFS_GETREAD_MASK) { 5974bc69f433Saguzovsk if (pp == NULL) { 5975bc69f433Saguzovsk mutex_exit(&ulp->ul_lock); 5976bc69f433Saguzovsk } 5977bc69f433Saguzovsk return (vmpss ? EIO : EINVAL); 5978bc69f433Saguzovsk } 59791a5e258fSJosef 'Jeff' Sipek atomic_inc_ulong(&ulp->ul_vnops_cnt); 5980bc69f433Saguzovsk if (pp == NULL) 5981bc69f433Saguzovsk mutex_exit(&ulp->ul_lock); 5982bc69f433Saguzovsk if (ufs_quiesce_pend) { 59831a5e258fSJosef 'Jeff' Sipek if (!atomic_dec_ulong_nv(&ulp->ul_vnops_cnt)) 5984bc69f433Saguzovsk cv_broadcast(&ulp->ul_cv); 5985bc69f433Saguzovsk return (vmpss ? EIO : EINVAL); 5986bc69f433Saguzovsk } 5987bc69f433Saguzovsk } 5988bc69f433Saguzovsk 59897c478bd9Sstevel@tonic-gate if (dolock) { 59907c478bd9Sstevel@tonic-gate /* 59917c478bd9Sstevel@tonic-gate * segvn may call VOP_PAGEIO() instead of VOP_GETPAGE() to 59927c478bd9Sstevel@tonic-gate * handle a fault against a segment that maps vnode pages with 59937c478bd9Sstevel@tonic-gate * large mappings. Segvn creates pages and holds them locked 59947c478bd9Sstevel@tonic-gate * SE_EXCL during VOP_PAGEIO() call. In this case we have to 59957c478bd9Sstevel@tonic-gate * use rw_tryenter() to avoid a potential deadlock since in 59967c478bd9Sstevel@tonic-gate * lock order i_contents needs to be taken first. 59977c478bd9Sstevel@tonic-gate * Segvn will retry via VOP_GETPAGE() if VOP_PAGEIO() fails. 59987c478bd9Sstevel@tonic-gate */ 59997c478bd9Sstevel@tonic-gate if (!vmpss) { 60007c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_READER); 60017c478bd9Sstevel@tonic-gate } else if (!rw_tryenter(&ip->i_contents, RW_READER)) { 60021a5e258fSJosef 'Jeff' Sipek if (!atomic_dec_ulong_nv(&ulp->ul_vnops_cnt)) 6003bc69f433Saguzovsk cv_broadcast(&ulp->ul_cv); 60047c478bd9Sstevel@tonic-gate return (EDEADLK); 60057c478bd9Sstevel@tonic-gate } 60067c478bd9Sstevel@tonic-gate } 60077c478bd9Sstevel@tonic-gate 600837fbc076Saguzovsk /* 600937fbc076Saguzovsk * Return an error to segvn because the pagefault request is beyond 601037fbc076Saguzovsk * PAGESIZE rounded EOF. 601137fbc076Saguzovsk */ 601237fbc076Saguzovsk if (vmpss && btopr(io_off + io_len) > btopr(ip->i_size)) { 601337fbc076Saguzovsk if (dolock) 601437fbc076Saguzovsk rw_exit(&ip->i_contents); 60151a5e258fSJosef 'Jeff' Sipek if (!atomic_dec_ulong_nv(&ulp->ul_vnops_cnt)) 6016bc69f433Saguzovsk cv_broadcast(&ulp->ul_cv); 601737fbc076Saguzovsk return (EFAULT); 601837fbc076Saguzovsk } 601937fbc076Saguzovsk 60207c478bd9Sstevel@tonic-gate if (pp == NULL) { 60217c478bd9Sstevel@tonic-gate if (bmap_has_holes(ip)) { 60227c478bd9Sstevel@tonic-gate err = ENOSYS; 60237c478bd9Sstevel@tonic-gate } else { 60247c478bd9Sstevel@tonic-gate err = EINVAL; 60257c478bd9Sstevel@tonic-gate } 60267c478bd9Sstevel@tonic-gate if (dolock) 60277c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 60281a5e258fSJosef 'Jeff' Sipek if (!atomic_dec_ulong_nv(&ulp->ul_vnops_cnt)) 6029bc69f433Saguzovsk cv_broadcast(&ulp->ul_cv); 60307c478bd9Sstevel@tonic-gate return (err); 60317c478bd9Sstevel@tonic-gate } 60327c478bd9Sstevel@tonic-gate 60337c478bd9Sstevel@tonic-gate /* 60347c478bd9Sstevel@tonic-gate * Break the io request into chunks, one for each contiguous 60357c478bd9Sstevel@tonic-gate * stretch of disk blocks in the target file. 60367c478bd9Sstevel@tonic-gate */ 60377c478bd9Sstevel@tonic-gate while (done_len < io_len) { 60387c478bd9Sstevel@tonic-gate ASSERT(cpp); 60397c478bd9Sstevel@tonic-gate contig = 0; 60407c478bd9Sstevel@tonic-gate if (err = bmap_read(ip, (u_offset_t)(io_off + done_len), 60417c478bd9Sstevel@tonic-gate &bn, &contig)) 60427c478bd9Sstevel@tonic-gate break; 60437c478bd9Sstevel@tonic-gate 60447c478bd9Sstevel@tonic-gate if (bn == UFS_HOLE) { /* No holey swapfiles */ 60457c478bd9Sstevel@tonic-gate if (vmpss) { 60467c478bd9Sstevel@tonic-gate err = EFAULT; 60477c478bd9Sstevel@tonic-gate break; 60487c478bd9Sstevel@tonic-gate } 60497c478bd9Sstevel@tonic-gate err = ufs_fault(ITOV(ip), "ufs_pageio: bn == UFS_HOLE"); 60507c478bd9Sstevel@tonic-gate break; 60517c478bd9Sstevel@tonic-gate } 60527c478bd9Sstevel@tonic-gate 60537c478bd9Sstevel@tonic-gate cur_len = MIN(io_len - done_len, contig); 60547c478bd9Sstevel@tonic-gate /* 60557c478bd9Sstevel@tonic-gate * Zero out a page beyond EOF, when the last block of 60567c478bd9Sstevel@tonic-gate * a file is a UFS fragment so that ufs_pageio() can be used 60577c478bd9Sstevel@tonic-gate * instead of ufs_getpage() to handle faults against 60587c478bd9Sstevel@tonic-gate * segvn segments that use large pages. 60597c478bd9Sstevel@tonic-gate */ 60607c478bd9Sstevel@tonic-gate page_list_break(&cpp, &npp, btopr(cur_len)); 60617c478bd9Sstevel@tonic-gate if ((flags & B_READ) && (cur_len & PAGEOFFSET)) { 60627c478bd9Sstevel@tonic-gate size_t xlen = cur_len & PAGEOFFSET; 60637c478bd9Sstevel@tonic-gate pagezero(cpp->p_prev, xlen, PAGESIZE - xlen); 60647c478bd9Sstevel@tonic-gate } 60657c478bd9Sstevel@tonic-gate 60667c478bd9Sstevel@tonic-gate bp = pageio_setup(cpp, cur_len, ip->i_devvp, flags); 60677c478bd9Sstevel@tonic-gate ASSERT(bp != NULL); 60687c478bd9Sstevel@tonic-gate 60697c478bd9Sstevel@tonic-gate bp->b_edev = ip->i_dev; 60707c478bd9Sstevel@tonic-gate bp->b_dev = cmpdev(ip->i_dev); 60717c478bd9Sstevel@tonic-gate bp->b_blkno = bn; 60727c478bd9Sstevel@tonic-gate bp->b_un.b_addr = (caddr_t)0; 60737c478bd9Sstevel@tonic-gate bp->b_file = ip->i_vnode; 60747c478bd9Sstevel@tonic-gate 6075d3d50737SRafael Vanoni ufsvfsp->vfs_iotstamp = ddi_get_lbolt(); 60767c478bd9Sstevel@tonic-gate ub.ub_pageios.value.ul++; 60777c478bd9Sstevel@tonic-gate if (ufsvfsp->vfs_snapshot) 60787c478bd9Sstevel@tonic-gate fssnap_strategy(&(ufsvfsp->vfs_snapshot), bp); 60797c478bd9Sstevel@tonic-gate else 60807c478bd9Sstevel@tonic-gate (void) bdev_strategy(bp); 60817c478bd9Sstevel@tonic-gate 60827c478bd9Sstevel@tonic-gate if (flags & B_READ) 60837c478bd9Sstevel@tonic-gate ufs_pageio_reads++; 60847c478bd9Sstevel@tonic-gate else 60857c478bd9Sstevel@tonic-gate ufs_pageio_writes++; 60867c478bd9Sstevel@tonic-gate if (flags & B_READ) 60877c478bd9Sstevel@tonic-gate lwp_stat_update(LWP_STAT_INBLK, 1); 60887c478bd9Sstevel@tonic-gate else 60897c478bd9Sstevel@tonic-gate lwp_stat_update(LWP_STAT_OUBLK, 1); 60907c478bd9Sstevel@tonic-gate /* 60917c478bd9Sstevel@tonic-gate * If the request is not B_ASYNC, wait for i/o to complete 60927c478bd9Sstevel@tonic-gate * and re-assemble the page list to return to the caller. 60937c478bd9Sstevel@tonic-gate * If it is B_ASYNC we leave the page list in pieces and 60947c478bd9Sstevel@tonic-gate * cleanup() will dispose of them. 60957c478bd9Sstevel@tonic-gate */ 60967c478bd9Sstevel@tonic-gate if ((flags & B_ASYNC) == 0) { 60977c478bd9Sstevel@tonic-gate err = biowait(bp); 60987c478bd9Sstevel@tonic-gate pageio_done(bp); 60997c478bd9Sstevel@tonic-gate if (err) 61007c478bd9Sstevel@tonic-gate break; 61017c478bd9Sstevel@tonic-gate page_list_concat(&opp, &cpp); 61027c478bd9Sstevel@tonic-gate } 61037c478bd9Sstevel@tonic-gate cpp = npp; 61047c478bd9Sstevel@tonic-gate npp = NULL; 61057c478bd9Sstevel@tonic-gate if (flags & B_READ) 61067c478bd9Sstevel@tonic-gate cur_len = P2ROUNDUP_TYPED(cur_len, PAGESIZE, size_t); 61077c478bd9Sstevel@tonic-gate done_len += cur_len; 61087c478bd9Sstevel@tonic-gate } 61097c478bd9Sstevel@tonic-gate ASSERT(err || (cpp == NULL && npp == NULL && done_len == io_len)); 61107c478bd9Sstevel@tonic-gate if (err) { 61117c478bd9Sstevel@tonic-gate if (flags & B_ASYNC) { 61127c478bd9Sstevel@tonic-gate /* Cleanup unprocessed parts of list */ 61137c478bd9Sstevel@tonic-gate page_list_concat(&cpp, &npp); 61147c478bd9Sstevel@tonic-gate if (flags & B_READ) 61157c478bd9Sstevel@tonic-gate pvn_read_done(cpp, B_ERROR); 61167c478bd9Sstevel@tonic-gate else 61177c478bd9Sstevel@tonic-gate pvn_write_done(cpp, B_ERROR); 61187c478bd9Sstevel@tonic-gate } else { 61197c478bd9Sstevel@tonic-gate /* Re-assemble list and let caller clean up */ 61207c478bd9Sstevel@tonic-gate page_list_concat(&opp, &cpp); 61217c478bd9Sstevel@tonic-gate page_list_concat(&opp, &npp); 61227c478bd9Sstevel@tonic-gate } 61237c478bd9Sstevel@tonic-gate } 6124bc69f433Saguzovsk 6125bc69f433Saguzovsk if (vmpss && !(ip->i_flag & IACC) && !ULOCKFS_IS_NOIACC(ulp) && 6126bc69f433Saguzovsk ufsvfsp->vfs_fs->fs_ronly == 0 && !ufsvfsp->vfs_noatime) { 6127bc69f433Saguzovsk mutex_enter(&ip->i_tlock); 6128bc69f433Saguzovsk ip->i_flag |= IACC; 6129bc69f433Saguzovsk ITIMES_NOLOCK(ip); 6130bc69f433Saguzovsk mutex_exit(&ip->i_tlock); 6131bc69f433Saguzovsk } 6132bc69f433Saguzovsk 61337c478bd9Sstevel@tonic-gate if (dolock) 61347c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 61351a5e258fSJosef 'Jeff' Sipek if (vmpss && !atomic_dec_ulong_nv(&ulp->ul_vnops_cnt)) 6136bc69f433Saguzovsk cv_broadcast(&ulp->ul_cv); 61377c478bd9Sstevel@tonic-gate return (err); 61387c478bd9Sstevel@tonic-gate } 61397c478bd9Sstevel@tonic-gate 61407c478bd9Sstevel@tonic-gate /* 61417c478bd9Sstevel@tonic-gate * Called when the kernel is in a frozen state to dump data 61427c478bd9Sstevel@tonic-gate * directly to the device. It uses a private dump data structure, 61437c478bd9Sstevel@tonic-gate * set up by dump_ctl, to locate the correct disk block to which to dump. 61447c478bd9Sstevel@tonic-gate */ 6145da6c28aaSamw /*ARGSUSED*/ 61467c478bd9Sstevel@tonic-gate static int 6147d7334e51Srm15945 ufs_dump(vnode_t *vp, caddr_t addr, offset_t ldbn, offset_t dblks, 6148d7334e51Srm15945 caller_context_t *ct) 61497c478bd9Sstevel@tonic-gate { 61507c478bd9Sstevel@tonic-gate u_offset_t file_size; 61517c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 61527c478bd9Sstevel@tonic-gate struct fs *fs = ip->i_fs; 61537c478bd9Sstevel@tonic-gate daddr_t dbn, lfsbn; 61547c478bd9Sstevel@tonic-gate int disk_blks = fs->fs_bsize >> DEV_BSHIFT; 61557c478bd9Sstevel@tonic-gate int error = 0; 61567c478bd9Sstevel@tonic-gate int ndbs, nfsbs; 61577c478bd9Sstevel@tonic-gate 61587c478bd9Sstevel@tonic-gate /* 61597c478bd9Sstevel@tonic-gate * forced unmount case 61607c478bd9Sstevel@tonic-gate */ 61617c478bd9Sstevel@tonic-gate if (ip->i_ufsvfs == NULL) 61627c478bd9Sstevel@tonic-gate return (EIO); 61637c478bd9Sstevel@tonic-gate /* 61647c478bd9Sstevel@tonic-gate * Validate the inode that it has not been modified since 61657c478bd9Sstevel@tonic-gate * the dump structure is allocated. 61667c478bd9Sstevel@tonic-gate */ 61677c478bd9Sstevel@tonic-gate mutex_enter(&ip->i_tlock); 61687c478bd9Sstevel@tonic-gate if ((dump_info == NULL) || 61697c478bd9Sstevel@tonic-gate (dump_info->ip != ip) || 61707c478bd9Sstevel@tonic-gate (dump_info->time.tv_sec != ip->i_mtime.tv_sec) || 61717c478bd9Sstevel@tonic-gate (dump_info->time.tv_usec != ip->i_mtime.tv_usec)) { 61727c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 61737c478bd9Sstevel@tonic-gate return (-1); 61747c478bd9Sstevel@tonic-gate } 61757c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 61767c478bd9Sstevel@tonic-gate 61777c478bd9Sstevel@tonic-gate /* 61787c478bd9Sstevel@tonic-gate * See that the file has room for this write 61797c478bd9Sstevel@tonic-gate */ 61807c478bd9Sstevel@tonic-gate UFS_GET_ISIZE(&file_size, ip); 61817c478bd9Sstevel@tonic-gate 6182d7334e51Srm15945 if (ldbtob(ldbn + dblks) > file_size) 61837c478bd9Sstevel@tonic-gate return (ENOSPC); 61847c478bd9Sstevel@tonic-gate 61857c478bd9Sstevel@tonic-gate /* 61867c478bd9Sstevel@tonic-gate * Find the physical disk block numbers from the dump 61877c478bd9Sstevel@tonic-gate * private data structure directly and write out the data 61887c478bd9Sstevel@tonic-gate * in contiguous block lumps 61897c478bd9Sstevel@tonic-gate */ 61907c478bd9Sstevel@tonic-gate while (dblks > 0 && !error) { 6191d7334e51Srm15945 lfsbn = (daddr_t)lblkno(fs, ldbtob(ldbn)); 61927c478bd9Sstevel@tonic-gate dbn = fsbtodb(fs, dump_info->dblk[lfsbn]) + ldbn % disk_blks; 61937c478bd9Sstevel@tonic-gate nfsbs = 1; 61947c478bd9Sstevel@tonic-gate ndbs = disk_blks - ldbn % disk_blks; 61957c478bd9Sstevel@tonic-gate while (ndbs < dblks && fsbtodb(fs, dump_info->dblk[lfsbn + 61967c478bd9Sstevel@tonic-gate nfsbs]) == dbn + ndbs) { 61977c478bd9Sstevel@tonic-gate nfsbs++; 61987c478bd9Sstevel@tonic-gate ndbs += disk_blks; 61997c478bd9Sstevel@tonic-gate } 62007c478bd9Sstevel@tonic-gate if (ndbs > dblks) 62017c478bd9Sstevel@tonic-gate ndbs = dblks; 62027c478bd9Sstevel@tonic-gate error = bdev_dump(ip->i_dev, addr, dbn, ndbs); 62037c478bd9Sstevel@tonic-gate addr += ldbtob((offset_t)ndbs); 62047c478bd9Sstevel@tonic-gate dblks -= ndbs; 62057c478bd9Sstevel@tonic-gate ldbn += ndbs; 62067c478bd9Sstevel@tonic-gate } 62077c478bd9Sstevel@tonic-gate return (error); 62087c478bd9Sstevel@tonic-gate 62097c478bd9Sstevel@tonic-gate } 62107c478bd9Sstevel@tonic-gate 62117c478bd9Sstevel@tonic-gate /* 62127c478bd9Sstevel@tonic-gate * Prepare the file system before and after the dump operation. 62137c478bd9Sstevel@tonic-gate * 62147c478bd9Sstevel@tonic-gate * action = DUMP_ALLOC: 62157c478bd9Sstevel@tonic-gate * Preparation before dump, allocate dump private data structure 62167c478bd9Sstevel@tonic-gate * to hold all the direct and indirect block info for dump. 62177c478bd9Sstevel@tonic-gate * 62187c478bd9Sstevel@tonic-gate * action = DUMP_FREE: 62197c478bd9Sstevel@tonic-gate * Clean up after dump, deallocate the dump private data structure. 62207c478bd9Sstevel@tonic-gate * 62217c478bd9Sstevel@tonic-gate * action = DUMP_SCAN: 62227c478bd9Sstevel@tonic-gate * Scan dump_info for *blkp DEV_BSIZE blocks of contig fs space; 62237c478bd9Sstevel@tonic-gate * if found, the starting file-relative DEV_BSIZE lbn is written 62247c478bd9Sstevel@tonic-gate * to *bklp; that lbn is intended for use with VOP_DUMP() 62257c478bd9Sstevel@tonic-gate */ 6226da6c28aaSamw /*ARGSUSED*/ 62277c478bd9Sstevel@tonic-gate static int 6228d7334e51Srm15945 ufs_dumpctl(vnode_t *vp, int action, offset_t *blkp, caller_context_t *ct) 62297c478bd9Sstevel@tonic-gate { 62307c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 62317c478bd9Sstevel@tonic-gate ufsvfs_t *ufsvfsp = ip->i_ufsvfs; 62327c478bd9Sstevel@tonic-gate struct fs *fs; 62337c478bd9Sstevel@tonic-gate daddr32_t *dblk, *storeblk; 62347c478bd9Sstevel@tonic-gate daddr32_t *nextblk, *endblk; 62357c478bd9Sstevel@tonic-gate struct buf *bp; 62367c478bd9Sstevel@tonic-gate int i, entry, entries; 62377c478bd9Sstevel@tonic-gate int n, ncontig; 62387c478bd9Sstevel@tonic-gate 62397c478bd9Sstevel@tonic-gate /* 62407c478bd9Sstevel@tonic-gate * check for forced unmount 62417c478bd9Sstevel@tonic-gate */ 62427c478bd9Sstevel@tonic-gate if (ufsvfsp == NULL) 62437c478bd9Sstevel@tonic-gate return (EIO); 62447c478bd9Sstevel@tonic-gate 62457c478bd9Sstevel@tonic-gate if (action == DUMP_ALLOC) { 62467c478bd9Sstevel@tonic-gate /* 62477c478bd9Sstevel@tonic-gate * alloc and record dump_info 62487c478bd9Sstevel@tonic-gate */ 62497c478bd9Sstevel@tonic-gate if (dump_info != NULL) 62507c478bd9Sstevel@tonic-gate return (EINVAL); 62517c478bd9Sstevel@tonic-gate 62527c478bd9Sstevel@tonic-gate ASSERT(vp->v_type == VREG); 62537c478bd9Sstevel@tonic-gate fs = ufsvfsp->vfs_fs; 62547c478bd9Sstevel@tonic-gate 62557c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_READER); 62567c478bd9Sstevel@tonic-gate 62577c478bd9Sstevel@tonic-gate if (bmap_has_holes(ip)) { 62587c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 62597c478bd9Sstevel@tonic-gate return (EFAULT); 62607c478bd9Sstevel@tonic-gate } 62617c478bd9Sstevel@tonic-gate 62627c478bd9Sstevel@tonic-gate /* 62637c478bd9Sstevel@tonic-gate * calculate and allocate space needed according to i_size 62647c478bd9Sstevel@tonic-gate */ 62657c478bd9Sstevel@tonic-gate entries = (int)lblkno(fs, blkroundup(fs, ip->i_size)); 626680d34432Sfrankho dump_info = kmem_alloc(sizeof (struct dump) + 626780d34432Sfrankho (entries - 1) * sizeof (daddr32_t), KM_NOSLEEP); 626880d34432Sfrankho if (dump_info == NULL) { 62697c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 62707c478bd9Sstevel@tonic-gate return (ENOMEM); 62717c478bd9Sstevel@tonic-gate } 62727c478bd9Sstevel@tonic-gate 62737c478bd9Sstevel@tonic-gate /* Start saving the info */ 62747c478bd9Sstevel@tonic-gate dump_info->fsbs = entries; 62757c478bd9Sstevel@tonic-gate dump_info->ip = ip; 62767c478bd9Sstevel@tonic-gate storeblk = &dump_info->dblk[0]; 62777c478bd9Sstevel@tonic-gate 62787c478bd9Sstevel@tonic-gate /* Direct Blocks */ 62797c478bd9Sstevel@tonic-gate for (entry = 0; entry < NDADDR && entry < entries; entry++) 62807c478bd9Sstevel@tonic-gate *storeblk++ = ip->i_db[entry]; 62817c478bd9Sstevel@tonic-gate 62827c478bd9Sstevel@tonic-gate /* Indirect Blocks */ 62837c478bd9Sstevel@tonic-gate for (i = 0; i < NIADDR; i++) { 62847c478bd9Sstevel@tonic-gate int error = 0; 62857c478bd9Sstevel@tonic-gate 62867c478bd9Sstevel@tonic-gate bp = UFS_BREAD(ufsvfsp, 628780d34432Sfrankho ip->i_dev, fsbtodb(fs, ip->i_ib[i]), fs->fs_bsize); 62887c478bd9Sstevel@tonic-gate if (bp->b_flags & B_ERROR) 62897c478bd9Sstevel@tonic-gate error = EIO; 62907c478bd9Sstevel@tonic-gate else { 62917c478bd9Sstevel@tonic-gate dblk = bp->b_un.b_daddr; 62927c478bd9Sstevel@tonic-gate if ((storeblk = save_dblks(ip, ufsvfsp, 62937c478bd9Sstevel@tonic-gate storeblk, dblk, i, entries)) == NULL) 62947c478bd9Sstevel@tonic-gate error = EIO; 62957c478bd9Sstevel@tonic-gate } 62967c478bd9Sstevel@tonic-gate 62977c478bd9Sstevel@tonic-gate brelse(bp); 62987c478bd9Sstevel@tonic-gate 62997c478bd9Sstevel@tonic-gate if (error != 0) { 63007c478bd9Sstevel@tonic-gate kmem_free(dump_info, sizeof (struct dump) + 63017c478bd9Sstevel@tonic-gate (entries - 1) * sizeof (daddr32_t)); 63027c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 63037c478bd9Sstevel@tonic-gate dump_info = NULL; 63047c478bd9Sstevel@tonic-gate return (error); 63057c478bd9Sstevel@tonic-gate } 63067c478bd9Sstevel@tonic-gate } 63077c478bd9Sstevel@tonic-gate /* and time stamp the information */ 63087c478bd9Sstevel@tonic-gate mutex_enter(&ip->i_tlock); 63097c478bd9Sstevel@tonic-gate dump_info->time = ip->i_mtime; 63107c478bd9Sstevel@tonic-gate mutex_exit(&ip->i_tlock); 63117c478bd9Sstevel@tonic-gate 63127c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 63137c478bd9Sstevel@tonic-gate } else if (action == DUMP_FREE) { 63147c478bd9Sstevel@tonic-gate /* 63157c478bd9Sstevel@tonic-gate * free dump_info 63167c478bd9Sstevel@tonic-gate */ 63177c478bd9Sstevel@tonic-gate if (dump_info == NULL) 63187c478bd9Sstevel@tonic-gate return (EINVAL); 63197c478bd9Sstevel@tonic-gate entries = dump_info->fsbs - 1; 63207c478bd9Sstevel@tonic-gate kmem_free(dump_info, sizeof (struct dump) + 63217c478bd9Sstevel@tonic-gate entries * sizeof (daddr32_t)); 63227c478bd9Sstevel@tonic-gate dump_info = NULL; 63237c478bd9Sstevel@tonic-gate } else if (action == DUMP_SCAN) { 63247c478bd9Sstevel@tonic-gate /* 63257c478bd9Sstevel@tonic-gate * scan dump_info 63267c478bd9Sstevel@tonic-gate */ 63277c478bd9Sstevel@tonic-gate if (dump_info == NULL) 63287c478bd9Sstevel@tonic-gate return (EINVAL); 63297c478bd9Sstevel@tonic-gate 63307c478bd9Sstevel@tonic-gate dblk = dump_info->dblk; 63317c478bd9Sstevel@tonic-gate nextblk = dblk + 1; 63327c478bd9Sstevel@tonic-gate endblk = dblk + dump_info->fsbs - 1; 63337c478bd9Sstevel@tonic-gate fs = ufsvfsp->vfs_fs; 63347c478bd9Sstevel@tonic-gate ncontig = *blkp >> (fs->fs_bshift - DEV_BSHIFT); 63357c478bd9Sstevel@tonic-gate 63367c478bd9Sstevel@tonic-gate /* 63377c478bd9Sstevel@tonic-gate * scan dblk[] entries; contig fs space is found when: 63387c478bd9Sstevel@tonic-gate * ((current blkno + frags per block) == next blkno) 63397c478bd9Sstevel@tonic-gate */ 63407c478bd9Sstevel@tonic-gate n = 0; 63417c478bd9Sstevel@tonic-gate while (n < ncontig && dblk < endblk) { 63427c478bd9Sstevel@tonic-gate if ((*dblk + fs->fs_frag) == *nextblk) 63437c478bd9Sstevel@tonic-gate n++; 63447c478bd9Sstevel@tonic-gate else 63457c478bd9Sstevel@tonic-gate n = 0; 63467c478bd9Sstevel@tonic-gate dblk++; 63477c478bd9Sstevel@tonic-gate nextblk++; 63487c478bd9Sstevel@tonic-gate } 63497c478bd9Sstevel@tonic-gate 63507c478bd9Sstevel@tonic-gate /* 63517c478bd9Sstevel@tonic-gate * index is where size bytes of contig space begins; 63527c478bd9Sstevel@tonic-gate * conversion from index to the file's DEV_BSIZE lbn 63537c478bd9Sstevel@tonic-gate * is equivalent to: (index * fs_bsize) / DEV_BSIZE 63547c478bd9Sstevel@tonic-gate */ 63557c478bd9Sstevel@tonic-gate if (n == ncontig) { 63567c478bd9Sstevel@tonic-gate i = (dblk - dump_info->dblk) - ncontig; 63577c478bd9Sstevel@tonic-gate *blkp = i << (fs->fs_bshift - DEV_BSHIFT); 63587c478bd9Sstevel@tonic-gate } else 63597c478bd9Sstevel@tonic-gate return (EFAULT); 63607c478bd9Sstevel@tonic-gate } 63617c478bd9Sstevel@tonic-gate return (0); 63627c478bd9Sstevel@tonic-gate } 63637c478bd9Sstevel@tonic-gate 63647c478bd9Sstevel@tonic-gate /* 63657c478bd9Sstevel@tonic-gate * Recursive helper function for ufs_dumpctl(). It follows the indirect file 63667c478bd9Sstevel@tonic-gate * system blocks until it reaches the the disk block addresses, which are 63677c478bd9Sstevel@tonic-gate * then stored into the given buffer, storeblk. 63687c478bd9Sstevel@tonic-gate */ 63697c478bd9Sstevel@tonic-gate static daddr32_t * 63707c478bd9Sstevel@tonic-gate save_dblks(struct inode *ip, struct ufsvfs *ufsvfsp, daddr32_t *storeblk, 63717c478bd9Sstevel@tonic-gate daddr32_t *dblk, int level, int entries) 63727c478bd9Sstevel@tonic-gate { 63737c478bd9Sstevel@tonic-gate struct fs *fs = ufsvfsp->vfs_fs; 63747c478bd9Sstevel@tonic-gate struct buf *bp; 63757c478bd9Sstevel@tonic-gate int i; 63767c478bd9Sstevel@tonic-gate 63777c478bd9Sstevel@tonic-gate if (level == 0) { 63787c478bd9Sstevel@tonic-gate for (i = 0; i < NINDIR(fs); i++) { 63797c478bd9Sstevel@tonic-gate if (storeblk - dump_info->dblk >= entries) 63807c478bd9Sstevel@tonic-gate break; 63817c478bd9Sstevel@tonic-gate *storeblk++ = dblk[i]; 63827c478bd9Sstevel@tonic-gate } 63837c478bd9Sstevel@tonic-gate return (storeblk); 63847c478bd9Sstevel@tonic-gate } 63857c478bd9Sstevel@tonic-gate for (i = 0; i < NINDIR(fs); i++) { 63867c478bd9Sstevel@tonic-gate if (storeblk - dump_info->dblk >= entries) 63877c478bd9Sstevel@tonic-gate break; 63887c478bd9Sstevel@tonic-gate bp = UFS_BREAD(ufsvfsp, 63897c478bd9Sstevel@tonic-gate ip->i_dev, fsbtodb(fs, dblk[i]), fs->fs_bsize); 63907c478bd9Sstevel@tonic-gate if (bp->b_flags & B_ERROR) { 63917c478bd9Sstevel@tonic-gate brelse(bp); 63927c478bd9Sstevel@tonic-gate return (NULL); 63937c478bd9Sstevel@tonic-gate } 63947c478bd9Sstevel@tonic-gate storeblk = save_dblks(ip, ufsvfsp, storeblk, bp->b_un.b_daddr, 63957c478bd9Sstevel@tonic-gate level - 1, entries); 63967c478bd9Sstevel@tonic-gate brelse(bp); 63977c478bd9Sstevel@tonic-gate 63987c478bd9Sstevel@tonic-gate if (storeblk == NULL) 63997c478bd9Sstevel@tonic-gate return (NULL); 64007c478bd9Sstevel@tonic-gate } 64017c478bd9Sstevel@tonic-gate return (storeblk); 64027c478bd9Sstevel@tonic-gate } 64037c478bd9Sstevel@tonic-gate 64047c478bd9Sstevel@tonic-gate /* ARGSUSED */ 64057c478bd9Sstevel@tonic-gate static int 64067c478bd9Sstevel@tonic-gate ufs_getsecattr(struct vnode *vp, vsecattr_t *vsap, int flag, 6407da6c28aaSamw struct cred *cr, caller_context_t *ct) 64087c478bd9Sstevel@tonic-gate { 64097c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 64107c478bd9Sstevel@tonic-gate struct ulockfs *ulp; 64117c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp = ip->i_ufsvfs; 64127c478bd9Sstevel@tonic-gate ulong_t vsa_mask = vsap->vsa_mask; 64137c478bd9Sstevel@tonic-gate int err = EINVAL; 64147c478bd9Sstevel@tonic-gate 64157c478bd9Sstevel@tonic-gate vsa_mask &= (VSA_ACL | VSA_ACLCNT | VSA_DFACL | VSA_DFACLCNT); 64167c478bd9Sstevel@tonic-gate 64177c478bd9Sstevel@tonic-gate /* 64187c478bd9Sstevel@tonic-gate * Only grab locks if needed - they're not needed to check vsa_mask 64197c478bd9Sstevel@tonic-gate * or if the mask contains no acl flags. 64207c478bd9Sstevel@tonic-gate */ 64217c478bd9Sstevel@tonic-gate if (vsa_mask != 0) { 64227c478bd9Sstevel@tonic-gate if (err = ufs_lockfs_begin(ufsvfsp, &ulp, 64237c478bd9Sstevel@tonic-gate ULOCKFS_GETATTR_MASK)) 64247c478bd9Sstevel@tonic-gate return (err); 64257c478bd9Sstevel@tonic-gate 64267c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_READER); 64277c478bd9Sstevel@tonic-gate err = ufs_acl_get(ip, vsap, flag, cr); 64287c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 64297c478bd9Sstevel@tonic-gate 64307c478bd9Sstevel@tonic-gate if (ulp) 64317c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 64327c478bd9Sstevel@tonic-gate } 64337c478bd9Sstevel@tonic-gate return (err); 64347c478bd9Sstevel@tonic-gate } 64357c478bd9Sstevel@tonic-gate 64367c478bd9Sstevel@tonic-gate /* ARGSUSED */ 64377c478bd9Sstevel@tonic-gate static int 6438da6c28aaSamw ufs_setsecattr(struct vnode *vp, vsecattr_t *vsap, int flag, struct cred *cr, 6439da6c28aaSamw caller_context_t *ct) 64407c478bd9Sstevel@tonic-gate { 64417c478bd9Sstevel@tonic-gate struct inode *ip = VTOI(vp); 64427c478bd9Sstevel@tonic-gate struct ulockfs *ulp = NULL; 64437c478bd9Sstevel@tonic-gate struct ufsvfs *ufsvfsp = VTOI(vp)->i_ufsvfs; 64447c478bd9Sstevel@tonic-gate ulong_t vsa_mask = vsap->vsa_mask; 64457c478bd9Sstevel@tonic-gate int err; 64467c478bd9Sstevel@tonic-gate int haverwlock = 1; 64477c478bd9Sstevel@tonic-gate int trans_size; 64487c478bd9Sstevel@tonic-gate int donetrans = 0; 64497c478bd9Sstevel@tonic-gate int retry = 1; 64507c478bd9Sstevel@tonic-gate 64517c478bd9Sstevel@tonic-gate ASSERT(RW_LOCK_HELD(&ip->i_rwlock)); 64527c478bd9Sstevel@tonic-gate 64537c478bd9Sstevel@tonic-gate /* Abort now if the request is either empty or invalid. */ 64547c478bd9Sstevel@tonic-gate vsa_mask &= (VSA_ACL | VSA_ACLCNT | VSA_DFACL | VSA_DFACLCNT); 64557c478bd9Sstevel@tonic-gate if ((vsa_mask == 0) || 64567c478bd9Sstevel@tonic-gate ((vsap->vsa_aclentp == NULL) && 64577c478bd9Sstevel@tonic-gate (vsap->vsa_dfaclentp == NULL))) { 64587c478bd9Sstevel@tonic-gate err = EINVAL; 64597c478bd9Sstevel@tonic-gate goto out; 64607c478bd9Sstevel@tonic-gate } 64617c478bd9Sstevel@tonic-gate 64627c478bd9Sstevel@tonic-gate /* 64637c478bd9Sstevel@tonic-gate * Following convention, if this is a directory then we acquire the 64647c478bd9Sstevel@tonic-gate * inode's i_rwlock after starting a UFS logging transaction; 64657c478bd9Sstevel@tonic-gate * otherwise, we acquire it beforehand. Since we were called (and 64667c478bd9Sstevel@tonic-gate * must therefore return) with the lock held, we will have to drop it, 64677c478bd9Sstevel@tonic-gate * and later reacquire it, if operating on a directory. 64687c478bd9Sstevel@tonic-gate */ 64697c478bd9Sstevel@tonic-gate if (vp->v_type == VDIR) { 64707c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 64717c478bd9Sstevel@tonic-gate haverwlock = 0; 64727c478bd9Sstevel@tonic-gate } else { 64737c478bd9Sstevel@tonic-gate /* Upgrade the lock if required. */ 64747c478bd9Sstevel@tonic-gate if (!rw_write_held(&ip->i_rwlock)) { 64757c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 64767c478bd9Sstevel@tonic-gate rw_enter(&ip->i_rwlock, RW_WRITER); 64777c478bd9Sstevel@tonic-gate } 64787c478bd9Sstevel@tonic-gate } 64797c478bd9Sstevel@tonic-gate 64807c478bd9Sstevel@tonic-gate again: 64817c478bd9Sstevel@tonic-gate ASSERT(!(vp->v_type == VDIR && haverwlock)); 64827c478bd9Sstevel@tonic-gate if (err = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_SETATTR_MASK)) { 64837c478bd9Sstevel@tonic-gate ulp = NULL; 64847c478bd9Sstevel@tonic-gate retry = 0; 64857c478bd9Sstevel@tonic-gate goto out; 64867c478bd9Sstevel@tonic-gate } 64877c478bd9Sstevel@tonic-gate 64887c478bd9Sstevel@tonic-gate /* 64897c478bd9Sstevel@tonic-gate * Check that the file system supports this operation. Note that 64907c478bd9Sstevel@tonic-gate * ufs_lockfs_begin() will have checked that the file system had 64917c478bd9Sstevel@tonic-gate * not been forcibly unmounted. 64927c478bd9Sstevel@tonic-gate */ 64937c478bd9Sstevel@tonic-gate if (ufsvfsp->vfs_fs->fs_ronly) { 64947c478bd9Sstevel@tonic-gate err = EROFS; 64957c478bd9Sstevel@tonic-gate goto out; 64967c478bd9Sstevel@tonic-gate } 64977c478bd9Sstevel@tonic-gate if (ufsvfsp->vfs_nosetsec) { 64987c478bd9Sstevel@tonic-gate err = ENOSYS; 64997c478bd9Sstevel@tonic-gate goto out; 65007c478bd9Sstevel@tonic-gate } 65017c478bd9Sstevel@tonic-gate 65027c478bd9Sstevel@tonic-gate if (ulp) { 65037c478bd9Sstevel@tonic-gate TRANS_BEGIN_ASYNC(ufsvfsp, TOP_SETSECATTR, 65047c478bd9Sstevel@tonic-gate trans_size = TOP_SETSECATTR_SIZE(VTOI(vp))); 65057c478bd9Sstevel@tonic-gate donetrans = 1; 65067c478bd9Sstevel@tonic-gate } 65077c478bd9Sstevel@tonic-gate 65087c478bd9Sstevel@tonic-gate if (vp->v_type == VDIR) { 65097c478bd9Sstevel@tonic-gate rw_enter(&ip->i_rwlock, RW_WRITER); 65107c478bd9Sstevel@tonic-gate haverwlock = 1; 65117c478bd9Sstevel@tonic-gate } 65127c478bd9Sstevel@tonic-gate 65137c478bd9Sstevel@tonic-gate ASSERT(haverwlock); 65147c478bd9Sstevel@tonic-gate 65157c478bd9Sstevel@tonic-gate /* Do the actual work. */ 65167c478bd9Sstevel@tonic-gate rw_enter(&ip->i_contents, RW_WRITER); 65177c478bd9Sstevel@tonic-gate /* 65187c478bd9Sstevel@tonic-gate * Suppress out of inodes messages if we will retry. 65197c478bd9Sstevel@tonic-gate */ 65207c478bd9Sstevel@tonic-gate if (retry) 65217c478bd9Sstevel@tonic-gate ip->i_flag |= IQUIET; 65227c478bd9Sstevel@tonic-gate err = ufs_acl_set(ip, vsap, flag, cr); 65237c478bd9Sstevel@tonic-gate ip->i_flag &= ~IQUIET; 65247c478bd9Sstevel@tonic-gate rw_exit(&ip->i_contents); 65257c478bd9Sstevel@tonic-gate 65267c478bd9Sstevel@tonic-gate out: 65277c478bd9Sstevel@tonic-gate if (ulp) { 65287c478bd9Sstevel@tonic-gate if (donetrans) { 65297c478bd9Sstevel@tonic-gate /* 65307c478bd9Sstevel@tonic-gate * top_end_async() can eventually call 65317c478bd9Sstevel@tonic-gate * top_end_sync(), which can block. We must 65327c478bd9Sstevel@tonic-gate * therefore observe the lock-ordering protocol 65337c478bd9Sstevel@tonic-gate * here as well. 65347c478bd9Sstevel@tonic-gate */ 65357c478bd9Sstevel@tonic-gate if (vp->v_type == VDIR) { 65367c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 65377c478bd9Sstevel@tonic-gate haverwlock = 0; 65387c478bd9Sstevel@tonic-gate } 65397c478bd9Sstevel@tonic-gate TRANS_END_ASYNC(ufsvfsp, TOP_SETSECATTR, trans_size); 65407c478bd9Sstevel@tonic-gate } 65417c478bd9Sstevel@tonic-gate ufs_lockfs_end(ulp); 65427c478bd9Sstevel@tonic-gate } 65437c478bd9Sstevel@tonic-gate /* 65447c478bd9Sstevel@tonic-gate * If no inodes available, try scaring a logically- 65457c478bd9Sstevel@tonic-gate * free one out of the delete queue to someplace 65467c478bd9Sstevel@tonic-gate * that we can find it. 65477c478bd9Sstevel@tonic-gate */ 65487c478bd9Sstevel@tonic-gate if ((err == ENOSPC) && retry && TRANS_ISTRANS(ufsvfsp)) { 65497c478bd9Sstevel@tonic-gate ufs_delete_drain_wait(ufsvfsp, 1); 65507c478bd9Sstevel@tonic-gate retry = 0; 65517c478bd9Sstevel@tonic-gate if (vp->v_type == VDIR && haverwlock) { 65527c478bd9Sstevel@tonic-gate rw_exit(&ip->i_rwlock); 65537c478bd9Sstevel@tonic-gate haverwlock = 0; 65547c478bd9Sstevel@tonic-gate } 65557c478bd9Sstevel@tonic-gate goto again; 65567c478bd9Sstevel@tonic-gate } 65577c478bd9Sstevel@tonic-gate /* 65587c478bd9Sstevel@tonic-gate * If we need to reacquire the lock then it is safe to do so 65597c478bd9Sstevel@tonic-gate * as a reader. This is because ufs_rwunlock(), which will be 65607c478bd9Sstevel@tonic-gate * called by our caller after we return, does not differentiate 65617c478bd9Sstevel@tonic-gate * between shared and exclusive locks. 65627c478bd9Sstevel@tonic-gate */ 65637c478bd9Sstevel@tonic-gate if (!haverwlock) { 65647c478bd9Sstevel@tonic-gate ASSERT(vp->v_type == VDIR); 65657c478bd9Sstevel@tonic-gate rw_enter(&ip->i_rwlock, RW_READER); 65667c478bd9Sstevel@tonic-gate } 65677c478bd9Sstevel@tonic-gate 65687c478bd9Sstevel@tonic-gate return (err); 65697c478bd9Sstevel@tonic-gate } 65709b5097eeSOwen Roberts 65719b5097eeSOwen Roberts /* 65729b5097eeSOwen Roberts * Locate the vnode to be used for an event notification. As this will 65739b5097eeSOwen Roberts * be called prior to the name space change perform basic verification 65749b5097eeSOwen Roberts * that the change will be allowed. 65759b5097eeSOwen Roberts */ 65769b5097eeSOwen Roberts 65779b5097eeSOwen Roberts static int 65789b5097eeSOwen Roberts ufs_eventlookup(struct vnode *dvp, char *nm, struct cred *cr, 65799b5097eeSOwen Roberts struct vnode **vpp) 65809b5097eeSOwen Roberts { 65819b5097eeSOwen Roberts int namlen; 65829b5097eeSOwen Roberts int error; 65839b5097eeSOwen Roberts struct vnode *vp; 65849b5097eeSOwen Roberts struct inode *ip; 65859b5097eeSOwen Roberts struct inode *xip; 65869b5097eeSOwen Roberts struct ufsvfs *ufsvfsp; 65879b5097eeSOwen Roberts struct ulockfs *ulp; 65889b5097eeSOwen Roberts 65899b5097eeSOwen Roberts ip = VTOI(dvp); 65909b5097eeSOwen Roberts *vpp = NULL; 65919b5097eeSOwen Roberts 65929b5097eeSOwen Roberts if ((namlen = strlen(nm)) == 0) 65939b5097eeSOwen Roberts return (EINVAL); 65949b5097eeSOwen Roberts 65959b5097eeSOwen Roberts if (nm[0] == '.') { 65969b5097eeSOwen Roberts if (namlen == 1) 65979b5097eeSOwen Roberts return (EINVAL); 65989b5097eeSOwen Roberts else if ((namlen == 2) && nm[1] == '.') { 65999b5097eeSOwen Roberts return (EEXIST); 66009b5097eeSOwen Roberts } 66019b5097eeSOwen Roberts } 66029b5097eeSOwen Roberts 66039b5097eeSOwen Roberts /* 66049b5097eeSOwen Roberts * Check accessibility and write access of parent directory as we 66059b5097eeSOwen Roberts * only want to post the event if we're able to make a change. 66069b5097eeSOwen Roberts */ 66079b5097eeSOwen Roberts if (error = ufs_diraccess(ip, IEXEC|IWRITE, cr)) 66089b5097eeSOwen Roberts return (error); 66099b5097eeSOwen Roberts 66109b5097eeSOwen Roberts if (vp = dnlc_lookup(dvp, nm)) { 66119b5097eeSOwen Roberts if (vp == DNLC_NO_VNODE) { 66129b5097eeSOwen Roberts VN_RELE(vp); 66139b5097eeSOwen Roberts return (ENOENT); 66149b5097eeSOwen Roberts } 66159b5097eeSOwen Roberts 66169b5097eeSOwen Roberts *vpp = vp; 66179b5097eeSOwen Roberts return (0); 66189b5097eeSOwen Roberts } 66199b5097eeSOwen Roberts 66209b5097eeSOwen Roberts /* 66219b5097eeSOwen Roberts * Keep the idle queue from getting too long by idling two 66229b5097eeSOwen Roberts * inodes before attempting to allocate another. 66239b5097eeSOwen Roberts * This operation must be performed before entering lockfs 66249b5097eeSOwen Roberts * or a transaction. 66259b5097eeSOwen Roberts */ 66269b5097eeSOwen Roberts if (ufs_idle_q.uq_ne > ufs_idle_q.uq_hiwat) 66279b5097eeSOwen Roberts if ((curthread->t_flag & T_DONTBLOCK) == 0) { 66289b5097eeSOwen Roberts ins.in_lidles.value.ul += ufs_lookup_idle_count; 66299b5097eeSOwen Roberts ufs_idle_some(ufs_lookup_idle_count); 66309b5097eeSOwen Roberts } 66319b5097eeSOwen Roberts 66329b5097eeSOwen Roberts ufsvfsp = ip->i_ufsvfs; 66339b5097eeSOwen Roberts 66349b5097eeSOwen Roberts retry_lookup: 66359b5097eeSOwen Roberts if (error = ufs_lockfs_begin(ufsvfsp, &ulp, ULOCKFS_LOOKUP_MASK)) 66369b5097eeSOwen Roberts return (error); 66379b5097eeSOwen Roberts 66389b5097eeSOwen Roberts if ((error = ufs_dirlook(ip, nm, &xip, cr, 1, 1)) == 0) { 66399b5097eeSOwen Roberts vp = ITOV(xip); 66409b5097eeSOwen Roberts *vpp = vp; 66419b5097eeSOwen Roberts } 66429b5097eeSOwen Roberts 66439b5097eeSOwen Roberts if (ulp) { 66449b5097eeSOwen Roberts ufs_lockfs_end(ulp); 66459b5097eeSOwen Roberts } 66469b5097eeSOwen Roberts 66479b5097eeSOwen Roberts if (error == EAGAIN) 66489b5097eeSOwen Roberts goto retry_lookup; 66499b5097eeSOwen Roberts 66509b5097eeSOwen Roberts return (error); 66519b5097eeSOwen Roberts } 6652