xref: /titanic_51/usr/src/uts/common/fs/ufs/ufs_inode.c (revision a799b1e741b6f59012a469e6b57c40cb8061127b)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2008 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  */
25 
26 /*	Copyright (c) 1983, 1984, 1985, 1986, 1987, 1988, 1989 AT&T	*/
27 /*	  All Rights Reserved  	*/
28 
29 /*
30  * University Copyright- Copyright (c) 1982, 1986, 1988
31  * The Regents of the University of California
32  * All Rights Reserved
33  *
34  * University Acknowledgment- Portions of this document are derived from
35  * software developed by the University of California, Berkeley, and its
36  * contributors.
37  */
38 
39 #include <sys/types.h>
40 #include <sys/t_lock.h>
41 #include <sys/param.h>
42 #include <sys/systm.h>
43 #include <sys/uio.h>
44 #include <sys/bitmap.h>
45 #include <sys/signal.h>
46 #include <sys/cred.h>
47 #include <sys/user.h>
48 #include <sys/vfs.h>
49 #include <sys/stat.h>
50 #include <sys/vnode.h>
51 #include <sys/buf.h>
52 #include <sys/proc.h>
53 #include <sys/disp.h>
54 #include <sys/dnlc.h>
55 #include <sys/mode.h>
56 #include <sys/cmn_err.h>
57 #include <sys/kstat.h>
58 #include <sys/acl.h>
59 #include <sys/var.h>
60 #include <sys/fs/ufs_inode.h>
61 #include <sys/fs/ufs_fs.h>
62 #include <sys/fs/ufs_trans.h>
63 #include <sys/fs/ufs_acl.h>
64 #include <sys/fs/ufs_bio.h>
65 #include <sys/fs/ufs_quota.h>
66 #include <sys/fs/ufs_log.h>
67 #include <vm/hat.h>
68 #include <vm/as.h>
69 #include <vm/pvn.h>
70 #include <vm/seg.h>
71 #include <sys/swap.h>
72 #include <sys/cpuvar.h>
73 #include <sys/sysmacros.h>
74 #include <sys/errno.h>
75 #include <sys/kmem.h>
76 #include <sys/debug.h>
77 #include <fs/fs_subr.h>
78 #include <sys/policy.h>
79 
80 struct kmem_cache *inode_cache;		/* cache of free inodes */
81 
82 /* UFS Inode Cache Stats -- Not protected */
83 struct	instats ins = {
84 	{ "size",		KSTAT_DATA_ULONG },
85 	{ "maxsize",		KSTAT_DATA_ULONG },
86 	{ "hits",		KSTAT_DATA_ULONG },
87 	{ "misses",		KSTAT_DATA_ULONG },
88 	{ "kmem allocs",	KSTAT_DATA_ULONG },
89 	{ "kmem frees",		KSTAT_DATA_ULONG },
90 	{ "maxsize reached",	KSTAT_DATA_ULONG },
91 	{ "puts at frontlist",	KSTAT_DATA_ULONG },
92 	{ "puts at backlist",	KSTAT_DATA_ULONG },
93 	{ "queues to free",	KSTAT_DATA_ULONG },
94 	{ "scans",		KSTAT_DATA_ULONG },
95 	{ "thread idles",	KSTAT_DATA_ULONG },
96 	{ "lookup idles",	KSTAT_DATA_ULONG },
97 	{ "vget idles",		KSTAT_DATA_ULONG },
98 	{ "cache allocs",	KSTAT_DATA_ULONG },
99 	{ "cache frees",	KSTAT_DATA_ULONG },
100 	{ "pushes at close",	KSTAT_DATA_ULONG }
101 };
102 
103 /* kstat data */
104 static kstat_t		*ufs_inode_kstat = NULL;
105 
106 union ihead *ihead;	/* inode LRU cache, Chris Maltby */
107 kmutex_t *ih_lock;	/* protect inode cache hash table */
108 static int ino_hashlen = 4;	/* desired average hash chain length */
109 int inohsz;		/* number of buckets in the hash table */
110 
111 kmutex_t	ufs_scan_lock;	/* stop racing multiple ufs_scan_inodes() */
112 kmutex_t	ufs_iuniqtime_lock; /* protect iuniqtime */
113 kmutex_t	ufsvfs_mutex;
114 struct ufsvfs	*oldufsvfslist, *ufsvfslist;
115 
116 /*
117  * time to wait after ufsvfsp->vfs_iotstamp before declaring that no
118  * I/Os are going on.
119  */
120 clock_t	ufs_iowait;
121 
122 /*
123  * the threads that process idle inodes and free (deleted) inodes
124  * have high water marks that are set in ufsinit().
125  * These values but can be no less then the minimum shown below
126  */
127 int	ufs_idle_max;	/* # of allowable idle inodes */
128 ulong_t	ufs_inode_max;	/* hard limit of allowable idle inodes */
129 #define	UFS_IDLE_MAX	(16)	/* min # of allowable idle inodes */
130 
131 /*
132  * Tunables for ufs write throttling.
133  * These are validated in ufs_iinit() since improper settings
134  * can lead to filesystem hangs.
135  */
136 #define	UFS_HW_DEFAULT	(16 * 1024 * 1024)
137 #define	UFS_LW_DEFAULT	(8 * 1024 * 1024)
138 int	ufs_HW = UFS_HW_DEFAULT;
139 int	ufs_LW = UFS_LW_DEFAULT;
140 
141 static void ihinit(void);
142 extern int hash2ints(int, int);
143 
144 static int ufs_iget_internal(struct vfs *, ino_t, struct inode **,
145     struct cred *, int);
146 
147 /* ARGSUSED */
148 static int
149 ufs_inode_kstat_update(kstat_t *ksp, int rw)
150 {
151 	if (rw == KSTAT_WRITE)
152 		return (EACCES);
153 
154 	ins.in_malloc.value.ul	= (ulong_t)kmem_cache_stat(inode_cache,
155 	    "slab_alloc");
156 	ins.in_mfree.value.ul	= (ulong_t)kmem_cache_stat(inode_cache,
157 	    "slab_free");
158 	ins.in_kcalloc.value.ul	= (ulong_t)kmem_cache_stat(inode_cache,
159 	    "alloc");
160 	ins.in_kcfree.value.ul	= (ulong_t)kmem_cache_stat(inode_cache,
161 	    "free");
162 	ins.in_size.value.ul	= (ulong_t)kmem_cache_stat(inode_cache,
163 	    "buf_inuse");
164 	ins.in_maxreached.value.ul = (ulong_t)kmem_cache_stat(inode_cache,
165 	    "buf_max");
166 	ins.in_misses.value.ul = ins.in_kcalloc.value.ul;
167 
168 	return (0);
169 }
170 
171 void
172 ufs_iinit(void)
173 {
174 	/*
175 	 * Validate that ufs_HW > ufs_LW.
176 	 * The default values for these two tunables have been increased.
177 	 * There is now a range of values for ufs_HW that used to be
178 	 * legal on previous Solaris versions but no longer is now.
179 	 * Upgrading a machine which has an /etc/system setting for ufs_HW
180 	 * from that range can lead to filesystem hangs unless the values
181 	 * are checked here.
182 	 */
183 	if (ufs_HW <= ufs_LW) {
184 		cmn_err(CE_WARN,
185 		    "ufs_HW (%d) <= ufs_LW (%d). Check /etc/system.",
186 		    ufs_HW, ufs_LW);
187 		ufs_LW = UFS_LW_DEFAULT;
188 		ufs_HW = UFS_HW_DEFAULT;
189 		cmn_err(CE_CONT, "using defaults, ufs_HW = %d, ufs_LW = %d\n",
190 		    ufs_HW, ufs_LW);
191 	}
192 
193 	/*
194 	 * Adjust the tunable `ufs_ninode' to a reasonable value
195 	 */
196 	if (ufs_ninode <= 0)
197 		ufs_ninode = ncsize;
198 	if (ufs_inode_max == 0)
199 		ufs_inode_max =
200 		    (ulong_t)((kmem_maxavail() >> 2) / sizeof (struct inode));
201 	if (ufs_ninode > ufs_inode_max || (ufs_ninode == 0 && ncsize == 0)) {
202 		cmn_err(CE_NOTE, "setting ufs_ninode to max value of %ld",
203 		    ufs_inode_max);
204 		ufs_ninode = ufs_inode_max;
205 	}
206 	/*
207 	 * Wait till third call of ufs_update to declare that no I/Os are
208 	 * going on. This allows deferred access times to be flushed to disk.
209 	 */
210 	ufs_iowait = v.v_autoup * hz * 2;
211 
212 	/*
213 	 * idle thread runs when 25% of ufs_ninode entries are on the queue
214 	 */
215 	if (ufs_idle_max == 0)
216 		ufs_idle_max = ufs_ninode >> 2;
217 	if (ufs_idle_max < UFS_IDLE_MAX)
218 		ufs_idle_max = UFS_IDLE_MAX;
219 	if (ufs_idle_max > ufs_ninode)
220 		ufs_idle_max = ufs_ninode;
221 	/*
222 	 * This is really a misnomer, it is ufs_queue_init
223 	 */
224 	ufs_thread_init(&ufs_idle_q, ufs_idle_max);
225 	ufs_thread_start(&ufs_idle_q, ufs_thread_idle, NULL);
226 
227 	/*
228 	 * global hlock thread
229 	 */
230 	ufs_thread_init(&ufs_hlock, 1);
231 	ufs_thread_start(&ufs_hlock, ufs_thread_hlock, NULL);
232 
233 	ihinit();
234 	qtinit();
235 	ins.in_maxsize.value.ul = ufs_ninode;
236 	if ((ufs_inode_kstat = kstat_create("ufs", 0, "inode_cache", "ufs",
237 	    KSTAT_TYPE_NAMED, sizeof (ins) / sizeof (kstat_named_t),
238 	    KSTAT_FLAG_VIRTUAL)) != NULL) {
239 		ufs_inode_kstat->ks_data = (void *)&ins;
240 		ufs_inode_kstat->ks_update = ufs_inode_kstat_update;
241 		kstat_install(ufs_inode_kstat);
242 	}
243 	ufsfx_init();		/* fix-on-panic initialization */
244 	si_cache_init();
245 	ufs_directio_init();
246 	lufs_init();
247 	mutex_init(&ufs_iuniqtime_lock, NULL, MUTEX_DEFAULT, NULL);
248 }
249 
250 /* ARGSUSED */
251 static int
252 ufs_inode_cache_constructor(void *buf, void *cdrarg, int kmflags)
253 {
254 	struct inode *ip = buf;
255 	struct vnode *vp;
256 
257 	vp = ip->i_vnode = vn_alloc(kmflags);
258 	if (vp == NULL) {
259 		return (-1);
260 	}
261 	vn_setops(vp, ufs_vnodeops);
262 	vp->v_data = ip;
263 
264 	rw_init(&ip->i_rwlock, NULL, RW_DEFAULT, NULL);
265 	rw_init(&ip->i_contents, NULL, RW_DEFAULT, NULL);
266 	mutex_init(&ip->i_tlock, NULL, MUTEX_DEFAULT, NULL);
267 	dnlc_dir_init(&ip->i_danchor);
268 
269 	cv_init(&ip->i_wrcv, NULL, CV_DRIVER, NULL);
270 
271 	return (0);
272 }
273 
274 /* ARGSUSED */
275 static void
276 ufs_inode_cache_destructor(void *buf, void *cdrarg)
277 {
278 	struct inode *ip = buf;
279 	struct vnode *vp;
280 
281 	vp = ITOV(ip);
282 
283 	rw_destroy(&ip->i_rwlock);
284 	rw_destroy(&ip->i_contents);
285 	mutex_destroy(&ip->i_tlock);
286 	if (vp->v_type == VDIR) {
287 		dnlc_dir_fini(&ip->i_danchor);
288 	}
289 
290 	cv_destroy(&ip->i_wrcv);
291 
292 	vn_free(vp);
293 }
294 
295 /*
296  * Initialize hash links for inodes
297  * and build inode free list.
298  */
299 void
300 ihinit(void)
301 {
302 	int i;
303 	union	ihead *ih = ihead;
304 
305 	mutex_init(&ufs_scan_lock, NULL, MUTEX_DEFAULT, NULL);
306 
307 	inohsz = 1 << highbit(ufs_ninode / ino_hashlen);
308 	ihead = kmem_zalloc(inohsz * sizeof (union ihead), KM_SLEEP);
309 	ih_lock = kmem_zalloc(inohsz * sizeof (kmutex_t), KM_SLEEP);
310 
311 	for (i = 0, ih = ihead; i < inohsz; i++,  ih++) {
312 		ih->ih_head[0] = ih;
313 		ih->ih_head[1] = ih;
314 		mutex_init(&ih_lock[i], NULL, MUTEX_DEFAULT, NULL);
315 	}
316 	inode_cache = kmem_cache_create("ufs_inode_cache",
317 	    sizeof (struct inode), 0, ufs_inode_cache_constructor,
318 	    ufs_inode_cache_destructor, ufs_inode_cache_reclaim,
319 	    NULL, NULL, 0);
320 }
321 
322 /*
323  * Free an inode structure
324  */
325 void
326 ufs_free_inode(struct inode *ip)
327 {
328 	vn_invalid(ITOV(ip));
329 	kmem_cache_free(inode_cache, ip);
330 }
331 
332 /*
333  * Allocate an inode structure
334  */
335 struct inode *
336 ufs_alloc_inode(ufsvfs_t *ufsvfsp, ino_t ino)
337 {
338 	struct inode *ip;
339 	vnode_t *vp;
340 
341 	ip = kmem_cache_alloc(inode_cache, KM_SLEEP);
342 	/*
343 	 * at this point we have a newly allocated inode
344 	 */
345 	ip->i_freef = ip;
346 	ip->i_freeb = ip;
347 	ip->i_flag = IREF;
348 	ip->i_seq = 0xFF;	/* Unique initial value */
349 	ip->i_dev = ufsvfsp->vfs_dev;
350 	ip->i_ufsvfs = ufsvfsp;
351 	ip->i_devvp = ufsvfsp->vfs_devvp;
352 	ip->i_number = ino;
353 	ip->i_diroff = 0;
354 	ip->i_nextr = 0;
355 	ip->i_map = NULL;
356 	ip->i_rdev = 0;
357 	ip->i_writes = 0;
358 	ip->i_mode = 0;
359 	ip->i_delaylen = 0;
360 	ip->i_delayoff = 0;
361 	ip->i_nextrio = 0;
362 	ip->i_ufs_acl = NULL;
363 	ip->i_cflags = 0;
364 	ip->i_mapcnt = 0;
365 	ip->i_dquot = NULL;
366 	ip->i_cachedir = CD_ENABLED;
367 	ip->i_writer = NULL;
368 
369 	/*
370 	 * the vnode for this inode was allocated by the constructor
371 	 */
372 	vp = ITOV(ip);
373 	vn_reinit(vp);
374 	if (ino == (ino_t)UFSROOTINO)
375 		vp->v_flag = VROOT;
376 	vp->v_vfsp = ufsvfsp->vfs_vfs;
377 	vn_exists(vp);
378 	return (ip);
379 }
380 
381 /*
382  * Look up an inode by device, inumber.  If it is in core (in the
383  * inode structure), honor the locking protocol.  If it is not in
384  * core, read it in from the specified device after freeing any pages.
385  * In all cases, a pointer to a VN_HELD inode structure is returned.
386  */
387 int
388 ufs_iget(struct vfs *vfsp, ino_t ino, struct inode **ipp, struct cred *cr)
389 {
390 	return (ufs_iget_internal(vfsp, ino, ipp, cr, 0));
391 }
392 
393 /*
394  * A version of ufs_iget which returns only allocated, linked inodes.
395  * This is appropriate for any callers who do not expect a free inode.
396  */
397 int
398 ufs_iget_alloced(struct vfs *vfsp, ino_t ino, struct inode **ipp,
399     struct cred *cr)
400 {
401 	return (ufs_iget_internal(vfsp, ino, ipp, cr, 1));
402 }
403 
404 /*
405  * Set vnode attributes based on v_type, this should be called whenever
406  * an inode's i_mode is changed.
407  */
408 void
409 ufs_reset_vnode(vnode_t *vp)
410 {
411 	/*
412 	 * an old DBE hack
413 	 */
414 	if ((VTOI(vp)->i_mode & (ISVTX | IEXEC | IFDIR)) == ISVTX)
415 		vp->v_flag |= VSWAPLIKE;
416 	else
417 		vp->v_flag &= ~VSWAPLIKE;
418 
419 	/*
420 	 * if not swap like and it's just a regular file, we want
421 	 * to maintain the vnode's pages sorted by clean/modified
422 	 * for faster sync'ing to disk
423 	 */
424 	if (vp->v_type == VREG)
425 		vp->v_flag |= VMODSORT;
426 	else
427 		vp->v_flag &= ~VMODSORT;
428 
429 	/*
430 	 * Is this an attribute hidden dir?
431 	 */
432 	if ((VTOI(vp)->i_mode & IFMT) == IFATTRDIR)
433 		vp->v_flag |= V_XATTRDIR;
434 	else
435 		vp->v_flag &= ~V_XATTRDIR;
436 }
437 
438 /*
439  * Shared implementation of ufs_iget and ufs_iget_alloced.  The 'validate'
440  * flag is used to distinguish the two; when true, we validate that the inode
441  * being retrieved looks like a linked and allocated inode.
442  */
443 /* ARGSUSED */
444 static int
445 ufs_iget_internal(struct vfs *vfsp, ino_t ino, struct inode **ipp,
446     struct cred *cr, int validate)
447 {
448 	struct inode *ip, *sp;
449 	union ihead *ih;
450 	kmutex_t *ihm;
451 	struct buf *bp;
452 	struct dinode *dp;
453 	struct vnode *vp;
454 	extern vfs_t EIO_vfs;
455 	int error;
456 	int ftype;	/* XXX - Remove later on */
457 	dev_t vfs_dev;
458 	struct ufsvfs *ufsvfsp;
459 	struct fs *fs;
460 	int hno;
461 	daddr_t bno;
462 	ulong_t ioff;
463 
464 	CPU_STATS_ADD_K(sys, ufsiget, 1);
465 
466 	/*
467 	 * Lookup inode in cache.
468 	 */
469 	vfs_dev = vfsp->vfs_dev;
470 	hno = INOHASH(ino);
471 	ih = &ihead[hno];
472 	ihm = &ih_lock[hno];
473 
474 again:
475 	mutex_enter(ihm);
476 	for (ip = ih->ih_chain[0]; ip != (struct inode *)ih; ip = ip->i_forw) {
477 		if (ino != ip->i_number || vfs_dev != ip->i_dev ||
478 		    (ip->i_flag & ISTALE))
479 			continue;
480 
481 		/*
482 		 * Found the interesting inode; hold it and drop the cache lock
483 		 */
484 		vp = ITOV(ip);	/* for locknest */
485 		VN_HOLD(vp);
486 		mutex_exit(ihm);
487 		rw_enter(&ip->i_contents, RW_READER);
488 
489 		/*
490 		 * if necessary, remove from idle list
491 		 */
492 		if ((ip->i_flag & IREF) == 0) {
493 			if (ufs_rmidle(ip))
494 				VN_RELE(vp);
495 		}
496 
497 		/*
498 		 * Could the inode be read from disk?
499 		 */
500 		if (ip->i_flag & ISTALE) {
501 			rw_exit(&ip->i_contents);
502 			VN_RELE(vp);
503 			goto again;
504 		}
505 
506 		ins.in_hits.value.ul++;
507 		*ipp = ip;
508 
509 		/*
510 		 * Reset the vnode's attribute flags
511 		 */
512 		mutex_enter(&vp->v_lock);
513 		ufs_reset_vnode(vp);
514 		mutex_exit(&vp->v_lock);
515 
516 		rw_exit(&ip->i_contents);
517 
518 		return (0);
519 	}
520 	mutex_exit(ihm);
521 
522 	/*
523 	 * Inode was not in cache.
524 	 *
525 	 * Allocate a new entry
526 	 */
527 	ufsvfsp = (struct ufsvfs *)vfsp->vfs_data;
528 	fs = ufsvfsp->vfs_fs;
529 
530 	ip = ufs_alloc_inode(ufsvfsp, ino);
531 	vp = ITOV(ip);
532 
533 	bno = fsbtodb(fs, itod(fs, ino));
534 	ioff = (sizeof (struct dinode)) * (itoo(fs, ino));
535 	ip->i_doff = (offset_t)ioff + ldbtob(bno);
536 
537 	/*
538 	 * put a place holder in the cache (if not already there)
539 	 */
540 	mutex_enter(ihm);
541 	for (sp = ih->ih_chain[0]; sp != (struct inode *)ih; sp = sp->i_forw)
542 		if (ino == sp->i_number && vfs_dev == sp->i_dev &&
543 		    ((sp->i_flag & ISTALE) == 0)) {
544 			mutex_exit(ihm);
545 			ufs_free_inode(ip);
546 			goto again;
547 		}
548 	/*
549 	 * It would be nice to ASSERT(RW_READ_HELD(&ufsvfsp->vfs_dqrwlock))
550 	 * here, but if we do, then shadow inode allocations panic the
551 	 * system.  We don't have to hold vfs_dqrwlock for shadow inodes
552 	 * and the ufs_iget() parameters don't tell us what we are getting
553 	 * so we have no way of knowing this is a ufs_iget() call from
554 	 * a ufs_ialloc() call for a shadow inode.
555 	 */
556 	rw_enter(&ip->i_contents, RW_WRITER);
557 	insque(ip, ih);
558 	mutex_exit(ihm);
559 	/*
560 	 * read the dinode
561 	 */
562 	bp = UFS_BREAD(ufsvfsp, ip->i_dev, bno, (int)fs->fs_bsize);
563 
564 	/*
565 	 * Check I/O errors
566 	 */
567 	error = ((bp->b_flags & B_ERROR) ? geterror(bp) : 0);
568 	if (error) {
569 		brelse(bp);
570 		ip->i_flag |= ISTALE;	/* in case someone is looking it up */
571 		rw_exit(&ip->i_contents);
572 		vp->v_vfsp = &EIO_vfs;
573 		VN_RELE(vp);
574 		return (error);
575 	}
576 	/*
577 	 * initialize the inode's dinode
578 	 */
579 	dp = (struct dinode *)(ioff + bp->b_un.b_addr);
580 	ip->i_ic = dp->di_ic;			/* structure assignment */
581 	brelse(bp);
582 
583 	/*
584 	 * Maintain compatibility with Solaris 1.x UFS
585 	 */
586 	if (ip->i_suid != UID_LONG)
587 		ip->i_uid = ip->i_suid;
588 	if (ip->i_sgid != GID_LONG)
589 		ip->i_gid = ip->i_sgid;
590 
591 	ftype = ip->i_mode & IFMT;
592 	if (ftype == IFBLK || ftype == IFCHR) {
593 		dev_t dv;
594 		uint_t top16 = ip->i_ordev & 0xffff0000u;
595 
596 		if (top16 == 0 || top16 == 0xffff0000u)
597 			dv = expdev(ip->i_ordev);
598 		else
599 			dv = expldev(ip->i_ordev);
600 		vp->v_rdev = ip->i_rdev = dv;
601 	}
602 
603 	/*
604 	 * if our caller only expects allocated inodes, verify that
605 	 * this inode looks good; throw it out if it's bad.
606 	 */
607 	if (validate) {
608 		if ((ftype == 0) || (ip->i_nlink <= 0)) {
609 			ip->i_flag |= ISTALE;
610 			rw_exit(&ip->i_contents);
611 			vp->v_vfsp = &EIO_vfs;
612 			VN_RELE(vp);
613 			cmn_err(CE_NOTE,
614 			    "%s: unexpected free inode %d, run fsck(1M)%s",
615 			    fs->fs_fsmnt, (int)ino,
616 			    (TRANS_ISTRANS(ufsvfsp) ? " -o f" : ""));
617 			return (EIO);
618 		}
619 	}
620 
621 	/*
622 	 * Finish initializing the vnode, special handling for shadow inodes
623 	 * because IFTOVT() will produce a v_type of VNON which is not what we
624 	 * want, set v_type to VREG explicitly in that case.
625 	 */
626 	if (ftype == IFSHAD) {
627 		vp->v_type = VREG;
628 	} else {
629 		vp->v_type = IFTOVT((mode_t)ip->i_mode);
630 	}
631 
632 	ufs_reset_vnode(vp);
633 
634 	/*
635 	 * read the shadow
636 	 */
637 	if (ftype != 0 && ip->i_shadow != 0) {
638 		if ((error = ufs_si_load(ip, cr)) != 0) {
639 			ip->i_flag |= ISTALE;
640 			ip->i_ufs_acl = NULL;
641 			rw_exit(&ip->i_contents);
642 			vp->v_vfsp = &EIO_vfs;
643 			VN_RELE(vp);
644 			return (error);
645 		}
646 	}
647 
648 	/*
649 	 * Only attach quota information if the inode has a type and if
650 	 * that type is not a shadow inode.
651 	 */
652 	if (ip->i_mode && ((ip->i_mode & IFMT) != IFSHAD) &&
653 	    ((ip->i_mode & IFMT) != IFATTRDIR)) {
654 		ip->i_dquot = getinoquota(ip);
655 	}
656 	TRANS_MATA_IGET(ufsvfsp, ip);
657 	*ipp = ip;
658 	rw_exit(&ip->i_contents);
659 
660 	return (0);
661 }
662 
663 /*
664  * Vnode is no longer referenced, write the inode out
665  * and if necessary, truncate and deallocate the file.
666  */
667 void
668 ufs_iinactive(struct inode *ip)
669 {
670 	int		front;
671 	struct inode	*iq;
672 	struct inode	*hip;
673 	struct ufs_q	*uq;
674 	struct vnode	*vp = ITOV(ip);
675 	struct ufsvfs   *ufsvfsp = ip->i_ufsvfs;
676 	struct ufs_delq_info *delq_info = &ufsvfsp->vfs_delete_info;
677 
678 	/*
679 	 * Because the vnode type might have been changed,
680 	 * the dnlc_dir_purge must be called unconditionally.
681 	 */
682 	dnlc_dir_purge(&ip->i_danchor);
683 
684 	/*
685 	 * Get exclusive access to inode data.
686 	 */
687 	rw_enter(&ip->i_contents, RW_WRITER);
688 	ASSERT(ip->i_flag & IREF);
689 
690 	/*
691 	 * Make sure no one reclaimed the inode before we put it on
692 	 * the freelist or destroy it. We keep our 'hold' on the vnode
693 	 * from vn_rele until we are ready to do something with the inode.
694 	 *
695 	 * Pageout may put a VN_HOLD/VN_RELE at anytime during this
696 	 * operation via an async putpage, so we must make sure
697 	 * we don't free/destroy the inode more than once. ufs_iget
698 	 * may also put a VN_HOLD on the inode before it grabs
699 	 * the i_contents lock. This is done so we don't free
700 	 * an inode that a thread is waiting on.
701 	 */
702 	mutex_enter(&vp->v_lock);
703 
704 	if (vp->v_count > 1) {
705 		vp->v_count--;  /* release our hold from vn_rele */
706 		mutex_exit(&vp->v_lock);
707 		rw_exit(&ip->i_contents);
708 		return;
709 	}
710 	mutex_exit(&vp->v_lock);
711 
712 	/*
713 	 * For umount case: if ufsvfs ptr is NULL, the inode is unhashed
714 	 * and clean.  It can be safely destroyed (cyf).
715 	 */
716 	if (ip->i_ufsvfs == NULL) {
717 		rw_exit(&ip->i_contents);
718 		ufs_si_del(ip);
719 		ASSERT((vp->v_type == VCHR) || !vn_has_cached_data(vp));
720 		ufs_free_inode(ip);
721 		return;
722 	}
723 
724 	/*
725 	 * queue idle inode to appropriate thread. Will check v_count == 1
726 	 * prior to putting this on the appropriate queue.
727 	 * Stale inodes will be unhashed and freed by the ufs idle thread
728 	 * in ufs_idle_free()
729 	 */
730 	front = 1;
731 	if ((ip->i_flag & ISTALE) == 0 && ip->i_fs->fs_ronly == 0 &&
732 	    ip->i_mode && ip->i_nlink <= 0) {
733 		/*
734 		 * Mark the i_flag to indicate that inode is being deleted.
735 		 * This flag will be cleared when the deletion is complete.
736 		 * This prevents nfs from sneaking in via ufs_vget() while
737 		 * the delete is in progress (bugid 1242481).
738 		 */
739 		ip->i_flag |= IDEL;
740 
741 		/*
742 		 * NOIDEL means that deletes are not allowed at this time;
743 		 * whoever resets NOIDEL will also send this inode back
744 		 * through ufs_iinactive.  IREF remains set.
745 		 */
746 		if (ULOCKFS_IS_NOIDEL(ITOUL(ip))) {
747 			mutex_enter(&vp->v_lock);
748 			vp->v_count--;
749 			mutex_exit(&vp->v_lock);
750 			rw_exit(&ip->i_contents);
751 			return;
752 		}
753 		if (!TRANS_ISTRANS(ip->i_ufsvfs)) {
754 			rw_exit(&ip->i_contents);
755 			ufs_delete(ip->i_ufsvfs, ip, 0);
756 			return;
757 		}
758 
759 		/* queue to delete thread; IREF remains set */
760 		ins.in_qfree.value.ul++;
761 		uq = &ip->i_ufsvfs->vfs_delete;
762 
763 		mutex_enter(&uq->uq_mutex);
764 
765 		/* add to q */
766 		if ((iq = uq->uq_ihead) != 0) {
767 			ip->i_freef = iq;
768 			ip->i_freeb = iq->i_freeb;
769 			iq->i_freeb->i_freef = ip;
770 			iq->i_freeb = ip;
771 			if (front)
772 				uq->uq_ihead = ip;
773 		} else {
774 			uq->uq_ihead = ip;
775 			ip->i_freef = ip;
776 			ip->i_freeb = ip;
777 		}
778 
779 		delq_info->delq_unreclaimed_files += 1;
780 		delq_info->delq_unreclaimed_blocks += ip->i_blocks;
781 	} else {
782 		/*
783 		 * queue to idle thread
784 		 *  Check the v_count == 1 again.
785 		 *
786 		 */
787 		mutex_enter(&vp->v_lock);
788 		if (vp->v_count > 1) {
789 			vp->v_count--;  /* release our hold from vn_rele */
790 			mutex_exit(&vp->v_lock);
791 			rw_exit(&ip->i_contents);
792 			return;
793 		}
794 		mutex_exit(&vp->v_lock);
795 		uq = &ufs_idle_q;
796 
797 		/*
798 		 * useful iff it has pages or is a fastsymlink; otherwise junk
799 		 */
800 		mutex_enter(&uq->uq_mutex);
801 
802 		/* clear IREF means `on idle list' */
803 		ip->i_flag &= ~(IREF | IDIRECTIO);
804 
805 		if (vn_has_cached_data(vp) || ip->i_flag & IFASTSYMLNK) {
806 			ins.in_frback.value.ul++;
807 			hip = (inode_t *)&ufs_useful_iq[IQHASH(ip)];
808 			ufs_nuseful_iq++;
809 		} else {
810 			ins.in_frfront.value.ul++;
811 			hip = (inode_t *)&ufs_junk_iq[IQHASH(ip)];
812 			ip->i_flag |= IJUNKIQ;
813 			ufs_njunk_iq++;
814 		}
815 		ip->i_freef = hip;
816 		ip->i_freeb = hip->i_freeb;
817 		hip->i_freeb->i_freef = ip;
818 		hip->i_freeb = ip;
819 	}
820 
821 	/* wakeup thread(s) if q is overfull */
822 	if (++uq->uq_ne == uq->uq_lowat)
823 		cv_broadcast(&uq->uq_cv);
824 
825 	/* all done, release the q and inode */
826 	mutex_exit(&uq->uq_mutex);
827 	rw_exit(&ip->i_contents);
828 }
829 
830 /*
831  * Check accessed and update flags on an inode structure.
832  * If any are on, update the inode with the (unique) current time.
833  * If waitfor is given, insure I/O order so wait for write to complete.
834  */
835 void
836 ufs_iupdat(struct inode *ip, int waitfor)
837 {
838 	struct buf	*bp;
839 	struct fs	*fp;
840 	struct dinode	*dp;
841 	struct ufsvfs	*ufsvfsp 	= ip->i_ufsvfs;
842 	int 		i;
843 	int		do_trans_times;
844 	ushort_t	flag;
845 	o_uid_t		suid;
846 	o_gid_t		sgid;
847 
848 	/*
849 	 * This function is now safe to be called with either the reader
850 	 * or writer i_contents lock.
851 	 */
852 	ASSERT(RW_LOCK_HELD(&ip->i_contents));
853 
854 	/*
855 	 * Return if file system has been forcibly umounted.
856 	 */
857 	if (ufsvfsp == NULL)
858 		return;
859 
860 	flag = ip->i_flag;	/* Atomic read */
861 	/*
862 	 * We better not update the disk inode from a stale inode.
863 	 */
864 	if (flag & ISTALE)
865 		return;
866 
867 	fp = ip->i_fs;
868 
869 	if ((flag & (IUPD|IACC|ICHG|IMOD|IMODACC|IATTCHG)) != 0) {
870 		if (fp->fs_ronly) {
871 			mutex_enter(&ip->i_tlock);
872 			ip->i_flag &= ~(IUPD|IACC|ICHG|IMOD|IMODACC|IATTCHG);
873 			mutex_exit(&ip->i_tlock);
874 			return;
875 		}
876 		/*
877 		 * fs is active while metadata is being written
878 		 */
879 		mutex_enter(&ufsvfsp->vfs_lock);
880 		ufs_notclean(ufsvfsp);
881 		/*
882 		 * get the dinode
883 		 */
884 		bp = UFS_BREAD(ufsvfsp, ip->i_dev,
885 		    (daddr_t)fsbtodb(fp, itod(fp, ip->i_number)),
886 		    (int)fp->fs_bsize);
887 		if (bp->b_flags & B_ERROR) {
888 			mutex_enter(&ip->i_tlock);
889 			ip->i_flag &=
890 			    ~(IUPD|IACC|ICHG|IMOD|IMODACC|IATTCHG);
891 			mutex_exit(&ip->i_tlock);
892 			brelse(bp);
893 			return;
894 		}
895 		/*
896 		 * munge inode fields
897 		 */
898 		mutex_enter(&ip->i_tlock);
899 		ITIMES_NOLOCK(ip);
900 		do_trans_times = ((ip->i_flag & (IMOD|IMODACC)) == IMODACC);
901 		ip->i_flag &= ~(IUPD|IACC|ICHG|IMOD|IMODACC|IATTCHG);
902 		mutex_exit(&ip->i_tlock);
903 
904 		/*
905 		 * For reads and concurrent re-writes, no deltas were
906 		 * entered for the access time changes - do it now.
907 		 */
908 		if (do_trans_times) {
909 			TRANS_INODE_TIMES(ufsvfsp, ip);
910 		}
911 
912 		/*
913 		 * For SunOS 5.0->5.4, these lines below read:
914 		 *
915 		 * suid = (ip->i_uid > MAXUID) ? UID_LONG : ip->i_uid;
916 		 * sgid = (ip->i_gid > MAXUID) ? GID_LONG : ip->i_gid;
917 		 *
918 		 * where MAXUID was set to 60002.  This was incorrect -
919 		 * the uids should have been constrained to what fitted into
920 		 * a 16-bit word.
921 		 *
922 		 * This means that files from 4.x filesystems that have an
923 		 * i_suid field larger than 60002 will have that field
924 		 * changed to 65535.
925 		 *
926 		 * Security note: 4.x UFS could never create a i_suid of
927 		 * UID_LONG since that would've corresponded to -1.
928 		 */
929 		suid = (ulong_t)ip->i_uid > (ulong_t)USHRT_MAX ?
930 		    UID_LONG : ip->i_uid;
931 		sgid = (ulong_t)ip->i_gid > (ulong_t)USHRT_MAX ?
932 		    GID_LONG : ip->i_gid;
933 
934 		if ((ip->i_suid != suid) || (ip->i_sgid != sgid)) {
935 			ip->i_suid = suid;
936 			ip->i_sgid = sgid;
937 			TRANS_INODE(ufsvfsp, ip);
938 		}
939 
940 		if ((ip->i_mode & IFMT) == IFBLK ||
941 		    (ip->i_mode & IFMT) == IFCHR) {
942 			dev_t d = ip->i_rdev;
943 			dev32_t dev32;
944 
945 			/*
946 			 * load first direct block only if special device
947 			 */
948 			if (!cmpldev(&dev32, d)) {
949 				/*
950 				 * We panic here because there's "no way"
951 				 * we should have been able to create a large
952 				 * inode with a large dev_t.  Earlier layers
953 				 * should've caught this.
954 				 */
955 				panic("ip %p: i_rdev too big", (void *)ip);
956 			}
957 
958 			if (dev32 & ~((O_MAXMAJ << L_BITSMINOR32) | O_MAXMIN)) {
959 				ip->i_ordev = dev32;	/* can't use old fmt. */
960 			} else {
961 				ip->i_ordev = cmpdev(d);
962 			}
963 		}
964 
965 		/*
966 		 * copy inode to dinode (zero fastsymlnk in dinode)
967 		 */
968 		dp = (struct dinode *)bp->b_un.b_addr + itoo(fp, ip->i_number);
969 		dp->di_ic = ip->i_ic;	/* structure assignment */
970 		if (flag & IFASTSYMLNK) {
971 			for (i = 1; i < NDADDR; i++)
972 				dp->di_db[i] = 0;
973 			for (i = 0; i < NIADDR; i++)
974 				dp->di_ib[i] = 0;
975 		}
976 		if (TRANS_ISTRANS(ufsvfsp)) {
977 			/*
978 			 * Pass only a sector size buffer containing
979 			 * the inode, otherwise when the buffer is copied
980 			 * into a cached roll buffer then too much memory
981 			 * gets consumed if 8KB inode buffers are passed.
982 			 */
983 			TRANS_LOG(ufsvfsp, (caddr_t)dp, ip->i_doff,
984 			    sizeof (struct dinode),
985 			    (caddr_t)P2ALIGN((uintptr_t)dp, DEV_BSIZE),
986 			    DEV_BSIZE);
987 
988 			brelse(bp);
989 		} else if (waitfor && (ip->i_ufsvfs->vfs_dio == 0)) {
990 			UFS_BRWRITE(ufsvfsp, bp);
991 
992 			/*
993 			 * Synchronous write has guaranteed that inode
994 			 * has been written on disk so clear the flag
995 			 */
996 			mutex_enter(&ip->i_tlock);
997 			ip->i_flag &= ~IBDWRITE;
998 			mutex_exit(&ip->i_tlock);
999 		} else {
1000 			bdrwrite(bp);
1001 
1002 			/*
1003 			 * This write hasn't guaranteed that inode has been
1004 			 * written on the disk.
1005 			 * Since, all updat flags on inode are cleared, we must
1006 			 * remember the condition in case inode is to be updated
1007 			 * synchronously later (e.g.- fsync()/fdatasync())
1008 			 * and inode has not been modified yet.
1009 			 */
1010 			mutex_enter(&ip->i_tlock);
1011 			ip->i_flag |= IBDWRITE;
1012 			mutex_exit(&ip->i_tlock);
1013 		}
1014 	} else {
1015 		/*
1016 		 * In case previous inode update was done asynchronously
1017 		 * (IBDWRITE) and this inode update request wants guaranteed
1018 		 * (synchronous) disk update, flush the inode.
1019 		 */
1020 		if (waitfor && (flag & IBDWRITE)) {
1021 			blkflush(ip->i_dev,
1022 			    (daddr_t)fsbtodb(fp, itod(fp, ip->i_number)));
1023 			mutex_enter(&ip->i_tlock);
1024 			ip->i_flag &= ~IBDWRITE;
1025 			mutex_exit(&ip->i_tlock);
1026 		}
1027 	}
1028 }
1029 
1030 #define	SINGLE	0	/* index of single indirect block */
1031 #define	DOUBLE	1	/* index of double indirect block */
1032 #define	TRIPLE	2	/* index of triple indirect block */
1033 
1034 /*
1035  * Release blocks associated with the inode ip and
1036  * stored in the indirect block bn.  Blocks are free'd
1037  * in LIFO order up to (but not including) lastbn.  If
1038  * level is greater than SINGLE, the block is an indirect
1039  * block and recursive calls to indirtrunc must be used to
1040  * cleanse other indirect blocks.
1041  *
1042  * N.B.: triple indirect blocks are untested.
1043  */
1044 static long
1045 indirtrunc(struct inode *ip, daddr_t bn, daddr_t lastbn, int level, int flags)
1046 {
1047 	int i;
1048 	struct buf *bp, *copy;
1049 	daddr32_t *bap;
1050 	struct ufsvfs *ufsvfsp = ip->i_ufsvfs;
1051 	struct fs *fs = ufsvfsp->vfs_fs;
1052 	daddr_t nb, last;
1053 	long factor;
1054 	int blocksreleased = 0, nblocks;
1055 
1056 	ASSERT(RW_WRITE_HELD(&ip->i_contents));
1057 	/*
1058 	 * Calculate index in current block of last
1059 	 * block to be kept.  -1 indicates the entire
1060 	 * block so we need not calculate the index.
1061 	 */
1062 	factor = 1;
1063 	for (i = SINGLE; i < level; i++)
1064 		factor *= NINDIR(fs);
1065 	last = lastbn;
1066 	if (lastbn > 0)
1067 		last /= factor;
1068 	nblocks = btodb(fs->fs_bsize);
1069 	/*
1070 	 * Get buffer of block pointers, zero those
1071 	 * entries corresponding to blocks to be free'd,
1072 	 * and update on disk copy first.
1073 	 * *Unless* the root pointer has been synchronously
1074 	 * written to disk.  If nothing points to this
1075 	 * indirect block then don't bother zero'ing and
1076 	 * writing it.
1077 	 */
1078 	bp = UFS_BREAD(ufsvfsp,
1079 	    ip->i_dev, (daddr_t)fsbtodb(fs, bn), (int)fs->fs_bsize);
1080 	if (bp->b_flags & B_ERROR) {
1081 		brelse(bp);
1082 		return (0);
1083 	}
1084 	bap = bp->b_un.b_daddr;
1085 	if ((flags & I_CHEAP) == 0) {
1086 		uint_t	zb;
1087 
1088 		zb = (uint_t)((NINDIR(fs) - (last + 1)) * sizeof (daddr32_t));
1089 
1090 		if (zb) {
1091 			/*
1092 			 * push any data into the log before we zero it
1093 			 */
1094 			if (bp->b_flags & B_DELWRI)
1095 				TRANS_LOG(ufsvfsp, (caddr_t)bap,
1096 				    ldbtob(bp->b_blkno), bp->b_bcount,
1097 				    bp->b_un.b_addr, bp->b_bcount);
1098 			copy = ngeteblk(fs->fs_bsize);
1099 			bcopy((caddr_t)bap, (caddr_t)copy->b_un.b_daddr,
1100 			    (uint_t)fs->fs_bsize);
1101 			bzero((caddr_t)&bap[last + 1], zb);
1102 
1103 			TRANS_BUF(ufsvfsp,
1104 			    (caddr_t)&bap[last + 1] - (caddr_t)bap,
1105 			    zb, bp, DT_ABZERO);
1106 
1107 			UFS_BRWRITE(ufsvfsp, bp);
1108 			bp = copy, bap = bp->b_un.b_daddr;
1109 		}
1110 	} else {
1111 		/* make sure write retries are also cleared */
1112 		bp->b_flags &= ~(B_DELWRI | B_RETRYWRI);
1113 		bp->b_flags |= B_STALE | B_AGE;
1114 	}
1115 
1116 	/*
1117 	 * Recursively free totally unused blocks.
1118 	 */
1119 	flags |= I_CHEAP;
1120 	for (i = NINDIR(fs) - 1; i > last; i--) {
1121 		nb = bap[i];
1122 		if (nb == 0)
1123 			continue;
1124 		if (level > SINGLE) {
1125 			blocksreleased +=
1126 			    indirtrunc(ip, nb, (daddr_t)-1, level - 1, flags);
1127 			free(ip, nb, (off_t)fs->fs_bsize, flags | I_IBLK);
1128 		} else
1129 			free(ip, nb, (off_t)fs->fs_bsize, flags);
1130 		blocksreleased += nblocks;
1131 	}
1132 	flags &= ~I_CHEAP;
1133 
1134 	/*
1135 	 * Recursively free last partial block.
1136 	 */
1137 	if (level > SINGLE && lastbn >= 0) {
1138 		last = lastbn % factor;
1139 		nb = bap[i];
1140 		if (nb != 0)
1141 			blocksreleased +=
1142 			    indirtrunc(ip, nb, last, level - 1, flags);
1143 	}
1144 	brelse(bp);
1145 	return (blocksreleased);
1146 }
1147 
1148 /*
1149  * Truncate the inode ip to at most length size.
1150  * Free affected disk blocks -- the blocks of the
1151  * file are removed in reverse order.
1152  *
1153  * N.B.: triple indirect blocks are untested.
1154  */
1155 static int i_genrand = 1234;
1156 int
1157 ufs_itrunc(struct inode *oip, u_offset_t length, int flags, cred_t *cr)
1158 {
1159 	struct fs *fs = oip->i_fs;
1160 	struct ufsvfs *ufsvfsp = oip->i_ufsvfs;
1161 	struct inode *ip;
1162 	daddr_t lastblock;
1163 	off_t bsize;
1164 	int boff;
1165 	daddr_t bn, lastiblock[NIADDR];
1166 	int level;
1167 	long nblocks, blocksreleased = 0;
1168 	int i;
1169 	ushort_t mode;
1170 	struct inode tip;
1171 	int err;
1172 	u_offset_t maxoffset = (ufsvfsp->vfs_lfflags & UFS_LARGEFILES) ?
1173 	    (UFS_MAXOFFSET_T) : (MAXOFF32_T);
1174 
1175 	/*
1176 	 * Shadow inodes do not need to hold the vfs_dqrwlock lock. Most
1177 	 * other uses need the reader lock. opendq() holds the writer lock.
1178 	 */
1179 	ASSERT((oip->i_mode & IFMT) == IFSHAD ||
1180 	    RW_LOCK_HELD(&ufsvfsp->vfs_dqrwlock));
1181 	ASSERT(RW_WRITE_HELD(&oip->i_contents));
1182 	/*
1183 	 * We only allow truncation of regular files and directories
1184 	 * to arbitrary lengths here.  In addition, we allow symbolic
1185 	 * links to be truncated only to zero length.  Other inode
1186 	 * types cannot have their length set here.  Disk blocks are
1187 	 * being dealt with - especially device inodes where
1188 	 * ip->i_ordev is actually being stored in ip->i_db[0]!
1189 	 */
1190 	TRANS_INODE(ufsvfsp, oip);
1191 	mode = oip->i_mode & IFMT;
1192 	if (flags & I_FREE) {
1193 		i_genrand *= 16843009;  /* turns into shift and adds */
1194 		i_genrand++;
1195 		oip->i_gen += ((i_genrand + lbolt) & 0xffff) + 1;
1196 		oip->i_flag |= ICHG |IUPD;
1197 		oip->i_seq++;
1198 		if (length == oip->i_size)
1199 			return (0);
1200 		flags |= I_CHEAP;
1201 	}
1202 	if (mode == IFIFO)
1203 		return (0);
1204 	if (mode != IFREG && mode != IFDIR && mode != IFATTRDIR &&
1205 	    !(mode == IFLNK && length == (offset_t)0) && mode != IFSHAD)
1206 		return (EINVAL);
1207 	if (length > maxoffset)
1208 		return (EFBIG);
1209 	if ((mode == IFDIR) || (mode == IFATTRDIR))
1210 		flags |= I_DIR;
1211 	if (mode == IFSHAD)
1212 		flags |= I_SHAD;
1213 	if (oip == ufsvfsp->vfs_qinod)
1214 		flags |= I_QUOTA;
1215 	if (length == oip->i_size) {
1216 		/* update ctime and mtime to please POSIX tests */
1217 		oip->i_flag |= ICHG |IUPD;
1218 		oip->i_seq++;
1219 		if (length == 0) {
1220 			/* nothing to cache so clear the flag */
1221 			oip->i_flag &= ~IFASTSYMLNK;
1222 		}
1223 		return (0);
1224 	}
1225 	/* wipe out fast symlink till next access */
1226 	if (oip->i_flag & IFASTSYMLNK) {
1227 		int j;
1228 
1229 		ASSERT(ITOV(oip)->v_type == VLNK);
1230 
1231 		oip->i_flag &= ~IFASTSYMLNK;
1232 
1233 		for (j = 1; j < NDADDR; j++)
1234 			oip->i_db[j] = 0;
1235 		for (j = 0; j < NIADDR; j++)
1236 			oip->i_ib[j] = 0;
1237 	}
1238 
1239 	boff = (int)blkoff(fs, length);
1240 
1241 	if (length > oip->i_size) {
1242 		/*
1243 		 * Trunc up case.  BMAPALLOC will insure that the right blocks
1244 		 * are allocated.  This includes extending the old frag to a
1245 		 * full block (if needed) in addition to doing any work
1246 		 * needed for allocating the last block.
1247 		 */
1248 		if (boff == 0)
1249 			err = BMAPALLOC(oip, length - 1, (int)fs->fs_bsize, cr);
1250 		else
1251 			err = BMAPALLOC(oip, length - 1, boff, cr);
1252 
1253 		if (err == 0) {
1254 			/*
1255 			 * Save old size and set inode's size now
1256 			 * so that we don't cause too much of the
1257 			 * file to be zero'd and pushed.
1258 			 */
1259 			u_offset_t osize = oip->i_size;
1260 			oip->i_size  = length;
1261 			/*
1262 			 * Make sure we zero out the remaining bytes of
1263 			 * the page in case a mmap scribbled on it. We
1264 			 * can't prevent a mmap from writing beyond EOF
1265 			 * on the last page of a file.
1266 			 *
1267 			 */
1268 			if ((boff = (int)blkoff(fs, osize)) != 0) {
1269 				bsize = (int)lblkno(fs, osize - 1) >= NDADDR ?
1270 				    fs->fs_bsize : fragroundup(fs, boff);
1271 				pvn_vpzero(ITOV(oip), osize,
1272 				    (size_t)(bsize - boff));
1273 			}
1274 			oip->i_flag |= ICHG|IATTCHG;
1275 			oip->i_seq++;
1276 			ITIMES_NOLOCK(oip);
1277 			/*
1278 			 * MAXOFF32_T is old 2GB size limit. If
1279 			 * this operation caused a large file to be
1280 			 * created, turn on the superblock flag
1281 			 * and update the superblock, if the flag
1282 			 * is not already on.
1283 			 */
1284 			if ((length > (u_offset_t)MAXOFF32_T) &&
1285 			    !(fs->fs_flags & FSLARGEFILES)) {
1286 				ASSERT(ufsvfsp->vfs_lfflags & UFS_LARGEFILES);
1287 				mutex_enter(&ufsvfsp->vfs_lock);
1288 				fs->fs_flags |= FSLARGEFILES;
1289 				ufs_sbwrite(ufsvfsp);
1290 				mutex_exit(&ufsvfsp->vfs_lock);
1291 			}
1292 		}
1293 
1294 		return (err);
1295 	}
1296 
1297 	/*
1298 	 * Update the pages of the file.  If the file is not being
1299 	 * truncated to a block boundary, the contents of the
1300 	 * pages following the end of the file must be zero'ed
1301 	 * in case it ever become accessible again because
1302 	 * of subsequent file growth.
1303 	 */
1304 	if (boff == 0) {
1305 		(void) pvn_vplist_dirty(ITOV(oip), length, ufs_putapage,
1306 		    B_INVAL | B_TRUNC, CRED());
1307 	} else {
1308 		/*
1309 		 * Make sure that the last block is properly allocated.
1310 		 * We only really have to do this if the last block is
1311 		 * actually allocated since ufs_bmap will now handle the case
1312 		 * of an fragment which has no block allocated.  Just to
1313 		 * be sure, we do it now independent of current allocation.
1314 		 */
1315 		err = BMAPALLOC(oip, length - 1, boff, cr);
1316 		if (err)
1317 			return (err);
1318 
1319 		/*
1320 		 * BMAPALLOC will call bmap_write which defers i_seq
1321 		 * processing.  If the timestamps were changed, update
1322 		 * i_seq before rdip drops i_contents or syncs the inode.
1323 		 */
1324 		if (oip->i_flag & (ICHG|IUPD))
1325 			oip->i_seq++;
1326 
1327 		/*
1328 		 * BugId 4069932
1329 		 * Make sure that the relevant partial page appears in
1330 		 * the v_pages list, so that pvn_vpzero() will do its
1331 		 * job.  Since doing this correctly requires everything
1332 		 * in rdip() except for the uiomove(), it's easier and
1333 		 * safer to do the uiomove() rather than duplicate the
1334 		 * rest of rdip() here.
1335 		 *
1336 		 * To get here, we know that length indicates a byte
1337 		 * that is not the first byte of a block.  (length - 1)
1338 		 * is the last actual byte known to exist.  Deduction
1339 		 * shows it is in the same block as byte (length).
1340 		 * Thus, this rdip() invocation should always succeed
1341 		 * except in the face of i/o errors, and give us the
1342 		 * block we care about.
1343 		 *
1344 		 * rdip() makes the same locking assertions and
1345 		 * assumptions as we do.  We do not acquire any locks
1346 		 * before calling it, so we have not changed the locking
1347 		 * situation.  Finally, there do not appear to be any
1348 		 * paths whereby rdip() ends up invoking us again.
1349 		 * Thus, infinite recursion is avoided.
1350 		 */
1351 		{
1352 			uio_t uio;
1353 			iovec_t iov[1];
1354 			char buffer;
1355 
1356 			uio.uio_iov = iov;
1357 			uio.uio_iovcnt = 1;
1358 			uio.uio_loffset = length - 1;
1359 			uio.uio_resid = 1;
1360 			uio.uio_segflg = UIO_SYSSPACE;
1361 			uio.uio_extflg = UIO_COPY_CACHED;
1362 
1363 			iov[0].iov_base = &buffer;
1364 			iov[0].iov_len = 1;
1365 
1366 			err = rdip(oip, &uio, UIO_READ, NULL);
1367 			if (err)
1368 				return (err);
1369 		}
1370 
1371 		bsize = (int)lblkno(fs, length - 1) >= NDADDR ?
1372 		    fs->fs_bsize : fragroundup(fs, boff);
1373 		pvn_vpzero(ITOV(oip), length, (size_t)(bsize - boff));
1374 		/*
1375 		 * Ensure full fs block is marked as dirty.
1376 		 */
1377 		(void) pvn_vplist_dirty(ITOV(oip), length + (bsize - boff),
1378 		    ufs_putapage, B_INVAL | B_TRUNC, CRED());
1379 	}
1380 
1381 	/*
1382 	 * Calculate index into inode's block list of
1383 	 * last direct and indirect blocks (if any)
1384 	 * which we want to keep.  Lastblock is -1 when
1385 	 * the file is truncated to 0.
1386 	 */
1387 	lastblock = lblkno(fs, length + fs->fs_bsize - 1) - 1;
1388 	lastiblock[SINGLE] = lastblock - NDADDR;
1389 	lastiblock[DOUBLE] = lastiblock[SINGLE] - NINDIR(fs);
1390 	lastiblock[TRIPLE] = lastiblock[DOUBLE] - NINDIR(fs) * NINDIR(fs);
1391 	nblocks = btodb(fs->fs_bsize);
1392 
1393 	/*
1394 	 * Update file and block pointers
1395 	 * on disk before we start freeing blocks.
1396 	 * If we crash before free'ing blocks below,
1397 	 * the blocks will be returned to the free list.
1398 	 * lastiblock values are also normalized to -1
1399 	 * for calls to indirtrunc below.
1400 	 */
1401 	tip = *oip;			/* structure copy */
1402 	ip = &tip;
1403 
1404 	for (level = TRIPLE; level >= SINGLE; level--)
1405 		if (lastiblock[level] < 0) {
1406 			oip->i_ib[level] = 0;
1407 			lastiblock[level] = -1;
1408 		}
1409 	for (i = NDADDR - 1; i > lastblock; i--) {
1410 		oip->i_db[i] = 0;
1411 		flags |= I_CHEAP;
1412 	}
1413 	oip->i_size = length;
1414 	oip->i_flag |= ICHG|IUPD|IATTCHG;
1415 	oip->i_seq++;
1416 	if (!TRANS_ISTRANS(ufsvfsp))
1417 		ufs_iupdat(oip, I_SYNC);	/* do sync inode update */
1418 
1419 	/*
1420 	 * Indirect blocks first.
1421 	 */
1422 	for (level = TRIPLE; level >= SINGLE; level--) {
1423 		bn = ip->i_ib[level];
1424 		if (bn != 0) {
1425 			blocksreleased +=
1426 			    indirtrunc(ip, bn, lastiblock[level], level, flags);
1427 			if (lastiblock[level] < 0) {
1428 				ip->i_ib[level] = 0;
1429 				free(ip, bn, (off_t)fs->fs_bsize,
1430 				    flags | I_IBLK);
1431 				blocksreleased += nblocks;
1432 			}
1433 		}
1434 		if (lastiblock[level] >= 0)
1435 			goto done;
1436 	}
1437 
1438 	/*
1439 	 * All whole direct blocks or frags.
1440 	 */
1441 	for (i = NDADDR - 1; i > lastblock; i--) {
1442 		bn = ip->i_db[i];
1443 		if (bn == 0)
1444 			continue;
1445 		ip->i_db[i] = 0;
1446 		bsize = (off_t)blksize(fs, ip, i);
1447 		free(ip, bn, bsize, flags);
1448 		blocksreleased += btodb(bsize);
1449 	}
1450 	if (lastblock < 0)
1451 		goto done;
1452 
1453 	/*
1454 	 * Finally, look for a change in size of the
1455 	 * last direct block; release any frags.
1456 	 */
1457 	bn = ip->i_db[lastblock];
1458 	if (bn != 0) {
1459 		off_t oldspace, newspace;
1460 
1461 		/*
1462 		 * Calculate amount of space we're giving
1463 		 * back as old block size minus new block size.
1464 		 */
1465 		oldspace = blksize(fs, ip, lastblock);
1466 		UFS_SET_ISIZE(length, ip);
1467 		newspace = blksize(fs, ip, lastblock);
1468 		if (newspace == 0) {
1469 			err = ufs_fault(ITOV(ip), "ufs_itrunc: newspace == 0");
1470 			return (err);
1471 		}
1472 		if (oldspace - newspace > 0) {
1473 			/*
1474 			 * Block number of space to be free'd is
1475 			 * the old block # plus the number of frags
1476 			 * required for the storage we're keeping.
1477 			 */
1478 			bn += numfrags(fs, newspace);
1479 			free(ip, bn, oldspace - newspace, flags);
1480 			blocksreleased += btodb(oldspace - newspace);
1481 		}
1482 	}
1483 done:
1484 /* BEGIN PARANOIA */
1485 	for (level = SINGLE; level <= TRIPLE; level++)
1486 		if (ip->i_ib[level] != oip->i_ib[level]) {
1487 			err = ufs_fault(ITOV(ip), "ufs_itrunc: indirect block");
1488 			return (err);
1489 		}
1490 
1491 	for (i = 0; i < NDADDR; i++)
1492 		if (ip->i_db[i] != oip->i_db[i]) {
1493 			err = ufs_fault(ITOV(ip), "ufs_itrunc: direct block");
1494 			return (err);
1495 		}
1496 /* END PARANOIA */
1497 	oip->i_blocks -= blocksreleased;
1498 
1499 	if (oip->i_blocks < 0) {		/* sanity */
1500 		cmn_err(CE_NOTE,
1501 		    "ufs_itrunc: %s/%d new size = %lld, blocks = %d\n",
1502 		    fs->fs_fsmnt, (int)oip->i_number, oip->i_size,
1503 		    (int)oip->i_blocks);
1504 		oip->i_blocks = 0;
1505 	}
1506 	oip->i_flag |= ICHG|IATTCHG;
1507 	oip->i_seq++;
1508 	/* blocksreleased is >= zero, so this can not fail */
1509 	(void) chkdq(oip, -blocksreleased, 0, cr, (char **)NULL,
1510 	    (size_t *)NULL);
1511 	return (0);
1512 }
1513 
1514 /*
1515  * Check mode permission on inode.  Mode is READ, WRITE or EXEC.
1516  * In the case of WRITE, the read-only status of the file system
1517  * is checked.  Depending on the calling user, the appropriate
1518  * mode bits are selected; privileges to override missing permission
1519  * bits are checked through secpolicy_vnode_access().
1520  * The i_contens lock must be held as reader here to prevent racing with
1521  * the acl subsystem removing/setting/changing acls on this inode.
1522  * The caller is responsible for indicating whether or not the i_contents
1523  * lock needs to be acquired here or if already held.
1524  */
1525 int
1526 ufs_iaccess(struct inode  *ip, int mode, struct cred *cr, int dolock)
1527 {
1528 	int shift = 0;
1529 	int ret = 0;
1530 
1531 	if (dolock)
1532 		rw_enter(&ip->i_contents, RW_READER);
1533 	ASSERT(RW_LOCK_HELD(&ip->i_contents));
1534 
1535 	if (mode & IWRITE) {
1536 		/*
1537 		 * Disallow write attempts on read-only
1538 		 * file systems, unless the file is a block
1539 		 * or character device or a FIFO.
1540 		 */
1541 		if (ip->i_fs->fs_ronly != 0) {
1542 			if ((ip->i_mode & IFMT) != IFCHR &&
1543 			    (ip->i_mode & IFMT) != IFBLK &&
1544 			    (ip->i_mode & IFMT) != IFIFO) {
1545 				ret = EROFS;
1546 				goto out;
1547 			}
1548 		}
1549 	}
1550 	/*
1551 	 * If there is an acl, check the acl and return.
1552 	 */
1553 	if (ip->i_ufs_acl && ip->i_ufs_acl->aowner) {
1554 		ret = ufs_acl_access(ip, mode, cr);
1555 		goto out;
1556 	}
1557 
1558 	/*
1559 	 * Access check is based on only one of owner, group, public.
1560 	 * If not owner, then check group.
1561 	 * If not a member of the group, then check public access.
1562 	 */
1563 	if (crgetuid(cr) != ip->i_uid) {
1564 		shift += 3;
1565 		if (!groupmember((uid_t)ip->i_gid, cr))
1566 			shift += 3;
1567 	}
1568 
1569 	mode &= ~(ip->i_mode << shift);
1570 
1571 	if (mode == 0)
1572 		goto out;
1573 
1574 	/* test missing privilege bits */
1575 	ret = secpolicy_vnode_access(cr, ITOV(ip), ip->i_uid, mode);
1576 out:
1577 	if (dolock)
1578 		rw_exit(&ip->i_contents);
1579 	return (ret);
1580 }
1581 
1582 /*
1583  * if necessary, remove an inode from the free list
1584  *	i_contents is held except at unmount
1585  *
1586  * Return 1 if the inode is taken off of the ufs_idle_q,
1587  * and the caller is expected to call VN_RELE.
1588  *
1589  * Return 0 otherwise.
1590  */
1591 int
1592 ufs_rmidle(struct inode *ip)
1593 {
1594 	int rval = 0;
1595 
1596 	mutex_enter(&ip->i_tlock);
1597 	if ((ip->i_flag & IREF) == 0) {
1598 		mutex_enter(&ufs_idle_q.uq_mutex);
1599 		ip->i_freef->i_freeb = ip->i_freeb;
1600 		ip->i_freeb->i_freef = ip->i_freef;
1601 		ip->i_freef = ip;
1602 		ip->i_freeb = ip;
1603 		ip->i_flag |= IREF;
1604 		ufs_idle_q.uq_ne--;
1605 		if (ip->i_flag & IJUNKIQ) {
1606 			ufs_njunk_iq--;
1607 			ip->i_flag &= ~IJUNKIQ;
1608 		} else {
1609 			ufs_nuseful_iq--;
1610 		}
1611 		mutex_exit(&ufs_idle_q.uq_mutex);
1612 		rval = 1;
1613 	}
1614 	mutex_exit(&ip->i_tlock);
1615 	return (rval);
1616 }
1617 
1618 /*
1619  * scan the hash of inodes and call func with the inode locked
1620  */
1621 int
1622 ufs_scan_inodes(int rwtry, int (*func)(struct inode *, void *), void *arg,
1623 		struct ufsvfs *ufsvfsp)
1624 {
1625 	struct inode		*ip;		/* current inode */
1626 	struct inode		*lip = NULL;	/* last/previous inode */
1627 	union ihead		*ih;		/* current hash chain */
1628 	int			error, i;
1629 	int			saverror = 0;
1630 	int			lip_held;	/* lip needs a VN_RELE() */
1631 
1632 	/*
1633 	 * If ufsvfsp is NULL, then our caller should be holding
1634 	 * ufs_scan_lock to avoid conflicts between ufs_unmount() and
1635 	 * ufs_update().  Otherwise, to avoid false-positives in
1636 	 * ufs_unmount()'s v_count-based EBUSY check, we only hold
1637 	 * those inodes that are in the file system our caller cares
1638 	 * about.
1639 	 *
1640 	 * We know that ip is a valid inode in the hash chain (and thus
1641 	 * we can trust i_ufsvfs) because the inode we chained from
1642 	 * (lip) is still in the hash chain.  This is true because either:
1643 	 *
1644 	 * 1. We did not drop the hash chain lock since the last
1645 	 *    iteration (because we were not interested in the last inode),
1646 	 * or
1647 	 * 2. We maintained a hold on the last inode while we
1648 	 *    we were processing it, so it could not be removed
1649 	 *    from the hash chain.
1650 	 *
1651 	 * The whole reason we're dropping and re-grabbing the chain
1652 	 * lock on every inode is so that we don't present a major
1653 	 * choke point on throughput, particularly when we've been
1654 	 * called on behalf of fsflush.
1655 	 */
1656 
1657 	for (i = 0, ih = ihead; i < inohsz; i++, ih++) {
1658 		mutex_enter(&ih_lock[i]);
1659 		for (ip = ih->ih_chain[0], lip_held = 0;
1660 		    ip != (struct inode *)ih;
1661 		    ip = lip->i_forw) {
1662 
1663 			ins.in_scan.value.ul++;
1664 
1665 			/*
1666 			 * Undo the previous iteration's VN_HOLD(), but
1667 			 * only if one was done.
1668 			 */
1669 			if (lip_held)
1670 				VN_RELE(ITOV(lip));
1671 
1672 			lip = ip;
1673 			if (ufsvfsp != NULL && ip->i_ufsvfs != ufsvfsp) {
1674 				/*
1675 				 * We're not processing all inodes, and
1676 				 * this inode is not in the filesystem of
1677 				 * interest, so skip it.  No need to do a
1678 				 * VN_HOLD() since we're not dropping the
1679 				 * hash chain lock until after we've
1680 				 * done the i_forw traversal above.
1681 				 */
1682 				lip_held = 0;
1683 				continue;
1684 			}
1685 			VN_HOLD(ITOV(ip));
1686 			lip_held = 1;
1687 			mutex_exit(&ih_lock[i]);
1688 
1689 			/*
1690 			 * Acquire the contents lock as writer to make
1691 			 * sure that the inode has been initialized in
1692 			 * the cache or removed from the idle list by
1693 			 * ufs_iget().  This works because ufs_iget()
1694 			 * acquires the contents lock before putting
1695 			 * the inode into the cache.  If we can lock
1696 			 * it, then he's done with it.
1697 			 */
1698 
1699 			if (rwtry) {
1700 				if (!rw_tryenter(&ip->i_contents, RW_WRITER)) {
1701 					mutex_enter(&ih_lock[i]);
1702 					continue;
1703 				}
1704 			} else {
1705 				rw_enter(&ip->i_contents, RW_WRITER);
1706 			}
1707 
1708 			rw_exit(&ip->i_contents);
1709 
1710 			/*
1711 			 * ISTALE means the inode couldn't be read
1712 			 *
1713 			 * We don't have to hold the i_contents lock
1714 			 * for this check for a couple of
1715 			 * reasons. First, if ISTALE is set then the
1716 			 * flag cannot be cleared until the inode is
1717 			 * removed from the cache and that cannot
1718 			 * happen until after we VN_RELE() it.
1719 			 * Second, if ISTALE is not set, then the
1720 			 * inode is in the cache and does not need to
1721 			 * be read from disk so ISTALE cannot be set
1722 			 * while we are not looking.
1723 			 */
1724 			if ((ip->i_flag & ISTALE) == 0) {
1725 				if ((error = (*func)(ip, arg)) != 0)
1726 					saverror = error;
1727 			}
1728 
1729 			mutex_enter(&ih_lock[i]);
1730 		}
1731 		if (lip_held)
1732 			VN_RELE(ITOV(lip));
1733 		mutex_exit(&ih_lock[i]);
1734 	}
1735 	return (saverror);
1736 }
1737 
1738 /*
1739  * Mark inode with the current time, plus a unique increment.
1740  *
1741  * Since we only keep 32-bit time on disk, if UFS is still alive
1742  * beyond 2038, filesystem times will simply stick at the last
1743  * possible second of 32-bit time. Not ideal, but probably better
1744  * than going into the remote past, or confusing applications with
1745  * negative time.
1746  */
1747 void
1748 ufs_imark(struct inode *ip)
1749 {
1750 	timestruc_t now;
1751 	int32_t usec, nsec;
1752 
1753 	/*
1754 	 * The update of i_seq may have been deferred, increase i_seq here
1755 	 * to make sure it is in sync with the timestamps.
1756 	 */
1757 	if (ip->i_flag & ISEQ) {
1758 		ASSERT(ip->i_flag & (IUPD|ICHG));
1759 		ip->i_seq++;
1760 		ip->i_flag &= ~ISEQ;
1761 	}
1762 
1763 	gethrestime(&now);
1764 
1765 	/*
1766 	 * Fast algorithm to convert nsec to usec -- see hrt2ts()
1767 	 * in common/os/timers.c for a full description.
1768 	 */
1769 	nsec = now.tv_nsec;
1770 	usec = nsec + (nsec >> 2);
1771 	usec = nsec + (usec >> 1);
1772 	usec = nsec + (usec >> 2);
1773 	usec = nsec + (usec >> 4);
1774 	usec = nsec - (usec >> 3);
1775 	usec = nsec + (usec >> 2);
1776 	usec = nsec + (usec >> 3);
1777 	usec = nsec + (usec >> 4);
1778 	usec = nsec + (usec >> 1);
1779 	usec = nsec + (usec >> 6);
1780 	usec = usec >> 10;
1781 
1782 	mutex_enter(&ufs_iuniqtime_lock);
1783 	if (now.tv_sec > (time_t)iuniqtime.tv_sec ||
1784 	    usec > iuniqtime.tv_usec) {
1785 		if (now.tv_sec < TIME32_MAX) {
1786 			iuniqtime.tv_sec = (time32_t)now.tv_sec;
1787 			iuniqtime.tv_usec = usec;
1788 		}
1789 	} else {
1790 		if (iuniqtime.tv_sec < TIME32_MAX) {
1791 			iuniqtime.tv_usec++;
1792 			/* Check for usec overflow */
1793 			if (iuniqtime.tv_usec >= MICROSEC) {
1794 				iuniqtime.tv_sec++;
1795 				iuniqtime.tv_usec = 0;
1796 			}
1797 		}
1798 	}
1799 
1800 	if ((ip->i_flag & IACC) && !(ip->i_ufsvfs->vfs_noatime)) {
1801 		ip->i_atime = iuniqtime;
1802 	}
1803 	if (ip->i_flag & IUPD) {
1804 		ip->i_mtime = iuniqtime;
1805 		ip->i_flag |= IMODTIME;
1806 	}
1807 	if (ip->i_flag & ICHG) {
1808 		ip->i_diroff = 0;
1809 		ip->i_ctime = iuniqtime;
1810 	}
1811 	mutex_exit(&ufs_iuniqtime_lock);
1812 }
1813 
1814 /*
1815  * Update timestamps in inode.
1816  */
1817 void
1818 ufs_itimes_nolock(struct inode *ip)
1819 {
1820 
1821 	/*
1822 	 * if noatime is set and the inode access time is the only field that
1823 	 * must be changed, exit immediately.
1824 	 */
1825 	if (((ip->i_flag & (IUPD|IACC|ICHG)) == IACC) &&
1826 	    (ip->i_ufsvfs->vfs_noatime)) {
1827 		return;
1828 	}
1829 
1830 	if (ip->i_flag & (IUPD|IACC|ICHG)) {
1831 		if (ip->i_flag & ICHG)
1832 			ip->i_flag |= IMOD;
1833 		else
1834 			ip->i_flag |= IMODACC;
1835 		ufs_imark(ip);
1836 		ip->i_flag &= ~(IACC|IUPD|ICHG);
1837 	}
1838 }
1839