xref: /titanic_50/usr/src/uts/common/fs/ufs/ufs_trans.c (revision c77a61a72b5ecdc507d6cf104142edd371a16c84)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License, Version 1.0 only
6  * (the "License").  You may not use this file except in compliance
7  * with the License.
8  *
9  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
10  * or http://www.opensolaris.org/os/licensing.
11  * See the License for the specific language governing permissions
12  * and limitations under the License.
13  *
14  * When distributing Covered Code, include this CDDL HEADER in each
15  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
16  * If applicable, add the following below this CDDL HEADER, with the
17  * fields enclosed by brackets "[]" replaced with your own identifying
18  * information: Portions Copyright [yyyy] [name of copyright owner]
19  *
20  * CDDL HEADER END
21  */
22 /*
23  * Copyright 2005 Sun Microsystems, Inc.  All rights reserved.
24  * Use is subject to license terms.
25  */
26 
27 /* Copyright (c) 1983, 1984, 1985, 1986, 1987, 1988, 1989 AT&T */
28 /* All Rights Reserved */
29 
30 /*
31  * Portions of this source code were derived from Berkeley 4.3 BSD
32  * under license from the Regents of the University of California.
33  */
34 
35 #pragma ident	"%Z%%M%	%I%	%E% SMI"
36 
37 #include <sys/sysmacros.h>
38 #include <sys/param.h>
39 #include <sys/types.h>
40 #include <sys/systm.h>
41 #include <sys/t_lock.h>
42 #include <sys/uio.h>
43 #include <sys/kmem.h>
44 #include <sys/thread.h>
45 #include <sys/vfs.h>
46 #include <sys/errno.h>
47 #include <sys/buf.h>
48 #include <sys/vnode.h>
49 #include <sys/fs/ufs_trans.h>
50 #include <sys/fs/ufs_inode.h>
51 #include <sys/fs/ufs_fs.h>
52 #include <sys/fs/ufs_fsdir.h>
53 #include <sys/fs/ufs_quota.h>
54 #include <sys/fs/ufs_panic.h>
55 #include <sys/fs/ufs_bio.h>
56 #include <sys/fs/ufs_log.h>
57 #include <sys/cmn_err.h>
58 #include <sys/file.h>
59 #include <sys/debug.h>
60 
61 
62 extern kmutex_t ufsvfs_mutex;
63 extern struct ufsvfs *ufs_instances;
64 
65 /*
66  * hlock any file systems w/errored logs
67  */
68 int
69 ufs_trans_hlock()
70 {
71 	struct ufsvfs	*ufsvfsp;
72 	struct lockfs	lockfs;
73 	int		error;
74 	int		retry	= 0;
75 
76 	/*
77 	 * find fs's that paniced or have errored logging devices
78 	 */
79 	mutex_enter(&ufsvfs_mutex);
80 	for (ufsvfsp = ufs_instances; ufsvfsp; ufsvfsp = ufsvfsp->vfs_next) {
81 		/*
82 		 * not mounted; continue
83 		 */
84 		if ((ufsvfsp->vfs_vfs == NULL) ||
85 		    (ufsvfsp->vfs_validfs == UT_UNMOUNTED))
86 			continue;
87 		/*
88 		 * disallow unmounts (hlock occurs below)
89 		 */
90 		if (TRANS_ISERROR(ufsvfsp))
91 			ufsvfsp->vfs_validfs = UT_HLOCKING;
92 	}
93 	mutex_exit(&ufsvfs_mutex);
94 
95 	/*
96 	 * hlock the fs's that paniced or have errored logging devices
97 	 */
98 again:
99 	mutex_enter(&ufsvfs_mutex);
100 	for (ufsvfsp = ufs_instances; ufsvfsp; ufsvfsp = ufsvfsp->vfs_next)
101 		if (ufsvfsp->vfs_validfs == UT_HLOCKING)
102 			break;
103 	mutex_exit(&ufsvfs_mutex);
104 	if (ufsvfsp == NULL)
105 		return (retry);
106 	/*
107 	 * hlock the file system
108 	 */
109 	(void) ufs_fiolfss(ufsvfsp->vfs_root, &lockfs);
110 	if (!LOCKFS_IS_ELOCK(&lockfs)) {
111 		lockfs.lf_lock = LOCKFS_HLOCK;
112 		lockfs.lf_flags = 0;
113 		lockfs.lf_comlen = 0;
114 		lockfs.lf_comment = NULL;
115 		error = ufs_fiolfs(ufsvfsp->vfs_root, &lockfs, 0);
116 		/*
117 		 * retry after awhile; another app currently doing lockfs
118 		 */
119 		if (error == EBUSY || error == EINVAL)
120 			retry = 1;
121 	} else {
122 		if (ufsfx_get_failure_qlen() > 0) {
123 			if (mutex_tryenter(&ufs_fix.uq_mutex)) {
124 				ufs_fix.uq_lowat = ufs_fix.uq_ne;
125 				cv_broadcast(&ufs_fix.uq_cv);
126 				mutex_exit(&ufs_fix.uq_mutex);
127 			}
128 		}
129 		retry = 1;
130 	}
131 
132 	/*
133 	 * allow unmounts
134 	 */
135 	ufsvfsp->vfs_validfs = UT_MOUNTED;
136 	goto again;
137 }
138 
139 /*ARGSUSED*/
140 void
141 ufs_trans_onerror()
142 {
143 	mutex_enter(&ufs_hlock.uq_mutex);
144 	ufs_hlock.uq_ne = ufs_hlock.uq_lowat;
145 	cv_broadcast(&ufs_hlock.uq_cv);
146 	mutex_exit(&ufs_hlock.uq_mutex);
147 }
148 
149 void
150 ufs_trans_sbupdate(struct ufsvfs *ufsvfsp, struct vfs *vfsp, top_t topid)
151 {
152 	if (curthread->t_flag & T_DONTBLOCK) {
153 		sbupdate(vfsp);
154 		return;
155 	} else {
156 
157 		if (panicstr && TRANS_ISTRANS(ufsvfsp))
158 			return;
159 
160 		curthread->t_flag |= T_DONTBLOCK;
161 		TRANS_BEGIN_ASYNC(ufsvfsp, topid, TOP_SBUPDATE_SIZE);
162 		sbupdate(vfsp);
163 		TRANS_END_ASYNC(ufsvfsp, topid, TOP_SBUPDATE_SIZE);
164 		curthread->t_flag &= ~T_DONTBLOCK;
165 	}
166 }
167 
168 void
169 ufs_trans_iupdat(struct inode *ip, int waitfor)
170 {
171 	struct ufsvfs	*ufsvfsp;
172 
173 	if (curthread->t_flag & T_DONTBLOCK) {
174 		rw_enter(&ip->i_contents, RW_READER);
175 		ufs_iupdat(ip, waitfor);
176 		rw_exit(&ip->i_contents);
177 		return;
178 	} else {
179 		ufsvfsp = ip->i_ufsvfs;
180 
181 		if (panicstr && TRANS_ISTRANS(ufsvfsp))
182 			return;
183 
184 		curthread->t_flag |= T_DONTBLOCK;
185 		TRANS_BEGIN_ASYNC(ufsvfsp, TOP_IUPDAT, TOP_IUPDAT_SIZE(ip));
186 		rw_enter(&ip->i_contents, RW_READER);
187 		ufs_iupdat(ip, waitfor);
188 		rw_exit(&ip->i_contents);
189 		TRANS_END_ASYNC(ufsvfsp, TOP_IUPDAT, TOP_IUPDAT_SIZE(ip));
190 		curthread->t_flag &= ~T_DONTBLOCK;
191 	}
192 }
193 
194 void
195 ufs_trans_sbwrite(struct ufsvfs *ufsvfsp, top_t topid)
196 {
197 	if (curthread->t_flag & T_DONTBLOCK) {
198 		mutex_enter(&ufsvfsp->vfs_lock);
199 		ufs_sbwrite(ufsvfsp);
200 		mutex_exit(&ufsvfsp->vfs_lock);
201 		return;
202 	} else {
203 
204 		if (panicstr && TRANS_ISTRANS(ufsvfsp))
205 			return;
206 
207 		curthread->t_flag |= T_DONTBLOCK;
208 		TRANS_BEGIN_ASYNC(ufsvfsp, topid, TOP_SBWRITE_SIZE);
209 		mutex_enter(&ufsvfsp->vfs_lock);
210 		ufs_sbwrite(ufsvfsp);
211 		mutex_exit(&ufsvfsp->vfs_lock);
212 		TRANS_END_ASYNC(ufsvfsp, topid, TOP_SBWRITE_SIZE);
213 		curthread->t_flag &= ~T_DONTBLOCK;
214 	}
215 }
216 
217 /*ARGSUSED*/
218 int
219 ufs_trans_push_si(ufsvfs_t *ufsvfsp, delta_t dtyp, int ignore)
220 {
221 	struct fs	*fs;
222 
223 	fs = ufsvfsp->vfs_fs;
224 	mutex_enter(&ufsvfsp->vfs_lock);
225 	TRANS_LOG(ufsvfsp, (char *)fs->fs_u.fs_csp,
226 		ldbtob(fsbtodb(fs, fs->fs_csaddr)), fs->fs_cssize,
227 		(caddr_t)fs->fs_u.fs_csp, fs->fs_cssize);
228 	mutex_exit(&ufsvfsp->vfs_lock);
229 	return (0);
230 }
231 
232 /*ARGSUSED*/
233 int
234 ufs_trans_push_buf(ufsvfs_t *ufsvfsp, delta_t dtyp, daddr_t bno)
235 {
236 	struct buf	*bp;
237 
238 	bp = (struct buf *)UFS_GETBLK(ufsvfsp, ufsvfsp->vfs_dev, bno, 1);
239 	if (bp == NULL)
240 		return (ENOENT);
241 
242 	if (bp->b_flags & B_DELWRI) {
243 		/*
244 		 * Do not use brwrite() here since the buffer is already
245 		 * marked for retry or not by the code that called
246 		 * TRANS_BUF().
247 		 */
248 		UFS_BWRITE(ufsvfsp, bp);
249 		return (0);
250 	}
251 	/*
252 	 * If we did not find the real buf for this block above then
253 	 * clear the dev so the buf won't be found by mistake
254 	 * for this block later.  We had to allocate at least a 1 byte
255 	 * buffer to keep brelse happy.
256 	 */
257 	if (bp->b_bufsize == 1) {
258 		bp->b_dev = (o_dev_t)NODEV;
259 		bp->b_edev = NODEV;
260 		bp->b_flags = 0;
261 	}
262 	brelse(bp);
263 	return (ENOENT);
264 }
265 
266 /*ARGSUSED*/
267 int
268 ufs_trans_push_inode(ufsvfs_t *ufsvfsp, delta_t dtyp, ino_t ino)
269 {
270 	int		error;
271 	struct inode	*ip;
272 
273 	/*
274 	 * Grab the quota lock (if the file system has not been forcibly
275 	 * unmounted).
276 	 */
277 	if (ufsvfsp)
278 		rw_enter(&ufsvfsp->vfs_dqrwlock, RW_READER);
279 
280 	error = ufs_iget(ufsvfsp->vfs_vfs, ino, &ip, kcred);
281 
282 	if (ufsvfsp)
283 		rw_exit(&ufsvfsp->vfs_dqrwlock);
284 	if (error)
285 		return (ENOENT);
286 
287 	if (ip->i_flag & (IUPD|IACC|ICHG|IMOD|IMODACC|IATTCHG)) {
288 		rw_enter(&ip->i_contents, RW_READER);
289 		ufs_iupdat(ip, 1);
290 		rw_exit(&ip->i_contents);
291 		VN_RELE(ITOV(ip));
292 		return (0);
293 	}
294 	VN_RELE(ITOV(ip));
295 	return (ENOENT);
296 }
297 
298 #ifdef DEBUG
299 /*
300  *	These routines maintain the metadata map (matamap)
301  */
302 
303 /*
304  * update the metadata map at mount
305  */
306 static int
307 ufs_trans_mata_mount_scan(struct inode *ip, void *arg)
308 {
309 	/*
310 	 * wrong file system; keep looking
311 	 */
312 	if (ip->i_ufsvfs != (struct ufsvfs *)arg)
313 		return (0);
314 
315 	/*
316 	 * load the metadata map
317 	 */
318 	rw_enter(&ip->i_contents, RW_WRITER);
319 	ufs_trans_mata_iget(ip);
320 	rw_exit(&ip->i_contents);
321 	return (0);
322 }
323 
324 void
325 ufs_trans_mata_mount(struct ufsvfs *ufsvfsp)
326 {
327 	struct fs	*fs	= ufsvfsp->vfs_fs;
328 	ino_t		ino;
329 	int		i;
330 
331 	/*
332 	 * put static metadata into matamap
333 	 *	superblock
334 	 *	cylinder groups
335 	 *	inode groups
336 	 *	existing inodes
337 	 */
338 	TRANS_MATAADD(ufsvfsp, ldbtob(SBLOCK), fs->fs_sbsize);
339 
340 	for (ino = i = 0; i < fs->fs_ncg; ++i, ino += fs->fs_ipg) {
341 		TRANS_MATAADD(ufsvfsp,
342 		    ldbtob(fsbtodb(fs, cgtod(fs, i))), fs->fs_cgsize);
343 		TRANS_MATAADD(ufsvfsp,
344 		    ldbtob(fsbtodb(fs, itod(fs, ino))),
345 		    fs->fs_ipg * sizeof (struct dinode));
346 	}
347 	(void) ufs_scan_inodes(0, ufs_trans_mata_mount_scan, ufsvfsp, ufsvfsp);
348 }
349 
350 /*
351  * clear the metadata map at umount
352  */
353 void
354 ufs_trans_mata_umount(struct ufsvfs *ufsvfsp)
355 {
356 	top_mataclr(ufsvfsp);
357 }
358 
359 /*
360  * summary info (may be extended during growfs test)
361  */
362 void
363 ufs_trans_mata_si(struct ufsvfs *ufsvfsp, struct fs *fs)
364 {
365 	TRANS_MATAADD(ufsvfsp, ldbtob(fsbtodb(fs, fs->fs_csaddr)),
366 			fs->fs_cssize);
367 }
368 
369 /*
370  * scan an allocation block (either inode or true block)
371  */
372 static void
373 ufs_trans_mata_direct(
374 	struct inode *ip,
375 	daddr_t *fragsp,
376 	daddr32_t *blkp,
377 	unsigned int nblk)
378 {
379 	int		i;
380 	daddr_t		frag;
381 	ulong_t		nb;
382 	struct ufsvfs	*ufsvfsp	= ip->i_ufsvfs;
383 	struct fs	*fs		= ufsvfsp->vfs_fs;
384 
385 	for (i = 0; i < nblk && *fragsp; ++i, ++blkp)
386 		if ((frag = *blkp) != 0) {
387 			if (*fragsp > fs->fs_frag) {
388 				nb = fs->fs_bsize;
389 				*fragsp -= fs->fs_frag;
390 			} else {
391 				nb = *fragsp * fs->fs_fsize;
392 				*fragsp = 0;
393 			}
394 			TRANS_MATAADD(ufsvfsp, ldbtob(fsbtodb(fs, frag)), nb);
395 		}
396 }
397 
398 /*
399  * scan an indirect allocation block (either inode or true block)
400  */
401 static void
402 ufs_trans_mata_indir(
403 	struct inode *ip,
404 	daddr_t *fragsp,
405 	daddr_t frag,
406 	int level)
407 {
408 	struct ufsvfs *ufsvfsp	= ip->i_ufsvfs;
409 	struct fs *fs = ufsvfsp->vfs_fs;
410 	int ne = fs->fs_bsize / (int)sizeof (daddr32_t);
411 	int i;
412 	struct buf *bp;
413 	daddr32_t *blkp;
414 	o_mode_t ifmt = ip->i_mode & IFMT;
415 
416 	bp = UFS_BREAD(ufsvfsp, ip->i_dev, fsbtodb(fs, frag), fs->fs_bsize);
417 	if (bp->b_flags & B_ERROR) {
418 		brelse(bp);
419 		return;
420 	}
421 	blkp = bp->b_un.b_daddr;
422 
423 	if (level || (ifmt == IFDIR) || (ifmt == IFSHAD) ||
424 	    (ifmt == IFATTRDIR) || (ip == ip->i_ufsvfs->vfs_qinod))
425 		ufs_trans_mata_direct(ip, fragsp, blkp, ne);
426 
427 	if (level)
428 		for (i = 0; i < ne && *fragsp; ++i, ++blkp)
429 			ufs_trans_mata_indir(ip, fragsp, *blkp, level-1);
430 	brelse(bp);
431 }
432 
433 /*
434  * put appropriate metadata into matamap for this inode
435  */
436 void
437 ufs_trans_mata_iget(struct inode *ip)
438 {
439 	int		i;
440 	daddr_t		frags	= dbtofsb(ip->i_fs, ip->i_blocks);
441 	o_mode_t	ifmt 	= ip->i_mode & IFMT;
442 
443 	if (frags && ((ifmt == IFDIR) || (ifmt == IFSHAD) ||
444 	    (ifmt == IFATTRDIR) || (ip == ip->i_ufsvfs->vfs_qinod)))
445 		ufs_trans_mata_direct(ip, &frags, &ip->i_db[0], NDADDR);
446 
447 	if (frags)
448 		ufs_trans_mata_direct(ip, &frags, &ip->i_ib[0], NIADDR);
449 
450 	for (i = 0; i < NIADDR && frags; ++i)
451 		if (ip->i_ib[i])
452 			ufs_trans_mata_indir(ip, &frags, ip->i_ib[i], i);
453 }
454 
455 /*
456  * freeing possible metadata (block of user data)
457  */
458 void
459 ufs_trans_mata_free(struct ufsvfs *ufsvfsp, offset_t mof, off_t nb)
460 {
461 	top_matadel(ufsvfsp, mof, nb);
462 
463 }
464 
465 /*
466  * allocating metadata
467  */
468 void
469 ufs_trans_mata_alloc(
470 	struct ufsvfs *ufsvfsp,
471 	struct inode *ip,
472 	daddr_t frag,
473 	ulong_t nb,
474 	int indir)
475 {
476 	struct fs	*fs	= ufsvfsp->vfs_fs;
477 	o_mode_t	ifmt 	= ip->i_mode & IFMT;
478 
479 	if (indir || ((ifmt == IFDIR) || (ifmt == IFSHAD) ||
480 	    (ifmt == IFATTRDIR) || (ip == ip->i_ufsvfs->vfs_qinod)))
481 		TRANS_MATAADD(ufsvfsp, ldbtob(fsbtodb(fs, frag)), nb);
482 }
483 
484 #endif /* DEBUG */
485 
486 /*
487  * ufs_trans_dir is used to declare a directory delta
488  */
489 int
490 ufs_trans_dir(struct inode *ip, off_t offset)
491 {
492 	daddr_t	bn;
493 	int	contig = 0, error;
494 
495 	ASSERT(ip);
496 	ASSERT(RW_WRITE_HELD(&ip->i_contents));
497 	error = bmap_read(ip, (u_offset_t)offset, &bn, &contig);
498 	if (error || (bn == UFS_HOLE)) {
499 		cmn_err(CE_WARN, "ufs_trans_dir - could not get block"
500 		    " number error = %d bn = %d\n", error, (int)bn);
501 		if (error == 0)	/* treat UFS_HOLE as an I/O error */
502 			error = EIO;
503 		return (error);
504 	}
505 	TRANS_DELTA(ip->i_ufsvfs, ldbtob(bn), DIRBLKSIZ, DT_DIR, 0, 0);
506 	return (error);
507 }
508 
509 /*ARGSUSED*/
510 int
511 ufs_trans_push_quota(ufsvfs_t *ufsvfsp, delta_t dtyp, struct dquot *dqp)
512 {
513 	/*
514 	 * Lock the quota subsystem (ufsvfsp can be NULL
515 	 * if the DQ_ERROR is set).
516 	 */
517 	if (ufsvfsp)
518 		rw_enter(&ufsvfsp->vfs_dqrwlock, RW_READER);
519 	mutex_enter(&dqp->dq_lock);
520 
521 	/*
522 	 * If this transaction has been cancelled by closedq_scan_inode(),
523 	 * then bail out now.  We don't call dqput() in this case because
524 	 * it has already been done.
525 	 */
526 	if ((dqp->dq_flags & DQ_TRANS) == 0) {
527 		mutex_exit(&dqp->dq_lock);
528 		if (ufsvfsp)
529 			rw_exit(&ufsvfsp->vfs_dqrwlock);
530 		return (0);
531 	}
532 
533 	if (dqp->dq_flags & DQ_ERROR) {
534 		/*
535 		 * Paranoia to make sure that there is at least one
536 		 * reference to the dquot struct.  We are done with
537 		 * the dquot (due to an error) so clear logging
538 		 * specific markers.
539 		 */
540 		ASSERT(dqp->dq_cnt >= 1);
541 		dqp->dq_flags &= ~DQ_TRANS;
542 		dqput(dqp);
543 		mutex_exit(&dqp->dq_lock);
544 		if (ufsvfsp)
545 			rw_exit(&ufsvfsp->vfs_dqrwlock);
546 		return (1);
547 	}
548 
549 	if (dqp->dq_flags & (DQ_MOD | DQ_BLKS | DQ_FILES)) {
550 		ASSERT((dqp->dq_mof != UFS_HOLE) && (dqp->dq_mof != 0));
551 		TRANS_LOG(ufsvfsp, (caddr_t)&dqp->dq_dqb,
552 		    dqp->dq_mof, (int)sizeof (struct dqblk), NULL, 0);
553 		/*
554 		 * Paranoia to make sure that there is at least one
555 		 * reference to the dquot struct.  Clear the
556 		 * modification flag because the operation is now in
557 		 * the log.  Also clear the logging specific markers
558 		 * that were set in ufs_trans_quota().
559 		 */
560 		ASSERT(dqp->dq_cnt >= 1);
561 		dqp->dq_flags &= ~(DQ_MOD | DQ_TRANS);
562 		dqput(dqp);
563 	}
564 
565 	/*
566 	 * At this point, the logging specific flag should be clear,
567 	 * but add paranoia just in case something has gone wrong.
568 	 */
569 	ASSERT((dqp->dq_flags & DQ_TRANS) == 0);
570 	mutex_exit(&dqp->dq_lock);
571 	if (ufsvfsp)
572 		rw_exit(&ufsvfsp->vfs_dqrwlock);
573 	return (0);
574 }
575 
576 /*
577  * ufs_trans_quota take in a uid, allocates the disk space, placing the
578  * quota record into the metamap, then declares the delta.
579  */
580 /*ARGSUSED*/
581 void
582 ufs_trans_quota(struct dquot *dqp)
583 {
584 
585 	struct inode	*qip = dqp->dq_ufsvfsp->vfs_qinod;
586 
587 	ASSERT(qip);
588 	ASSERT(MUTEX_HELD(&dqp->dq_lock));
589 	ASSERT(dqp->dq_flags & DQ_MOD);
590 	ASSERT(dqp->dq_mof != 0);
591 	ASSERT(dqp->dq_mof != UFS_HOLE);
592 
593 	/*
594 	 * Mark this dquot to indicate that we are starting a logging
595 	 * file system operation for this dquot.  Also increment the
596 	 * reference count so that the dquot does not get reused while
597 	 * it is on the mapentry_t list.  DQ_TRANS is cleared and the
598 	 * reference count is decremented by ufs_trans_push_quota.
599 	 *
600 	 * If the file system is force-unmounted while there is a
601 	 * pending quota transaction, then closedq_scan_inode() will
602 	 * clear the DQ_TRANS flag and decrement the reference count.
603 	 *
604 	 * Since deltamap_add() drops multiple transactions to the
605 	 * same dq_mof and ufs_trans_push_quota() won't get called,
606 	 * we use DQ_TRANS to prevent repeat transactions from
607 	 * incrementing the reference count (or calling TRANS_DELTA()).
608 	 */
609 	if ((dqp->dq_flags & DQ_TRANS) == 0) {
610 		dqp->dq_flags |= DQ_TRANS;
611 		dqp->dq_cnt++;
612 		TRANS_DELTA(qip->i_ufsvfs, dqp->dq_mof, sizeof (struct dqblk),
613 		    DT_QR, ufs_trans_push_quota, (ulong_t)dqp);
614 	}
615 }
616 
617 void
618 ufs_trans_dqrele(struct dquot *dqp)
619 {
620 	struct ufsvfs	*ufsvfsp = dqp->dq_ufsvfsp;
621 
622 	curthread->t_flag |= T_DONTBLOCK;
623 	TRANS_BEGIN_ASYNC(ufsvfsp, TOP_QUOTA, TOP_QUOTA_SIZE);
624 	rw_enter(&ufsvfsp->vfs_dqrwlock, RW_READER);
625 	dqrele(dqp);
626 	rw_exit(&ufsvfsp->vfs_dqrwlock);
627 	TRANS_END_ASYNC(ufsvfsp, TOP_QUOTA, TOP_QUOTA_SIZE);
628 	curthread->t_flag &= ~T_DONTBLOCK;
629 }
630 
631 int ufs_trans_max_resv = TOP_MAX_RESV;	/* will be adjusted for testing */
632 long ufs_trans_avgbfree = 0;		/* will be adjusted for testing */
633 #define	TRANS_MAX_WRITE	(1024 * 1024)
634 size_t ufs_trans_max_resid = TRANS_MAX_WRITE;
635 
636 /*
637  * Calculate the log reservation for the given write or truncate
638  */
639 static ulong_t
640 ufs_log_amt(struct inode *ip, offset_t offset, ssize_t resid, int trunc)
641 {
642 	long		ncg, last2blk;
643 	long		niblk		= 0;
644 	u_offset_t	writeend, offblk;
645 	int		resv;
646 	daddr_t		nblk, maxfblk;
647 	long		avgbfree;
648 	struct ufsvfs	*ufsvfsp	= ip->i_ufsvfs;
649 	struct fs	*fs		= ufsvfsp->vfs_fs;
650 	long		fni		= NINDIR(fs);
651 	int		bsize		= fs->fs_bsize;
652 
653 	/*
654 	 * Assume that the request will fit in 1 or 2 cg's,
655 	 * resv is the amount of log space to reserve (in bytes).
656 	 */
657 	resv = SIZECG(ip) * 2 + INODESIZE + 1024;
658 
659 	/*
660 	 * get max position of write in fs blocks
661 	 */
662 	writeend = offset + resid;
663 	maxfblk = lblkno(fs, writeend);
664 	offblk = lblkno(fs, offset);
665 	/*
666 	 * request size in fs blocks
667 	 */
668 	nblk = lblkno(fs, blkroundup(fs, resid));
669 	/*
670 	 * Adjust for sparse files
671 	 */
672 	if (trunc)
673 		nblk = MIN(nblk, ip->i_blocks);
674 
675 	/*
676 	 * Adjust avgbfree (for testing)
677 	 */
678 	avgbfree = (ufs_trans_avgbfree) ? 1 : ufsvfsp->vfs_avgbfree + 1;
679 
680 	/*
681 	 * Calculate maximum number of blocks of triple indirect
682 	 * pointers to write.
683 	 */
684 	last2blk = NDADDR + fni + fni * fni;
685 	if (maxfblk > last2blk) {
686 		long nl2ptr;
687 		long n3blk;
688 
689 		if (offblk > last2blk)
690 			n3blk = maxfblk - offblk;
691 		else
692 			n3blk = maxfblk - last2blk;
693 		niblk += roundup(n3blk * sizeof (daddr_t), bsize) / bsize + 1;
694 		nl2ptr = roundup(niblk, fni) / fni + 1;
695 		niblk += roundup(nl2ptr * sizeof (daddr_t), bsize) / bsize + 2;
696 		maxfblk -= n3blk;
697 	}
698 	/*
699 	 * calculate maximum number of blocks of double indirect
700 	 * pointers to write.
701 	 */
702 	if (maxfblk > NDADDR + fni) {
703 		long n2blk;
704 
705 		if (offblk > NDADDR + fni)
706 			n2blk = maxfblk - offblk;
707 		else
708 			n2blk = maxfblk - NDADDR + fni;
709 		niblk += roundup(n2blk * sizeof (daddr_t), bsize) / bsize + 2;
710 		maxfblk -= n2blk;
711 	}
712 	/*
713 	 * Add in indirect pointer block write
714 	 */
715 	if (maxfblk > NDADDR) {
716 		niblk += 1;
717 	}
718 	/*
719 	 * Calculate deltas for indirect pointer writes
720 	 */
721 	resv += niblk * (fs->fs_bsize + sizeof (struct delta));
722 	/*
723 	 * maximum number of cg's needed for request
724 	 */
725 	ncg = nblk / avgbfree;
726 	if (ncg > fs->fs_ncg)
727 		ncg = fs->fs_ncg;
728 
729 	/*
730 	 * maximum amount of log space needed for request
731 	 */
732 	if (ncg > 2)
733 		resv += (ncg - 2) * SIZECG(ip);
734 
735 	return (resv);
736 }
737 
738 /*
739  * Calculate the amount of log space that needs to be reserved for this
740  * trunc request.  If the amount of log space is too large, then
741  * calculate the the size that the requests needs to be split into.
742  */
743 void
744 ufs_trans_trunc_resv(
745 	struct inode *ip,
746 	u_offset_t length,
747 	int *resvp,
748 	u_offset_t *residp)
749 {
750 	ulong_t		resv;
751 	u_offset_t	size, offset, resid;
752 	int		nchunks;
753 
754 	/*
755 	 *    *resvp is the amount of log space to reserve (in bytes).
756 	 *    when nonzero, *residp is the number of bytes to truncate.
757 	 */
758 	*residp = 0;
759 
760 	if (length < ip->i_size) {
761 		size = ip->i_size - length;
762 	} else {
763 		resv = SIZECG(ip) * 2 + INODESIZE + 1024;
764 		/*
765 		 * truncate up, doesn't really use much space,
766 		 * the default above should be sufficient.
767 		 */
768 		goto done;
769 	}
770 
771 	offset = length;
772 	resid = size;
773 	nchunks = 1;
774 	for (; (resv = ufs_log_amt(ip, offset, resid, 1)) > ufs_trans_max_resv;
775 		offset = length + (nchunks - 1) * resid) {
776 		nchunks++;
777 		resid = size / nchunks;
778 	}
779 	/*
780 	 * If this request takes too much log space, it will be split
781 	 */
782 	if (nchunks > 1) {
783 		*residp = resid;
784 	}
785 done:
786 	*resvp = resv;
787 }
788 
789 int
790 ufs_trans_itrunc(struct inode *ip, u_offset_t length, int flags, cred_t *cr)
791 {
792 	int 		err, issync, resv;
793 	u_offset_t	resid;
794 	int		do_block	= 0;
795 	struct ufsvfs	*ufsvfsp	= ip->i_ufsvfs;
796 	struct fs	*fs		= ufsvfsp->vfs_fs;
797 
798 	/*
799 	 * Not logging; just do the trunc
800 	 */
801 	if (!TRANS_ISTRANS(ufsvfsp)) {
802 		rw_enter(&ufsvfsp->vfs_dqrwlock, RW_READER);
803 		rw_enter(&ip->i_contents, RW_WRITER);
804 		err = ufs_itrunc(ip, length, flags, cr);
805 		rw_exit(&ip->i_contents);
806 		rw_exit(&ufsvfsp->vfs_dqrwlock);
807 		return (err);
808 	}
809 
810 	/*
811 	 * within the lockfs protocol but *not* part of a transaction
812 	 */
813 	do_block = curthread->t_flag & T_DONTBLOCK;
814 	curthread->t_flag |= T_DONTBLOCK;
815 
816 	/*
817 	 * Trunc the file (in pieces, if necessary)
818 	 */
819 again:
820 	ufs_trans_trunc_resv(ip, length, &resv, &resid);
821 	TRANS_BEGIN_CSYNC(ufsvfsp, issync, TOP_ITRUNC, resv);
822 	rw_enter(&ufsvfsp->vfs_dqrwlock, RW_READER);
823 	rw_enter(&ip->i_contents, RW_WRITER);
824 	if (resid) {
825 		/*
826 		 * resid is only set if we have to truncate in chunks
827 		 */
828 		ASSERT(length + resid < ip->i_size);
829 
830 		/*
831 		 * Partially trunc file down to desired size (length).
832 		 * Only retain I_FREE on the last partial trunc.
833 		 * Round up size to a block boundary, to ensure the truncate
834 		 * doesn't have to allocate blocks. This is done both for
835 		 * performance and to fix a bug where if the block can't be
836 		 * allocated then the inode delete fails, but the inode
837 		 * is still freed with attached blocks and non-zero size
838 		 * (bug 4348738).
839 		 */
840 		err = ufs_itrunc(ip, blkroundup(fs, (ip->i_size - resid)),
841 		    flags & ~I_FREE, cr);
842 		ASSERT(ip->i_size != length);
843 	} else
844 		err = ufs_itrunc(ip, length, flags, cr);
845 	if (!do_block)
846 		curthread->t_flag &= ~T_DONTBLOCK;
847 	rw_exit(&ip->i_contents);
848 	rw_exit(&ufsvfsp->vfs_dqrwlock);
849 	TRANS_END_CSYNC(ufsvfsp, err, issync, TOP_ITRUNC, resv);
850 
851 	if ((err == 0) && resid) {
852 		ufsvfsp->vfs_avgbfree = fs->fs_cstotal.cs_nbfree / fs->fs_ncg;
853 		goto again;
854 	}
855 	return (err);
856 }
857 
858 /*
859  * Fault in the pages of the first n bytes specified by the uio structure.
860  * 1 byte in each page is touched and the uio struct is unmodified.
861  * Any error will terminate the process as this is only a best
862  * attempt to get the pages resident.
863  */
864 static void
865 ufs_trans_touch(ssize_t n, struct uio *uio)
866 {
867 	struct iovec *iov;
868 	ulong_t cnt, incr;
869 	caddr_t p;
870 	uint8_t tmp;
871 
872 	iov = uio->uio_iov;
873 
874 	while (n) {
875 		cnt = MIN(iov->iov_len, n);
876 		if (cnt == 0) {
877 			/* empty iov entry */
878 			iov++;
879 			continue;
880 		}
881 		n -= cnt;
882 		/*
883 		 * touch each page in this segment.
884 		 */
885 		p = iov->iov_base;
886 		while (cnt) {
887 			switch (uio->uio_segflg) {
888 			case UIO_USERSPACE:
889 			case UIO_USERISPACE:
890 				if (fuword8(p, &tmp))
891 					return;
892 				break;
893 			case UIO_SYSSPACE:
894 				if (kcopy(p, &tmp, 1))
895 					return;
896 				break;
897 			}
898 			incr = MIN(cnt, PAGESIZE);
899 			p += incr;
900 			cnt -= incr;
901 		}
902 		/*
903 		 * touch the last byte in case it straddles a page.
904 		 */
905 		p--;
906 		switch (uio->uio_segflg) {
907 		case UIO_USERSPACE:
908 		case UIO_USERISPACE:
909 			if (fuword8(p, &tmp))
910 				return;
911 			break;
912 		case UIO_SYSSPACE:
913 			if (kcopy(p, &tmp, 1))
914 				return;
915 			break;
916 		}
917 		iov++;
918 	}
919 }
920 
921 /*
922  * Calculate the amount of log space that needs to be reserved for this
923  * write request.  If the amount of log space is too large, then
924  * calculate the size that the requests needs to be split into.
925  * First try fixed chunks of size ufs_trans_max_resid. If that
926  * is too big, iterate down to the largest size that will fit.
927  * Pagein the pages in the first chunk here, so that the pagein is
928  * avoided later when the transaction is open.
929  */
930 void
931 ufs_trans_write_resv(
932 	struct inode *ip,
933 	struct uio *uio,
934 	int *resvp,
935 	int *residp)
936 {
937 	ulong_t		resv;
938 	offset_t	offset;
939 	ssize_t		resid;
940 	int		nchunks;
941 
942 	*residp = 0;
943 	offset = uio->uio_offset;
944 	resid = MIN(uio->uio_resid, ufs_trans_max_resid);
945 	resv = ufs_log_amt(ip, offset, resid, 0);
946 	if (resv <= ufs_trans_max_resv) {
947 		ufs_trans_touch(resid, uio);
948 		if (resid != uio->uio_resid)
949 			*residp = resid;
950 		*resvp = resv;
951 		return;
952 	}
953 
954 	resid = uio->uio_resid;
955 	nchunks = 1;
956 	for (; (resv = ufs_log_amt(ip, offset, resid, 0)) > ufs_trans_max_resv;
957 		offset = uio->uio_offset + (nchunks - 1) * resid) {
958 		nchunks++;
959 		resid = uio->uio_resid / nchunks;
960 	}
961 	ufs_trans_touch(resid, uio);
962 	/*
963 	 * If this request takes too much log space, it will be split
964 	 */
965 	if (nchunks > 1)
966 		*residp = resid;
967 	*resvp = resv;
968 }
969 
970 /*
971  * Issue write request.
972  *
973  * Split a large request into smaller chunks.
974  */
975 int
976 ufs_trans_write(
977 	struct inode *ip,
978 	struct uio *uio,
979 	int ioflag,
980 	cred_t *cr,
981 	int resv,
982 	long resid)
983 {
984 	long		realresid;
985 	int		err;
986 	struct ufsvfs	*ufsvfsp = ip->i_ufsvfs;
987 
988 	/*
989 	 * since the write is too big and would "HOG THE LOG" it needs to
990 	 * be broken up and done in pieces.  NOTE, the caller will
991 	 * issue the EOT after the request has been completed
992 	 */
993 	realresid = uio->uio_resid;
994 
995 again:
996 	/*
997 	 * Perform partial request (uiomove will update uio for us)
998 	 *	Request is split up into "resid" size chunks until
999 	 *	"realresid" bytes have been transferred.
1000 	 */
1001 	uio->uio_resid = MIN(resid, realresid);
1002 	realresid -= uio->uio_resid;
1003 	err = wrip(ip, uio, ioflag, cr);
1004 
1005 	/*
1006 	 * Error or request is done; caller issues final EOT
1007 	 */
1008 	if (err || uio->uio_resid || (realresid == 0)) {
1009 		uio->uio_resid += realresid;
1010 		return (err);
1011 	}
1012 
1013 	/*
1014 	 * Generate EOT for this part of the request
1015 	 */
1016 	rw_exit(&ip->i_contents);
1017 	rw_exit(&ufsvfsp->vfs_dqrwlock);
1018 	if (ioflag & (FSYNC|FDSYNC)) {
1019 		TRANS_END_SYNC(ufsvfsp, err, TOP_WRITE_SYNC, resv);
1020 	} else {
1021 		TRANS_END_ASYNC(ufsvfsp, TOP_WRITE, resv);
1022 	}
1023 
1024 	/*
1025 	 * Make sure the input buffer is resident before starting
1026 	 * the next transaction.
1027 	 */
1028 	ufs_trans_touch(MIN(resid, realresid), uio);
1029 
1030 	/*
1031 	 * Generate BOT for next part of the request
1032 	 */
1033 	if (ioflag & (FSYNC|FDSYNC)) {
1034 		int error;
1035 		TRANS_BEGIN_SYNC(ufsvfsp, TOP_WRITE_SYNC, resv, error);
1036 		ASSERT(!error);
1037 	} else {
1038 		TRANS_BEGIN_ASYNC(ufsvfsp, TOP_WRITE, resv);
1039 	}
1040 	rw_enter(&ufsvfsp->vfs_dqrwlock, RW_READER);
1041 	rw_enter(&ip->i_contents, RW_WRITER);
1042 	/*
1043 	 * Error during EOT (probably device error while writing commit rec)
1044 	 */
1045 	if (err)
1046 		return (err);
1047 	goto again;
1048 }
1049