1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright 2007 Sun Microsystems, Inc. All rights reserved. 23 * Use is subject to license terms. 24 */ 25 26 /* Copyright (c) 1983, 1984, 1985, 1986, 1987, 1988, 1989 AT&T */ 27 /* All Rights Reserved */ 28 29 /* 30 * University Copyright- Copyright (c) 1982, 1986, 1988 31 * The Regents of the University of California 32 * All Rights Reserved 33 * 34 * University Acknowledgment- Portions of this document are derived from 35 * software developed by the University of California, Berkeley, and its 36 * contributors. 37 */ 38 39 40 #pragma ident "%Z%%M% %I% %E% SMI" 41 42 #include <sys/types.h> 43 #include <sys/t_lock.h> 44 #include <sys/param.h> 45 #include <sys/systm.h> 46 #include <sys/signal.h> 47 #include <sys/user.h> 48 #include <sys/vnode.h> 49 #include <sys/buf.h> 50 #include <sys/disp.h> 51 #include <sys/proc.h> 52 #include <sys/conf.h> 53 #include <sys/fs/ufs_inode.h> 54 #include <sys/fs/ufs_fs.h> 55 #include <sys/fs/ufs_quota.h> 56 #include <sys/fs/ufs_trans.h> 57 #include <sys/fs/ufs_bio.h> 58 #include <vm/seg.h> 59 #include <sys/errno.h> 60 #include <sys/sysmacros.h> 61 #include <sys/vfs.h> 62 #include <sys/debug.h> 63 #include <sys/kmem.h> 64 #include <sys/cmn_err.h> 65 66 /* 67 * This structure is used to track blocks as we allocate them, so that 68 * we can free them if we encounter an error during allocation. We 69 * keep track of five pieces of information for each allocated block: 70 * - The number of the newly allocated block 71 * - The size of the block (lets us deal with fragments if we want) 72 * - The number of the block containing a pointer to it; or whether 73 * the pointer is in the inode 74 * - The offset within the block (or inode) containing a pointer to it. 75 * - A flag indicating the usage of the block. (Logging needs to know 76 * this to avoid overwriting a data block if it was previously used 77 * for metadata.) 78 */ 79 80 enum ufs_owner_type { 81 ufs_no_owner, /* Owner has not yet been updated */ 82 ufs_inode_direct, /* Listed in inode's direct block table */ 83 ufs_inode_indirect, /* Listed in inode's indirect block table */ 84 ufs_indirect_block /* Listed in an indirect block */ 85 }; 86 87 struct ufs_allocated_block { 88 daddr_t this_block; /* Number of this block */ 89 off_t block_size; /* Size of this block, in bytes */ 90 enum ufs_owner_type owner; /* Who points to this block? */ 91 daddr_t owner_block; /* Number of the owning block */ 92 uint_t owner_offset; /* Offset within that block or inode */ 93 int usage_flags; /* Usage flags, as expected by free() */ 94 }; 95 96 97 static int findextent(struct fs *fs, daddr32_t *sbp, int n, int *lenp, 98 int maxtrans); 99 100 static void ufs_undo_allocation(inode_t *ip, int block_count, 101 struct ufs_allocated_block table[], int inode_sector_adjust); 102 103 /* 104 * Find the extent and the matching block number. 105 * 106 * bsize > PAGESIZE 107 * boff indicates that we want a page in the middle 108 * min expression is supposed to make sure no extra page[s] after EOF 109 * PAGESIZE >= bsize 110 * we assume that a page is a multiple of bsize, i.e., 111 * boff always == 0 112 * 113 * We always return a length that is suitable for a disk transfer. 114 */ 115 #define DOEXTENT(fs, lbn, boff, bnp, lenp, size, tblp, n, chkfrag, maxtrans) {\ 116 register daddr32_t *dp = (tblp); \ 117 register int _chkfrag = chkfrag; /* for lint. sigh */ \ 118 \ 119 if (*dp == 0) { \ 120 *(bnp) = UFS_HOLE; \ 121 } else { \ 122 register int len; \ 123 \ 124 len = findextent(fs, dp, (int)(n), lenp, maxtrans) << \ 125 (fs)->fs_bshift; \ 126 if (_chkfrag) { \ 127 register u_offset_t tmp; \ 128 \ 129 tmp = fragroundup((fs), size) - \ 130 (((u_offset_t)lbn) << fs->fs_bshift); \ 131 len = (int)MIN(tmp, len); \ 132 } \ 133 len -= (boff); \ 134 if (len <= 0) { \ 135 *(bnp) = UFS_HOLE; \ 136 } else { \ 137 *(bnp) = fsbtodb(fs, *dp) + btodb(boff); \ 138 *(lenp) = len; \ 139 } \ 140 } \ 141 } 142 143 /* 144 * The maximum supported file size is actually somewhat less that 1 145 * terabyte. This is because the total number of blocks used for the 146 * file and its metadata must fit into the ic_blocks field of the 147 * inode, which is a signed 32-bit quantity. The metadata allocated 148 * for a file (that is, the single, double, and triple indirect blocks 149 * used to reference the file blocks) is actually quite small, 150 * but just to make sure, we check for overflow in the ic_blocks 151 * ic_blocks fields for all files whose total block count is 152 * within 1 GB of a terabyte. VERYLARGEFILESIZE below is the number of 153 * 512-byte blocks in a terabyte (2^31), minus the number of 512-byte blocks 154 * in a gigabyte (2^21). We only check for overflow in the ic_blocks 155 * field if the number of blocks currently allocated to the file is 156 * greater than VERYLARGEFILESIZE. 157 * 158 * Note that file "size" is the not the same as file "length". A 159 * file's "size" is the number of blocks allocated to it. A file's 160 * "length" is the maximum offset in the file. A UFS FILE can have a 161 * length of a terabyte, but the size is limited to somewhat less than 162 * a terabyte, as described above. 163 */ 164 #define VERYLARGEFILESIZE 0x7FE00000 165 166 /* 167 * bmap{read,write} define the structure of file system storage by mapping 168 * a logical offset in a file to a physical block number on the device. 169 * It should be called with a locked inode when allocation is to be 170 * done (bmap_write). Note this strangeness: bmap_write is always called from 171 * getpage(), not putpage(), since getpage() is where all the allocation 172 * is done. 173 * 174 * S_READ, S_OTHER -> bmap_read; S_WRITE -> bmap_write. 175 * 176 * NOTICE: the block number returned is the disk block number, not the 177 * file system block number. All the worries about block offsets and 178 * page/block sizes are hidden inside of bmap. Well, not quite, 179 * unfortunately. It's impossible to find one place to hide all this 180 * mess. There are 3 cases: 181 * 182 * PAGESIZE < bsize 183 * In this case, the {get,put}page routines will attempt to align to 184 * a file system block boundry (XXX - maybe this is a mistake?). Since 185 * the kluster routines may be out of memory, we don't always get all 186 * the pages we wanted. If we called bmap first, to find out how much 187 * to kluster, we handed in the block aligned offset. If we didn't get 188 * all the pages, we have to chop off the amount we didn't get from the 189 * amount handed back by bmap. 190 * 191 * PAGESIZE == bsize 192 * Life is quite pleasant here, no extra work needed, mainly because we 193 * (probably?) won't kluster backwards, just forwards. 194 * 195 * PAGESIZE > bsize 196 * This one has a different set of problems, specifically, we may have to 197 * do N reads to fill one page. Let us hope that Sun will stay with small 198 * pages. 199 * 200 * Returns 0 on success, or a non-zero errno if an error occurs. 201 * 202 * TODO 203 * LMXXX - add a bmap cache. This could be a couple of extents in the 204 * inode. Two is nice for PAGESIZE > bsize. 205 */ 206 207 int 208 bmap_read(struct inode *ip, u_offset_t off, daddr_t *bnp, int *lenp) 209 { 210 daddr_t lbn; 211 ufsvfs_t *ufsvfsp = ip->i_ufsvfs; 212 struct fs *fs = ufsvfsp->vfs_fs; 213 struct buf *bp; 214 int i, j, boff; 215 int shft; /* we maintain sh = 1 << shft */ 216 daddr_t ob, nb, tbn; 217 daddr32_t *bap; 218 int nindirshift, nindiroffset; 219 220 ASSERT(RW_LOCK_HELD(&ip->i_contents)); 221 lbn = (daddr_t)lblkno(fs, off); 222 boff = (int)blkoff(fs, off); 223 if (lbn < 0) 224 return (EFBIG); 225 226 /* 227 * The first NDADDR blocks are direct blocks. 228 */ 229 if (lbn < NDADDR) { 230 DOEXTENT(fs, lbn, boff, bnp, lenp, 231 ip->i_size, &ip->i_db[lbn], NDADDR - lbn, 1, 232 ufsvfsp->vfs_iotransz); 233 return (0); 234 } 235 236 nindirshift = ufsvfsp->vfs_nindirshift; 237 nindiroffset = ufsvfsp->vfs_nindiroffset; 238 /* 239 * Determine how many levels of indirection. 240 */ 241 shft = 0; /* sh = 1 */ 242 tbn = lbn - NDADDR; 243 for (j = NIADDR; j > 0; j--) { 244 longlong_t sh; 245 246 shft += nindirshift; /* sh *= nindir */ 247 sh = 1LL << shft; 248 if (tbn < sh) 249 break; 250 tbn -= sh; 251 } 252 if (j == 0) 253 return (EFBIG); 254 255 /* 256 * Fetch the first indirect block. 257 */ 258 nb = ip->i_ib[NIADDR - j]; 259 if (nb == 0) { 260 *bnp = UFS_HOLE; 261 return (0); 262 } 263 264 /* 265 * Fetch through the indirect blocks. 266 */ 267 for (; j <= NIADDR; j++) { 268 ob = nb; 269 bp = UFS_BREAD(ufsvfsp, 270 ip->i_dev, fsbtodb(fs, ob), fs->fs_bsize); 271 if (bp->b_flags & B_ERROR) { 272 brelse(bp); 273 return (EIO); 274 } 275 bap = bp->b_un.b_daddr; 276 277 ASSERT(!ufs_indir_badblock(ip, bap)); 278 279 shft -= nindirshift; /* sh / nindir */ 280 i = (tbn >> shft) & nindiroffset; /* (tbn / sh) % nindir */ 281 nb = bap[i]; 282 if (nb == 0) { 283 *bnp = UFS_HOLE; 284 brelse(bp); 285 return (0); 286 } 287 if (j != NIADDR) 288 brelse(bp); 289 } 290 DOEXTENT(fs, lbn, boff, bnp, lenp, ip->i_size, &bap[i], 291 MIN(NINDIR(fs) - i, (daddr_t)lblkno(fs, ip->i_size - 1) - lbn + 1), 292 0, ufsvfsp->vfs_iotransz); 293 brelse(bp); 294 return (0); 295 } 296 297 /* 298 * See bmap_read for general notes. 299 * 300 * The block must be at least size bytes and will be extended or 301 * allocated as needed. If alloc_type is of type BI_ALLOC_ONLY, then bmap 302 * will not create any in-core pages that correspond to the new disk allocation. 303 * If alloc_type is of BI_FALLOCATE, blocks will be stored as (-1) * block addr 304 * and security is maintained b/c upon reading a negative block number pages 305 * are zeroed. For all other allocation types (BI_NORMAL) the in-core pages will 306 * be created and initialized as needed. 307 * 308 * Returns 0 on success, or a non-zero errno if an error occurs. 309 */ 310 int 311 bmap_write(struct inode *ip, u_offset_t off, int size, 312 enum bi_type alloc_type, daddr_t *allocblk, struct cred *cr) 313 { 314 struct fs *fs; 315 struct buf *bp; 316 int i; 317 struct buf *nbp; 318 int j; 319 int shft; /* we maintain sh = 1 << shft */ 320 daddr_t ob, nb, pref, lbn, llbn, tbn; 321 daddr32_t *bap; 322 struct vnode *vp = ITOV(ip); 323 long bsize = VBSIZE(vp); 324 long osize, nsize; 325 int issync, metaflag, isdirquota; 326 int err; 327 dev_t dev; 328 struct fbuf *fbp; 329 int nindirshift; 330 int nindiroffset; 331 struct ufsvfs *ufsvfsp; 332 int added_sectors; /* sectors added to this inode */ 333 int alloced_blocks; /* fs blocks newly allocated */ 334 struct ufs_allocated_block undo_table[NIADDR+1]; 335 int verylargefile = 0; 336 337 ASSERT(RW_WRITE_HELD(&ip->i_contents)); 338 339 if (allocblk) 340 *allocblk = 0; 341 342 ufsvfsp = ip->i_ufsvfs; 343 fs = ufsvfsp->vfs_bufp->b_un.b_fs; 344 lbn = (daddr_t)lblkno(fs, off); 345 if (lbn < 0) 346 return (EFBIG); 347 if (ip->i_blocks >= VERYLARGEFILESIZE) 348 verylargefile = 1; 349 llbn = (daddr_t)((ip->i_size) ? lblkno(fs, ip->i_size - 1) : 0); 350 metaflag = isdirquota = 0; 351 if (((ip->i_mode & IFMT) == IFDIR) || 352 ((ip->i_mode & IFMT) == IFATTRDIR)) 353 isdirquota = metaflag = I_DIR; 354 else if ((ip->i_mode & IFMT) == IFSHAD) 355 metaflag = I_SHAD; 356 else if (ip->i_ufsvfs->vfs_qinod == ip) 357 isdirquota = metaflag = I_QUOTA; 358 359 issync = ((ip->i_flag & ISYNC) != 0); 360 361 if (isdirquota || issync) { 362 alloc_type = BI_NORMAL; /* make sure */ 363 } 364 365 /* 366 * If the next write will extend the file into a new block, 367 * and the file is currently composed of a fragment 368 * this fragment has to be extended to be a full block. 369 */ 370 if (llbn < NDADDR && llbn < lbn && (ob = ip->i_db[llbn]) != 0) { 371 osize = blksize(fs, ip, llbn); 372 if (osize < bsize && osize > 0) { 373 /* 374 * Check to see if doing this will make the file too 375 * big. Only check if we are dealing with a very 376 * large file. 377 */ 378 if (verylargefile == 1) { 379 if (((unsigned)ip->i_blocks + 380 btodb(bsize - osize)) > INT_MAX) { 381 return (EFBIG); 382 } 383 } 384 /* 385 * Make sure we have all needed pages setup correctly. 386 * 387 * We pass S_OTHER to fbread here because we want 388 * an exclusive lock on the page in question 389 * (see ufs_getpage). I/O to the old block location 390 * may still be in progress and we are about to free 391 * the old block. We don't want anyone else to get 392 * a hold of the old block once we free it until 393 * the I/O is complete. 394 */ 395 err = fbread(ITOV(ip), 396 ((offset_t)llbn << fs->fs_bshift), 397 (uint_t)bsize, S_OTHER, &fbp); 398 if (err) 399 return (err); 400 pref = blkpref(ip, llbn, (int)llbn, &ip->i_db[0]); 401 err = realloccg(ip, ob, pref, (int)osize, (int)bsize, 402 &nb, cr); 403 if (err) { 404 if (fbp) 405 fbrelse(fbp, S_OTHER); 406 return (err); 407 } 408 ASSERT(!ufs_badblock(ip, nb)); 409 410 /* 411 * Update the inode before releasing the 412 * lock on the page. If we released the page 413 * lock first, the data could be written to it's 414 * old address and then destroyed. 415 */ 416 TRANS_MATA_ALLOC(ufsvfsp, ip, nb, bsize, 0); 417 ip->i_db[llbn] = nb; 418 UFS_SET_ISIZE(((u_offset_t)(llbn + 1)) << fs->fs_bshift, 419 ip); 420 ip->i_blocks += btodb(bsize - osize); 421 ASSERT((unsigned)ip->i_blocks <= INT_MAX); 422 TRANS_INODE(ufsvfsp, ip); 423 ip->i_flag |= IUPD | ICHG | IATTCHG; 424 425 /* Caller is responsible for updating i_seq */ 426 /* 427 * Don't check metaflag here, directories won't do this 428 * 429 */ 430 if (issync) { 431 (void) ufs_fbiwrite(fbp, ip, nb, fs->fs_fsize); 432 } else { 433 ASSERT(fbp); 434 fbrelse(fbp, S_WRITE); 435 } 436 437 if (nb != ob) { 438 (void) free(ip, ob, (off_t)osize, metaflag); 439 } 440 } 441 } 442 443 /* 444 * The first NDADDR blocks are direct blocks. 445 */ 446 if (lbn < NDADDR) { 447 nb = ip->i_db[lbn]; 448 if (nb == 0 || 449 ip->i_size < ((u_offset_t)(lbn + 1)) << fs->fs_bshift) { 450 if (nb != 0) { 451 /* consider need to reallocate a frag */ 452 osize = fragroundup(fs, blkoff(fs, ip->i_size)); 453 nsize = fragroundup(fs, size); 454 if (nsize <= osize) 455 goto gotit; 456 /* 457 * Check to see if doing this will make the 458 * file too big. Only check if we are dealing 459 * with a very large file. 460 */ 461 if (verylargefile == 1) { 462 if (((unsigned)ip->i_blocks + 463 btodb(nsize - osize)) > INT_MAX) { 464 return (EFBIG); 465 } 466 } 467 /* 468 * need to re-allocate a block or frag 469 */ 470 ob = nb; 471 pref = blkpref(ip, lbn, (int)lbn, 472 &ip->i_db[0]); 473 err = realloccg(ip, ob, pref, (int)osize, 474 (int)nsize, &nb, cr); 475 if (err) 476 return (err); 477 if (allocblk) 478 *allocblk = nb; 479 ASSERT(!ufs_badblock(ip, nb)); 480 481 } else { 482 /* 483 * need to allocate a block or frag 484 */ 485 osize = 0; 486 if (ip->i_size < 487 ((u_offset_t)(lbn + 1)) << fs->fs_bshift) 488 nsize = fragroundup(fs, size); 489 else 490 nsize = bsize; 491 /* 492 * Check to see if doing this will make the 493 * file too big. Only check if we are dealing 494 * with a very large file. 495 */ 496 if (verylargefile == 1) { 497 if (((unsigned)ip->i_blocks + 498 btodb(nsize - osize)) > INT_MAX) { 499 return (EFBIG); 500 } 501 } 502 pref = blkpref(ip, lbn, (int)lbn, &ip->i_db[0]); 503 err = alloc(ip, pref, (int)nsize, &nb, cr); 504 if (err) 505 return (err); 506 if (allocblk) 507 *allocblk = nb; 508 ASSERT(!ufs_badblock(ip, nb)); 509 ob = nb; 510 } 511 512 /* 513 * Read old/create new zero pages 514 */ 515 fbp = NULL; 516 if (osize == 0) { 517 /* 518 * mmap S_WRITE faults always enter here 519 */ 520 /* 521 * We zero it if its also BI_FALLOCATE, but 522 * only for direct blocks! 523 */ 524 if (alloc_type == BI_NORMAL || 525 alloc_type == BI_FALLOCATE || 526 P2ROUNDUP_TYPED(size, 527 PAGESIZE, u_offset_t) < nsize) { 528 /* fbzero doesn't cause a pagefault */ 529 fbzero(ITOV(ip), 530 ((offset_t)lbn << fs->fs_bshift), 531 (uint_t)nsize, &fbp); 532 } 533 } else { 534 err = fbread(vp, 535 ((offset_t)lbn << fs->fs_bshift), 536 (uint_t)nsize, S_OTHER, &fbp); 537 if (err) { 538 if (nb != ob) { 539 (void) free(ip, nb, 540 (off_t)nsize, metaflag); 541 } else { 542 (void) free(ip, 543 ob + numfrags(fs, osize), 544 (off_t)(nsize - osize), 545 metaflag); 546 } 547 ASSERT(nsize >= osize); 548 (void) chkdq(ip, 549 -(long)btodb(nsize - osize), 550 0, cr, (char **)NULL, 551 (size_t *)NULL); 552 return (err); 553 } 554 } 555 TRANS_MATA_ALLOC(ufsvfsp, ip, nb, nsize, 0); 556 ip->i_db[lbn] = nb; 557 ip->i_blocks += btodb(nsize - osize); 558 ASSERT((unsigned)ip->i_blocks <= INT_MAX); 559 TRANS_INODE(ufsvfsp, ip); 560 ip->i_flag |= IUPD | ICHG | IATTCHG; 561 562 /* Caller is responsible for updating i_seq */ 563 564 /* 565 * Write directory and shadow blocks synchronously so 566 * that they never appear with garbage in them on the 567 * disk. 568 * 569 */ 570 if (isdirquota && (ip->i_size || 571 TRANS_ISTRANS(ufsvfsp))) { 572 /* 573 * XXX man not be necessary with harpy trans 574 * bug id 1130055 575 */ 576 (void) ufs_fbiwrite(fbp, ip, nb, fs->fs_fsize); 577 } else if (fbp) { 578 fbrelse(fbp, S_WRITE); 579 } 580 581 if (nb != ob) 582 (void) free(ip, ob, (off_t)osize, metaflag); 583 } 584 gotit: 585 return (0); 586 } 587 588 added_sectors = alloced_blocks = 0; /* No blocks alloced yet */ 589 590 /* 591 * Determine how many levels of indirection. 592 */ 593 nindirshift = ip->i_ufsvfs->vfs_nindirshift; 594 nindiroffset = ip->i_ufsvfs->vfs_nindiroffset; 595 pref = 0; 596 shft = 0; /* sh = 1 */ 597 tbn = lbn - NDADDR; 598 for (j = NIADDR; j > 0; j--) { 599 longlong_t sh; 600 601 shft += nindirshift; /* sh *= nindir */ 602 sh = 1LL << shft; 603 if (tbn < sh) 604 break; 605 tbn -= sh; 606 } 607 608 if (j == 0) 609 return (EFBIG); 610 611 /* 612 * Fetch the first indirect block. 613 */ 614 dev = ip->i_dev; 615 nb = ip->i_ib[NIADDR - j]; 616 if (nb == 0) { 617 /* 618 * Check to see if doing this will make the 619 * file too big. Only check if we are dealing 620 * with a very large file. 621 */ 622 if (verylargefile == 1) { 623 if (((unsigned)ip->i_blocks + btodb(bsize)) 624 > INT_MAX) { 625 return (EFBIG); 626 } 627 } 628 /* 629 * Need to allocate an indirect block. 630 */ 631 pref = blkpref(ip, lbn, 0, (daddr32_t *)0); 632 err = alloc(ip, pref, (int)bsize, &nb, cr); 633 if (err) 634 return (err); 635 TRANS_MATA_ALLOC(ufsvfsp, ip, nb, bsize, 1); 636 ASSERT(!ufs_badblock(ip, nb)); 637 638 /* 639 * Keep track of this allocation so we can undo it if we 640 * get an error later. 641 */ 642 643 ASSERT(alloced_blocks <= NIADDR); 644 645 undo_table[alloced_blocks].this_block = nb; 646 undo_table[alloced_blocks].block_size = bsize; 647 undo_table[alloced_blocks].owner = ufs_no_owner; 648 undo_table[alloced_blocks].usage_flags = metaflag | I_IBLK; 649 650 alloced_blocks++; 651 652 /* 653 * Write zero block synchronously so that 654 * indirect blocks never point at garbage. 655 */ 656 bp = UFS_GETBLK(ufsvfsp, dev, fsbtodb(fs, nb), bsize); 657 658 clrbuf(bp); 659 /* XXX Maybe special-case this? */ 660 TRANS_BUF(ufsvfsp, 0, bsize, bp, DT_ABZERO); 661 UFS_BWRITE2(ufsvfsp, bp); 662 if (bp->b_flags & B_ERROR) { 663 err = geterror(bp); 664 brelse(bp); 665 ufs_undo_allocation(ip, alloced_blocks, 666 undo_table, added_sectors); 667 return (err); 668 } 669 brelse(bp); 670 671 ip->i_ib[NIADDR - j] = nb; 672 added_sectors += btodb(bsize); 673 ip->i_blocks += btodb(bsize); 674 ASSERT((unsigned)ip->i_blocks <= INT_MAX); 675 TRANS_INODE(ufsvfsp, ip); 676 ip->i_flag |= IUPD | ICHG | IATTCHG; 677 /* Caller is responsible for updating i_seq */ 678 679 /* 680 * Update the 'undo table' now that we've linked this block 681 * to an inode. 682 */ 683 684 undo_table[alloced_blocks-1].owner = ufs_inode_indirect; 685 undo_table[alloced_blocks-1].owner_offset = NIADDR - j; 686 687 /* 688 * In the ISYNC case, wrip will notice that the block 689 * count on the inode has changed and will be sure to 690 * ufs_iupdat the inode at the end of wrip. 691 */ 692 } 693 694 /* 695 * Fetch through the indirect blocks. 696 */ 697 for (; j <= NIADDR; j++) { 698 ob = nb; 699 bp = UFS_BREAD(ufsvfsp, ip->i_dev, fsbtodb(fs, ob), bsize); 700 701 if (bp->b_flags & B_ERROR) { 702 err = geterror(bp); 703 brelse(bp); 704 /* 705 * Return any partial allocations. 706 * 707 * It is possible that we have not yet made any 708 * allocations at this point (if this is the first 709 * pass through the loop and we didn't have to 710 * allocate the first indirect block, above). 711 * In this case, alloced_blocks and added_sectors will 712 * be zero, and ufs_undo_allocation will do nothing. 713 */ 714 ufs_undo_allocation(ip, alloced_blocks, 715 undo_table, added_sectors); 716 return (err); 717 } 718 bap = bp->b_un.b_daddr; 719 shft -= nindirshift; /* sh /= nindir */ 720 i = (tbn >> shft) & nindiroffset; /* (tbn / sh) % nindir */ 721 nb = bap[i]; 722 723 if (nb == 0) { 724 /* 725 * Check to see if doing this will make the 726 * file too big. Only check if we are dealing 727 * with a very large file. 728 */ 729 if (verylargefile == 1) { 730 if (((unsigned)ip->i_blocks + btodb(bsize)) 731 > INT_MAX) { 732 brelse(bp); 733 ufs_undo_allocation(ip, alloced_blocks, 734 undo_table, added_sectors); 735 return (EFBIG); 736 } 737 } 738 if (pref == 0) { 739 if (j < NIADDR) { 740 /* Indirect block */ 741 pref = blkpref(ip, lbn, 0, 742 (daddr32_t *)0); 743 } else { 744 /* Data block */ 745 pref = blkpref(ip, lbn, i, &bap[0]); 746 } 747 } 748 749 /* 750 * release "bp" buf to avoid deadlock (re-bread later) 751 */ 752 brelse(bp); 753 754 err = alloc(ip, pref, (int)bsize, &nb, cr); 755 if (err) { 756 /* 757 * Return any partial allocations. 758 */ 759 ufs_undo_allocation(ip, alloced_blocks, 760 undo_table, added_sectors); 761 return (err); 762 } 763 764 ASSERT(!ufs_badblock(ip, nb)); 765 ASSERT(alloced_blocks <= NIADDR); 766 767 if (allocblk) 768 *allocblk = nb; 769 770 undo_table[alloced_blocks].this_block = nb; 771 undo_table[alloced_blocks].block_size = bsize; 772 undo_table[alloced_blocks].owner = ufs_no_owner; 773 undo_table[alloced_blocks].usage_flags = metaflag | 774 ((j < NIADDR) ? I_IBLK : 0); 775 776 alloced_blocks++; 777 778 if (j < NIADDR) { 779 TRANS_MATA_ALLOC(ufsvfsp, ip, nb, bsize, 1); 780 /* 781 * Write synchronously so indirect 782 * blocks never point at garbage. 783 */ 784 nbp = UFS_GETBLK( 785 ufsvfsp, dev, fsbtodb(fs, nb), bsize); 786 787 clrbuf(nbp); 788 /* XXX Maybe special-case this? */ 789 TRANS_BUF(ufsvfsp, 0, bsize, nbp, DT_ABZERO); 790 UFS_BWRITE2(ufsvfsp, nbp); 791 if (nbp->b_flags & B_ERROR) { 792 err = geterror(nbp); 793 brelse(nbp); 794 /* 795 * Return any partial 796 * allocations. 797 */ 798 ufs_undo_allocation(ip, 799 alloced_blocks, 800 undo_table, added_sectors); 801 return (err); 802 } 803 brelse(nbp); 804 } else if (alloc_type == BI_NORMAL || 805 P2ROUNDUP_TYPED(size, 806 PAGESIZE, u_offset_t) < bsize) { 807 TRANS_MATA_ALLOC(ufsvfsp, ip, nb, bsize, 0); 808 fbzero(ITOV(ip), 809 ((offset_t)lbn << fs->fs_bshift), 810 (uint_t)bsize, &fbp); 811 812 /* 813 * Cases which we need to do a synchronous 814 * write of the zeroed data pages: 815 * 816 * 1) If we are writing a directory then we 817 * want to write synchronously so blocks in 818 * directories never contain garbage. 819 * 820 * 2) If we are filling in a hole and the 821 * indirect block is going to be synchronously 822 * written back below we need to make sure 823 * that the zeroes are written here before 824 * the indirect block is updated so that if 825 * we crash before the real data is pushed 826 * we will not end up with random data is 827 * the middle of the file. 828 * 829 * 3) If the size of the request rounded up 830 * to the system page size is smaller than 831 * the file system block size, we want to 832 * write out all the pages now so that 833 * they are not aborted before they actually 834 * make it to ufs_putpage since the length 835 * of the inode will not include the pages. 836 */ 837 838 if (isdirquota || (issync && 839 lbn < llbn)) 840 (void) ufs_fbiwrite(fbp, ip, nb, 841 fs->fs_fsize); 842 else 843 fbrelse(fbp, S_WRITE); 844 } 845 846 /* 847 * re-acquire "bp" buf 848 */ 849 bp = UFS_BREAD(ufsvfsp, 850 ip->i_dev, fsbtodb(fs, ob), bsize); 851 if (bp->b_flags & B_ERROR) { 852 err = geterror(bp); 853 brelse(bp); 854 /* 855 * Return any partial allocations. 856 */ 857 ufs_undo_allocation(ip, 858 alloced_blocks, 859 undo_table, added_sectors); 860 return (err); 861 } 862 bap = bp->b_un.b_daddr; 863 bap[i] = nb; 864 865 /* 866 * The magic explained: j will be equal to NIADDR 867 * when we are at the lowest level, this is where the 868 * array entries point directly to data blocks. Since 869 * we will be 'fallocate'ing we will go ahead and negate 870 * the addresses. 871 */ 872 if (alloc_type == BI_FALLOCATE && j == NIADDR) 873 bap[i] = -bap[i]; 874 875 TRANS_BUF_ITEM_128(ufsvfsp, bap[i], bap, bp, DT_AB); 876 added_sectors += btodb(bsize); 877 ip->i_blocks += btodb(bsize); 878 ASSERT((unsigned)ip->i_blocks <= INT_MAX); 879 TRANS_INODE(ufsvfsp, ip); 880 ip->i_flag |= IUPD | ICHG | IATTCHG; 881 882 /* Caller is responsible for updating i_seq */ 883 884 undo_table[alloced_blocks-1].owner = 885 ufs_indirect_block; 886 undo_table[alloced_blocks-1].owner_block = ob; 887 undo_table[alloced_blocks-1].owner_offset = i; 888 889 if (issync) { 890 UFS_BWRITE2(ufsvfsp, bp); 891 if (bp->b_flags & B_ERROR) { 892 err = geterror(bp); 893 brelse(bp); 894 /* 895 * Return any partial 896 * allocations. 897 */ 898 ufs_undo_allocation(ip, 899 alloced_blocks, 900 undo_table, added_sectors); 901 return (err); 902 } 903 brelse(bp); 904 } else { 905 bdrwrite(bp); 906 } 907 } else { 908 brelse(bp); 909 } 910 } 911 return (0); 912 } 913 914 /* 915 * Return 1 if inode has unmapped blocks (UFS holes). 916 */ 917 int 918 bmap_has_holes(struct inode *ip) 919 { 920 struct fs *fs = ip->i_fs; 921 uint_t dblks; /* # of data blocks */ 922 uint_t mblks; /* # of data + metadata blocks */ 923 int nindirshift; 924 int nindiroffset; 925 uint_t cnt; 926 int n, j, shft; 927 uint_t nindirblks; 928 929 int fsbshift = fs->fs_bshift; 930 int fsboffset = (1 << fsbshift) - 1; 931 932 dblks = (ip->i_size + fsboffset) >> fsbshift; 933 mblks = (ldbtob((u_offset_t)ip->i_blocks) + fsboffset) >> fsbshift; 934 935 /* 936 * File has only direct blocks. 937 */ 938 if (dblks <= NDADDR) 939 return (mblks < dblks); 940 nindirshift = ip->i_ufsvfs->vfs_nindirshift; 941 942 nindiroffset = ip->i_ufsvfs->vfs_nindiroffset; 943 nindirblks = nindiroffset + 1; 944 945 dblks -= NDADDR; 946 shft = 0; 947 /* 948 * Determine how many levels of indirection. 949 */ 950 for (j = NIADDR; j > 0; j--) { 951 longlong_t sh; 952 953 shft += nindirshift; /* sh *= nindir */ 954 sh = 1LL << shft; 955 if (dblks <= sh) 956 break; 957 dblks -= sh; 958 } 959 /* LINTED: warning: logical expression always true: op "||" */ 960 ASSERT(NIADDR <= 3); 961 ASSERT(j <= NIADDR); 962 if (j == NIADDR) /* single level indirection */ 963 cnt = NDADDR + 1 + dblks; 964 else if (j == NIADDR-1) /* double indirection */ 965 cnt = NDADDR + 1 + nindirblks + 966 1 + (dblks + nindiroffset)/nindirblks + dblks; 967 else if (j == NIADDR-2) { /* triple indirection */ 968 n = (dblks + nindiroffset)/nindirblks; 969 cnt = NDADDR + 1 + nindirblks + 970 1 + nindirblks + nindirblks*nindirblks + 971 1 + (n + nindiroffset)/nindirblks + n + dblks; 972 } 973 974 return (mblks < cnt); 975 } 976 977 /* 978 * find some contig blocks starting at *sbp and going for min(n, max_contig) 979 * return the number of blocks (not frags) found. 980 * The array passed in must be at least [0..n-1]. 981 */ 982 static int 983 findextent(struct fs *fs, daddr32_t *sbp, int n, int *lenp, int maxtransfer) 984 { 985 register daddr_t bn, nextbn; 986 register daddr32_t *bp; 987 register int diff; 988 int maxtransblk; 989 990 if (n <= 0) 991 return (0); 992 bn = *sbp; 993 if (bn == 0) 994 return (0); 995 996 diff = fs->fs_frag; 997 if (*lenp) { 998 n = MIN(n, lblkno(fs, *lenp)); 999 } else { 1000 /* 1001 * If the user has set the value for maxcontig lower than 1002 * the drive transfer size, then assume they want this 1003 * to be the maximum value for the size of the data transfer. 1004 */ 1005 maxtransblk = maxtransfer >> DEV_BSHIFT; 1006 if (fs->fs_maxcontig < maxtransblk) { 1007 n = MIN(n, fs->fs_maxcontig); 1008 } else { 1009 n = MIN(n, maxtransblk); 1010 } 1011 } 1012 bp = sbp; 1013 while (--n > 0) { 1014 nextbn = *(bp + 1); 1015 if (nextbn == 0 || bn + diff != nextbn) 1016 break; 1017 bn = nextbn; 1018 bp++; 1019 } 1020 return ((int)(bp - sbp) + 1); 1021 } 1022 1023 /* 1024 * Free any blocks which had been successfully allocated. Always called 1025 * as a result of an error, so we don't bother returning an error code 1026 * from here. 1027 * 1028 * If block_count and inode_sector_adjust are both zero, we'll do nothing. 1029 * Thus it is safe to call this as part of error handling, whether or not 1030 * any blocks have been allocated. 1031 * 1032 * The ufs_inode_direct case is currently unused. 1033 */ 1034 1035 static void 1036 ufs_undo_allocation( 1037 inode_t *ip, 1038 int block_count, 1039 struct ufs_allocated_block table[], 1040 int inode_sector_adjust) 1041 { 1042 int i; 1043 int inode_changed; 1044 int error_updating_pointers; 1045 struct ufsvfs *ufsvfsp; 1046 1047 inode_changed = 0; 1048 error_updating_pointers = 0; 1049 1050 ufsvfsp = ip->i_ufsvfs; 1051 1052 /* 1053 * Update pointers on disk before freeing blocks. If we fail, 1054 * some blocks may remain busy; but they will be reclaimed by 1055 * an fsck. (This is better than letting a block wind up with 1056 * two owners if we successfully freed it but could not remove 1057 * the pointer to it.) 1058 */ 1059 1060 for (i = 0; i < block_count; i++) { 1061 switch (table[i].owner) { 1062 case ufs_no_owner: 1063 /* Nothing to do here, nobody points to us */ 1064 break; 1065 case ufs_inode_direct: 1066 ASSERT(table[i].owner_offset < NDADDR); 1067 ip->i_db[table[i].owner_offset] = 0; 1068 inode_changed = 1; 1069 break; 1070 case ufs_inode_indirect: 1071 ASSERT(table[i].owner_offset < NIADDR); 1072 ip->i_ib[table[i].owner_offset] = 0; 1073 inode_changed = 1; 1074 break; 1075 case ufs_indirect_block: { 1076 buf_t *bp; 1077 daddr32_t *block_data; 1078 1079 /* Read/modify/log/write. */ 1080 1081 ASSERT(table[i].owner_offset < 1082 (VBSIZE(ITOV(ip)) / sizeof (daddr32_t))); 1083 1084 bp = UFS_BREAD(ufsvfsp, ip->i_dev, 1085 fsbtodb(ufsvfsp->vfs_fs, table[i].owner_block), 1086 VBSIZE(ITOV(ip))); 1087 1088 if (bp->b_flags & B_ERROR) { 1089 /* Couldn't read this block; give up. */ 1090 error_updating_pointers = 1; 1091 brelse(bp); 1092 break; /* out of SWITCH */ 1093 } 1094 1095 block_data = bp->b_un.b_daddr; 1096 block_data[table[i].owner_offset] = 0; 1097 1098 /* Write a log entry which includes the zero. */ 1099 /* It might be possible to optimize this by using */ 1100 /* TRANS_BUF directly and zeroing only the four */ 1101 /* bytes involved, but an attempt to do that led */ 1102 /* to panics in the logging code. The attempt was */ 1103 /* TRANS_BUF(ufsvfsp, */ 1104 /* table[i].owner_offset * sizeof (daddr32_t), */ 1105 /* sizeof (daddr32_t), */ 1106 /* bp, */ 1107 /* DT_ABZERO); */ 1108 1109 TRANS_BUF_ITEM_128(ufsvfsp, 1110 block_data[table[i].owner_offset], 1111 block_data, bp, DT_AB); 1112 1113 /* Now we can write the buffer itself. */ 1114 1115 UFS_BWRITE2(ufsvfsp, bp); 1116 1117 if (bp->b_flags & B_ERROR) { 1118 error_updating_pointers = 1; 1119 } 1120 1121 brelse(bp); 1122 break; 1123 } 1124 default: 1125 (void) ufs_fault(ITOV(ip), 1126 "ufs_undo_allocation failure\n"); 1127 break; 1128 } 1129 } 1130 1131 /* 1132 * If the inode changed, or if we need to update its block count, 1133 * then do that now. We update the inode synchronously on disk 1134 * to ensure that it won't transiently point at a block we've 1135 * freed (only necessary if we're not logging). 1136 * 1137 * NOTE: Currently ufs_iupdat() does not check for errors. When 1138 * it is fixed, we should verify that we successfully updated the 1139 * inode before freeing blocks below. 1140 */ 1141 1142 if (inode_changed || (inode_sector_adjust != 0)) { 1143 ip->i_blocks -= inode_sector_adjust; 1144 ASSERT((unsigned)ip->i_blocks <= INT_MAX); 1145 TRANS_INODE(ufsvfsp, ip); 1146 ip->i_flag |= IUPD | ICHG | IATTCHG; 1147 ip->i_seq++; 1148 if (!TRANS_ISTRANS(ufsvfsp)) 1149 ufs_iupdat(ip, I_SYNC); 1150 } 1151 1152 /* 1153 * Now we go through and actually free the blocks, but only if we 1154 * successfully removed the pointers to them. 1155 */ 1156 1157 if (!error_updating_pointers) { 1158 for (i = 0; i < block_count; i++) { 1159 free(ip, table[i].this_block, table[i].block_size, 1160 table[i].usage_flags); 1161 } 1162 } 1163 } 1164 1165 /* 1166 * Find the next hole or data block in file starting at *off 1167 * Return found offset in *off, which can be less than the 1168 * starting offset if not block aligned. 1169 * This code is based on bmap_read(). 1170 * Errors: ENXIO for end of file 1171 * EIO for block read error. 1172 */ 1173 int 1174 bmap_find(struct inode *ip, boolean_t hole, u_offset_t *off) 1175 { 1176 ufsvfs_t *ufsvfsp = ip->i_ufsvfs; 1177 struct fs *fs = ufsvfsp->vfs_fs; 1178 buf_t *bp[NIADDR]; 1179 int i, j; 1180 int shft; /* we maintain sh = 1 << shft */ 1181 int nindirshift, nindiroffset; 1182 daddr_t ob, nb, tbn, lbn, skip; 1183 daddr32_t *bap; 1184 u_offset_t isz = (offset_t)ip->i_size; 1185 int32_t bs = fs->fs_bsize; /* file system block size */ 1186 int32_t nindir = fs->fs_nindir; 1187 dev_t dev; 1188 int error = 0; 1189 daddr_t limits[NIADDR]; 1190 1191 ASSERT(*off < isz); 1192 ASSERT(RW_LOCK_HELD(&ip->i_contents)); 1193 lbn = (daddr_t)lblkno(fs, *off); 1194 ASSERT(lbn >= 0); 1195 1196 for (i = 0; i < NIADDR; i++) 1197 bp[i] = NULL; 1198 1199 /* 1200 * The first NDADDR blocks are direct blocks. 1201 */ 1202 if (lbn < NDADDR) { 1203 for (; lbn < NDADDR; lbn++) { 1204 if ((hole && (ip->i_db[lbn] == 0)) || 1205 (!hole && (ip->i_db[lbn] != 0))) { 1206 goto out; 1207 } 1208 } 1209 if ((u_offset_t)lbn << fs->fs_bshift >= isz) 1210 goto out; 1211 } 1212 1213 nindir = fs->fs_nindir; 1214 nindirshift = ufsvfsp->vfs_nindirshift; 1215 nindiroffset = ufsvfsp->vfs_nindiroffset; 1216 dev = ip->i_dev; 1217 1218 /* Set up limits array */ 1219 for (limits[0] = NDADDR, j = 1; j < NIADDR; j++) 1220 limits[j] = limits[j-1] + (1ULL << (nindirshift * j)); 1221 1222 loop: 1223 /* 1224 * Determine how many levels of indirection. 1225 */ 1226 shft = 0; /* sh = 1 */ 1227 tbn = lbn - NDADDR; 1228 for (j = NIADDR; j > 0; j--) { 1229 longlong_t sh; 1230 1231 shft += nindirshift; /* sh *= nindir */ 1232 sh = 1LL << shft; 1233 if (tbn < sh) 1234 break; 1235 tbn -= sh; 1236 } 1237 if (j == 0) { 1238 /* must have passed end of file */ 1239 ASSERT(((u_offset_t)lbn << fs->fs_bshift) >= isz); 1240 goto out; 1241 } 1242 1243 /* 1244 * Fetch the first indirect block. 1245 */ 1246 nb = ip->i_ib[NIADDR - j]; 1247 if (nb == 0) { 1248 if (hole) { 1249 lbn = limits[NIADDR - j]; 1250 goto out; 1251 } else { 1252 lbn = limits[NIADDR - j + 1]; 1253 if ((u_offset_t)lbn << fs->fs_bshift >= isz) 1254 goto out; 1255 goto loop; 1256 } 1257 } 1258 1259 /* 1260 * Fetch through the indirect blocks. 1261 */ 1262 for (; ((j <= NIADDR) && (nb != 0)); j++) { 1263 ob = nb; 1264 /* 1265 * if there's a different block at this level then release 1266 * the old one and in with the new. 1267 */ 1268 if ((bp[j-1] == NULL) || bp[j-1]->b_blkno != fsbtodb(fs, ob)) { 1269 if (bp[j-1] != NULL) 1270 brelse(bp[j-1]); 1271 bp[j-1] = UFS_BREAD(ufsvfsp, dev, fsbtodb(fs, ob), bs); 1272 if (bp[j-1]->b_flags & B_ERROR) { 1273 error = EIO; 1274 goto out; 1275 } 1276 } 1277 bap = bp[j-1]->b_un.b_daddr; 1278 1279 shft -= nindirshift; /* sh / nindir */ 1280 i = (tbn >> shft) & nindiroffset; /* (tbn / sh) % nindir */ 1281 nb = bap[i]; 1282 skip = 1LL << (nindirshift * (NIADDR - j)); 1283 } 1284 1285 /* 1286 * Scan through the blocks in this array. 1287 */ 1288 for (; i < nindir; i++, lbn += skip) { 1289 if (hole && (bap[i] == 0)) 1290 goto out; 1291 if (!hole && (bap[i] != 0)) { 1292 if (skip == 1) { 1293 /* we're at the lowest level */ 1294 goto out; 1295 } else { 1296 goto loop; 1297 } 1298 } 1299 } 1300 if (((u_offset_t)lbn << fs->fs_bshift) < isz) 1301 goto loop; 1302 out: 1303 for (i = 0; i < NIADDR; i++) { 1304 if (bp[i]) 1305 brelse(bp[i]); 1306 } 1307 if (error == 0) { 1308 if (((u_offset_t)lbn << fs->fs_bshift) >= isz) { 1309 error = ENXIO; 1310 } else { 1311 /* success */ 1312 *off = (u_offset_t)lbn << fs->fs_bshift; 1313 } 1314 } 1315 return (error); 1316 } 1317 1318 /* 1319 * Set a particular offset in the inode list to be a certain block. 1320 * User is responsible for calling TRANS* functions 1321 */ 1322 int 1323 bmap_set_bn(struct vnode *vp, u_offset_t off, daddr32_t bn) 1324 { 1325 daddr_t lbn; 1326 struct inode *ip; 1327 ufsvfs_t *ufsvfsp; 1328 struct fs *fs; 1329 struct buf *bp; 1330 int i, j; 1331 int shft; /* we maintain sh = 1 << shft */ 1332 int err; 1333 daddr_t ob, nb, tbn; 1334 daddr32_t *bap; 1335 int nindirshift, nindiroffset; 1336 1337 ip = VTOI(vp); 1338 ufsvfsp = ip->i_ufsvfs; 1339 fs = ufsvfsp->vfs_fs; 1340 lbn = (daddr_t)lblkno(fs, off); 1341 1342 ASSERT(RW_LOCK_HELD(&ip->i_contents)); 1343 1344 if (lbn < 0) 1345 return (EFBIG); 1346 1347 /* 1348 * Take care of direct block assignment 1349 */ 1350 if (lbn < NDADDR) { 1351 ip->i_db[lbn] = bn; 1352 return (0); 1353 } 1354 1355 nindirshift = ip->i_ufsvfs->vfs_nindirshift; 1356 nindiroffset = ip->i_ufsvfs->vfs_nindiroffset; 1357 /* 1358 * Determine how many levels of indirection. 1359 */ 1360 shft = 0; /* sh = 1 */ 1361 tbn = lbn - NDADDR; 1362 for (j = NIADDR; j > 0; j--) { 1363 longlong_t sh; 1364 1365 shft += nindirshift; /* sh *= nindir */ 1366 sh = 1LL << shft; 1367 if (tbn < sh) 1368 break; 1369 tbn -= sh; 1370 } 1371 if (j == 0) 1372 return (EFBIG); 1373 1374 /* 1375 * Fetch the first indirect block. 1376 */ 1377 nb = ip->i_ib[NIADDR - j]; 1378 if (nb == 0) { 1379 err = ufs_fault(ITOV(ip), "ufs_set_bn: nb == UFS_HOLE"); 1380 return (err); 1381 } 1382 1383 /* 1384 * Fetch through the indirect blocks. 1385 */ 1386 for (; j <= NIADDR; j++) { 1387 ob = nb; 1388 bp = UFS_BREAD(ufsvfsp, 1389 ip->i_dev, fsbtodb(fs, ob), fs->fs_bsize); 1390 if (bp->b_flags & B_ERROR) { 1391 err = geterror(bp); 1392 brelse(bp); 1393 return (err); 1394 } 1395 bap = bp->b_un.b_daddr; 1396 1397 ASSERT(!ufs_indir_badblock(ip, bap)); 1398 1399 shft -= nindirshift; /* sh / nindir */ 1400 i = (tbn >> shft) & nindiroffset; /* (tbn / sh) % nindir */ 1401 1402 nb = bap[i]; 1403 if (nb == 0) { 1404 err = ufs_fault(ITOV(ip), "ufs_set_bn: nb == UFS_HOLE"); 1405 return (err); 1406 } 1407 1408 if (j == NIADDR) { 1409 bap[i] = bn; 1410 bdrwrite(bp); 1411 return (0); 1412 } 1413 1414 brelse(bp); 1415 } 1416 return (0); 1417 } 1418