1 /* -*- mode: c; c-basic-offset: 8; -*- 2 * vim: noexpandtab sw=8 ts=8 sts=0: 3 * 4 * file.c 5 * 6 * File open, close, extend, truncate 7 * 8 * Copyright (C) 2002, 2004 Oracle. All rights reserved. 9 * 10 * This program is free software; you can redistribute it and/or 11 * modify it under the terms of the GNU General Public 12 * License as published by the Free Software Foundation; either 13 * version 2 of the License, or (at your option) any later version. 14 * 15 * This program is distributed in the hope that it will be useful, 16 * but WITHOUT ANY WARRANTY; without even the implied warranty of 17 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 18 * General Public License for more details. 19 * 20 * You should have received a copy of the GNU General Public 21 * License along with this program; if not, write to the 22 * Free Software Foundation, Inc., 59 Temple Place - Suite 330, 23 * Boston, MA 021110-1307, USA. 24 */ 25 26 #include <linux/capability.h> 27 #include <linux/fs.h> 28 #include <linux/types.h> 29 #include <linux/slab.h> 30 #include <linux/highmem.h> 31 #include <linux/pagemap.h> 32 #include <linux/uio.h> 33 #include <linux/sched.h> 34 #include <linux/splice.h> 35 #include <linux/mount.h> 36 #include <linux/writeback.h> 37 #include <linux/falloc.h> 38 39 #define MLOG_MASK_PREFIX ML_INODE 40 #include <cluster/masklog.h> 41 42 #include "ocfs2.h" 43 44 #include "alloc.h" 45 #include "aops.h" 46 #include "dir.h" 47 #include "dlmglue.h" 48 #include "extent_map.h" 49 #include "file.h" 50 #include "sysfile.h" 51 #include "inode.h" 52 #include "ioctl.h" 53 #include "journal.h" 54 #include "locks.h" 55 #include "mmap.h" 56 #include "suballoc.h" 57 #include "super.h" 58 #include "xattr.h" 59 60 #include "buffer_head_io.h" 61 62 static int ocfs2_sync_inode(struct inode *inode) 63 { 64 filemap_fdatawrite(inode->i_mapping); 65 return sync_mapping_buffers(inode->i_mapping); 66 } 67 68 static int ocfs2_init_file_private(struct inode *inode, struct file *file) 69 { 70 struct ocfs2_file_private *fp; 71 72 fp = kzalloc(sizeof(struct ocfs2_file_private), GFP_KERNEL); 73 if (!fp) 74 return -ENOMEM; 75 76 fp->fp_file = file; 77 mutex_init(&fp->fp_mutex); 78 ocfs2_file_lock_res_init(&fp->fp_flock, fp); 79 file->private_data = fp; 80 81 return 0; 82 } 83 84 static void ocfs2_free_file_private(struct inode *inode, struct file *file) 85 { 86 struct ocfs2_file_private *fp = file->private_data; 87 struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); 88 89 if (fp) { 90 ocfs2_simple_drop_lockres(osb, &fp->fp_flock); 91 ocfs2_lock_res_free(&fp->fp_flock); 92 kfree(fp); 93 file->private_data = NULL; 94 } 95 } 96 97 static int ocfs2_file_open(struct inode *inode, struct file *file) 98 { 99 int status; 100 int mode = file->f_flags; 101 struct ocfs2_inode_info *oi = OCFS2_I(inode); 102 103 mlog_entry("(0x%p, 0x%p, '%.*s')\n", inode, file, 104 file->f_path.dentry->d_name.len, file->f_path.dentry->d_name.name); 105 106 spin_lock(&oi->ip_lock); 107 108 /* Check that the inode hasn't been wiped from disk by another 109 * node. If it hasn't then we're safe as long as we hold the 110 * spin lock until our increment of open count. */ 111 if (OCFS2_I(inode)->ip_flags & OCFS2_INODE_DELETED) { 112 spin_unlock(&oi->ip_lock); 113 114 status = -ENOENT; 115 goto leave; 116 } 117 118 if (mode & O_DIRECT) 119 oi->ip_flags |= OCFS2_INODE_OPEN_DIRECT; 120 121 oi->ip_open_count++; 122 spin_unlock(&oi->ip_lock); 123 124 status = ocfs2_init_file_private(inode, file); 125 if (status) { 126 /* 127 * We want to set open count back if we're failing the 128 * open. 129 */ 130 spin_lock(&oi->ip_lock); 131 oi->ip_open_count--; 132 spin_unlock(&oi->ip_lock); 133 } 134 135 leave: 136 mlog_exit(status); 137 return status; 138 } 139 140 static int ocfs2_file_release(struct inode *inode, struct file *file) 141 { 142 struct ocfs2_inode_info *oi = OCFS2_I(inode); 143 144 mlog_entry("(0x%p, 0x%p, '%.*s')\n", inode, file, 145 file->f_path.dentry->d_name.len, 146 file->f_path.dentry->d_name.name); 147 148 spin_lock(&oi->ip_lock); 149 if (!--oi->ip_open_count) 150 oi->ip_flags &= ~OCFS2_INODE_OPEN_DIRECT; 151 spin_unlock(&oi->ip_lock); 152 153 ocfs2_free_file_private(inode, file); 154 155 mlog_exit(0); 156 157 return 0; 158 } 159 160 static int ocfs2_dir_open(struct inode *inode, struct file *file) 161 { 162 return ocfs2_init_file_private(inode, file); 163 } 164 165 static int ocfs2_dir_release(struct inode *inode, struct file *file) 166 { 167 ocfs2_free_file_private(inode, file); 168 return 0; 169 } 170 171 static int ocfs2_sync_file(struct file *file, 172 struct dentry *dentry, 173 int datasync) 174 { 175 int err = 0; 176 journal_t *journal; 177 struct inode *inode = dentry->d_inode; 178 struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); 179 180 mlog_entry("(0x%p, 0x%p, %d, '%.*s')\n", file, dentry, datasync, 181 dentry->d_name.len, dentry->d_name.name); 182 183 err = ocfs2_sync_inode(dentry->d_inode); 184 if (err) 185 goto bail; 186 187 journal = osb->journal->j_journal; 188 err = jbd2_journal_force_commit(journal); 189 190 bail: 191 mlog_exit(err); 192 193 return (err < 0) ? -EIO : 0; 194 } 195 196 int ocfs2_should_update_atime(struct inode *inode, 197 struct vfsmount *vfsmnt) 198 { 199 struct timespec now; 200 struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); 201 202 if (ocfs2_is_hard_readonly(osb) || ocfs2_is_soft_readonly(osb)) 203 return 0; 204 205 if ((inode->i_flags & S_NOATIME) || 206 ((inode->i_sb->s_flags & MS_NODIRATIME) && S_ISDIR(inode->i_mode))) 207 return 0; 208 209 /* 210 * We can be called with no vfsmnt structure - NFSD will 211 * sometimes do this. 212 * 213 * Note that our action here is different than touch_atime() - 214 * if we can't tell whether this is a noatime mount, then we 215 * don't know whether to trust the value of s_atime_quantum. 216 */ 217 if (vfsmnt == NULL) 218 return 0; 219 220 if ((vfsmnt->mnt_flags & MNT_NOATIME) || 221 ((vfsmnt->mnt_flags & MNT_NODIRATIME) && S_ISDIR(inode->i_mode))) 222 return 0; 223 224 if (vfsmnt->mnt_flags & MNT_RELATIME) { 225 if ((timespec_compare(&inode->i_atime, &inode->i_mtime) <= 0) || 226 (timespec_compare(&inode->i_atime, &inode->i_ctime) <= 0)) 227 return 1; 228 229 return 0; 230 } 231 232 now = CURRENT_TIME; 233 if ((now.tv_sec - inode->i_atime.tv_sec <= osb->s_atime_quantum)) 234 return 0; 235 else 236 return 1; 237 } 238 239 int ocfs2_update_inode_atime(struct inode *inode, 240 struct buffer_head *bh) 241 { 242 int ret; 243 struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); 244 handle_t *handle; 245 struct ocfs2_dinode *di = (struct ocfs2_dinode *) bh->b_data; 246 247 mlog_entry_void(); 248 249 handle = ocfs2_start_trans(osb, OCFS2_INODE_UPDATE_CREDITS); 250 if (handle == NULL) { 251 ret = -ENOMEM; 252 mlog_errno(ret); 253 goto out; 254 } 255 256 ret = ocfs2_journal_access(handle, inode, bh, 257 OCFS2_JOURNAL_ACCESS_WRITE); 258 if (ret) { 259 mlog_errno(ret); 260 goto out_commit; 261 } 262 263 /* 264 * Don't use ocfs2_mark_inode_dirty() here as we don't always 265 * have i_mutex to guard against concurrent changes to other 266 * inode fields. 267 */ 268 inode->i_atime = CURRENT_TIME; 269 di->i_atime = cpu_to_le64(inode->i_atime.tv_sec); 270 di->i_atime_nsec = cpu_to_le32(inode->i_atime.tv_nsec); 271 272 ret = ocfs2_journal_dirty(handle, bh); 273 if (ret < 0) 274 mlog_errno(ret); 275 276 out_commit: 277 ocfs2_commit_trans(OCFS2_SB(inode->i_sb), handle); 278 out: 279 mlog_exit(ret); 280 return ret; 281 } 282 283 static int ocfs2_set_inode_size(handle_t *handle, 284 struct inode *inode, 285 struct buffer_head *fe_bh, 286 u64 new_i_size) 287 { 288 int status; 289 290 mlog_entry_void(); 291 i_size_write(inode, new_i_size); 292 inode->i_blocks = ocfs2_inode_sector_count(inode); 293 inode->i_ctime = inode->i_mtime = CURRENT_TIME; 294 295 status = ocfs2_mark_inode_dirty(handle, inode, fe_bh); 296 if (status < 0) { 297 mlog_errno(status); 298 goto bail; 299 } 300 301 bail: 302 mlog_exit(status); 303 return status; 304 } 305 306 static int ocfs2_simple_size_update(struct inode *inode, 307 struct buffer_head *di_bh, 308 u64 new_i_size) 309 { 310 int ret; 311 struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); 312 handle_t *handle = NULL; 313 314 handle = ocfs2_start_trans(osb, OCFS2_INODE_UPDATE_CREDITS); 315 if (handle == NULL) { 316 ret = -ENOMEM; 317 mlog_errno(ret); 318 goto out; 319 } 320 321 ret = ocfs2_set_inode_size(handle, inode, di_bh, 322 new_i_size); 323 if (ret < 0) 324 mlog_errno(ret); 325 326 ocfs2_commit_trans(osb, handle); 327 out: 328 return ret; 329 } 330 331 static int ocfs2_orphan_for_truncate(struct ocfs2_super *osb, 332 struct inode *inode, 333 struct buffer_head *fe_bh, 334 u64 new_i_size) 335 { 336 int status; 337 handle_t *handle; 338 struct ocfs2_dinode *di; 339 u64 cluster_bytes; 340 341 mlog_entry_void(); 342 343 /* TODO: This needs to actually orphan the inode in this 344 * transaction. */ 345 346 handle = ocfs2_start_trans(osb, OCFS2_INODE_UPDATE_CREDITS); 347 if (IS_ERR(handle)) { 348 status = PTR_ERR(handle); 349 mlog_errno(status); 350 goto out; 351 } 352 353 status = ocfs2_journal_access(handle, inode, fe_bh, 354 OCFS2_JOURNAL_ACCESS_WRITE); 355 if (status < 0) { 356 mlog_errno(status); 357 goto out_commit; 358 } 359 360 /* 361 * Do this before setting i_size. 362 */ 363 cluster_bytes = ocfs2_align_bytes_to_clusters(inode->i_sb, new_i_size); 364 status = ocfs2_zero_range_for_truncate(inode, handle, new_i_size, 365 cluster_bytes); 366 if (status) { 367 mlog_errno(status); 368 goto out_commit; 369 } 370 371 i_size_write(inode, new_i_size); 372 inode->i_ctime = inode->i_mtime = CURRENT_TIME; 373 374 di = (struct ocfs2_dinode *) fe_bh->b_data; 375 di->i_size = cpu_to_le64(new_i_size); 376 di->i_ctime = di->i_mtime = cpu_to_le64(inode->i_ctime.tv_sec); 377 di->i_ctime_nsec = di->i_mtime_nsec = cpu_to_le32(inode->i_ctime.tv_nsec); 378 379 status = ocfs2_journal_dirty(handle, fe_bh); 380 if (status < 0) 381 mlog_errno(status); 382 383 out_commit: 384 ocfs2_commit_trans(osb, handle); 385 out: 386 387 mlog_exit(status); 388 return status; 389 } 390 391 static int ocfs2_truncate_file(struct inode *inode, 392 struct buffer_head *di_bh, 393 u64 new_i_size) 394 { 395 int status = 0; 396 struct ocfs2_dinode *fe = NULL; 397 struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); 398 struct ocfs2_truncate_context *tc = NULL; 399 400 mlog_entry("(inode = %llu, new_i_size = %llu\n", 401 (unsigned long long)OCFS2_I(inode)->ip_blkno, 402 (unsigned long long)new_i_size); 403 404 fe = (struct ocfs2_dinode *) di_bh->b_data; 405 if (!OCFS2_IS_VALID_DINODE(fe)) { 406 OCFS2_RO_ON_INVALID_DINODE(inode->i_sb, fe); 407 status = -EIO; 408 goto bail; 409 } 410 411 mlog_bug_on_msg(le64_to_cpu(fe->i_size) != i_size_read(inode), 412 "Inode %llu, inode i_size = %lld != di " 413 "i_size = %llu, i_flags = 0x%x\n", 414 (unsigned long long)OCFS2_I(inode)->ip_blkno, 415 i_size_read(inode), 416 (unsigned long long)le64_to_cpu(fe->i_size), 417 le32_to_cpu(fe->i_flags)); 418 419 if (new_i_size > le64_to_cpu(fe->i_size)) { 420 mlog(0, "asked to truncate file with size (%llu) to size (%llu)!\n", 421 (unsigned long long)le64_to_cpu(fe->i_size), 422 (unsigned long long)new_i_size); 423 status = -EINVAL; 424 mlog_errno(status); 425 goto bail; 426 } 427 428 mlog(0, "inode %llu, i_size = %llu, new_i_size = %llu\n", 429 (unsigned long long)le64_to_cpu(fe->i_blkno), 430 (unsigned long long)le64_to_cpu(fe->i_size), 431 (unsigned long long)new_i_size); 432 433 /* lets handle the simple truncate cases before doing any more 434 * cluster locking. */ 435 if (new_i_size == le64_to_cpu(fe->i_size)) 436 goto bail; 437 438 down_write(&OCFS2_I(inode)->ip_alloc_sem); 439 440 /* 441 * The inode lock forced other nodes to sync and drop their 442 * pages, which (correctly) happens even if we have a truncate 443 * without allocation change - ocfs2 cluster sizes can be much 444 * greater than page size, so we have to truncate them 445 * anyway. 446 */ 447 unmap_mapping_range(inode->i_mapping, new_i_size + PAGE_SIZE - 1, 0, 1); 448 truncate_inode_pages(inode->i_mapping, new_i_size); 449 450 if (OCFS2_I(inode)->ip_dyn_features & OCFS2_INLINE_DATA_FL) { 451 status = ocfs2_truncate_inline(inode, di_bh, new_i_size, 452 i_size_read(inode), 1); 453 if (status) 454 mlog_errno(status); 455 456 goto bail_unlock_sem; 457 } 458 459 /* alright, we're going to need to do a full blown alloc size 460 * change. Orphan the inode so that recovery can complete the 461 * truncate if necessary. This does the task of marking 462 * i_size. */ 463 status = ocfs2_orphan_for_truncate(osb, inode, di_bh, new_i_size); 464 if (status < 0) { 465 mlog_errno(status); 466 goto bail_unlock_sem; 467 } 468 469 status = ocfs2_prepare_truncate(osb, inode, di_bh, &tc); 470 if (status < 0) { 471 mlog_errno(status); 472 goto bail_unlock_sem; 473 } 474 475 status = ocfs2_commit_truncate(osb, inode, di_bh, tc); 476 if (status < 0) { 477 mlog_errno(status); 478 goto bail_unlock_sem; 479 } 480 481 /* TODO: orphan dir cleanup here. */ 482 bail_unlock_sem: 483 up_write(&OCFS2_I(inode)->ip_alloc_sem); 484 485 bail: 486 487 mlog_exit(status); 488 return status; 489 } 490 491 /* 492 * extend file allocation only here. 493 * we'll update all the disk stuff, and oip->alloc_size 494 * 495 * expect stuff to be locked, a transaction started and enough data / 496 * metadata reservations in the contexts. 497 * 498 * Will return -EAGAIN, and a reason if a restart is needed. 499 * If passed in, *reason will always be set, even in error. 500 */ 501 int ocfs2_add_inode_data(struct ocfs2_super *osb, 502 struct inode *inode, 503 u32 *logical_offset, 504 u32 clusters_to_add, 505 int mark_unwritten, 506 struct buffer_head *fe_bh, 507 handle_t *handle, 508 struct ocfs2_alloc_context *data_ac, 509 struct ocfs2_alloc_context *meta_ac, 510 enum ocfs2_alloc_restarted *reason_ret) 511 { 512 int ret; 513 struct ocfs2_extent_tree et; 514 515 ocfs2_init_dinode_extent_tree(&et, inode, fe_bh); 516 ret = ocfs2_add_clusters_in_btree(osb, inode, logical_offset, 517 clusters_to_add, mark_unwritten, 518 &et, handle, 519 data_ac, meta_ac, reason_ret); 520 521 return ret; 522 } 523 524 static int __ocfs2_extend_allocation(struct inode *inode, u32 logical_start, 525 u32 clusters_to_add, int mark_unwritten) 526 { 527 int status = 0; 528 int restart_func = 0; 529 int credits; 530 u32 prev_clusters; 531 struct buffer_head *bh = NULL; 532 struct ocfs2_dinode *fe = NULL; 533 handle_t *handle = NULL; 534 struct ocfs2_alloc_context *data_ac = NULL; 535 struct ocfs2_alloc_context *meta_ac = NULL; 536 enum ocfs2_alloc_restarted why; 537 struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); 538 struct ocfs2_extent_tree et; 539 540 mlog_entry("(clusters_to_add = %u)\n", clusters_to_add); 541 542 /* 543 * This function only exists for file systems which don't 544 * support holes. 545 */ 546 BUG_ON(mark_unwritten && !ocfs2_sparse_alloc(osb)); 547 548 status = ocfs2_read_block(inode, OCFS2_I(inode)->ip_blkno, &bh); 549 if (status < 0) { 550 mlog_errno(status); 551 goto leave; 552 } 553 554 fe = (struct ocfs2_dinode *) bh->b_data; 555 if (!OCFS2_IS_VALID_DINODE(fe)) { 556 OCFS2_RO_ON_INVALID_DINODE(inode->i_sb, fe); 557 status = -EIO; 558 goto leave; 559 } 560 561 restart_all: 562 BUG_ON(le32_to_cpu(fe->i_clusters) != OCFS2_I(inode)->ip_clusters); 563 564 mlog(0, "extend inode %llu, i_size = %lld, di->i_clusters = %u, " 565 "clusters_to_add = %u\n", 566 (unsigned long long)OCFS2_I(inode)->ip_blkno, 567 (long long)i_size_read(inode), le32_to_cpu(fe->i_clusters), 568 clusters_to_add); 569 ocfs2_init_dinode_extent_tree(&et, inode, bh); 570 status = ocfs2_lock_allocators(inode, &et, clusters_to_add, 0, 571 &data_ac, &meta_ac); 572 if (status) { 573 mlog_errno(status); 574 goto leave; 575 } 576 577 credits = ocfs2_calc_extend_credits(osb->sb, &fe->id2.i_list, 578 clusters_to_add); 579 handle = ocfs2_start_trans(osb, credits); 580 if (IS_ERR(handle)) { 581 status = PTR_ERR(handle); 582 handle = NULL; 583 mlog_errno(status); 584 goto leave; 585 } 586 587 restarted_transaction: 588 /* reserve a write to the file entry early on - that we if we 589 * run out of credits in the allocation path, we can still 590 * update i_size. */ 591 status = ocfs2_journal_access(handle, inode, bh, 592 OCFS2_JOURNAL_ACCESS_WRITE); 593 if (status < 0) { 594 mlog_errno(status); 595 goto leave; 596 } 597 598 prev_clusters = OCFS2_I(inode)->ip_clusters; 599 600 status = ocfs2_add_inode_data(osb, 601 inode, 602 &logical_start, 603 clusters_to_add, 604 mark_unwritten, 605 bh, 606 handle, 607 data_ac, 608 meta_ac, 609 &why); 610 if ((status < 0) && (status != -EAGAIN)) { 611 if (status != -ENOSPC) 612 mlog_errno(status); 613 goto leave; 614 } 615 616 status = ocfs2_journal_dirty(handle, bh); 617 if (status < 0) { 618 mlog_errno(status); 619 goto leave; 620 } 621 622 spin_lock(&OCFS2_I(inode)->ip_lock); 623 clusters_to_add -= (OCFS2_I(inode)->ip_clusters - prev_clusters); 624 spin_unlock(&OCFS2_I(inode)->ip_lock); 625 626 if (why != RESTART_NONE && clusters_to_add) { 627 if (why == RESTART_META) { 628 mlog(0, "restarting function.\n"); 629 restart_func = 1; 630 } else { 631 BUG_ON(why != RESTART_TRANS); 632 633 mlog(0, "restarting transaction.\n"); 634 /* TODO: This can be more intelligent. */ 635 credits = ocfs2_calc_extend_credits(osb->sb, 636 &fe->id2.i_list, 637 clusters_to_add); 638 status = ocfs2_extend_trans(handle, credits); 639 if (status < 0) { 640 /* handle still has to be committed at 641 * this point. */ 642 status = -ENOMEM; 643 mlog_errno(status); 644 goto leave; 645 } 646 goto restarted_transaction; 647 } 648 } 649 650 mlog(0, "fe: i_clusters = %u, i_size=%llu\n", 651 le32_to_cpu(fe->i_clusters), 652 (unsigned long long)le64_to_cpu(fe->i_size)); 653 mlog(0, "inode: ip_clusters=%u, i_size=%lld\n", 654 OCFS2_I(inode)->ip_clusters, (long long)i_size_read(inode)); 655 656 leave: 657 if (handle) { 658 ocfs2_commit_trans(osb, handle); 659 handle = NULL; 660 } 661 if (data_ac) { 662 ocfs2_free_alloc_context(data_ac); 663 data_ac = NULL; 664 } 665 if (meta_ac) { 666 ocfs2_free_alloc_context(meta_ac); 667 meta_ac = NULL; 668 } 669 if ((!status) && restart_func) { 670 restart_func = 0; 671 goto restart_all; 672 } 673 brelse(bh); 674 bh = NULL; 675 676 mlog_exit(status); 677 return status; 678 } 679 680 /* Some parts of this taken from generic_cont_expand, which turned out 681 * to be too fragile to do exactly what we need without us having to 682 * worry about recursive locking in ->prepare_write() and 683 * ->commit_write(). */ 684 static int ocfs2_write_zero_page(struct inode *inode, 685 u64 size) 686 { 687 struct address_space *mapping = inode->i_mapping; 688 struct page *page; 689 unsigned long index; 690 unsigned int offset; 691 handle_t *handle = NULL; 692 int ret; 693 694 offset = (size & (PAGE_CACHE_SIZE-1)); /* Within page */ 695 /* ugh. in prepare/commit_write, if from==to==start of block, we 696 ** skip the prepare. make sure we never send an offset for the start 697 ** of a block 698 */ 699 if ((offset & (inode->i_sb->s_blocksize - 1)) == 0) { 700 offset++; 701 } 702 index = size >> PAGE_CACHE_SHIFT; 703 704 page = grab_cache_page(mapping, index); 705 if (!page) { 706 ret = -ENOMEM; 707 mlog_errno(ret); 708 goto out; 709 } 710 711 ret = ocfs2_prepare_write_nolock(inode, page, offset, offset); 712 if (ret < 0) { 713 mlog_errno(ret); 714 goto out_unlock; 715 } 716 717 if (ocfs2_should_order_data(inode)) { 718 handle = ocfs2_start_walk_page_trans(inode, page, offset, 719 offset); 720 if (IS_ERR(handle)) { 721 ret = PTR_ERR(handle); 722 handle = NULL; 723 goto out_unlock; 724 } 725 } 726 727 /* must not update i_size! */ 728 ret = block_commit_write(page, offset, offset); 729 if (ret < 0) 730 mlog_errno(ret); 731 else 732 ret = 0; 733 734 if (handle) 735 ocfs2_commit_trans(OCFS2_SB(inode->i_sb), handle); 736 out_unlock: 737 unlock_page(page); 738 page_cache_release(page); 739 out: 740 return ret; 741 } 742 743 static int ocfs2_zero_extend(struct inode *inode, 744 u64 zero_to_size) 745 { 746 int ret = 0; 747 u64 start_off; 748 struct super_block *sb = inode->i_sb; 749 750 start_off = ocfs2_align_bytes_to_blocks(sb, i_size_read(inode)); 751 while (start_off < zero_to_size) { 752 ret = ocfs2_write_zero_page(inode, start_off); 753 if (ret < 0) { 754 mlog_errno(ret); 755 goto out; 756 } 757 758 start_off += sb->s_blocksize; 759 760 /* 761 * Very large extends have the potential to lock up 762 * the cpu for extended periods of time. 763 */ 764 cond_resched(); 765 } 766 767 out: 768 return ret; 769 } 770 771 int ocfs2_extend_no_holes(struct inode *inode, u64 new_i_size, u64 zero_to) 772 { 773 int ret; 774 u32 clusters_to_add; 775 struct ocfs2_inode_info *oi = OCFS2_I(inode); 776 777 clusters_to_add = ocfs2_clusters_for_bytes(inode->i_sb, new_i_size); 778 if (clusters_to_add < oi->ip_clusters) 779 clusters_to_add = 0; 780 else 781 clusters_to_add -= oi->ip_clusters; 782 783 if (clusters_to_add) { 784 ret = __ocfs2_extend_allocation(inode, oi->ip_clusters, 785 clusters_to_add, 0); 786 if (ret) { 787 mlog_errno(ret); 788 goto out; 789 } 790 } 791 792 /* 793 * Call this even if we don't add any clusters to the tree. We 794 * still need to zero the area between the old i_size and the 795 * new i_size. 796 */ 797 ret = ocfs2_zero_extend(inode, zero_to); 798 if (ret < 0) 799 mlog_errno(ret); 800 801 out: 802 return ret; 803 } 804 805 static int ocfs2_extend_file(struct inode *inode, 806 struct buffer_head *di_bh, 807 u64 new_i_size) 808 { 809 int ret = 0; 810 struct ocfs2_inode_info *oi = OCFS2_I(inode); 811 812 BUG_ON(!di_bh); 813 814 /* setattr sometimes calls us like this. */ 815 if (new_i_size == 0) 816 goto out; 817 818 if (i_size_read(inode) == new_i_size) 819 goto out; 820 BUG_ON(new_i_size < i_size_read(inode)); 821 822 /* 823 * Fall through for converting inline data, even if the fs 824 * supports sparse files. 825 * 826 * The check for inline data here is legal - nobody can add 827 * the feature since we have i_mutex. We must check it again 828 * after acquiring ip_alloc_sem though, as paths like mmap 829 * might have raced us to converting the inode to extents. 830 */ 831 if (!(oi->ip_dyn_features & OCFS2_INLINE_DATA_FL) 832 && ocfs2_sparse_alloc(OCFS2_SB(inode->i_sb))) 833 goto out_update_size; 834 835 /* 836 * The alloc sem blocks people in read/write from reading our 837 * allocation until we're done changing it. We depend on 838 * i_mutex to block other extend/truncate calls while we're 839 * here. 840 */ 841 down_write(&oi->ip_alloc_sem); 842 843 if (oi->ip_dyn_features & OCFS2_INLINE_DATA_FL) { 844 /* 845 * We can optimize small extends by keeping the inodes 846 * inline data. 847 */ 848 if (ocfs2_size_fits_inline_data(di_bh, new_i_size)) { 849 up_write(&oi->ip_alloc_sem); 850 goto out_update_size; 851 } 852 853 ret = ocfs2_convert_inline_data_to_extents(inode, di_bh); 854 if (ret) { 855 up_write(&oi->ip_alloc_sem); 856 857 mlog_errno(ret); 858 goto out; 859 } 860 } 861 862 if (!ocfs2_sparse_alloc(OCFS2_SB(inode->i_sb))) 863 ret = ocfs2_extend_no_holes(inode, new_i_size, new_i_size); 864 865 up_write(&oi->ip_alloc_sem); 866 867 if (ret < 0) { 868 mlog_errno(ret); 869 goto out; 870 } 871 872 out_update_size: 873 ret = ocfs2_simple_size_update(inode, di_bh, new_i_size); 874 if (ret < 0) 875 mlog_errno(ret); 876 877 out: 878 return ret; 879 } 880 881 int ocfs2_setattr(struct dentry *dentry, struct iattr *attr) 882 { 883 int status = 0, size_change; 884 struct inode *inode = dentry->d_inode; 885 struct super_block *sb = inode->i_sb; 886 struct ocfs2_super *osb = OCFS2_SB(sb); 887 struct buffer_head *bh = NULL; 888 handle_t *handle = NULL; 889 890 mlog_entry("(0x%p, '%.*s')\n", dentry, 891 dentry->d_name.len, dentry->d_name.name); 892 893 /* ensuring we don't even attempt to truncate a symlink */ 894 if (S_ISLNK(inode->i_mode)) 895 attr->ia_valid &= ~ATTR_SIZE; 896 897 if (attr->ia_valid & ATTR_MODE) 898 mlog(0, "mode change: %d\n", attr->ia_mode); 899 if (attr->ia_valid & ATTR_UID) 900 mlog(0, "uid change: %d\n", attr->ia_uid); 901 if (attr->ia_valid & ATTR_GID) 902 mlog(0, "gid change: %d\n", attr->ia_gid); 903 if (attr->ia_valid & ATTR_SIZE) 904 mlog(0, "size change...\n"); 905 if (attr->ia_valid & (ATTR_ATIME | ATTR_MTIME | ATTR_CTIME)) 906 mlog(0, "time change...\n"); 907 908 #define OCFS2_VALID_ATTRS (ATTR_ATIME | ATTR_MTIME | ATTR_CTIME | ATTR_SIZE \ 909 | ATTR_GID | ATTR_UID | ATTR_MODE) 910 if (!(attr->ia_valid & OCFS2_VALID_ATTRS)) { 911 mlog(0, "can't handle attrs: 0x%x\n", attr->ia_valid); 912 return 0; 913 } 914 915 status = inode_change_ok(inode, attr); 916 if (status) 917 return status; 918 919 size_change = S_ISREG(inode->i_mode) && attr->ia_valid & ATTR_SIZE; 920 if (size_change) { 921 status = ocfs2_rw_lock(inode, 1); 922 if (status < 0) { 923 mlog_errno(status); 924 goto bail; 925 } 926 } 927 928 status = ocfs2_inode_lock(inode, &bh, 1); 929 if (status < 0) { 930 if (status != -ENOENT) 931 mlog_errno(status); 932 goto bail_unlock_rw; 933 } 934 935 if (size_change && attr->ia_size != i_size_read(inode)) { 936 if (attr->ia_size > sb->s_maxbytes) { 937 status = -EFBIG; 938 goto bail_unlock; 939 } 940 941 if (i_size_read(inode) > attr->ia_size) { 942 if (ocfs2_should_order_data(inode)) { 943 status = ocfs2_begin_ordered_truncate(inode, 944 attr->ia_size); 945 if (status) 946 goto bail_unlock; 947 } 948 status = ocfs2_truncate_file(inode, bh, attr->ia_size); 949 } else 950 status = ocfs2_extend_file(inode, bh, attr->ia_size); 951 if (status < 0) { 952 if (status != -ENOSPC) 953 mlog_errno(status); 954 status = -ENOSPC; 955 goto bail_unlock; 956 } 957 } 958 959 handle = ocfs2_start_trans(osb, OCFS2_INODE_UPDATE_CREDITS); 960 if (IS_ERR(handle)) { 961 status = PTR_ERR(handle); 962 mlog_errno(status); 963 goto bail_unlock; 964 } 965 966 /* 967 * This will intentionally not wind up calling vmtruncate(), 968 * since all the work for a size change has been done above. 969 * Otherwise, we could get into problems with truncate as 970 * ip_alloc_sem is used there to protect against i_size 971 * changes. 972 */ 973 status = inode_setattr(inode, attr); 974 if (status < 0) { 975 mlog_errno(status); 976 goto bail_commit; 977 } 978 979 status = ocfs2_mark_inode_dirty(handle, inode, bh); 980 if (status < 0) 981 mlog_errno(status); 982 983 bail_commit: 984 ocfs2_commit_trans(osb, handle); 985 bail_unlock: 986 ocfs2_inode_unlock(inode, 1); 987 bail_unlock_rw: 988 if (size_change) 989 ocfs2_rw_unlock(inode, 1); 990 bail: 991 brelse(bh); 992 993 mlog_exit(status); 994 return status; 995 } 996 997 int ocfs2_getattr(struct vfsmount *mnt, 998 struct dentry *dentry, 999 struct kstat *stat) 1000 { 1001 struct inode *inode = dentry->d_inode; 1002 struct super_block *sb = dentry->d_inode->i_sb; 1003 struct ocfs2_super *osb = sb->s_fs_info; 1004 int err; 1005 1006 mlog_entry_void(); 1007 1008 err = ocfs2_inode_revalidate(dentry); 1009 if (err) { 1010 if (err != -ENOENT) 1011 mlog_errno(err); 1012 goto bail; 1013 } 1014 1015 generic_fillattr(inode, stat); 1016 1017 /* We set the blksize from the cluster size for performance */ 1018 stat->blksize = osb->s_clustersize; 1019 1020 bail: 1021 mlog_exit(err); 1022 1023 return err; 1024 } 1025 1026 int ocfs2_permission(struct inode *inode, int mask) 1027 { 1028 int ret; 1029 1030 mlog_entry_void(); 1031 1032 ret = ocfs2_inode_lock(inode, NULL, 0); 1033 if (ret) { 1034 if (ret != -ENOENT) 1035 mlog_errno(ret); 1036 goto out; 1037 } 1038 1039 ret = generic_permission(inode, mask, NULL); 1040 1041 ocfs2_inode_unlock(inode, 0); 1042 out: 1043 mlog_exit(ret); 1044 return ret; 1045 } 1046 1047 static int __ocfs2_write_remove_suid(struct inode *inode, 1048 struct buffer_head *bh) 1049 { 1050 int ret; 1051 handle_t *handle; 1052 struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); 1053 struct ocfs2_dinode *di; 1054 1055 mlog_entry("(Inode %llu, mode 0%o)\n", 1056 (unsigned long long)OCFS2_I(inode)->ip_blkno, inode->i_mode); 1057 1058 handle = ocfs2_start_trans(osb, OCFS2_INODE_UPDATE_CREDITS); 1059 if (handle == NULL) { 1060 ret = -ENOMEM; 1061 mlog_errno(ret); 1062 goto out; 1063 } 1064 1065 ret = ocfs2_journal_access(handle, inode, bh, 1066 OCFS2_JOURNAL_ACCESS_WRITE); 1067 if (ret < 0) { 1068 mlog_errno(ret); 1069 goto out_trans; 1070 } 1071 1072 inode->i_mode &= ~S_ISUID; 1073 if ((inode->i_mode & S_ISGID) && (inode->i_mode & S_IXGRP)) 1074 inode->i_mode &= ~S_ISGID; 1075 1076 di = (struct ocfs2_dinode *) bh->b_data; 1077 di->i_mode = cpu_to_le16(inode->i_mode); 1078 1079 ret = ocfs2_journal_dirty(handle, bh); 1080 if (ret < 0) 1081 mlog_errno(ret); 1082 1083 out_trans: 1084 ocfs2_commit_trans(osb, handle); 1085 out: 1086 mlog_exit(ret); 1087 return ret; 1088 } 1089 1090 /* 1091 * Will look for holes and unwritten extents in the range starting at 1092 * pos for count bytes (inclusive). 1093 */ 1094 static int ocfs2_check_range_for_holes(struct inode *inode, loff_t pos, 1095 size_t count) 1096 { 1097 int ret = 0; 1098 unsigned int extent_flags; 1099 u32 cpos, clusters, extent_len, phys_cpos; 1100 struct super_block *sb = inode->i_sb; 1101 1102 cpos = pos >> OCFS2_SB(sb)->s_clustersize_bits; 1103 clusters = ocfs2_clusters_for_bytes(sb, pos + count) - cpos; 1104 1105 while (clusters) { 1106 ret = ocfs2_get_clusters(inode, cpos, &phys_cpos, &extent_len, 1107 &extent_flags); 1108 if (ret < 0) { 1109 mlog_errno(ret); 1110 goto out; 1111 } 1112 1113 if (phys_cpos == 0 || (extent_flags & OCFS2_EXT_UNWRITTEN)) { 1114 ret = 1; 1115 break; 1116 } 1117 1118 if (extent_len > clusters) 1119 extent_len = clusters; 1120 1121 clusters -= extent_len; 1122 cpos += extent_len; 1123 } 1124 out: 1125 return ret; 1126 } 1127 1128 static int ocfs2_write_remove_suid(struct inode *inode) 1129 { 1130 int ret; 1131 struct buffer_head *bh = NULL; 1132 struct ocfs2_inode_info *oi = OCFS2_I(inode); 1133 1134 ret = ocfs2_read_block(inode, oi->ip_blkno, &bh); 1135 if (ret < 0) { 1136 mlog_errno(ret); 1137 goto out; 1138 } 1139 1140 ret = __ocfs2_write_remove_suid(inode, bh); 1141 out: 1142 brelse(bh); 1143 return ret; 1144 } 1145 1146 /* 1147 * Allocate enough extents to cover the region starting at byte offset 1148 * start for len bytes. Existing extents are skipped, any extents 1149 * added are marked as "unwritten". 1150 */ 1151 static int ocfs2_allocate_unwritten_extents(struct inode *inode, 1152 u64 start, u64 len) 1153 { 1154 int ret; 1155 u32 cpos, phys_cpos, clusters, alloc_size; 1156 u64 end = start + len; 1157 struct buffer_head *di_bh = NULL; 1158 1159 if (OCFS2_I(inode)->ip_dyn_features & OCFS2_INLINE_DATA_FL) { 1160 ret = ocfs2_read_block(inode, OCFS2_I(inode)->ip_blkno, 1161 &di_bh); 1162 if (ret) { 1163 mlog_errno(ret); 1164 goto out; 1165 } 1166 1167 /* 1168 * Nothing to do if the requested reservation range 1169 * fits within the inode. 1170 */ 1171 if (ocfs2_size_fits_inline_data(di_bh, end)) 1172 goto out; 1173 1174 ret = ocfs2_convert_inline_data_to_extents(inode, di_bh); 1175 if (ret) { 1176 mlog_errno(ret); 1177 goto out; 1178 } 1179 } 1180 1181 /* 1182 * We consider both start and len to be inclusive. 1183 */ 1184 cpos = start >> OCFS2_SB(inode->i_sb)->s_clustersize_bits; 1185 clusters = ocfs2_clusters_for_bytes(inode->i_sb, start + len); 1186 clusters -= cpos; 1187 1188 while (clusters) { 1189 ret = ocfs2_get_clusters(inode, cpos, &phys_cpos, 1190 &alloc_size, NULL); 1191 if (ret) { 1192 mlog_errno(ret); 1193 goto out; 1194 } 1195 1196 /* 1197 * Hole or existing extent len can be arbitrary, so 1198 * cap it to our own allocation request. 1199 */ 1200 if (alloc_size > clusters) 1201 alloc_size = clusters; 1202 1203 if (phys_cpos) { 1204 /* 1205 * We already have an allocation at this 1206 * region so we can safely skip it. 1207 */ 1208 goto next; 1209 } 1210 1211 ret = __ocfs2_extend_allocation(inode, cpos, alloc_size, 1); 1212 if (ret) { 1213 if (ret != -ENOSPC) 1214 mlog_errno(ret); 1215 goto out; 1216 } 1217 1218 next: 1219 cpos += alloc_size; 1220 clusters -= alloc_size; 1221 } 1222 1223 ret = 0; 1224 out: 1225 1226 brelse(di_bh); 1227 return ret; 1228 } 1229 1230 static int __ocfs2_remove_inode_range(struct inode *inode, 1231 struct buffer_head *di_bh, 1232 u32 cpos, u32 phys_cpos, u32 len, 1233 struct ocfs2_cached_dealloc_ctxt *dealloc) 1234 { 1235 int ret; 1236 u64 phys_blkno = ocfs2_clusters_to_blocks(inode->i_sb, phys_cpos); 1237 struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); 1238 struct inode *tl_inode = osb->osb_tl_inode; 1239 handle_t *handle; 1240 struct ocfs2_alloc_context *meta_ac = NULL; 1241 struct ocfs2_dinode *di = (struct ocfs2_dinode *)di_bh->b_data; 1242 struct ocfs2_extent_tree et; 1243 1244 ocfs2_init_dinode_extent_tree(&et, inode, di_bh); 1245 1246 ret = ocfs2_lock_allocators(inode, &et, 0, 1, NULL, &meta_ac); 1247 if (ret) { 1248 mlog_errno(ret); 1249 return ret; 1250 } 1251 1252 mutex_lock(&tl_inode->i_mutex); 1253 1254 if (ocfs2_truncate_log_needs_flush(osb)) { 1255 ret = __ocfs2_flush_truncate_log(osb); 1256 if (ret < 0) { 1257 mlog_errno(ret); 1258 goto out; 1259 } 1260 } 1261 1262 handle = ocfs2_start_trans(osb, OCFS2_REMOVE_EXTENT_CREDITS); 1263 if (handle == NULL) { 1264 ret = -ENOMEM; 1265 mlog_errno(ret); 1266 goto out; 1267 } 1268 1269 ret = ocfs2_journal_access(handle, inode, di_bh, 1270 OCFS2_JOURNAL_ACCESS_WRITE); 1271 if (ret) { 1272 mlog_errno(ret); 1273 goto out; 1274 } 1275 1276 ret = ocfs2_remove_extent(inode, &et, cpos, len, handle, meta_ac, 1277 dealloc); 1278 if (ret) { 1279 mlog_errno(ret); 1280 goto out_commit; 1281 } 1282 1283 OCFS2_I(inode)->ip_clusters -= len; 1284 di->i_clusters = cpu_to_le32(OCFS2_I(inode)->ip_clusters); 1285 1286 ret = ocfs2_journal_dirty(handle, di_bh); 1287 if (ret) { 1288 mlog_errno(ret); 1289 goto out_commit; 1290 } 1291 1292 ret = ocfs2_truncate_log_append(osb, handle, phys_blkno, len); 1293 if (ret) 1294 mlog_errno(ret); 1295 1296 out_commit: 1297 ocfs2_commit_trans(osb, handle); 1298 out: 1299 mutex_unlock(&tl_inode->i_mutex); 1300 1301 if (meta_ac) 1302 ocfs2_free_alloc_context(meta_ac); 1303 1304 return ret; 1305 } 1306 1307 /* 1308 * Truncate a byte range, avoiding pages within partial clusters. This 1309 * preserves those pages for the zeroing code to write to. 1310 */ 1311 static void ocfs2_truncate_cluster_pages(struct inode *inode, u64 byte_start, 1312 u64 byte_len) 1313 { 1314 struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); 1315 loff_t start, end; 1316 struct address_space *mapping = inode->i_mapping; 1317 1318 start = (loff_t)ocfs2_align_bytes_to_clusters(inode->i_sb, byte_start); 1319 end = byte_start + byte_len; 1320 end = end & ~(osb->s_clustersize - 1); 1321 1322 if (start < end) { 1323 unmap_mapping_range(mapping, start, end - start, 0); 1324 truncate_inode_pages_range(mapping, start, end - 1); 1325 } 1326 } 1327 1328 static int ocfs2_zero_partial_clusters(struct inode *inode, 1329 u64 start, u64 len) 1330 { 1331 int ret = 0; 1332 u64 tmpend, end = start + len; 1333 struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); 1334 unsigned int csize = osb->s_clustersize; 1335 handle_t *handle; 1336 1337 /* 1338 * The "start" and "end" values are NOT necessarily part of 1339 * the range whose allocation is being deleted. Rather, this 1340 * is what the user passed in with the request. We must zero 1341 * partial clusters here. There's no need to worry about 1342 * physical allocation - the zeroing code knows to skip holes. 1343 */ 1344 mlog(0, "byte start: %llu, end: %llu\n", 1345 (unsigned long long)start, (unsigned long long)end); 1346 1347 /* 1348 * If both edges are on a cluster boundary then there's no 1349 * zeroing required as the region is part of the allocation to 1350 * be truncated. 1351 */ 1352 if ((start & (csize - 1)) == 0 && (end & (csize - 1)) == 0) 1353 goto out; 1354 1355 handle = ocfs2_start_trans(osb, OCFS2_INODE_UPDATE_CREDITS); 1356 if (handle == NULL) { 1357 ret = -ENOMEM; 1358 mlog_errno(ret); 1359 goto out; 1360 } 1361 1362 /* 1363 * We want to get the byte offset of the end of the 1st cluster. 1364 */ 1365 tmpend = (u64)osb->s_clustersize + (start & ~(osb->s_clustersize - 1)); 1366 if (tmpend > end) 1367 tmpend = end; 1368 1369 mlog(0, "1st range: start: %llu, tmpend: %llu\n", 1370 (unsigned long long)start, (unsigned long long)tmpend); 1371 1372 ret = ocfs2_zero_range_for_truncate(inode, handle, start, tmpend); 1373 if (ret) 1374 mlog_errno(ret); 1375 1376 if (tmpend < end) { 1377 /* 1378 * This may make start and end equal, but the zeroing 1379 * code will skip any work in that case so there's no 1380 * need to catch it up here. 1381 */ 1382 start = end & ~(osb->s_clustersize - 1); 1383 1384 mlog(0, "2nd range: start: %llu, end: %llu\n", 1385 (unsigned long long)start, (unsigned long long)end); 1386 1387 ret = ocfs2_zero_range_for_truncate(inode, handle, start, end); 1388 if (ret) 1389 mlog_errno(ret); 1390 } 1391 1392 ocfs2_commit_trans(osb, handle); 1393 out: 1394 return ret; 1395 } 1396 1397 static int ocfs2_remove_inode_range(struct inode *inode, 1398 struct buffer_head *di_bh, u64 byte_start, 1399 u64 byte_len) 1400 { 1401 int ret = 0; 1402 u32 trunc_start, trunc_len, cpos, phys_cpos, alloc_size; 1403 struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); 1404 struct ocfs2_cached_dealloc_ctxt dealloc; 1405 struct address_space *mapping = inode->i_mapping; 1406 1407 ocfs2_init_dealloc_ctxt(&dealloc); 1408 1409 if (byte_len == 0) 1410 return 0; 1411 1412 if (OCFS2_I(inode)->ip_dyn_features & OCFS2_INLINE_DATA_FL) { 1413 ret = ocfs2_truncate_inline(inode, di_bh, byte_start, 1414 byte_start + byte_len, 0); 1415 if (ret) { 1416 mlog_errno(ret); 1417 goto out; 1418 } 1419 /* 1420 * There's no need to get fancy with the page cache 1421 * truncate of an inline-data inode. We're talking 1422 * about less than a page here, which will be cached 1423 * in the dinode buffer anyway. 1424 */ 1425 unmap_mapping_range(mapping, 0, 0, 0); 1426 truncate_inode_pages(mapping, 0); 1427 goto out; 1428 } 1429 1430 trunc_start = ocfs2_clusters_for_bytes(osb->sb, byte_start); 1431 trunc_len = (byte_start + byte_len) >> osb->s_clustersize_bits; 1432 if (trunc_len >= trunc_start) 1433 trunc_len -= trunc_start; 1434 else 1435 trunc_len = 0; 1436 1437 mlog(0, "Inode: %llu, start: %llu, len: %llu, cstart: %u, clen: %u\n", 1438 (unsigned long long)OCFS2_I(inode)->ip_blkno, 1439 (unsigned long long)byte_start, 1440 (unsigned long long)byte_len, trunc_start, trunc_len); 1441 1442 ret = ocfs2_zero_partial_clusters(inode, byte_start, byte_len); 1443 if (ret) { 1444 mlog_errno(ret); 1445 goto out; 1446 } 1447 1448 cpos = trunc_start; 1449 while (trunc_len) { 1450 ret = ocfs2_get_clusters(inode, cpos, &phys_cpos, 1451 &alloc_size, NULL); 1452 if (ret) { 1453 mlog_errno(ret); 1454 goto out; 1455 } 1456 1457 if (alloc_size > trunc_len) 1458 alloc_size = trunc_len; 1459 1460 /* Only do work for non-holes */ 1461 if (phys_cpos != 0) { 1462 ret = __ocfs2_remove_inode_range(inode, di_bh, cpos, 1463 phys_cpos, alloc_size, 1464 &dealloc); 1465 if (ret) { 1466 mlog_errno(ret); 1467 goto out; 1468 } 1469 } 1470 1471 cpos += alloc_size; 1472 trunc_len -= alloc_size; 1473 } 1474 1475 ocfs2_truncate_cluster_pages(inode, byte_start, byte_len); 1476 1477 out: 1478 ocfs2_schedule_truncate_log_flush(osb, 1); 1479 ocfs2_run_deallocs(osb, &dealloc); 1480 1481 return ret; 1482 } 1483 1484 /* 1485 * Parts of this function taken from xfs_change_file_space() 1486 */ 1487 static int __ocfs2_change_file_space(struct file *file, struct inode *inode, 1488 loff_t f_pos, unsigned int cmd, 1489 struct ocfs2_space_resv *sr, 1490 int change_size) 1491 { 1492 int ret; 1493 s64 llen; 1494 loff_t size; 1495 struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); 1496 struct buffer_head *di_bh = NULL; 1497 handle_t *handle; 1498 unsigned long long max_off = inode->i_sb->s_maxbytes; 1499 1500 if (ocfs2_is_hard_readonly(osb) || ocfs2_is_soft_readonly(osb)) 1501 return -EROFS; 1502 1503 mutex_lock(&inode->i_mutex); 1504 1505 /* 1506 * This prevents concurrent writes on other nodes 1507 */ 1508 ret = ocfs2_rw_lock(inode, 1); 1509 if (ret) { 1510 mlog_errno(ret); 1511 goto out; 1512 } 1513 1514 ret = ocfs2_inode_lock(inode, &di_bh, 1); 1515 if (ret) { 1516 mlog_errno(ret); 1517 goto out_rw_unlock; 1518 } 1519 1520 if (inode->i_flags & (S_IMMUTABLE|S_APPEND)) { 1521 ret = -EPERM; 1522 goto out_inode_unlock; 1523 } 1524 1525 switch (sr->l_whence) { 1526 case 0: /*SEEK_SET*/ 1527 break; 1528 case 1: /*SEEK_CUR*/ 1529 sr->l_start += f_pos; 1530 break; 1531 case 2: /*SEEK_END*/ 1532 sr->l_start += i_size_read(inode); 1533 break; 1534 default: 1535 ret = -EINVAL; 1536 goto out_inode_unlock; 1537 } 1538 sr->l_whence = 0; 1539 1540 llen = sr->l_len > 0 ? sr->l_len - 1 : sr->l_len; 1541 1542 if (sr->l_start < 0 1543 || sr->l_start > max_off 1544 || (sr->l_start + llen) < 0 1545 || (sr->l_start + llen) > max_off) { 1546 ret = -EINVAL; 1547 goto out_inode_unlock; 1548 } 1549 size = sr->l_start + sr->l_len; 1550 1551 if (cmd == OCFS2_IOC_RESVSP || cmd == OCFS2_IOC_RESVSP64) { 1552 if (sr->l_len <= 0) { 1553 ret = -EINVAL; 1554 goto out_inode_unlock; 1555 } 1556 } 1557 1558 if (file && should_remove_suid(file->f_path.dentry)) { 1559 ret = __ocfs2_write_remove_suid(inode, di_bh); 1560 if (ret) { 1561 mlog_errno(ret); 1562 goto out_inode_unlock; 1563 } 1564 } 1565 1566 down_write(&OCFS2_I(inode)->ip_alloc_sem); 1567 switch (cmd) { 1568 case OCFS2_IOC_RESVSP: 1569 case OCFS2_IOC_RESVSP64: 1570 /* 1571 * This takes unsigned offsets, but the signed ones we 1572 * pass have been checked against overflow above. 1573 */ 1574 ret = ocfs2_allocate_unwritten_extents(inode, sr->l_start, 1575 sr->l_len); 1576 break; 1577 case OCFS2_IOC_UNRESVSP: 1578 case OCFS2_IOC_UNRESVSP64: 1579 ret = ocfs2_remove_inode_range(inode, di_bh, sr->l_start, 1580 sr->l_len); 1581 break; 1582 default: 1583 ret = -EINVAL; 1584 } 1585 up_write(&OCFS2_I(inode)->ip_alloc_sem); 1586 if (ret) { 1587 mlog_errno(ret); 1588 goto out_inode_unlock; 1589 } 1590 1591 /* 1592 * We update c/mtime for these changes 1593 */ 1594 handle = ocfs2_start_trans(osb, OCFS2_INODE_UPDATE_CREDITS); 1595 if (IS_ERR(handle)) { 1596 ret = PTR_ERR(handle); 1597 mlog_errno(ret); 1598 goto out_inode_unlock; 1599 } 1600 1601 if (change_size && i_size_read(inode) < size) 1602 i_size_write(inode, size); 1603 1604 inode->i_ctime = inode->i_mtime = CURRENT_TIME; 1605 ret = ocfs2_mark_inode_dirty(handle, inode, di_bh); 1606 if (ret < 0) 1607 mlog_errno(ret); 1608 1609 ocfs2_commit_trans(osb, handle); 1610 1611 out_inode_unlock: 1612 brelse(di_bh); 1613 ocfs2_inode_unlock(inode, 1); 1614 out_rw_unlock: 1615 ocfs2_rw_unlock(inode, 1); 1616 1617 out: 1618 mutex_unlock(&inode->i_mutex); 1619 return ret; 1620 } 1621 1622 int ocfs2_change_file_space(struct file *file, unsigned int cmd, 1623 struct ocfs2_space_resv *sr) 1624 { 1625 struct inode *inode = file->f_path.dentry->d_inode; 1626 struct ocfs2_super *osb = OCFS2_SB(inode->i_sb);; 1627 1628 if ((cmd == OCFS2_IOC_RESVSP || cmd == OCFS2_IOC_RESVSP64) && 1629 !ocfs2_writes_unwritten_extents(osb)) 1630 return -ENOTTY; 1631 else if ((cmd == OCFS2_IOC_UNRESVSP || cmd == OCFS2_IOC_UNRESVSP64) && 1632 !ocfs2_sparse_alloc(osb)) 1633 return -ENOTTY; 1634 1635 if (!S_ISREG(inode->i_mode)) 1636 return -EINVAL; 1637 1638 if (!(file->f_mode & FMODE_WRITE)) 1639 return -EBADF; 1640 1641 return __ocfs2_change_file_space(file, inode, file->f_pos, cmd, sr, 0); 1642 } 1643 1644 static long ocfs2_fallocate(struct inode *inode, int mode, loff_t offset, 1645 loff_t len) 1646 { 1647 struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); 1648 struct ocfs2_space_resv sr; 1649 int change_size = 1; 1650 1651 if (!ocfs2_writes_unwritten_extents(osb)) 1652 return -EOPNOTSUPP; 1653 1654 if (S_ISDIR(inode->i_mode)) 1655 return -ENODEV; 1656 1657 if (mode & FALLOC_FL_KEEP_SIZE) 1658 change_size = 0; 1659 1660 sr.l_whence = 0; 1661 sr.l_start = (s64)offset; 1662 sr.l_len = (s64)len; 1663 1664 return __ocfs2_change_file_space(NULL, inode, offset, 1665 OCFS2_IOC_RESVSP64, &sr, change_size); 1666 } 1667 1668 static int ocfs2_prepare_inode_for_write(struct dentry *dentry, 1669 loff_t *ppos, 1670 size_t count, 1671 int appending, 1672 int *direct_io) 1673 { 1674 int ret = 0, meta_level = 0; 1675 struct inode *inode = dentry->d_inode; 1676 loff_t saved_pos, end; 1677 1678 /* 1679 * We start with a read level meta lock and only jump to an ex 1680 * if we need to make modifications here. 1681 */ 1682 for(;;) { 1683 ret = ocfs2_inode_lock(inode, NULL, meta_level); 1684 if (ret < 0) { 1685 meta_level = -1; 1686 mlog_errno(ret); 1687 goto out; 1688 } 1689 1690 /* Clear suid / sgid if necessary. We do this here 1691 * instead of later in the write path because 1692 * remove_suid() calls ->setattr without any hint that 1693 * we may have already done our cluster locking. Since 1694 * ocfs2_setattr() *must* take cluster locks to 1695 * proceeed, this will lead us to recursively lock the 1696 * inode. There's also the dinode i_size state which 1697 * can be lost via setattr during extending writes (we 1698 * set inode->i_size at the end of a write. */ 1699 if (should_remove_suid(dentry)) { 1700 if (meta_level == 0) { 1701 ocfs2_inode_unlock(inode, meta_level); 1702 meta_level = 1; 1703 continue; 1704 } 1705 1706 ret = ocfs2_write_remove_suid(inode); 1707 if (ret < 0) { 1708 mlog_errno(ret); 1709 goto out_unlock; 1710 } 1711 } 1712 1713 /* work on a copy of ppos until we're sure that we won't have 1714 * to recalculate it due to relocking. */ 1715 if (appending) { 1716 saved_pos = i_size_read(inode); 1717 mlog(0, "O_APPEND: inode->i_size=%llu\n", saved_pos); 1718 } else { 1719 saved_pos = *ppos; 1720 } 1721 1722 end = saved_pos + count; 1723 1724 /* 1725 * Skip the O_DIRECT checks if we don't need 1726 * them. 1727 */ 1728 if (!direct_io || !(*direct_io)) 1729 break; 1730 1731 /* 1732 * There's no sane way to do direct writes to an inode 1733 * with inline data. 1734 */ 1735 if (OCFS2_I(inode)->ip_dyn_features & OCFS2_INLINE_DATA_FL) { 1736 *direct_io = 0; 1737 break; 1738 } 1739 1740 /* 1741 * Allowing concurrent direct writes means 1742 * i_size changes wouldn't be synchronized, so 1743 * one node could wind up truncating another 1744 * nodes writes. 1745 */ 1746 if (end > i_size_read(inode)) { 1747 *direct_io = 0; 1748 break; 1749 } 1750 1751 /* 1752 * We don't fill holes during direct io, so 1753 * check for them here. If any are found, the 1754 * caller will have to retake some cluster 1755 * locks and initiate the io as buffered. 1756 */ 1757 ret = ocfs2_check_range_for_holes(inode, saved_pos, count); 1758 if (ret == 1) { 1759 *direct_io = 0; 1760 ret = 0; 1761 } else if (ret < 0) 1762 mlog_errno(ret); 1763 break; 1764 } 1765 1766 if (appending) 1767 *ppos = saved_pos; 1768 1769 out_unlock: 1770 ocfs2_inode_unlock(inode, meta_level); 1771 1772 out: 1773 return ret; 1774 } 1775 1776 static ssize_t ocfs2_file_aio_write(struct kiocb *iocb, 1777 const struct iovec *iov, 1778 unsigned long nr_segs, 1779 loff_t pos) 1780 { 1781 int ret, direct_io, appending, rw_level, have_alloc_sem = 0; 1782 int can_do_direct; 1783 ssize_t written = 0; 1784 size_t ocount; /* original count */ 1785 size_t count; /* after file limit checks */ 1786 loff_t old_size, *ppos = &iocb->ki_pos; 1787 u32 old_clusters; 1788 struct file *file = iocb->ki_filp; 1789 struct inode *inode = file->f_path.dentry->d_inode; 1790 struct ocfs2_super *osb = OCFS2_SB(inode->i_sb); 1791 1792 mlog_entry("(0x%p, %u, '%.*s')\n", file, 1793 (unsigned int)nr_segs, 1794 file->f_path.dentry->d_name.len, 1795 file->f_path.dentry->d_name.name); 1796 1797 if (iocb->ki_left == 0) 1798 return 0; 1799 1800 vfs_check_frozen(inode->i_sb, SB_FREEZE_WRITE); 1801 1802 appending = file->f_flags & O_APPEND ? 1 : 0; 1803 direct_io = file->f_flags & O_DIRECT ? 1 : 0; 1804 1805 mutex_lock(&inode->i_mutex); 1806 1807 relock: 1808 /* to match setattr's i_mutex -> i_alloc_sem -> rw_lock ordering */ 1809 if (direct_io) { 1810 down_read(&inode->i_alloc_sem); 1811 have_alloc_sem = 1; 1812 } 1813 1814 /* concurrent O_DIRECT writes are allowed */ 1815 rw_level = !direct_io; 1816 ret = ocfs2_rw_lock(inode, rw_level); 1817 if (ret < 0) { 1818 mlog_errno(ret); 1819 goto out_sems; 1820 } 1821 1822 can_do_direct = direct_io; 1823 ret = ocfs2_prepare_inode_for_write(file->f_path.dentry, ppos, 1824 iocb->ki_left, appending, 1825 &can_do_direct); 1826 if (ret < 0) { 1827 mlog_errno(ret); 1828 goto out; 1829 } 1830 1831 /* 1832 * We can't complete the direct I/O as requested, fall back to 1833 * buffered I/O. 1834 */ 1835 if (direct_io && !can_do_direct) { 1836 ocfs2_rw_unlock(inode, rw_level); 1837 up_read(&inode->i_alloc_sem); 1838 1839 have_alloc_sem = 0; 1840 rw_level = -1; 1841 1842 direct_io = 0; 1843 goto relock; 1844 } 1845 1846 /* 1847 * To later detect whether a journal commit for sync writes is 1848 * necessary, we sample i_size, and cluster count here. 1849 */ 1850 old_size = i_size_read(inode); 1851 old_clusters = OCFS2_I(inode)->ip_clusters; 1852 1853 /* communicate with ocfs2_dio_end_io */ 1854 ocfs2_iocb_set_rw_locked(iocb, rw_level); 1855 1856 if (direct_io) { 1857 ret = generic_segment_checks(iov, &nr_segs, &ocount, 1858 VERIFY_READ); 1859 if (ret) 1860 goto out_dio; 1861 1862 ret = generic_write_checks(file, ppos, &count, 1863 S_ISBLK(inode->i_mode)); 1864 if (ret) 1865 goto out_dio; 1866 1867 written = generic_file_direct_write(iocb, iov, &nr_segs, *ppos, 1868 ppos, count, ocount); 1869 if (written < 0) { 1870 ret = written; 1871 goto out_dio; 1872 } 1873 } else { 1874 written = generic_file_aio_write_nolock(iocb, iov, nr_segs, 1875 *ppos); 1876 } 1877 1878 out_dio: 1879 /* buffered aio wouldn't have proper lock coverage today */ 1880 BUG_ON(ret == -EIOCBQUEUED && !(file->f_flags & O_DIRECT)); 1881 1882 if ((file->f_flags & O_SYNC && !direct_io) || IS_SYNC(inode)) { 1883 /* 1884 * The generic write paths have handled getting data 1885 * to disk, but since we don't make use of the dirty 1886 * inode list, a manual journal commit is necessary 1887 * here. 1888 */ 1889 if (old_size != i_size_read(inode) || 1890 old_clusters != OCFS2_I(inode)->ip_clusters) { 1891 ret = jbd2_journal_force_commit(osb->journal->j_journal); 1892 if (ret < 0) 1893 written = ret; 1894 } 1895 } 1896 1897 /* 1898 * deep in g_f_a_w_n()->ocfs2_direct_IO we pass in a ocfs2_dio_end_io 1899 * function pointer which is called when o_direct io completes so that 1900 * it can unlock our rw lock. (it's the clustered equivalent of 1901 * i_alloc_sem; protects truncate from racing with pending ios). 1902 * Unfortunately there are error cases which call end_io and others 1903 * that don't. so we don't have to unlock the rw_lock if either an 1904 * async dio is going to do it in the future or an end_io after an 1905 * error has already done it. 1906 */ 1907 if (ret == -EIOCBQUEUED || !ocfs2_iocb_is_rw_locked(iocb)) { 1908 rw_level = -1; 1909 have_alloc_sem = 0; 1910 } 1911 1912 out: 1913 if (rw_level != -1) 1914 ocfs2_rw_unlock(inode, rw_level); 1915 1916 out_sems: 1917 if (have_alloc_sem) 1918 up_read(&inode->i_alloc_sem); 1919 1920 mutex_unlock(&inode->i_mutex); 1921 1922 mlog_exit(ret); 1923 return written ? written : ret; 1924 } 1925 1926 static ssize_t ocfs2_file_splice_write(struct pipe_inode_info *pipe, 1927 struct file *out, 1928 loff_t *ppos, 1929 size_t len, 1930 unsigned int flags) 1931 { 1932 int ret; 1933 struct inode *inode = out->f_path.dentry->d_inode; 1934 1935 mlog_entry("(0x%p, 0x%p, %u, '%.*s')\n", out, pipe, 1936 (unsigned int)len, 1937 out->f_path.dentry->d_name.len, 1938 out->f_path.dentry->d_name.name); 1939 1940 inode_double_lock(inode, pipe->inode); 1941 1942 ret = ocfs2_rw_lock(inode, 1); 1943 if (ret < 0) { 1944 mlog_errno(ret); 1945 goto out; 1946 } 1947 1948 ret = ocfs2_prepare_inode_for_write(out->f_path.dentry, ppos, len, 0, 1949 NULL); 1950 if (ret < 0) { 1951 mlog_errno(ret); 1952 goto out_unlock; 1953 } 1954 1955 ret = generic_file_splice_write_nolock(pipe, out, ppos, len, flags); 1956 1957 out_unlock: 1958 ocfs2_rw_unlock(inode, 1); 1959 out: 1960 inode_double_unlock(inode, pipe->inode); 1961 1962 mlog_exit(ret); 1963 return ret; 1964 } 1965 1966 static ssize_t ocfs2_file_splice_read(struct file *in, 1967 loff_t *ppos, 1968 struct pipe_inode_info *pipe, 1969 size_t len, 1970 unsigned int flags) 1971 { 1972 int ret = 0; 1973 struct inode *inode = in->f_path.dentry->d_inode; 1974 1975 mlog_entry("(0x%p, 0x%p, %u, '%.*s')\n", in, pipe, 1976 (unsigned int)len, 1977 in->f_path.dentry->d_name.len, 1978 in->f_path.dentry->d_name.name); 1979 1980 /* 1981 * See the comment in ocfs2_file_aio_read() 1982 */ 1983 ret = ocfs2_inode_lock(inode, NULL, 0); 1984 if (ret < 0) { 1985 mlog_errno(ret); 1986 goto bail; 1987 } 1988 ocfs2_inode_unlock(inode, 0); 1989 1990 ret = generic_file_splice_read(in, ppos, pipe, len, flags); 1991 1992 bail: 1993 mlog_exit(ret); 1994 return ret; 1995 } 1996 1997 static ssize_t ocfs2_file_aio_read(struct kiocb *iocb, 1998 const struct iovec *iov, 1999 unsigned long nr_segs, 2000 loff_t pos) 2001 { 2002 int ret = 0, rw_level = -1, have_alloc_sem = 0, lock_level = 0; 2003 struct file *filp = iocb->ki_filp; 2004 struct inode *inode = filp->f_path.dentry->d_inode; 2005 2006 mlog_entry("(0x%p, %u, '%.*s')\n", filp, 2007 (unsigned int)nr_segs, 2008 filp->f_path.dentry->d_name.len, 2009 filp->f_path.dentry->d_name.name); 2010 2011 if (!inode) { 2012 ret = -EINVAL; 2013 mlog_errno(ret); 2014 goto bail; 2015 } 2016 2017 /* 2018 * buffered reads protect themselves in ->readpage(). O_DIRECT reads 2019 * need locks to protect pending reads from racing with truncate. 2020 */ 2021 if (filp->f_flags & O_DIRECT) { 2022 down_read(&inode->i_alloc_sem); 2023 have_alloc_sem = 1; 2024 2025 ret = ocfs2_rw_lock(inode, 0); 2026 if (ret < 0) { 2027 mlog_errno(ret); 2028 goto bail; 2029 } 2030 rw_level = 0; 2031 /* communicate with ocfs2_dio_end_io */ 2032 ocfs2_iocb_set_rw_locked(iocb, rw_level); 2033 } 2034 2035 /* 2036 * We're fine letting folks race truncates and extending 2037 * writes with read across the cluster, just like they can 2038 * locally. Hence no rw_lock during read. 2039 * 2040 * Take and drop the meta data lock to update inode fields 2041 * like i_size. This allows the checks down below 2042 * generic_file_aio_read() a chance of actually working. 2043 */ 2044 ret = ocfs2_inode_lock_atime(inode, filp->f_vfsmnt, &lock_level); 2045 if (ret < 0) { 2046 mlog_errno(ret); 2047 goto bail; 2048 } 2049 ocfs2_inode_unlock(inode, lock_level); 2050 2051 ret = generic_file_aio_read(iocb, iov, nr_segs, iocb->ki_pos); 2052 if (ret == -EINVAL) 2053 mlog(0, "generic_file_aio_read returned -EINVAL\n"); 2054 2055 /* buffered aio wouldn't have proper lock coverage today */ 2056 BUG_ON(ret == -EIOCBQUEUED && !(filp->f_flags & O_DIRECT)); 2057 2058 /* see ocfs2_file_aio_write */ 2059 if (ret == -EIOCBQUEUED || !ocfs2_iocb_is_rw_locked(iocb)) { 2060 rw_level = -1; 2061 have_alloc_sem = 0; 2062 } 2063 2064 bail: 2065 if (have_alloc_sem) 2066 up_read(&inode->i_alloc_sem); 2067 if (rw_level != -1) 2068 ocfs2_rw_unlock(inode, rw_level); 2069 mlog_exit(ret); 2070 2071 return ret; 2072 } 2073 2074 const struct inode_operations ocfs2_file_iops = { 2075 .setattr = ocfs2_setattr, 2076 .getattr = ocfs2_getattr, 2077 .permission = ocfs2_permission, 2078 .setxattr = generic_setxattr, 2079 .getxattr = generic_getxattr, 2080 .listxattr = ocfs2_listxattr, 2081 .removexattr = generic_removexattr, 2082 .fallocate = ocfs2_fallocate, 2083 .fiemap = ocfs2_fiemap, 2084 }; 2085 2086 const struct inode_operations ocfs2_special_file_iops = { 2087 .setattr = ocfs2_setattr, 2088 .getattr = ocfs2_getattr, 2089 .permission = ocfs2_permission, 2090 }; 2091 2092 /* 2093 * Other than ->lock, keep ocfs2_fops and ocfs2_dops in sync with 2094 * ocfs2_fops_no_plocks and ocfs2_dops_no_plocks! 2095 */ 2096 const struct file_operations ocfs2_fops = { 2097 .llseek = generic_file_llseek, 2098 .read = do_sync_read, 2099 .write = do_sync_write, 2100 .mmap = ocfs2_mmap, 2101 .fsync = ocfs2_sync_file, 2102 .release = ocfs2_file_release, 2103 .open = ocfs2_file_open, 2104 .aio_read = ocfs2_file_aio_read, 2105 .aio_write = ocfs2_file_aio_write, 2106 .unlocked_ioctl = ocfs2_ioctl, 2107 #ifdef CONFIG_COMPAT 2108 .compat_ioctl = ocfs2_compat_ioctl, 2109 #endif 2110 .lock = ocfs2_lock, 2111 .flock = ocfs2_flock, 2112 .splice_read = ocfs2_file_splice_read, 2113 .splice_write = ocfs2_file_splice_write, 2114 }; 2115 2116 const struct file_operations ocfs2_dops = { 2117 .llseek = generic_file_llseek, 2118 .read = generic_read_dir, 2119 .readdir = ocfs2_readdir, 2120 .fsync = ocfs2_sync_file, 2121 .release = ocfs2_dir_release, 2122 .open = ocfs2_dir_open, 2123 .unlocked_ioctl = ocfs2_ioctl, 2124 #ifdef CONFIG_COMPAT 2125 .compat_ioctl = ocfs2_compat_ioctl, 2126 #endif 2127 .lock = ocfs2_lock, 2128 .flock = ocfs2_flock, 2129 }; 2130 2131 /* 2132 * POSIX-lockless variants of our file_operations. 2133 * 2134 * These will be used if the underlying cluster stack does not support 2135 * posix file locking, if the user passes the "localflocks" mount 2136 * option, or if we have a local-only fs. 2137 * 2138 * ocfs2_flock is in here because all stacks handle UNIX file locks, 2139 * so we still want it in the case of no stack support for 2140 * plocks. Internally, it will do the right thing when asked to ignore 2141 * the cluster. 2142 */ 2143 const struct file_operations ocfs2_fops_no_plocks = { 2144 .llseek = generic_file_llseek, 2145 .read = do_sync_read, 2146 .write = do_sync_write, 2147 .mmap = ocfs2_mmap, 2148 .fsync = ocfs2_sync_file, 2149 .release = ocfs2_file_release, 2150 .open = ocfs2_file_open, 2151 .aio_read = ocfs2_file_aio_read, 2152 .aio_write = ocfs2_file_aio_write, 2153 .unlocked_ioctl = ocfs2_ioctl, 2154 #ifdef CONFIG_COMPAT 2155 .compat_ioctl = ocfs2_compat_ioctl, 2156 #endif 2157 .flock = ocfs2_flock, 2158 .splice_read = ocfs2_file_splice_read, 2159 .splice_write = ocfs2_file_splice_write, 2160 }; 2161 2162 const struct file_operations ocfs2_dops_no_plocks = { 2163 .llseek = generic_file_llseek, 2164 .read = generic_read_dir, 2165 .readdir = ocfs2_readdir, 2166 .fsync = ocfs2_sync_file, 2167 .release = ocfs2_dir_release, 2168 .open = ocfs2_dir_open, 2169 .unlocked_ioctl = ocfs2_ioctl, 2170 #ifdef CONFIG_COMPAT 2171 .compat_ioctl = ocfs2_compat_ioctl, 2172 #endif 2173 .flock = ocfs2_flock, 2174 }; 2175