Lines Matching +full:deep +full:- +full:touch

1 // SPDX-License-Identifier: GPL-2.0-only
10 #include <linux/backing-dev.h>
35 * inode->i_lock protects:
36 * inode->i_state, inode->i_hash, __iget(), inode->i_io_list
38 * inode->i_sb->s_inode_lru, inode->i_lru
39 * inode->i_sb->s_inode_list_lock protects:
40 * inode->i_sb->s_inodes, inode->i_sb_list
41 * bdi->wb.list_lock protects:
42 * bdi->wb.b_{dirty,io,more_io,dirty_time}, inode->i_io_list
44 * inode_hashtable, inode->i_hash
48 * inode->i_sb->s_inode_list_lock
49 * inode->i_lock
52 * bdi->wb.list_lock
53 * inode->i_lock
56 * inode->i_sb->s_inode_list_lock
57 * inode->i_lock
102 long nr_dirty = get_nr_inodes() - get_nr_inodes_unused(); in get_nr_dirty_inodes()
189 .procname = "inode-nr",
196 .procname = "inode-state",
214 return -ENXIO; in no_open()
218 * inode_init_always_gfp - perform inode structure initialisation
231 struct address_space *const mapping = &inode->i_data; in inode_init_always_gfp()
233 inode->i_sb = sb; in inode_init_always_gfp()
234 inode->i_blkbits = sb->s_blocksize_bits; in inode_init_always_gfp()
235 inode->i_flags = 0; in inode_init_always_gfp()
236 inode->i_state = 0; in inode_init_always_gfp()
237 atomic64_set(&inode->i_sequence, 0); in inode_init_always_gfp()
238 atomic_set(&inode->i_count, 1); in inode_init_always_gfp()
239 inode->i_op = &empty_iops; in inode_init_always_gfp()
240 inode->i_fop = &no_open_fops; in inode_init_always_gfp()
241 inode->i_ino = 0; in inode_init_always_gfp()
242 inode->__i_nlink = 1; in inode_init_always_gfp()
243 inode->i_opflags = 0; in inode_init_always_gfp()
244 if (sb->s_xattr) in inode_init_always_gfp()
245 inode->i_opflags |= IOP_XATTR; in inode_init_always_gfp()
246 if (sb->s_type->fs_flags & FS_MGTIME) in inode_init_always_gfp()
247 inode->i_opflags |= IOP_MGTIME; in inode_init_always_gfp()
250 atomic_set(&inode->i_writecount, 0); in inode_init_always_gfp()
251 inode->i_size = 0; in inode_init_always_gfp()
252 inode->i_write_hint = WRITE_LIFE_NOT_SET; in inode_init_always_gfp()
253 inode->i_blocks = 0; in inode_init_always_gfp()
254 inode->i_bytes = 0; in inode_init_always_gfp()
255 inode->i_generation = 0; in inode_init_always_gfp()
256 inode->i_pipe = NULL; in inode_init_always_gfp()
257 inode->i_cdev = NULL; in inode_init_always_gfp()
258 inode->i_link = NULL; in inode_init_always_gfp()
259 inode->i_dir_seq = 0; in inode_init_always_gfp()
260 inode->i_rdev = 0; in inode_init_always_gfp()
261 inode->dirtied_when = 0; in inode_init_always_gfp()
264 inode->i_wb_frn_winner = 0; in inode_init_always_gfp()
265 inode->i_wb_frn_avg_time = 0; in inode_init_always_gfp()
266 inode->i_wb_frn_history = 0; in inode_init_always_gfp()
269 spin_lock_init(&inode->i_lock); in inode_init_always_gfp()
270 lockdep_set_class(&inode->i_lock, &sb->s_type->i_lock_key); in inode_init_always_gfp()
272 init_rwsem(&inode->i_rwsem); in inode_init_always_gfp()
273 lockdep_set_class(&inode->i_rwsem, &sb->s_type->i_mutex_key); in inode_init_always_gfp()
275 atomic_set(&inode->i_dio_count, 0); in inode_init_always_gfp()
277 mapping->a_ops = &empty_aops; in inode_init_always_gfp()
278 mapping->host = inode; in inode_init_always_gfp()
279 mapping->flags = 0; in inode_init_always_gfp()
280 mapping->wb_err = 0; in inode_init_always_gfp()
281 atomic_set(&mapping->i_mmap_writable, 0); in inode_init_always_gfp()
283 atomic_set(&mapping->nr_thps, 0); in inode_init_always_gfp()
286 mapping->i_private_data = NULL; in inode_init_always_gfp()
287 mapping->writeback_index = 0; in inode_init_always_gfp()
288 init_rwsem(&mapping->invalidate_lock); in inode_init_always_gfp()
289 lockdep_set_class_and_name(&mapping->invalidate_lock, in inode_init_always_gfp()
290 &sb->s_type->invalidate_lock_key, in inode_init_always_gfp()
292 if (sb->s_iflags & SB_I_STABLE_WRITES) in inode_init_always_gfp()
294 inode->i_private = NULL; in inode_init_always_gfp()
295 inode->i_mapping = mapping; in inode_init_always_gfp()
296 INIT_HLIST_HEAD(&inode->i_dentry); /* buggered by rcu freeing */ in inode_init_always_gfp()
298 inode->i_acl = inode->i_default_acl = ACL_NOT_CACHED; in inode_init_always_gfp()
302 inode->i_fsnotify_mask = 0; in inode_init_always_gfp()
304 inode->i_flctx = NULL; in inode_init_always_gfp()
307 return -ENOMEM; in inode_init_always_gfp()
324 if (inode->free_inode) in i_callback()
325 inode->free_inode(inode); in i_callback()
331 * alloc_inode - obtain an inode
337 * - fs can't be unmount
338 * - quotas, fsnotify, writeback can't work
342 const struct super_operations *ops = sb->s_op; in alloc_inode()
345 if (ops->alloc_inode) in alloc_inode()
346 inode = ops->alloc_inode(sb); in alloc_inode()
354 if (ops->destroy_inode) { in alloc_inode()
355 ops->destroy_inode(inode); in alloc_inode()
356 if (!ops->free_inode) in alloc_inode()
359 inode->free_inode = ops->free_inode; in alloc_inode()
360 i_callback(&inode->i_rcu); in alloc_inode()
374 if (!inode->i_nlink) { in __destroy_inode()
375 WARN_ON(atomic_long_read(&inode->i_sb->s_remove_count) == 0); in __destroy_inode()
376 atomic_long_dec(&inode->i_sb->s_remove_count); in __destroy_inode()
380 if (inode->i_acl && !is_uncached_acl(inode->i_acl)) in __destroy_inode()
381 posix_acl_release(inode->i_acl); in __destroy_inode()
382 if (inode->i_default_acl && !is_uncached_acl(inode->i_default_acl)) in __destroy_inode()
383 posix_acl_release(inode->i_default_acl); in __destroy_inode()
391 const struct super_operations *ops = inode->i_sb->s_op; in destroy_inode()
393 BUG_ON(!list_empty(&inode->i_lru)); in destroy_inode()
395 if (ops->destroy_inode) { in destroy_inode()
396 ops->destroy_inode(inode); in destroy_inode()
397 if (!ops->free_inode) in destroy_inode()
400 inode->free_inode = ops->free_inode; in destroy_inode()
401 call_rcu(&inode->i_rcu, i_callback); in destroy_inode()
405 * drop_nlink - directly drop an inode's link count
408 * This is a low-level filesystem helper to replace any
417 WARN_ON(inode->i_nlink == 0); in drop_nlink()
418 inode->__i_nlink--; in drop_nlink()
419 if (!inode->i_nlink) in drop_nlink()
420 atomic_long_inc(&inode->i_sb->s_remove_count); in drop_nlink()
425 * clear_nlink - directly zero an inode's link count
428 * This is a low-level filesystem helper to replace any
434 if (inode->i_nlink) { in clear_nlink()
435 inode->__i_nlink = 0; in clear_nlink()
436 atomic_long_inc(&inode->i_sb->s_remove_count); in clear_nlink()
442 * set_nlink - directly set an inode's link count
444 * @nlink: new nlink (should be non-zero)
446 * This is a low-level filesystem helper to replace any
455 if (inode->i_nlink == 0) in set_nlink()
456 atomic_long_dec(&inode->i_sb->s_remove_count); in set_nlink()
458 inode->__i_nlink = nlink; in set_nlink()
464 * inc_nlink - directly increment an inode's link count
467 * This is a low-level filesystem helper to replace any
473 if (unlikely(inode->i_nlink == 0)) { in inc_nlink()
474 WARN_ON(!(inode->i_state & I_LINKABLE)); in inc_nlink()
475 atomic_long_dec(&inode->i_sb->s_remove_count); in inc_nlink()
478 inode->__i_nlink++; in inc_nlink()
484 xa_init_flags(&mapping->i_pages, XA_FLAGS_LOCK_IRQ | XA_FLAGS_ACCOUNT); in __address_space_init_once()
485 init_rwsem(&mapping->i_mmap_rwsem); in __address_space_init_once()
486 INIT_LIST_HEAD(&mapping->i_private_list); in __address_space_init_once()
487 spin_lock_init(&mapping->i_private_lock); in __address_space_init_once()
488 mapping->i_mmap = RB_ROOT_CACHED; in __address_space_init_once()
506 INIT_HLIST_NODE(&inode->i_hash); in inode_init_once()
507 INIT_LIST_HEAD(&inode->i_devices); in inode_init_once()
508 INIT_LIST_HEAD(&inode->i_io_list); in inode_init_once()
509 INIT_LIST_HEAD(&inode->i_wb_list); in inode_init_once()
510 INIT_LIST_HEAD(&inode->i_lru); in inode_init_once()
511 INIT_LIST_HEAD(&inode->i_sb_list); in inode_init_once()
512 __address_space_init_once(&inode->i_data); in inode_init_once()
529 WARN_ON(atomic_inc_return(&inode->i_count) < 2); in ihold()
535 if (inode->i_state & (I_DIRTY_ALL | I_SYNC | I_FREEING | I_WILL_FREE)) in __inode_add_lru()
539 if (!(inode->i_sb->s_flags & SB_ACTIVE)) in __inode_add_lru()
541 if (!mapping_shrinkable(&inode->i_data)) in __inode_add_lru()
544 if (list_lru_add_obj(&inode->i_sb->s_inode_lru, &inode->i_lru)) in __inode_add_lru()
547 inode->i_state |= I_REFERENCED; in __inode_add_lru()
564 * Needs inode->i_lock held.
573 if (list_lru_del_obj(&inode->i_sb->s_inode_lru, &inode->i_lru)) in inode_lru_list_del()
579 lockdep_assert_held(&inode->i_lock); in inode_pin_lru_isolating()
580 WARN_ON(inode->i_state & (I_LRU_ISOLATING | I_FREEING | I_WILL_FREE)); in inode_pin_lru_isolating()
581 inode->i_state |= I_LRU_ISOLATING; in inode_pin_lru_isolating()
586 spin_lock(&inode->i_lock); in inode_unpin_lru_isolating()
587 WARN_ON(!(inode->i_state & I_LRU_ISOLATING)); in inode_unpin_lru_isolating()
588 inode->i_state &= ~I_LRU_ISOLATING; in inode_unpin_lru_isolating()
589 /* Called with inode->i_lock which ensures memory ordering. */ in inode_unpin_lru_isolating()
591 spin_unlock(&inode->i_lock); in inode_unpin_lru_isolating()
599 lockdep_assert_held(&inode->i_lock); in inode_wait_for_lru_isolating()
600 if (!(inode->i_state & I_LRU_ISOLATING)) in inode_wait_for_lru_isolating()
607 * Checking I_LRU_ISOLATING with inode->i_lock guarantees in inode_wait_for_lru_isolating()
610 if (!(inode->i_state & I_LRU_ISOLATING)) in inode_wait_for_lru_isolating()
612 spin_unlock(&inode->i_lock); in inode_wait_for_lru_isolating()
614 spin_lock(&inode->i_lock); in inode_wait_for_lru_isolating()
617 WARN_ON(inode->i_state & I_LRU_ISOLATING); in inode_wait_for_lru_isolating()
621 * inode_sb_list_add - add inode to the superblock list of inodes
626 struct super_block *sb = inode->i_sb; in inode_sb_list_add()
628 spin_lock(&sb->s_inode_list_lock); in inode_sb_list_add()
629 list_add(&inode->i_sb_list, &sb->s_inodes); in inode_sb_list_add()
630 spin_unlock(&sb->s_inode_list_lock); in inode_sb_list_add()
636 struct super_block *sb = inode->i_sb; in inode_sb_list_del()
638 if (!list_empty(&inode->i_sb_list)) { in inode_sb_list_del()
639 spin_lock(&sb->s_inode_list_lock); in inode_sb_list_del()
640 list_del_init(&inode->i_sb_list); in inode_sb_list_del()
641 spin_unlock(&sb->s_inode_list_lock); in inode_sb_list_del()
656 * __insert_inode_hash - hash an inode
665 struct hlist_head *b = inode_hashtable + hash(inode->i_sb, hashval); in __insert_inode_hash()
668 spin_lock(&inode->i_lock); in __insert_inode_hash()
669 hlist_add_head_rcu(&inode->i_hash, b); in __insert_inode_hash()
670 spin_unlock(&inode->i_lock); in __insert_inode_hash()
676 * __remove_inode_hash - remove an inode from the hash
684 spin_lock(&inode->i_lock); in __remove_inode_hash()
685 hlist_del_init_rcu(&inode->i_hash); in __remove_inode_hash()
686 spin_unlock(&inode->i_lock); in __remove_inode_hash()
705 if (get_kernel_nofault(host, &mapping->host) || in dump_mapping()
706 get_kernel_nofault(a_ops, &mapping->a_ops)) { in dump_mapping()
716 if (get_kernel_nofault(dentry_first, &host->i_dentry.first) || in dump_mapping()
717 get_kernel_nofault(ino, &host->i_ino)) { in dump_mapping()
752 xa_lock_irq(&inode->i_data.i_pages); in clear_inode()
753 BUG_ON(inode->i_data.nrpages); in clear_inode()
755 * Almost always, mapping_empty(&inode->i_data) here; but there are in clear_inode()
756 * two known and long-standing ways in which nodes may get left behind in clear_inode()
757 * (when deep radix-tree node allocation failed partway; or when THP in clear_inode()
762 xa_unlock_irq(&inode->i_data.i_pages); in clear_inode()
763 BUG_ON(!list_empty(&inode->i_data.i_private_list)); in clear_inode()
764 BUG_ON(!(inode->i_state & I_FREEING)); in clear_inode()
765 BUG_ON(inode->i_state & I_CLEAR); in clear_inode()
766 BUG_ON(!list_empty(&inode->i_wb_list)); in clear_inode()
768 inode->i_state = I_FREEING | I_CLEAR; in clear_inode()
787 const struct super_operations *op = inode->i_sb->s_op; in evict()
789 BUG_ON(!(inode->i_state & I_FREEING)); in evict()
790 BUG_ON(!list_empty(&inode->i_lru)); in evict()
792 if (!list_empty(&inode->i_io_list)) in evict()
797 spin_lock(&inode->i_lock); in evict()
807 spin_unlock(&inode->i_lock); in evict()
809 if (op->evict_inode) { in evict()
810 op->evict_inode(inode); in evict()
812 truncate_inode_pages_final(&inode->i_data); in evict()
815 if (S_ISCHR(inode->i_mode) && inode->i_cdev) in evict()
824 * accounted for before remove_inode_hash() acquires ->i_lock -- both in evict()
832 BUG_ON(inode->i_state != (I_FREEING | I_CLEAR)); in evict()
838 * dispose_list - dispose of the contents of a local list
841 * Dispose-list gets a local list with local inodes in it, so it doesn't
850 list_del_init(&inode->i_lru); in dispose_list()
858 * evict_inodes - evict all evictable inodes for a superblock
872 spin_lock(&sb->s_inode_list_lock); in evict_inodes()
873 list_for_each_entry(inode, &sb->s_inodes, i_sb_list) { in evict_inodes()
877 spin_lock(&inode->i_lock); in evict_inodes()
879 spin_unlock(&inode->i_lock); in evict_inodes()
882 if (inode->i_state & (I_NEW | I_FREEING | I_WILL_FREE)) { in evict_inodes()
883 spin_unlock(&inode->i_lock); in evict_inodes()
887 inode->i_state |= I_FREEING; in evict_inodes()
889 spin_unlock(&inode->i_lock); in evict_inodes()
890 list_add(&inode->i_lru, &dispose); in evict_inodes()
898 spin_unlock(&sb->s_inode_list_lock); in evict_inodes()
904 spin_unlock(&sb->s_inode_list_lock); in evict_inodes()
914 * used recently - the flag is set in iput_final(). When we encounter such an
928 * We are inverting the lru lock/inode->i_lock here, so use a in inode_lru_isolate()
931 if (!spin_trylock(&inode->i_lock)) in inode_lru_isolate()
941 (inode->i_state & ~I_REFERENCED) || in inode_lru_isolate()
942 !mapping_shrinkable(&inode->i_data)) { in inode_lru_isolate()
943 list_lru_isolate(lru, &inode->i_lru); in inode_lru_isolate()
944 spin_unlock(&inode->i_lock); in inode_lru_isolate()
950 if (inode->i_state & I_REFERENCED) { in inode_lru_isolate()
951 inode->i_state &= ~I_REFERENCED; in inode_lru_isolate()
952 spin_unlock(&inode->i_lock); in inode_lru_isolate()
961 if (inode_has_buffers(inode) || !mapping_empty(&inode->i_data)) { in inode_lru_isolate()
963 spin_unlock(&inode->i_lock); in inode_lru_isolate()
964 spin_unlock(&lru->lock); in inode_lru_isolate()
967 reap = invalidate_mapping_pages(&inode->i_data, 0, -1); in inode_lru_isolate()
978 WARN_ON(inode->i_state & I_NEW); in inode_lru_isolate()
979 inode->i_state |= I_FREEING; in inode_lru_isolate()
980 list_lru_isolate_move(lru, &inode->i_lru, freeable); in inode_lru_isolate()
981 spin_unlock(&inode->i_lock); in inode_lru_isolate()
998 freed = list_lru_shrink_walk(&sb->s_inode_lru, sc, in prune_icache_sb()
1023 if (inode->i_sb != sb) in find_inode()
1027 spin_lock(&inode->i_lock); in find_inode()
1028 if (inode->i_state & (I_FREEING|I_WILL_FREE)) { in find_inode()
1032 if (unlikely(inode->i_state & I_CREATING)) { in find_inode()
1033 spin_unlock(&inode->i_lock); in find_inode()
1035 return ERR_PTR(-ESTALE); in find_inode()
1038 spin_unlock(&inode->i_lock); in find_inode()
1064 if (inode->i_ino != ino) in find_inode_fast()
1066 if (inode->i_sb != sb) in find_inode_fast()
1068 spin_lock(&inode->i_lock); in find_inode_fast()
1069 if (inode->i_state & (I_FREEING|I_WILL_FREE)) { in find_inode_fast()
1073 if (unlikely(inode->i_state & I_CREATING)) { in find_inode_fast()
1074 spin_unlock(&inode->i_lock); in find_inode_fast()
1076 return ERR_PTR(-ESTALE); in find_inode_fast()
1079 spin_unlock(&inode->i_lock); in find_inode_fast()
1093 * consume at most LAST_INO_BATCH-1 unused inode numbers. So there is
1094 * NR_CPUS*(LAST_INO_BATCH-1) wastage. At 4096 and 1024, this is ~0.1% of the
1095 * 2^32 range, and is a worst-case. Even a 50% wastage would only increase
1111 if (unlikely((res & (LAST_INO_BATCH-1)) == 0)) { in get_next_ino()
1115 res = next - LAST_INO_BATCH; in get_next_ino()
1130 * new_inode - obtain an inode
1134 * for allocations related to inode->i_mapping is GFP_HIGHUSER_MOVABLE.
1155 if (S_ISDIR(inode->i_mode)) { in lockdep_annotate_inode_mutex_key()
1156 struct file_system_type *type = inode->i_sb->s_type; in lockdep_annotate_inode_mutex_key()
1159 if (lockdep_match_class(&inode->i_rwsem, &type->i_mutex_key)) { in lockdep_annotate_inode_mutex_key()
1163 init_rwsem(&inode->i_rwsem); in lockdep_annotate_inode_mutex_key()
1164 lockdep_set_class(&inode->i_rwsem, in lockdep_annotate_inode_mutex_key()
1165 &type->i_mutex_dir_key); in lockdep_annotate_inode_mutex_key()
1173 * unlock_new_inode - clear the I_NEW state and wake up any waiters
1182 spin_lock(&inode->i_lock); in unlock_new_inode()
1183 WARN_ON(!(inode->i_state & I_NEW)); in unlock_new_inode()
1184 inode->i_state &= ~I_NEW & ~I_CREATING; in unlock_new_inode()
1192 spin_unlock(&inode->i_lock); in unlock_new_inode()
1199 spin_lock(&inode->i_lock); in discard_new_inode()
1200 WARN_ON(!(inode->i_state & I_NEW)); in discard_new_inode()
1201 inode->i_state &= ~I_NEW; in discard_new_inode()
1209 spin_unlock(&inode->i_lock); in discard_new_inode()
1215 * lock_two_nondirectories - take two i_mutexes on non-directory objects
1217 * Lock any non-NULL argument. Passed objects must not be directories.
1226 WARN_ON_ONCE(S_ISDIR(inode1->i_mode)); in lock_two_nondirectories()
1228 WARN_ON_ONCE(S_ISDIR(inode2->i_mode)); in lock_two_nondirectories()
1239 * unlock_two_nondirectories - release locks from lock_two_nondirectories()
1246 WARN_ON_ONCE(S_ISDIR(inode1->i_mode)); in unlock_two_nondirectories()
1250 WARN_ON_ONCE(S_ISDIR(inode2->i_mode)); in unlock_two_nondirectories()
1257 * inode_insert5 - obtain an inode from a mounted file system
1258 * @inode: pre-allocated inode to use for insert to cache
1268 * If the inode is not present in the cache, insert the pre-allocated inode and
1279 struct hlist_head *head = inode_hashtable + hash(inode->i_sb, hashval); in inode_insert5()
1286 old = find_inode(inode->i_sb, head, test, data, true); in inode_insert5()
1312 spin_lock(&inode->i_lock); in inode_insert5()
1313 inode->i_state |= I_NEW; in inode_insert5()
1314 hlist_add_head_rcu(&inode->i_hash, head); in inode_insert5()
1315 spin_unlock(&inode->i_lock); in inode_insert5()
1323 if (list_empty(&inode->i_sb_list)) in inode_insert5()
1331 * iget5_locked - obtain an inode from a mounted file system
1370 * iget5_locked_rcu - obtain an inode from a mounted file system
1378 * tolerate the inode not being stable, including being mid-teardown.
1413 * iget_locked - obtain an inode from a mounted file system
1453 inode->i_ino = ino; in iget_locked()
1454 spin_lock(&inode->i_lock); in iget_locked()
1455 inode->i_state = I_NEW; in iget_locked()
1456 hlist_add_head_rcu(&inode->i_hash, head); in iget_locked()
1457 spin_unlock(&inode->i_lock); in iget_locked()
1500 if (inode->i_ino == ino && inode->i_sb == sb) in test_inode_iunique()
1507 * iunique - get a unique inode number
1547 spin_lock(&inode->i_lock); in igrab()
1548 if (!(inode->i_state & (I_FREEING|I_WILL_FREE))) { in igrab()
1550 spin_unlock(&inode->i_lock); in igrab()
1552 spin_unlock(&inode->i_lock); in igrab()
1554 * Handle the case where s_op->clear_inode is not been in igrab()
1565 * ilookup5_nowait - search for an inode in the inode cache
1595 * ilookup5 - search for an inode in the inode cache
1632 * ilookup - search for an inode in the inode cache
1663 * find_inode_nowait - find an inode in the inode cache
1671 * does not match, 1 if the inode does match, and -1 if the search
1679 * function must never block --- find_inode() can block in
1680 * __wait_on_freeing_inode() --- or when the caller can not increment
1697 if (inode->i_sb != sb) in find_inode_nowait()
1713 * find_inode_rcu - find an inode in the inode cache
1743 if (inode->i_sb == sb && in find_inode_rcu()
1744 !(READ_ONCE(inode->i_state) & (I_FREEING | I_WILL_FREE)) && in find_inode_rcu()
1753 * find_inode_by_ino_rcu - Find an inode in the inode cache
1781 if (inode->i_ino == ino && in find_inode_by_ino_rcu()
1782 inode->i_sb == sb && in find_inode_by_ino_rcu()
1783 !(READ_ONCE(inode->i_state) & (I_FREEING | I_WILL_FREE))) in find_inode_by_ino_rcu()
1792 struct super_block *sb = inode->i_sb; in insert_inode_locked()
1793 ino_t ino = inode->i_ino; in insert_inode_locked()
1802 if (old->i_ino != ino) in insert_inode_locked()
1804 if (old->i_sb != sb) in insert_inode_locked()
1806 spin_lock(&old->i_lock); in insert_inode_locked()
1807 if (old->i_state & (I_FREEING|I_WILL_FREE)) { in insert_inode_locked()
1808 spin_unlock(&old->i_lock); in insert_inode_locked()
1814 spin_lock(&inode->i_lock); in insert_inode_locked()
1815 inode->i_state |= I_NEW | I_CREATING; in insert_inode_locked()
1816 hlist_add_head_rcu(&inode->i_hash, head); in insert_inode_locked()
1817 spin_unlock(&inode->i_lock); in insert_inode_locked()
1821 if (unlikely(old->i_state & I_CREATING)) { in insert_inode_locked()
1822 spin_unlock(&old->i_lock); in insert_inode_locked()
1824 return -EBUSY; in insert_inode_locked()
1827 spin_unlock(&old->i_lock); in insert_inode_locked()
1832 return -EBUSY; in insert_inode_locked()
1846 inode->i_state |= I_CREATING; in insert_inode_locked4()
1851 return -EBUSY; in insert_inode_locked4()
1876 struct super_block *sb = inode->i_sb; in iput_final()
1877 const struct super_operations *op = inode->i_sb->s_op; in iput_final()
1881 WARN_ON(inode->i_state & I_NEW); in iput_final()
1883 if (op->drop_inode) in iput_final()
1884 drop = op->drop_inode(inode); in iput_final()
1889 !(inode->i_state & I_DONTCACHE) && in iput_final()
1890 (sb->s_flags & SB_ACTIVE)) { in iput_final()
1892 spin_unlock(&inode->i_lock); in iput_final()
1896 state = inode->i_state; in iput_final()
1898 WRITE_ONCE(inode->i_state, state | I_WILL_FREE); in iput_final()
1899 spin_unlock(&inode->i_lock); in iput_final()
1903 spin_lock(&inode->i_lock); in iput_final()
1904 state = inode->i_state; in iput_final()
1909 WRITE_ONCE(inode->i_state, state | I_FREEING); in iput_final()
1910 if (!list_empty(&inode->i_lru)) in iput_final()
1912 spin_unlock(&inode->i_lock); in iput_final()
1918 * iput - put an inode
1933 lockdep_assert_not_held(&inode->i_lock); in iput()
1934 VFS_BUG_ON_INODE(inode->i_state & I_CLEAR, inode); in iput()
1940 VFS_BUG_ON_INODE(atomic_read(&inode->i_count) < 1, inode); in iput()
1942 if (atomic_add_unless(&inode->i_count, -1, 1)) in iput()
1945 if ((inode->i_state & I_DIRTY_TIME) && inode->i_nlink) { in iput()
1951 spin_lock(&inode->i_lock); in iput()
1952 if (unlikely((inode->i_state & I_DIRTY_TIME) && inode->i_nlink)) { in iput()
1953 spin_unlock(&inode->i_lock); in iput()
1957 if (!atomic_dec_and_test(&inode->i_count)) { in iput()
1958 spin_unlock(&inode->i_lock); in iput()
1963 * iput_final() drops ->i_lock, we can't assert on it as the inode may in iput()
1972 * bmap - find a block number in a file
1982 * Returns -EINVAL in case of error, 0 otherwise. If mapping falls into a
1987 if (!inode->i_mapping->a_ops->bmap) in bmap()
1988 return -EINVAL; in bmap()
1990 *block = inode->i_mapping->a_ops->bmap(inode->i_mapping, *block); in bmap()
2006 if (!(mnt->mnt_flags & MNT_RELATIME)) in relatime_need_update()
2026 if ((long)(now.tv_sec - atime.tv_sec) >= 24*60*60) in relatime_need_update()
2035 * inode_update_timestamps - update the timestamps on the inode
2085 * generic_update_time - update the timestamps on the inode
2102 dirty_flags = inode->i_sb->s_flags & SB_LAZYTIME ? I_DIRTY_TIME : I_DIRTY_SYNC; in generic_update_time()
2116 if (inode->i_op->update_time) in inode_update_time()
2117 return inode->i_op->update_time(inode, flags); in inode_update_time()
2124 * atime_needs_update - update the access time
2134 struct vfsmount *mnt = path->mnt; in atime_needs_update()
2137 if (inode->i_flags & S_NOATIME) in atime_needs_update()
2148 if ((inode->i_sb->s_flags & SB_NODIRATIME) && S_ISDIR(inode->i_mode)) in atime_needs_update()
2151 if (mnt->mnt_flags & MNT_NOATIME) in atime_needs_update()
2153 if ((mnt->mnt_flags & MNT_NODIRATIME) && S_ISDIR(inode->i_mode)) in atime_needs_update()
2170 struct vfsmount *mnt = path->mnt; in touch_atime()
2171 struct inode *inode = d_inode(path->dentry); in touch_atime()
2176 if (!sb_start_write_trylock(inode->i_sb)) in touch_atime()
2184 * Btrfs), but since we touch atime while walking down the path we in touch_atime()
2193 sb_end_write(inode->i_sb); in touch_atime()
2241 if (IS_NOSEC(inode) || !S_ISREG(inode->i_mode)) in file_remove_privs_flags()
2250 return -EAGAIN; in file_remove_privs_flags()
2261 * file_remove_privs - remove special file privileges (suid, capabilities)
2276 * current_time - Return FS time (possibly fine-grained)
2281 * as having been QUERIED, get a fine-grained timestamp, but don't update
2299 cns = smp_load_acquire(&inode->i_ctime_nsec); in current_time()
2302 * If there is no apparent change, then get a fine-grained in current_time()
2353 * file_update_time - update mtime and ctime time
2380 * file_modified_flags - handle mandated vfs changes when modifying a file
2388 * time settings will not be updated. It will return -EAGAIN.
2407 if (unlikely(file->f_mode & FMODE_NOCMTIME)) in file_modified_flags()
2414 return -EAGAIN; in file_modified_flags()
2420 * file_modified - handle mandated vfs changes when modifying a file
2437 * kiocb_modified - handle mandated vfs changes when modifying a file
2449 return file_modified_flags(iocb->ki_filp, iocb->ki_flags); in kiocb_modified()
2457 if (S_ISDIR(inode->i_mode) && IS_DIRSYNC(inode)) in inode_needs_sync()
2471 * wake_up_bit(&inode->i_state, __I_NEW) after removing from the hash list
2484 spin_unlock(&inode->i_lock); in __wait_on_freeing_inode()
2490 spin_unlock(&inode->i_lock); in __wait_on_freeing_inode()
2523 alloc_large_system_hash("Inode-cache", in inode_init_early()
2549 alloc_large_system_hash("Inode-cache", in inode_init()
2562 inode->i_mode = mode; in init_special_inode()
2563 switch (inode->i_mode & S_IFMT) { in init_special_inode()
2565 inode->i_fop = &def_chr_fops; in init_special_inode()
2566 inode->i_rdev = rdev; in init_special_inode()
2570 inode->i_fop = &def_blk_fops; in init_special_inode()
2571 inode->i_rdev = rdev; in init_special_inode()
2574 inode->i_fop = &pipefifo_fops; in init_special_inode()
2581 " inode %s:%lu\n", mode, inode->i_sb->s_id, in init_special_inode()
2582 inode->i_ino); in init_special_inode()
2589 * inode_init_owner - Init uid,gid,mode for new inode according to posix standards
2598 * and initializing i_uid and i_gid. On non-idmapped mounts or if permission
2605 if (dir && dir->i_mode & S_ISGID) { in inode_init_owner()
2606 inode->i_gid = dir->i_gid; in inode_init_owner()
2613 inode->i_mode = mode; in inode_init_owner()
2618 * inode_owner_or_capable - check current task permissions to inode
2628 * On non-idmapped mounts or if permission checking is to be performed on the
2653 return atomic_read(&inode->i_dio_count) == 0; in inode_dio_finished()
2658 * inode_dio_wait - wait for outstanding DIO requests to finish
2665 * to i_dio_count, usually by inode->i_rwsem.
2669 wait_var_event(&inode->i_dio_count, inode_dio_finished(inode)); in inode_dio_wait()
2675 wait_var_event_interruptible(&inode->i_dio_count, in inode_dio_wait_interruptible()
2681 * inode_set_flags - atomically set some inode flags
2686 * --- which wouldn't be necessary if all code paths which modify
2700 set_mask_bits(&inode->i_flags, mask, flags); in inode_set_flags()
2706 mapping_set_gfp_mask(inode->i_mapping, GFP_USER); in inode_nohighmem()
2714 inode->i_ctime_sec = ts.tv_sec; in inode_set_ctime_to_ts()
2715 inode->i_ctime_nsec = ts.tv_nsec; in inode_set_ctime_to_ts()
2721 * timestamp_truncate - Truncate timespec to a granularity
2731 struct super_block *sb = inode->i_sb; in timestamp_truncate()
2732 unsigned int gran = sb->s_time_gran; in timestamp_truncate()
2734 t.tv_sec = clamp(t.tv_sec, sb->s_time_min, sb->s_time_max); in timestamp_truncate()
2735 if (unlikely(t.tv_sec == sb->s_time_max || t.tv_sec == sb->s_time_min)) in timestamp_truncate()
2744 t.tv_nsec -= t.tv_nsec % gran; in timestamp_truncate()
2752 * inode_set_ctime_current - set the ctime to current_time
2759 * If it is multigrain, then we first see if the coarse-grained timestamp is
2761 * fine-grained timestamp.
2783 * A fine-grained time is only needed if someone has queried in inode_set_ctime_current()
2787 cns = smp_load_acquire(&inode->i_ctime_nsec); in inode_set_ctime_current()
2789 struct timespec64 ctime = { .tv_sec = inode->i_ctime_sec, in inode_set_ctime_current()
2801 if (cns == now.tv_nsec && inode->i_ctime_sec == now.tv_sec) { in inode_set_ctime_current()
2808 if (try_cmpxchg(&inode->i_ctime_nsec, &cur, now.tv_nsec)) { in inode_set_ctime_current()
2810 inode->i_ctime_sec = now.tv_sec; in inode_set_ctime_current()
2825 now.tv_sec = inode->i_ctime_sec; in inode_set_ctime_current()
2834 * inode_set_ctime_deleg - try to update the ctime on a delegated inode
2856 cur = smp_load_acquire(&inode->i_ctime_nsec); in inode_set_ctime_deleg()
2858 cur_ts.tv_sec = inode->i_ctime_sec; in inode_set_ctime_deleg()
2883 if (try_cmpxchg(&inode->i_ctime_nsec, &cur, update.tv_nsec)) { in inode_set_ctime_deleg()
2884 inode->i_ctime_sec = update.tv_sec; in inode_set_ctime_deleg()
2900 cur_ts.tv_sec = inode->i_ctime_sec; in inode_set_ctime_deleg()
2907 * in_group_or_capable - check whether caller is CAP_FSETID privileged
2930 * mode_strip_sgid - handle the sgid bit for non-directories
2948 if (S_ISDIR(mode) || !dir || !(dir->i_mode & S_ISGID)) in mode_strip_sgid()
2968 struct super_block *sb = inode->i_sb; in dump_inode()
2972 reason, inode, sb->s_type->name, inode->i_mode, inode->i_opflags, in dump_inode()
2973 inode->i_flags, inode->i_state, atomic_read(&inode->i_count)); in dump_inode()