| /linux/fs/xfs/scrub/ |
| H A D | dir.c | 439 xfs_dablk_t lblk, in xchk_directory_check_free_entry() argument 457 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); in xchk_directory_check_free_entry() 464 xfs_dablk_t lblk, in xchk_directory_data_bestfree() argument 483 if (lblk != XFS_B_TO_FSBT(mp, XFS_DIR2_DATA_OFFSET)) in xchk_directory_data_bestfree() 484 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); in xchk_directory_data_bestfree() 488 error = xfs_dir3_data_read(sc->tp, sc->ip, sc->ip->i_ino, lblk, in xchk_directory_data_bestfree() 491 if (!xchk_fblock_process_error(sc, XFS_DATA_FORK, lblk, &error)) in xchk_directory_data_bestfree() 508 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); in xchk_directory_data_bestfree() 518 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); in xchk_directory_data_bestfree() 524 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, lblk); in xchk_directory_data_bestfree() [all …]
|
| H A D | trace.h | 2204 __field(xfs_fileoff_t, lblk) 2213 __entry->lblk = irec->br_startoff; 2222 __entry->lblk, 2846 __field(xfs_fileoff_t, lblk) 2855 __entry->lblk = irec->br_startoff; 2864 __entry->lblk,
|
| /linux/fs/ext4/ |
| H A D | extents.c | 333 struct ext4_ext_path *path, ext4_lblk_t lblk, in ext4_force_split_extent_at() argument 341 return ext4_split_extent_at(handle, inode, path, lblk, flags); in ext4_force_split_extent_at() 390 ext4_lblk_t lblk, ext4_fsblk_t *pblk, in ext4_valid_extent_entries() argument 411 lblk != le32_to_cpu(ext->ee_block)) in ext4_valid_extent_entries() 435 lblk != le32_to_cpu(ext_idx->ei_block)) in ext4_valid_extent_entries() 457 int depth, ext4_fsblk_t pblk, ext4_lblk_t lblk) in __ext4_ext_check() argument 487 if (!ext4_valid_extent_entries(inode, eh, lblk, &pblk, depth)) { in __ext4_ext_check() 535 ext4_lblk_t lblk = le32_to_cpu(ex->ee_block); in ext4_cache_extents() local 538 if (prev && (prev != lblk)) in ext4_cache_extents() 539 ext4_es_cache_extent(inode, prev, lblk in ext4_cache_extents() 2282 ext4_ext_find_hole(struct inode * inode,struct ext4_ext_path * path,ext4_lblk_t * lblk) ext4_ext_find_hole() argument 2462 ext4_rereserve_cluster(struct inode * inode,ext4_lblk_t lblk) ext4_rereserve_cluster() argument 2873 ext4_lblk_t ee_block, ex_end, lblk; ext4_ext_remove_space() local 3313 uint64_t lblk, pblk, len; ext4_split_extent_zeroout() local 4193 ext4_ext_determine_insert_hole(struct inode * inode,struct ext4_ext_path * path,ext4_lblk_t lblk) ext4_ext_determine_insert_hole() argument [all...] |
| H A D | inode.c | 412 int ext4_issue_zeroout(struct inode *inode, ext4_lblk_t lblk, ext4_fsblk_t pblk, in ext4_issue_zeroout() 417 KUNIT_STATIC_STUB_REDIRECT(ext4_issue_zeroout, inode, lblk, pblk, len); in ext4_issue_zeroout() 421 (loff_t)lblk << inode->i_blkbits, in ext4_issue_zeroout() 1797 * Check whether the cluster containing lblk has been allocated or has 1804 static int ext4_clu_alloc_state(struct inode *inode, ext4_lblk_t lblk) in ext4_clu_alloc_state() 1810 if (ext4_es_scan_clu(inode, &ext4_es_is_delayed, lblk)) in ext4_clu_alloc_state() 1814 if (ext4_es_scan_clu(inode, &ext4_es_is_mapped, lblk)) in ext4_clu_alloc_state() 1816 ret = ext4_clu_mapped(inode, EXT4_B2C(sbi, lblk)); in ext4_clu_alloc_state() 1832 * @lblk - start logical block to be added 1837 static int ext4_insert_delayed_blocks(struct inode *inode, ext4_lblk_t lblk, in ext4_insert_delayed_blocks() 410 ext4_issue_zeroout(struct inode * inode,ext4_lblk_t lblk,ext4_fsblk_t pblk,ext4_lblk_t len) ext4_issue_zeroout() argument 1802 ext4_clu_alloc_state(struct inode * inode,ext4_lblk_t lblk) ext4_clu_alloc_state() argument 1835 ext4_insert_delayed_blocks(struct inode * inode,ext4_lblk_t lblk,ext4_lblk_t len) ext4_insert_delayed_blocks() argument 2118 mpage_add_bh_to_extent(struct mpage_da_data * mpd,ext4_lblk_t lblk,struct buffer_head * bh) mpage_add_bh_to_extent() argument 2175 mpage_process_page_bufs(struct mpage_da_data * mpd,struct buffer_head * head,struct buffer_head * bh,ext4_lblk_t lblk) mpage_process_page_bufs() argument 2235 ext4_lblk_t lblk = *m_lblk; mpage_process_folio() local 2305 ext4_lblk_t lblk; mpage_map_and_submit_buffers() local 2625 ext4_lblk_t lblk; mpage_prepare_extent_to_map() local [all...] |
| H A D | extents-test.c | 195 static int ext4_issue_zeroout_stub(struct inode *inode, ext4_lblk_t lblk, in ext4_issue_zeroout_stub() argument 203 "%s: lblk=%u pblk=%llu len=%u", __func__, lblk, pblk, len); in ext4_issue_zeroout_stub() 204 KUNIT_EXPECT_EQ(test, 1, lblk >= EXT_DATA_LBLK); in ext4_issue_zeroout_stub() 205 KUNIT_EXPECT_EQ(test, 1, lblk + len <= EXT_DATA_LBLK + EXT_DATA_LEN); in ext4_issue_zeroout_stub() 206 KUNIT_EXPECT_EQ(test, 1, lblk - EXT_DATA_LBLK == pblk - EXT_DATA_PBLK); in ext4_issue_zeroout_stub() 208 off_blk = lblk - EXT_DATA_LBLK; in ext4_issue_zeroout_stub()
|
| H A D | ext4.h | 371 #define EXT4_LBLK_TO_B(inode, lblk) ((loff_t)(lblk) << (inode)->i_blkbits) argument 374 #define EXT4_LBLK_TO_PG(inode, lblk) (EXT4_LBLK_TO_B((inode), (lblk)) >> \ argument 389 #define EXT4_LBLK_CMASK(s, lblk) ((lblk) & \ argument 392 #define EXT4_LBLK_CFILL(sbi, lblk) ((lblk) | \ argument 397 #define EXT4_LBLK_COFF(s, lblk) ((lblk) argument [all...] |
| H A D | ext4_extents.h | 124 ext4_lblk_t lblk; /* logical block number within logical cluster */ member
|
| H A D | fast_commit.c | 1723 ext4_lblk_t lblk, ext4_fsblk_t pblk, int len, int replay) in ext4_fc_record_regions() argument 1751 region->lblk = lblk; in ext4_fc_record_regions() 1798 ext4_debug("ADD_RANGE, lblk %d, pblk %lld, len %d, unwritten %d, inode %llu\n", in ext4_fc_replay_add_range() 1909 ext4_debug("DEL_RANGE, inode %llu, lblk %d, len %d\n", in ext4_fc_replay_del_range()
|
| /linux/include/linux/ |
| H A D | fscrypt.h | 876 u64 fscrypt_limit_io_blocks(const struct inode *inode, u64 lblk, u64 nr_blocks); 901 static inline u64 fscrypt_limit_io_blocks(const struct inode *inode, u64 lblk, 758 fscrypt_zeroout_range(const struct inode * inode,pgoff_t lblk,sector_t pblk,unsigned int len) fscrypt_zeroout_range() argument 920 fscrypt_limit_io_blocks(const struct inode * inode,u64 lblk,u64 nr_blocks) fscrypt_limit_io_blocks() argument
|
| /linux/fs/crypto/ |
| H A D | inline_crypt.c | 408 * @lblk: the block at which the I/O is being started from 409 * @nr_blocks: the number of blocks we want to submit starting at @lblk 412 * targeting @lblk without causing a data unit number (DUN) discontiguity. 425 u64 fscrypt_limit_io_blocks(const struct inode *inode, u64 lblk, u64 nr_blocks) in fscrypt_mergeable_bio_bh() 443 dun = ci->ci_hashed_ino + lblk; 495 fscrypt_limit_io_blocks(const struct inode * inode,u64 lblk,u64 nr_blocks) fscrypt_limit_io_blocks() argument
|
| /linux/fs/udf/ |
| H A D | inode.c | 405 sector_t lblk; 428 ret = inode_bmap(inode, map->lblk, &epos, &eloc, &elen, &offset, in udf_map_block() 450 if (((loff_t)map->lblk) << inode->i_blkbits >= iinfo->i_lenExtents) in udf_map_block() 463 .lblk = block, in __udf_get_block() 745 b_off = (loff_t)map->lblk << inode->i_sb->s_blocksize_bits; in inode_getblk() 893 if (iinfo->i_next_alloc_block == map->lblk) in inode_getblk() 936 iinfo->i_next_alloc_block = map->lblk + 1; in inode_getblk() 1245 .lblk = block, in udf_bread() 403 sector_t lblk; global() member
|
| /linux/fs/xfs/ |
| H A D | xfs_trace.h | 4053 __field(xfs_fileoff_t, lblk) 4061 __entry->lblk = irec->br_startoff; 4069 __entry->lblk,
|