| /linux/fs/btrfs/ |
| H A D | bio.c | 24 struct btrfs_bio *bbio; member 30 static inline bool is_data_bbio(const struct btrfs_bio *bbio) in is_data_bbio() argument 32 return bbio->inode && is_data_inode(bbio->inode); in is_data_bbio() 35 static bool bbio_has_ordered_extent(const struct btrfs_bio *bbio) in bbio_has_ordered_extent() argument 37 return is_data_bbio(bbio) && btrfs_op(&bbio->bio) == BTRFS_MAP_WRITE; in bbio_has_ordered_extent() 44 void btrfs_bio_init(struct btrfs_bio *bbio, struct btrfs_inode *inode, u64 file_offset, in btrfs_bio_init() argument 50 memset(bbio, 0, offsetof(struct btrfs_bio, bio)); in btrfs_bio_init() 51 bbio->inode = inode; in btrfs_bio_init() 52 bbio->end_io = end_io; in btrfs_bio_init() 53 bbio->private = private; in btrfs_bio_init() [all …]
|
| H A D | file-item.c | 339 int btrfs_lookup_bio_sums(struct btrfs_bio *bbio) in btrfs_lookup_bio_sums() argument 341 struct btrfs_inode *inode = bbio->inode; in btrfs_lookup_bio_sums() 343 struct bio *bio = &bbio->bio; in btrfs_lookup_bio_sums() 375 bbio->csum = kvcalloc(nblocks, csum_size, GFP_NOFS); in btrfs_lookup_bio_sums() 376 if (!bbio->csum) in btrfs_lookup_bio_sums() 379 bbio->csum = bbio->csum_inline; in btrfs_lookup_bio_sums() 424 if (bbio->csum_search_commit_root) { in btrfs_lookup_bio_sums() 433 u8 *csum_dst = bbio->csum + in btrfs_lookup_bio_sums() 440 if (bbio->csum != bbio->csum_inline) in btrfs_lookup_bio_sums() 441 kvfree(bbio->csum); in btrfs_lookup_bio_sums() [all …]
|
| H A D | compression.c | 56 static inline struct compressed_bio *to_compressed_bio(struct btrfs_bio *bbio) in to_compressed_bio() argument 58 return container_of(bbio, struct compressed_bio, bbio); in to_compressed_bio() 65 struct btrfs_bio *bbio; in alloc_compressed_bio() local 67 bbio = btrfs_bio(bio_alloc_bioset(NULL, BTRFS_MAX_COMPRESSED_PAGES, op, in alloc_compressed_bio() 69 btrfs_bio_init(bbio, inode, start, end_io, NULL); in alloc_compressed_bio() 70 return to_compressed_bio(bbio); in alloc_compressed_bio() 231 static void end_bbio_compressed_read(struct btrfs_bio *bbio) in end_bbio_compressed_read() argument 233 struct compressed_bio *cb = to_compressed_bio(bbio); in end_bbio_compressed_read() 234 blk_status_t status = bbio in end_bbio_compressed_read() 290 end_bbio_compressed_write(struct btrfs_bio * bbio) end_bbio_compressed_write() argument 522 btrfs_submit_compressed_read(struct btrfs_bio * bbio) btrfs_submit_compressed_read() argument [all...] |
| H A D | scrub.c | 890 static void scrub_repair_read_endio(struct btrfs_bio *bbio) in scrub_repair_read_endio() argument 892 struct scrub_stripe *stripe = bbio->private; in scrub_repair_read_endio() 895 int sector_nr = calc_sector_number(stripe, bio_first_bvec_all(&bbio->bio)); in scrub_repair_read_endio() 901 bio_for_each_bvec_all(bvec, &bbio->bio, i) in scrub_repair_read_endio() 904 if (bbio->bio.bi_status) { in scrub_repair_read_endio() 913 bio_put(&bbio->bio); in scrub_repair_read_endio() 924 static void scrub_bio_add_sector(struct btrfs_bio *bbio, struct scrub_stripe *stripe, in scrub_bio_add_sector() argument 927 struct btrfs_fs_info *fs_info = bbio->inode->root->fs_info; in scrub_bio_add_sector() 931 ret = bio_add_page(&bbio->bio, virt_to_page(kaddr), fs_info->sectorsize, in scrub_bio_add_sector() 934 * Caller should ensure the bbio ha in scrub_bio_add_sector() 949 struct btrfs_bio *bbio; alloc_scrub_bbio() local 962 struct btrfs_bio *bbio = NULL; scrub_stripe_submit_repair_read() local 1249 scrub_read_endio(struct btrfs_bio * bbio) scrub_read_endio() argument 1277 scrub_write_endio(struct btrfs_bio * bbio) scrub_write_endio() argument 1308 scrub_submit_write_bio(struct scrub_ctx * sctx,struct scrub_stripe * stripe,struct btrfs_bio * bbio,bool dev_replace) scrub_submit_write_bio() argument 1352 struct btrfs_bio *bbio = NULL; scrub_write_sectors() local 1804 struct btrfs_bio *bbio = NULL; scrub_submit_extent_sector_read() local 1884 struct btrfs_bio *bbio; scrub_submit_initial_read() local [all...] |
| H A D | extent_io.c | 98 struct btrfs_bio *bbio; member 162 struct btrfs_bio *bbio = bio_ctrl->bbio; in bio_set_csum_search_commit_root() local 164 ASSERT(bbio); in bio_set_csum_search_commit_root() 166 if (!(btrfs_op(&bbio->bio) == BTRFS_MAP_READ && is_data_inode(bbio->inode))) in bio_set_csum_search_commit_root() 169 bio_ctrl->bbio->csum_search_commit_root = in bio_set_csum_search_commit_root() 171 bio_ctrl->generation < btrfs_get_fs_generation(bbio->inode->root->fs_info)); in bio_set_csum_search_commit_root() 176 struct btrfs_bio *bbio = bio_ctrl->bbio; in submit_one_bio() local 178 if (!bbio) in submit_one_bio() 182 ASSERT(bbio->bio.bi_iter.bi_size); in submit_one_bio() 186 if (btrfs_op(&bbio->bio) == BTRFS_MAP_READ && in submit_one_bio() [all …]
|
| H A D | lzo.c | 278 struct btrfs_inode *inode = cb->bbio.inode; in lzo_compress_bio() 281 struct bio *bio = &cb->bbio.bio; in lzo_compress_bio() 396 bio_next_folio(fi, &cb->bbio.bio); in get_current_folio() 432 const struct btrfs_fs_info *fs_info = cb->bbio.inode->root->fs_info; in lzo_decompress_bio() 446 bio_first_folio(&fi, &cb->bbio.bio, 0); in lzo_decompress_bio() 465 struct btrfs_inode *inode = cb->bbio.inode; in lzo_decompress_bio() 496 struct btrfs_inode *inode = cb->bbio.inode; in lzo_decompress_bio() 517 struct btrfs_inode *inode = cb->bbio.inode; in lzo_decompress_bio()
|
| H A D | zlib.c | 150 struct btrfs_inode *inode = cb->bbio.inode; in zlib_compress_bio() 154 struct bio *bio = &cb->bbio.bio; in zlib_compress_bio() 355 bio_first_folio(&fi, &cb->bbio.bio, 0); in zlib_decompress_bio() 384 struct btrfs_inode *inode = cb->bbio.inode; in zlib_decompress_bio() 417 bio_next_folio(&fi, &cb->bbio.bio); in zlib_decompress_bio() 430 btrfs_err(cb->bbio.inode->root->fs_info, in zlib_decompress_bio() 432 ret, btrfs_root_id(cb->bbio.inode->root), in zlib_decompress_bio() 433 btrfs_ino(cb->bbio.inode), cb->start); in zlib_decompress_bio()
|
| H A D | zstd.c | 401 struct btrfs_inode *inode = cb->bbio.inode; in zstd_compress_bio() 405 struct bio *bio = &cb->bbio.bio; in zstd_compress_bio() 600 bio_first_folio(&fi, &cb->bbio.bio, 0); in zstd_decompress_bio() 608 struct btrfs_inode *inode = cb->bbio.inode; in zstd_decompress_bio() 631 struct btrfs_inode *inode = cb->bbio.inode; in zstd_decompress_bio() 665 bio_next_folio(&fi, &cb->bbio.bio); in zstd_decompress_bio()
|
| H A D | zoned.c | 2017 bool btrfs_use_zone_append(struct btrfs_bio *bbio) 2019 u64 start = (bbio->bio.bi_iter.bi_sector << SECTOR_SHIFT); in btrfs_use_zone_append() 2020 struct btrfs_inode *inode = bbio->inode; in btrfs_use_zone_append() 2031 if (btrfs_op(&bbio->bio) != BTRFS_MAP_WRITE) in btrfs_use_zone_append() 2056 void btrfs_record_physical_zoned(struct btrfs_bio *bbio) 2058 const u64 physical = bbio->bio.bi_iter.bi_sector << SECTOR_SHIFT; in btrfs_record_physical_zoned() 2059 struct btrfs_ordered_sum *sum = bbio->sums; in btrfs_record_physical_zoned() 2061 if (physical < bbio->orig_physical) in btrfs_record_physical_zoned() 2062 sum->logical -= bbio->orig_physical - physical; in btrfs_record_physical_zoned() 2064 sum->logical += physical - bbio in btrfs_record_physical_zoned() 2018 btrfs_use_zone_append(struct btrfs_bio * bbio) btrfs_use_zone_append() argument 2057 btrfs_record_physical_zoned(struct btrfs_bio * bbio) btrfs_record_physical_zoned() argument [all...] |
| H A D | inode.c | 866 struct bio *bio = &cb->bbio.bio; in compressed_bio_last_folio() 895 struct bio *bio = &cb->bbio.bio; in zero_last_folio() 905 struct bio *bio = &cb->bbio.bio; in round_up_last_block() 1030 total_compressed = cb->bbio.bio.bi_iter.bi_size; in compress_file_range() 1056 bio_first_folio_all(&cb->bbio.bio), false); in compress_file_range() 1173 compressed_size = async_extent->cb->bbio.bio.bi_iter.bi_size; in submit_one_async_extent() 1200 async_extent->cb->bbio.bio.bi_iter.bi_sector = ins.objectid >> SECTOR_SHIFT; in submit_one_async_extent() 1216 async_extent->cb->bbio.ordered = ordered; in submit_one_async_extent() 1223 btrfs_submit_bbio(&async_extent->cb->bbio, 0); in submit_one_async_extent() 3499 * @bbio 3509 btrfs_data_csum_ok(struct btrfs_bio * bbio,struct btrfs_device * dev,u32 bio_offset,const phys_addr_t paddrs[]) btrfs_data_csum_ok() argument 9499 btrfs_encoded_read_endio(struct btrfs_bio * bbio) btrfs_encoded_read_endio() argument 9533 struct btrfs_bio *bbio; btrfs_encoded_read_regular_fill_pages() local [all...] |
| H A D | relocation.c | 3976 static void reloc_endio(struct btrfs_bio *bbio) in reloc_endio() argument 3978 struct reloc_io_private *priv = bbio->private; in reloc_endio() 3980 if (bbio->bio.bi_status) in reloc_endio() 3981 WRITE_ONCE(priv->status, bbio->bio.bi_status); in reloc_endio() 3986 bio_put(&bbio->bio); in reloc_endio() 3994 struct btrfs_bio *bbio; in copy_remapped_data_io() local 4001 bbio = btrfs_bio_alloc(BIO_MAX_VECS, op, BTRFS_I(fs_info->btree_inode), in copy_remapped_data_io() 4003 bbio->bio.bi_iter.bi_sector = (addr >> SECTOR_SHIFT); in copy_remapped_data_io() 4004 bbio->is_remap = true; in copy_remapped_data_io() 4010 if (bio_add_page(&bbio->bio, pages[i], bytes, 0) < bytes) { in copy_remapped_data_io() [all …]
|
| H A D | disk-io.c | 255 int btree_csum_one_bio(struct btrfs_bio *bbio) in btree_csum_one_bio() argument 257 struct extent_buffer *eb = bbio->private; in btree_csum_one_bio() 265 if (WARN_ON_ONCE(bbio->file_offset != eb->start)) in btree_csum_one_bio() 267 if (WARN_ON_ONCE(bbio->bio.bi_iter.bi_size != eb->len)) in btree_csum_one_bio()
|
| /linux/drivers/md/bcache/ |
| H A D | request.c | 481 struct bbio bio; 501 struct bbio *b = container_of(bio, struct bbio, bio); in bch_cache_read_endio() 568 bio_key = &container_of(n, struct bbio, bio)->key; in cache_lookup_fn()
|
| H A D | bcache.h | 745 struct bbio { struct
|
| H A D | super.c | 949 if (bioset_init(&d->bio_split, 4, offsetof(struct bbio, bio), in bcache_device_init() 1934 sizeof(struct bbio) + in bch_cache_set_alloc() 1941 if (bioset_init(&c->bio_split, 4, offsetof(struct bbio, bio), in bch_cache_set_alloc()
|