Home
last modified time | relevance | path

Searched refs:bbio (Results 1 – 15 of 15) sorted by relevance

/linux/fs/btrfs/
H A Dbio.c24 struct btrfs_bio *bbio; member
30 static inline bool is_data_bbio(const struct btrfs_bio *bbio) in is_data_bbio() argument
32 return bbio->inode && is_data_inode(bbio->inode); in is_data_bbio()
35 static bool bbio_has_ordered_extent(const struct btrfs_bio *bbio) in bbio_has_ordered_extent() argument
37 return is_data_bbio(bbio) && btrfs_op(&bbio->bio) == BTRFS_MAP_WRITE; in bbio_has_ordered_extent()
44 void btrfs_bio_init(struct btrfs_bio *bbio, struct btrfs_inode *inode, u64 file_offset, in btrfs_bio_init() argument
50 memset(bbio, 0, offsetof(struct btrfs_bio, bio)); in btrfs_bio_init()
51 bbio->inode = inode; in btrfs_bio_init()
52 bbio->end_io = end_io; in btrfs_bio_init()
53 bbio->private = private; in btrfs_bio_init()
[all …]
H A Dfile-item.c339 int btrfs_lookup_bio_sums(struct btrfs_bio *bbio) in btrfs_lookup_bio_sums() argument
341 struct btrfs_inode *inode = bbio->inode; in btrfs_lookup_bio_sums()
343 struct bio *bio = &bbio->bio; in btrfs_lookup_bio_sums()
375 bbio->csum = kvcalloc(nblocks, csum_size, GFP_NOFS); in btrfs_lookup_bio_sums()
376 if (!bbio->csum) in btrfs_lookup_bio_sums()
379 bbio->csum = bbio->csum_inline; in btrfs_lookup_bio_sums()
424 if (bbio->csum_search_commit_root) { in btrfs_lookup_bio_sums()
433 u8 *csum_dst = bbio->csum + in btrfs_lookup_bio_sums()
440 if (bbio->csum != bbio->csum_inline) in btrfs_lookup_bio_sums()
441 kvfree(bbio->csum); in btrfs_lookup_bio_sums()
[all …]
H A Dcompression.c56 static inline struct compressed_bio *to_compressed_bio(struct btrfs_bio *bbio) in to_compressed_bio() argument
58 return container_of(bbio, struct compressed_bio, bbio); in to_compressed_bio()
65 struct btrfs_bio *bbio; in alloc_compressed_bio() local
67 bbio = btrfs_bio(bio_alloc_bioset(NULL, BTRFS_MAX_COMPRESSED_PAGES, op, in alloc_compressed_bio()
69 btrfs_bio_init(bbio, inode, start, end_io, NULL); in alloc_compressed_bio()
70 return to_compressed_bio(bbio); in alloc_compressed_bio()
231 static void end_bbio_compressed_read(struct btrfs_bio *bbio) in end_bbio_compressed_read() argument
233 struct compressed_bio *cb = to_compressed_bio(bbio); in end_bbio_compressed_read()
234 blk_status_t status = bbio in end_bbio_compressed_read()
290 end_bbio_compressed_write(struct btrfs_bio * bbio) end_bbio_compressed_write() argument
522 btrfs_submit_compressed_read(struct btrfs_bio * bbio) btrfs_submit_compressed_read() argument
[all...]
H A Dscrub.c890 static void scrub_repair_read_endio(struct btrfs_bio *bbio) in scrub_repair_read_endio() argument
892 struct scrub_stripe *stripe = bbio->private; in scrub_repair_read_endio()
895 int sector_nr = calc_sector_number(stripe, bio_first_bvec_all(&bbio->bio)); in scrub_repair_read_endio()
901 bio_for_each_bvec_all(bvec, &bbio->bio, i) in scrub_repair_read_endio()
904 if (bbio->bio.bi_status) { in scrub_repair_read_endio()
913 bio_put(&bbio->bio); in scrub_repair_read_endio()
924 static void scrub_bio_add_sector(struct btrfs_bio *bbio, struct scrub_stripe *stripe, in scrub_bio_add_sector() argument
927 struct btrfs_fs_info *fs_info = bbio->inode->root->fs_info; in scrub_bio_add_sector()
931 ret = bio_add_page(&bbio->bio, virt_to_page(kaddr), fs_info->sectorsize, in scrub_bio_add_sector()
934 * Caller should ensure the bbio ha in scrub_bio_add_sector()
949 struct btrfs_bio *bbio; alloc_scrub_bbio() local
962 struct btrfs_bio *bbio = NULL; scrub_stripe_submit_repair_read() local
1249 scrub_read_endio(struct btrfs_bio * bbio) scrub_read_endio() argument
1277 scrub_write_endio(struct btrfs_bio * bbio) scrub_write_endio() argument
1308 scrub_submit_write_bio(struct scrub_ctx * sctx,struct scrub_stripe * stripe,struct btrfs_bio * bbio,bool dev_replace) scrub_submit_write_bio() argument
1352 struct btrfs_bio *bbio = NULL; scrub_write_sectors() local
1804 struct btrfs_bio *bbio = NULL; scrub_submit_extent_sector_read() local
1884 struct btrfs_bio *bbio; scrub_submit_initial_read() local
[all...]
H A Dextent_io.c98 struct btrfs_bio *bbio; member
162 struct btrfs_bio *bbio = bio_ctrl->bbio; in bio_set_csum_search_commit_root() local
164 ASSERT(bbio); in bio_set_csum_search_commit_root()
166 if (!(btrfs_op(&bbio->bio) == BTRFS_MAP_READ && is_data_inode(bbio->inode))) in bio_set_csum_search_commit_root()
169 bio_ctrl->bbio->csum_search_commit_root = in bio_set_csum_search_commit_root()
171 bio_ctrl->generation < btrfs_get_fs_generation(bbio->inode->root->fs_info)); in bio_set_csum_search_commit_root()
176 struct btrfs_bio *bbio = bio_ctrl->bbio; in submit_one_bio() local
178 if (!bbio) in submit_one_bio()
182 ASSERT(bbio->bio.bi_iter.bi_size); in submit_one_bio()
186 if (btrfs_op(&bbio->bio) == BTRFS_MAP_READ && in submit_one_bio()
[all …]
H A Dlzo.c278 struct btrfs_inode *inode = cb->bbio.inode; in lzo_compress_bio()
281 struct bio *bio = &cb->bbio.bio; in lzo_compress_bio()
396 bio_next_folio(fi, &cb->bbio.bio); in get_current_folio()
432 const struct btrfs_fs_info *fs_info = cb->bbio.inode->root->fs_info; in lzo_decompress_bio()
446 bio_first_folio(&fi, &cb->bbio.bio, 0); in lzo_decompress_bio()
465 struct btrfs_inode *inode = cb->bbio.inode; in lzo_decompress_bio()
496 struct btrfs_inode *inode = cb->bbio.inode; in lzo_decompress_bio()
517 struct btrfs_inode *inode = cb->bbio.inode; in lzo_decompress_bio()
H A Dzlib.c150 struct btrfs_inode *inode = cb->bbio.inode; in zlib_compress_bio()
154 struct bio *bio = &cb->bbio.bio; in zlib_compress_bio()
355 bio_first_folio(&fi, &cb->bbio.bio, 0); in zlib_decompress_bio()
384 struct btrfs_inode *inode = cb->bbio.inode; in zlib_decompress_bio()
417 bio_next_folio(&fi, &cb->bbio.bio); in zlib_decompress_bio()
430 btrfs_err(cb->bbio.inode->root->fs_info, in zlib_decompress_bio()
432 ret, btrfs_root_id(cb->bbio.inode->root), in zlib_decompress_bio()
433 btrfs_ino(cb->bbio.inode), cb->start); in zlib_decompress_bio()
H A Dzstd.c401 struct btrfs_inode *inode = cb->bbio.inode; in zstd_compress_bio()
405 struct bio *bio = &cb->bbio.bio; in zstd_compress_bio()
600 bio_first_folio(&fi, &cb->bbio.bio, 0); in zstd_decompress_bio()
608 struct btrfs_inode *inode = cb->bbio.inode; in zstd_decompress_bio()
631 struct btrfs_inode *inode = cb->bbio.inode; in zstd_decompress_bio()
665 bio_next_folio(&fi, &cb->bbio.bio); in zstd_decompress_bio()
H A Dzoned.c2017 bool btrfs_use_zone_append(struct btrfs_bio *bbio)
2019 u64 start = (bbio->bio.bi_iter.bi_sector << SECTOR_SHIFT); in btrfs_use_zone_append()
2020 struct btrfs_inode *inode = bbio->inode; in btrfs_use_zone_append()
2031 if (btrfs_op(&bbio->bio) != BTRFS_MAP_WRITE) in btrfs_use_zone_append()
2056 void btrfs_record_physical_zoned(struct btrfs_bio *bbio)
2058 const u64 physical = bbio->bio.bi_iter.bi_sector << SECTOR_SHIFT; in btrfs_record_physical_zoned()
2059 struct btrfs_ordered_sum *sum = bbio->sums; in btrfs_record_physical_zoned()
2061 if (physical < bbio->orig_physical) in btrfs_record_physical_zoned()
2062 sum->logical -= bbio->orig_physical - physical; in btrfs_record_physical_zoned()
2064 sum->logical += physical - bbio in btrfs_record_physical_zoned()
2018 btrfs_use_zone_append(struct btrfs_bio * bbio) btrfs_use_zone_append() argument
2057 btrfs_record_physical_zoned(struct btrfs_bio * bbio) btrfs_record_physical_zoned() argument
[all...]
H A Dinode.c866 struct bio *bio = &cb->bbio.bio; in compressed_bio_last_folio()
895 struct bio *bio = &cb->bbio.bio; in zero_last_folio()
905 struct bio *bio = &cb->bbio.bio; in round_up_last_block()
1030 total_compressed = cb->bbio.bio.bi_iter.bi_size; in compress_file_range()
1056 bio_first_folio_all(&cb->bbio.bio), false); in compress_file_range()
1173 compressed_size = async_extent->cb->bbio.bio.bi_iter.bi_size; in submit_one_async_extent()
1200 async_extent->cb->bbio.bio.bi_iter.bi_sector = ins.objectid >> SECTOR_SHIFT; in submit_one_async_extent()
1216 async_extent->cb->bbio.ordered = ordered; in submit_one_async_extent()
1223 btrfs_submit_bbio(&async_extent->cb->bbio, 0); in submit_one_async_extent()
3499 * @bbio
3509 btrfs_data_csum_ok(struct btrfs_bio * bbio,struct btrfs_device * dev,u32 bio_offset,const phys_addr_t paddrs[]) btrfs_data_csum_ok() argument
9499 btrfs_encoded_read_endio(struct btrfs_bio * bbio) btrfs_encoded_read_endio() argument
9533 struct btrfs_bio *bbio; btrfs_encoded_read_regular_fill_pages() local
[all...]
H A Drelocation.c3976 static void reloc_endio(struct btrfs_bio *bbio) in reloc_endio() argument
3978 struct reloc_io_private *priv = bbio->private; in reloc_endio()
3980 if (bbio->bio.bi_status) in reloc_endio()
3981 WRITE_ONCE(priv->status, bbio->bio.bi_status); in reloc_endio()
3986 bio_put(&bbio->bio); in reloc_endio()
3994 struct btrfs_bio *bbio; in copy_remapped_data_io() local
4001 bbio = btrfs_bio_alloc(BIO_MAX_VECS, op, BTRFS_I(fs_info->btree_inode), in copy_remapped_data_io()
4003 bbio->bio.bi_iter.bi_sector = (addr >> SECTOR_SHIFT); in copy_remapped_data_io()
4004 bbio->is_remap = true; in copy_remapped_data_io()
4010 if (bio_add_page(&bbio->bio, pages[i], bytes, 0) < bytes) { in copy_remapped_data_io()
[all …]
H A Ddisk-io.c255 int btree_csum_one_bio(struct btrfs_bio *bbio) in btree_csum_one_bio() argument
257 struct extent_buffer *eb = bbio->private; in btree_csum_one_bio()
265 if (WARN_ON_ONCE(bbio->file_offset != eb->start)) in btree_csum_one_bio()
267 if (WARN_ON_ONCE(bbio->bio.bi_iter.bi_size != eb->len)) in btree_csum_one_bio()
/linux/drivers/md/bcache/
H A Drequest.c481 struct bbio bio;
501 struct bbio *b = container_of(bio, struct bbio, bio); in bch_cache_read_endio()
568 bio_key = &container_of(n, struct bbio, bio)->key; in cache_lookup_fn()
H A Dbcache.h745 struct bbio { struct
H A Dsuper.c949 if (bioset_init(&d->bio_split, 4, offsetof(struct bbio, bio), in bcache_device_init()
1934 sizeof(struct bbio) + in bch_cache_set_alloc()
1941 if (bioset_init(&c->bio_split, 4, offsetof(struct bbio, bio), in bch_cache_set_alloc()