| /linux/fs/btrfs/ |
| H A D | bio.c | 25 struct btrfs_bio *bbio; 31 static inline bool is_data_bbio(const struct btrfs_bio *bbio) in is_data_bbio() 33 return bbio->inode && is_data_inode(bbio->inode); in is_data_bbio() 36 static bool bbio_has_ordered_extent(const struct btrfs_bio *bbio) in bbio_has_ordered_extent() 38 return is_data_bbio(bbio) && btrfs_op(&bbio->bio) == BTRFS_MAP_WRITE; in bbio_has_ordered_extent() 45 void btrfs_bio_init(struct btrfs_bio *bbio, struct btrfs_inode *inode, u64 file_offset, in btrfs_bio_init() 51 memset(bbio, 0, offsetof(struct btrfs_bio, bio)); in btrfs_bio_init() 52 bbio in btrfs_bio_init() 24 struct btrfs_bio *bbio; global() member 30 is_data_bbio(const struct btrfs_bio * bbio) is_data_bbio() argument 35 bbio_has_ordered_extent(const struct btrfs_bio * bbio) bbio_has_ordered_extent() argument 44 btrfs_bio_init(struct btrfs_bio * bbio,struct btrfs_inode * inode,u64 file_offset,btrfs_bio_end_io_t end_io,void * private) btrfs_bio_init() argument 70 struct btrfs_bio *bbio; btrfs_bio_alloc() local 83 struct btrfs_bio *bbio; btrfs_split_bio() local 111 btrfs_bio_end_io(struct btrfs_bio * bbio,blk_status_t status) btrfs_bio_end_io() argument 302 btrfs_check_read_bio(struct btrfs_bio * bbio,struct btrfs_device * dev) btrfs_check_read_bio() argument 376 struct btrfs_bio *bbio = container_of(work, struct btrfs_bio, end_io_work); simple_end_io_work() local 392 struct btrfs_bio *bbio = btrfs_bio(bio); btrfs_simple_end_io() local 408 struct btrfs_bio *bbio = btrfs_bio(bio); btrfs_raid56_end_io() local 425 struct btrfs_bio *bbio = container_of(work, struct btrfs_bio, end_io_work); orig_write_end_io_work() local 455 struct btrfs_bio *bbio = btrfs_bio(bio); btrfs_orig_write_end_io() local 463 struct btrfs_bio *bbio = container_of(work, struct btrfs_bio, end_io_work); clone_write_end_io_work() local 481 struct btrfs_bio *bbio = btrfs_bio(bio); btrfs_clone_write_end_io() local 590 btrfs_bio_csum(struct btrfs_bio * bbio) btrfs_bio_csum() argument 606 struct btrfs_bio *bbio; global() member 668 should_async_write(struct btrfs_bio * bbio) should_async_write() argument 705 btrfs_wq_submit_bio(struct btrfs_bio * bbio,struct btrfs_io_context * bioc,struct btrfs_io_stripe * smap,int mirror_num) btrfs_wq_submit_bio() argument 726 btrfs_append_map_length(struct btrfs_bio * bbio,u64 map_length) btrfs_append_map_length() argument 746 btrfs_submit_chunk(struct btrfs_bio * bbio,int mirror_num) btrfs_submit_chunk() argument 872 assert_bbio_alignment(struct btrfs_bio * bbio) assert_bbio_alignment() argument 899 btrfs_submit_bbio(struct btrfs_bio * bbio,int mirror_num) btrfs_submit_bbio() argument 1005 btrfs_submit_repair_write(struct btrfs_bio * bbio,int mirror_num,bool dev_replace) btrfs_submit_repair_write() argument [all...] |
| H A D | file-item.c | 346 int btrfs_lookup_bio_sums(struct btrfs_bio *bbio) in btrfs_lookup_bio_sums() argument 348 struct btrfs_inode *inode = bbio->inode; in btrfs_lookup_bio_sums() 350 struct bio *bio = &bbio->bio; in btrfs_lookup_bio_sums() 382 bbio->csum = kvcalloc(nblocks, csum_size, GFP_NOFS); in btrfs_lookup_bio_sums() 383 if (!bbio->csum) in btrfs_lookup_bio_sums() 386 bbio->csum = bbio->csum_inline; in btrfs_lookup_bio_sums() 431 if (bbio->csum_search_commit_root) { in btrfs_lookup_bio_sums() 440 u8 *csum_dst = bbio->csum + in btrfs_lookup_bio_sums() 447 if (bbio in btrfs_lookup_bio_sums() 774 csum_one_bio(struct btrfs_bio * bbio,struct bvec_iter * src) csum_one_bio() argument 802 struct btrfs_bio *bbio = container_of(work, struct btrfs_bio, csum_work); csum_one_bio_work() local 813 btrfs_csum_one_bio(struct btrfs_bio * bbio,bool async) btrfs_csum_one_bio() argument 853 btrfs_alloc_dummy_sum(struct btrfs_bio * bbio) btrfs_alloc_dummy_sum() argument [all...] |
| H A D | compression.c | 56 static inline struct compressed_bio *to_compressed_bio(struct btrfs_bio *bbio) in to_compressed_bio() argument 58 return container_of(bbio, struct compressed_bio, bbio); in to_compressed_bio() 65 struct btrfs_bio *bbio; in alloc_compressed_bio() local 67 bbio = btrfs_bio(bio_alloc_bioset(NULL, BTRFS_MAX_COMPRESSED_PAGES, op, in alloc_compressed_bio() 69 btrfs_bio_init(bbio, inode, start, end_io, NULL); in alloc_compressed_bio() 70 return to_compressed_bio(bbio); in alloc_compressed_bio() 231 static void end_bbio_compressed_read(struct btrfs_bio *bbio) in end_bbio_compressed_read() argument 233 struct compressed_bio *cb = to_compressed_bio(bbio); in end_bbio_compressed_read() 234 blk_status_t status = bbio in end_bbio_compressed_read() 290 end_bbio_compressed_write(struct btrfs_bio * bbio) end_bbio_compressed_write() argument 527 btrfs_submit_compressed_read(struct btrfs_bio * bbio) btrfs_submit_compressed_read() argument [all...] |
| H A D | scrub.c | 890 static void scrub_repair_read_endio(struct btrfs_bio *bbio) in scrub_repair_read_endio() argument 892 struct scrub_stripe *stripe = bbio->private; in scrub_repair_read_endio() 894 int sector_nr = calc_sector_number(stripe, bio_first_bvec_all(&bbio->bio)); in scrub_repair_read_endio() 895 const u32 bio_size = bio_get_size(&bbio->bio); in scrub_repair_read_endio() 899 if (bbio->bio.bi_status) { in scrub_repair_read_endio() 908 bio_put(&bbio->bio); in scrub_repair_read_endio() 919 static void scrub_bio_add_sector(struct btrfs_bio *bbio, struct scrub_stripe *stripe, in calc_next_mirror() 922 struct btrfs_fs_info *fs_info = bbio->inode->root->fs_info; in calc_next_mirror() 926 ret = bio_add_page(&bbio->bio, virt_to_page(kaddr), fs_info->sectorsize, in scrub_bio_add_sector() 929 * Caller should ensure the bbio ha in scrub_bio_add_sector() 924 scrub_bio_add_sector(struct btrfs_bio * bbio,struct scrub_stripe * stripe,int sector_nr) scrub_bio_add_sector() argument 949 struct btrfs_bio *bbio; alloc_scrub_bbio() local 962 struct btrfs_bio *bbio = NULL; scrub_stripe_submit_repair_read() local 1249 scrub_read_endio(struct btrfs_bio * bbio) scrub_read_endio() argument 1277 scrub_write_endio(struct btrfs_bio * bbio) scrub_write_endio() argument 1308 scrub_submit_write_bio(struct scrub_ctx * sctx,struct scrub_stripe * stripe,struct btrfs_bio * bbio,bool dev_replace) scrub_submit_write_bio() argument 1352 struct btrfs_bio *bbio = NULL; scrub_write_sectors() local 1804 struct btrfs_bio *bbio = NULL; scrub_submit_extent_sector_read() local 1884 struct btrfs_bio *bbio; scrub_submit_initial_read() local [all...] |
| H A D | extent_io.c | 99 struct btrfs_bio *bbio; 100 /* Last byte contained in bbio + 1 . */ 118 * generation with E's only when we are sure E is added to bio_ctrl->bbio 156 * Helper to set the csum search commit root option for a bio_ctrl's bbio 163 struct btrfs_bio *bbio = bio_ctrl->bbio; in bio_set_csum_search_commit_root() 165 ASSERT(bbio); in bio_set_csum_search_commit_root() 167 if (!(btrfs_op(&bbio->bio) == BTRFS_MAP_READ && is_data_inode(bbio->inode))) in bio_set_csum_search_commit_root() 170 bio_ctrl->bbio in bio_set_csum_search_commit_root() 98 struct btrfs_bio *bbio; global() member 162 struct btrfs_bio *bbio = bio_ctrl->bbio; bio_set_csum_search_commit_root() local 176 struct btrfs_bio *bbio = bio_ctrl->bbio; submit_one_bio() local 207 struct btrfs_bio *bbio = bio_ctrl->bbio; submit_write_bio() local 517 end_bbio_data_write(struct btrfs_bio * bbio) end_bbio_data_write() argument 573 end_bbio_data_read(struct btrfs_bio * bbio) end_bbio_data_read() argument 743 struct btrfs_bio *bbio; alloc_new_bio() local 2198 end_bbio_meta_write(struct btrfs_bio * bbio) end_bbio_meta_write() argument 2247 struct btrfs_bio *bbio; write_one_eb() local 3839 end_bbio_meta_read(struct btrfs_bio * bbio) end_bbio_meta_read() argument 3872 struct btrfs_bio *bbio; read_extent_buffer_pages_nowait() local [all...] |
| H A D | lzo.c | 262 struct btrfs_inode *inode = cb->bbio.inode; in copy_compressed_data_to_bio() 265 struct bio *bio = &cb->bbio.bio; in copy_compressed_data_to_bio() 380 bio_next_folio(fi, &cb->bbio.bio); in lzo_compress_bio() 416 struct btrfs_fs_info *fs_info = cb->bbio.inode->root->fs_info; in copy_compressed_segment() 418 const u32 compressed_len = bio_get_size(&cb->bbio.bio); in copy_compressed_segment() 431 bio_first_folio(&fi, &cb->bbio.bio, 0); in lzo_decompress_bio() 450 struct btrfs_inode *inode = cb->bbio.inode; in lzo_decompress_bio() 481 struct btrfs_inode *inode = cb->bbio.inode; in lzo_decompress_bio() 502 struct btrfs_inode *inode = cb->bbio.inode; in lzo_decompress_bio()
|
| H A D | zlib.c | 149 struct btrfs_inode *inode = cb->bbio.inode; in zlib_compress_bio() 153 struct bio *bio = &cb->bbio.bio; in zlib_compress_bio() 348 const size_t srclen = bio_get_size(&cb->bbio.bio); in zlib_decompress_bio() 351 bio_first_folio(&fi, &cb->bbio.bio, 0); in zlib_decompress_bio() 380 struct btrfs_inode *inode = cb->bbio.inode; in zlib_decompress_bio() 413 bio_next_folio(&fi, &cb->bbio.bio); in zlib_decompress_bio() 426 btrfs_err(cb->bbio.inode->root->fs_info, in zlib_decompress_bio() 428 ret, btrfs_root_id(cb->bbio.inode->root), in zlib_decompress_bio() 429 btrfs_ino(cb->bbio.inode), cb->start); in zlib_decompress_bio()
|
| H A D | zstd.c | 400 struct btrfs_inode *inode = cb->bbio.inode; in zstd_compress_bio() 404 struct bio *bio = &cb->bbio.bio; in zstd_compress_bio() 588 size_t srclen = bio_get_size(&cb->bbio.bio); in zstd_decompress_bio() 597 bio_first_folio(&fi, &cb->bbio.bio, 0); in zstd_decompress_bio() 605 struct btrfs_inode *inode = cb->bbio.inode; in zstd_decompress_bio() 628 struct btrfs_inode *inode = cb->bbio.inode; in zstd_decompress_bio() 662 bio_next_folio(&fi, &cb->bbio.bio); in zstd_decompress_bio()
|
| H A D | zoned.c | 2025 bool btrfs_use_zone_append(struct btrfs_bio *bbio) 2027 u64 start = (bbio->bio.bi_iter.bi_sector << SECTOR_SHIFT); in btrfs_use_zone_append() 2028 struct btrfs_inode *inode = bbio->inode; in btrfs_use_zone_append() 2039 if (btrfs_op(&bbio->bio) != BTRFS_MAP_WRITE) in btrfs_use_zone_append() 2064 void btrfs_record_physical_zoned(struct btrfs_bio *bbio) 2066 const u64 physical = bbio->bio.bi_iter.bi_sector << SECTOR_SHIFT; in btrfs_record_physical_zoned() 2067 struct btrfs_ordered_sum *sum = bbio->sums; in btrfs_record_physical_zoned() 2069 if (physical < bbio->orig_physical) in btrfs_record_physical_zoned() 2070 sum->logical -= bbio->orig_physical - physical; in btrfs_record_physical_zoned() 2072 sum->logical += physical - bbio in btrfs_record_physical_zoned() 2026 btrfs_use_zone_append(struct btrfs_bio * bbio) btrfs_use_zone_append() argument 2065 btrfs_record_physical_zoned(struct btrfs_bio * bbio) btrfs_record_physical_zoned() argument [all...] |
| H A D | inode.c | 822 struct bio *bio = &cb->bbio.bio; in inode_need_compress() 851 struct bio *bio = &cb->bbio.bio; in extent_range_clear_dirty_for_io() 975 total_compressed = cb->bbio.bio.bi_iter.bi_size; in compress_file_range() 983 total_compressed = cb->bbio.bio.bi_iter.bi_size; in compress_file_range() 1088 compressed_size = async_extent->cb->bbio.bio.bi_iter.bi_size; in compress_file_range() 1115 async_extent->cb->bbio.bio.bi_iter.bi_sector = ins.objectid >> SECTOR_SHIFT; in submit_uncompressed_range() 1131 async_extent->cb->bbio.ordered = ordered; in submit_uncompressed_range() 1138 btrfs_submit_bbio(&async_extent->cb->bbio, 0); in submit_one_async_extent() 2361 compressed_size = cb->bbio.bio.bi_iter.bi_size; in run_delalloc_nocow() 2373 bio_first_folio_all(&cb->bbio in run_delalloc_nocow() 3538 btrfs_data_csum_ok(struct btrfs_bio * bbio,struct btrfs_device * dev,u32 bio_offset,const phys_addr_t paddrs[]) btrfs_data_csum_ok() argument 9547 btrfs_encoded_read_endio(struct btrfs_bio * bbio) btrfs_encoded_read_endio() argument 9581 struct btrfs_bio *bbio; btrfs_encoded_read_regular_fill_pages() local [all...] |
| H A D | relocation.c | 3968 static void reloc_endio(struct btrfs_bio *bbio) 3970 struct reloc_io_private *priv = bbio->private; 3972 if (bbio->bio.bi_status) 3973 WRITE_ONCE(priv->status, bbio->bio.bi_status); 3978 bio_put(&bbio->bio); in reloc_endio() 3986 struct btrfs_bio *bbio; in reloc_endio() 3993 bbio = btrfs_bio_alloc(BIO_MAX_VECS, op, BTRFS_I(fs_info->btree_inode), in copy_remapped_data_io() 3995 bbio->bio.bi_iter.bi_sector = (addr >> SECTOR_SHIFT); in copy_remapped_data_io() 3996 bbio->is_remap = true; in copy_remapped_data_io() 4002 if (bio_add_page(&bbio in copy_remapped_data_io() 3976 reloc_endio(struct btrfs_bio * bbio) reloc_endio() argument 3994 struct btrfs_bio *bbio; copy_remapped_data_io() local [all...] |
| H A D | disk-io.c | 258 int btree_csum_one_bio(struct btrfs_bio *bbio) in btree_csum_one_bio() 260 struct extent_buffer *eb = bbio->private; in btree_csum_one_bio() 268 if (WARN_ON_ONCE(bbio->file_offset != eb->start)) in btree_csum_one_bio() 270 if (WARN_ON_ONCE(bbio->bio.bi_iter.bi_size != eb->len)) in btree_csum_one_bio() 255 btree_csum_one_bio(struct btrfs_bio * bbio) btree_csum_one_bio() argument
|
| /linux/drivers/md/bcache/ |
| H A D | request.c | 481 struct bbio bio; 501 struct bbio *b = container_of(bio, struct bbio, bio); in bch_cache_read_endio() 568 bio_key = &container_of(n, struct bbio, bio)->key; in cache_lookup_fn()
|
| H A D | bcache.h | 745 struct bbio { struct
|
| H A D | super.c | 949 if (bioset_init(&d->bio_split, 4, offsetof(struct bbio, bio), in bcache_device_init() 1942 sizeof(struct bbio) + in bch_cache_set_alloc() 1949 if (bioset_init(&c->bio_split, 4, offsetof(struct bbio, bio), in bch_cache_set_alloc()
|