/linux/fs/crypto/ |
H A D | bio.c | 71 ret = bio_add_page(bio, ZERO_PAGE(0), bytes_this_page, 0); in fscrypt_zeroout_range_inline_crypt() 177 ret = bio_add_page(bio, pages[i++], offset, 0); in fscrypt_zeroout_range()
|
/linux/fs/xfs/ |
H A D | xfs_bio_io.c | 38 while (bio_add_page(bio, page, len, off) != len) { in xfs_rw_bdev()
|
/linux/fs/gfs2/ |
H A D | lops.c | 337 ret = bio_add_page(bio, page, size, offset); in gfs2_log_write() 341 ret = bio_add_page(bio, page, size, offset); in gfs2_log_write() 544 sz = bio_add_page(bio, page, bsize, off); in gfs2_find_jhead() 565 sz = bio_add_page(bio, page, bsize, off); in gfs2_find_jhead()
|
/linux/drivers/md/ |
H A D | raid1-10.c | 106 if (WARN_ON(!bio_add_page(bio, page, len, 0))) { in md_bio_reset_resync_pages()
|
H A D | dm-io.c | 371 if (!bio_add_page(bio, page, len, offset)) in do_region()
|
H A D | raid5-ppl.c | 492 if (!bio_add_page(bio, sh->ppl_page, PAGE_SIZE, 0)) { in ppl_submit_iounit()
|
/linux/fs/hfsplus/ |
H A D | wrapper.c | 79 ret = bio_add_page(bio, virt_to_page(buf), len, page_offset); in hfsplus_submit_bio()
|
/linux/block/ |
H A D | blk-lib.c | 215 added = bio_add_page(bio, ZERO_PAGE(0), len, 0); in __blkdev_issue_zero_pages()
|
H A D | bio.c | 1132 int bio_add_page(struct bio *bio, struct page *page, in bio_add_page() function 1154 EXPORT_SYMBOL(bio_add_page); 1184 return bio_add_page(bio, &folio->page, len, off) > 0; in bio_add_folio()
|
/linux/fs/bcachefs/ |
H A D | util.c | 630 BUG_ON(!bio_add_page(bio, page, len, offset)); in bch2_bio_map() 645 if (unlikely(!bio_add_page(bio, page, len, 0))) { in bch2_bio_alloc_pages()
|
H A D | io_write.c | 137 BUG_ON(!bio_add_page(bio, page, len, 0)); in bch2_bio_alloc_pages_pool()
|
/linux/drivers/target/ |
H A D | target_core_iblock.c | 546 while (bio_add_page(bio, sg_page(sg), sg->length, sg->offset) in iblock_execute_write_same() 801 while (bio_add_page(bio, sg_page(sg), sg->length, sg->offset) in iblock_execute_rw()
|
/linux/drivers/md/dm-vdo/ |
H A D | vio.c | 225 bytes_added = bio_add_page(bio, page, bytes, offset); in vio_reset_bio()
|
/linux/fs/nilfs2/ |
H A D | segbuf.c | 388 len = bio_add_page(wi->bio, bh->b_page, bh->b_size, bh_offset(bh)); in nilfs_segbuf_submit_bh()
|
/linux/drivers/nvme/target/ |
H A D | io-cmd-bdev.c | 298 while (bio_add_page(bio, sg_page(sg), sg->length, sg->offset) in nvmet_bdev_execute_rw()
|
/linux/include/linux/ |
H A D | bio.h | 415 int __must_check bio_add_page(struct bio *bio, struct page *page, unsigned len,
|
/linux/drivers/block/rnbd/ |
H A D | rnbd-srv.c | 150 if (bio_add_page(bio, virt_to_page(data), datalen, in process_rdma()
|
/linux/fs/btrfs/ |
H A D | scrub.c | 854 ret = bio_add_page(&bbio->bio, page, fs_info->sectorsize, pgoff); in scrub_stripe_submit_repair_read() 1221 ret = bio_add_page(&bbio->bio, page, fs_info->sectorsize, pgoff); in scrub_write_sectors() 1760 ret = bio_add_page(&bbio->bio, page, fs_info->sectorsize, pgoff); in scrub_submit_initial_read()
|
/linux/fs/ |
H A D | direct-io.c | 692 ret = bio_add_page(sdio->bio, sdio->cur_page, in dio_bio_add_page()
|
/linux/fs/nfs/blocklayout/ |
H A D | blocklayout.c | 162 if (bio_add_page(bio, page, *len, offset) < *len) { in do_add_page_to_bio()
|
/linux/fs/ntfs3/ |
H A D | fsntfs.c | 1548 if (bio_add_page(bio, page, add, off) < add) in ntfs_bio_pages() 1644 if (bio_add_page(bio, fill, add, 0) < add) in ntfs_bio_fill_1()
|
/linux/drivers/block/drbd/ |
H A D | drbd_actlog.c | 146 if (bio_add_page(bio, device->md_io.page, size, 0) != size) in _drbd_md_sync_page_io()
|
/linux/kernel/power/ |
H A D | swap.c | 281 if (bio_add_page(bio, page, PAGE_SIZE, 0) < PAGE_SIZE) { in hib_submit_io()
|
/linux/drivers/block/xen-blkback/ |
H A D | blkback.c | 1430 (bio_add_page(bio, in dispatch_rw_block_io()
|
/linux/fs/erofs/ |
H A D | zdata.c | 1673 if (!bio_add_page(bio, bvec.bv_page, bvec.bv_len, in z_erofs_submit_queue()
|