Home
last modified time | relevance | path

Searched refs:bv_len (Results 1 – 25 of 76) sorted by relevance

1234

/linux/drivers/block/
H A Dn64cart.c69 (bv->bv_len & (MIN_ALIGNMENT - 1))); in n64cart_do_bvec()
79 n64cart_write_reg(PI_WRITE_REG, bv->bv_len - 1); in n64cart_do_bvec()
83 dma_unmap_page(dev, dma_addr, bv->bv_len, DMA_FROM_DEVICE); in n64cart_do_bvec()
99 pos += bvec.bv_len; in n64cart_submit_bio()
H A Dbrd.c147 bv.bv_len = min_t(u32, bv.bv_len, PAGE_SIZE - offset); in brd_rw_bvec()
158 memcpy_to_page(page, offset, kaddr, bv.bv_len); in brd_rw_bvec()
161 memcpy_from_page(kaddr, page, offset, bv.bv_len); in brd_rw_bvec()
163 memset(kaddr, 0, bv.bv_len); in brd_rw_bvec()
167 bio_advance_iter_single(bio, &bio->bi_iter, bv.bv_len); in brd_rw_bvec()
/linux/drivers/md/
H A Ddm-ebs-target.c71 unsigned int bv_len = bv->bv_len; in __ebs_rw_bvec() local
76 if (unlikely(!bv->bv_page || !bv_len)) in __ebs_rw_bvec()
82 while (bv_len) { in __ebs_rw_bvec()
83 cur_len = min(dm_bufio_get_block_size(ec->bufio) - buf_off, bv_len); in __ebs_rw_bvec()
86 if (op == REQ_OP_READ || buf_off || bv_len < dm_bufio_get_block_size(ec->bufio)) in __ebs_rw_bvec()
113 bv_len -= cur_len; in __ebs_rw_bvec()
H A Ddm-io-rewind.c27 while (idx >= 0 && bytes && bytes > bv[idx].bv_len) { in dm_bvec_iter_rewind()
28 bytes -= bv[idx].bv_len; in dm_bvec_iter_rewind()
41 iter->bi_bvec_done = bv[idx].bv_len - bytes; in dm_bvec_iter_rewind()
H A Ddm-log-writes.c367 block->vecs[i].bv_len, 0); in log_one_block()
368 if (ret != block->vecs[i].bv_len) { in log_one_block()
380 block->vecs[i].bv_len, 0); in log_one_block()
381 if (ret != block->vecs[i].bv_len) { in log_one_block()
387 sector += block->vecs[i].bv_len >> SECTOR_SHIFT; in log_one_block()
751 block->vecs[i].bv_len = bv.bv_len; in log_writes_map()
/linux/block/
H A Dt10-pi.c148 for (j = 0; j < iv.bv_len; j += tuple_sz) { in t10_pi_type1_prepare()
196 for (j = 0; j < iv.bv_len && intervals; j += tuple_sz) { in t10_pi_type1_complete()
322 for (j = 0; j < iv.bv_len; j += tuple_sz) { in ext_pi_type1_prepare()
359 for (j = 0; j < iv.bv_len && intervals; j += tuple_sz) { in ext_pi_type1_complete()
391 iter.data_size = bv.bv_len; in blk_integrity_generate()
428 iter.data_size = bv.bv_len; in blk_integrity_verify_iter()
H A Dblk-crypto-fallback.c281 if (!IS_ALIGNED(src_bv.bv_len | src_bv.bv_offset, in __blk_crypto_fallback_encrypt_bio()
287 __bio_add_page(enc_bio, enc_page, src_bv.bv_len, in __blk_crypto_fallback_encrypt_bio()
303 for (i = 0; i < src_bv.bv_len; i += data_unit_size) { in __blk_crypto_fallback_encrypt_bio()
315 src_bv.bv_len); in __blk_crypto_fallback_encrypt_bio()
398 if (!IS_ALIGNED(bv.bv_len | bv.bv_offset, data_unit_size)) in __blk_crypto_fallback_decrypt_bio()
404 for (i = 0; i < bv.bv_len; i += data_unit_size) { in __blk_crypto_fallback_decrypt_bio()
H A Dblk-map.c77 bvec->bv_len, in bio_copy_from_iter()
83 if (ret < bvec->bv_len) in bio_copy_from_iter()
108 bvec->bv_len, in bio_copy_to_iter()
114 if (ret < bvec->bv_len) in bio_copy_to_iter()
299 len += bio->bi_io_vec[i].bv_len; in bio_invalidate_vmalloc_pages()
348 p += bvec->bv_len; in bio_copy_kern_endio_read()
H A Dbio.c344 bio->bi_iter.bi_size += bio->bi_io_vec[i].bv_len; in bio_reuse()
699 if (done + bv.bv_len > new_size) { in bio_truncate()
707 bv.bv_len - offset); in bio_truncate()
710 done += bv.bv_len; in bio_truncate()
969 size_t bv_end = bv->bv_offset + bv->bv_len; in bvec_try_merge_page()
985 bv->bv_len += len; in bvec_try_merge_page()
1006 if (len > queue_max_segment_size(q) - bv->bv_len) in bvec_try_merge_hw_page()
1227 if (nbytes < bv->bv_len) { in bio_iov_iter_align_down()
1228 bv->bv_len -= nbytes; in bio_iov_iter_align_down()
1236 nbytes -= bv->bv_len; in bio_iov_iter_align_down()
[all …]
H A Dblk-integrity.c41 if (seg_size + iv.bv_len > queue_max_segment_size(q)) in blk_rq_count_integrity_sg()
44 seg_size += iv.bv_len; in blk_rq_count_integrity_sg()
48 seg_size = iv.bv_len; in blk_rq_count_integrity_sg()
/linux/drivers/md/bcache/
H A Dutil.c244 start: bv->bv_len = min_t(size_t, PAGE_SIZE - bv->bv_offset, in bch_bio_map()
251 base += bv->bv_len; in bch_bio_map()
254 size -= bv->bv_len; in bch_bio_map()
H A Ddebug.c136 cache_set_err_on(memcmp(p1, p2, bv.bv_len), in bch_data_verify()
144 bio_advance_iter(check, &citer, bv.bv_len); in bch_data_verify()
/linux/drivers/s390/block/
H A Ddasd_fba.c457 if (bv.bv_len & (blksize - 1)) in dasd_fba_build_cp_regular()
460 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_fba_build_cp_regular()
461 if (idal_is_needed (page_address(bv.bv_page), bv.bv_len)) in dasd_fba_build_cp_regular()
462 cidaw += bv.bv_len / blksize; in dasd_fba_build_cp_regular()
504 memcpy(copy + bv.bv_offset, dst, bv.bv_len); in dasd_fba_build_cp_regular()
508 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_fba_build_cp_regular()
582 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_fba_free_cp()
593 memcpy(dst, cda, bv.bv_len); in dasd_fba_free_cp()
H A Ddasd_diag.c537 if (bv.bv_len & (blksize - 1)) in dasd_diag_build_cp()
540 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_diag_build_cp()
557 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_diag_build_cp()
H A Ddcssblk.c918 !IS_ALIGNED(bvec.bv_len, PAGE_SIZE))) in dcssblk_submit_bio()
922 memcpy(page_addr, __va(source_addr), bvec.bv_len); in dcssblk_submit_bio()
924 memcpy(__va(source_addr), page_addr, bvec.bv_len); in dcssblk_submit_bio()
925 bytes_done += bvec.bv_len; in dcssblk_submit_bio()
/linux/net/ceph/
H A Dmessenger_v2.c156 it->bvec->bv_len - it->iov_offset), in do_try_sendpage()
173 iov_iter_bvec(&msg.msg_iter, ITER_SOURCE, &bv, 1, bv.bv_len); in do_try_sendpage()
237 iov_iter_bvec(&con->v2.in_iter, ITER_DEST, &con->v2.in_bvec, 1, bv->bv_len); in set_in_bvec()
282 con->v2.out_bvec.bv_len); in set_out_bvec()
294 con->v2.out_bvec.bv_len); in set_out_bvec_zero()
881 ceph_msg_data_advance(cursor, bv.bv_len); in calc_sg_cnt_cursor()
929 sg_set_page(*sg, bv.bv_page, bv.bv_len, bv.bv_offset); in init_sgs_cursor()
932 ceph_msg_data_advance(cursor, bv.bv_len); in init_sgs_cursor()
1091 len = min_t(int, len, bv.bv_len); in process_v2_sparse_read()
1866 con->v2.in_bvec.bv_len); in prepare_read_data_cont()
[all …]
/linux/fs/squashfs/
H A Dblock.c50 int bytes_to_copy = min_t(int, bvec->bv_len - offset, in copy_bio_to_actor()
69 if (offset >= bvec->bv_len) { in copy_bio_to_actor()
330 if (offset < bvec->bv_len - 1) { in squashfs_read_data()
H A Dlzo_wrapper.c77 int avail = min(bytes, ((int)bvec->bv_len) - offset); in lzo_uncompress()
H A Dzlib_wrapper.c83 avail = min(length, ((int)bvec->bv_len) - offset); in zlib_uncompress()
/linux/lib/
H A Diov_iter.c513 if (likely(size < bvec->bv_len)) in iov_iter_bvec_advance()
515 size -= bvec->bv_len; in iov_iter_bvec_advance()
642 size_t n = (--bvec)->bv_len; in iov_iter_revert()
679 return min(i->count, i->bvec->bv_len - i->iov_offset); in iov_iter_single_seg_count()
829 size_t len = bvec->bv_len - skip; in iov_iter_alignment_bvec()
1054 len = i->bvec->bv_len - skip; in first_bvec_segment()
1116 if (i->iov_offset == i->bvec->bv_len) { in __iov_iter_get_pages_alloc()
1185 size_t len = min(p->bv_len - skip, size); in bvec_npages()
1620 if (i->iov_offset == i->bvec->bv_len) { in iov_iter_extract_bvec_pages()
1649 size += bv.bv_len; in iov_iter_extract_bvec_pages()
[all …]
/linux/drivers/infiniband/core/
H A Drw.c238 sg_set_page(sg, bv.bv_page, bv.bv_len, bv.bv_offset); in rdma_rw_init_mr_wrs_bvec()
239 bvec_iter_advance(bvecs, iter, bv.bv_len); in rdma_rw_init_mr_wrs_bvec()
387 ctx->single.sge.length = bv.bv_len; in rdma_rw_init_single_wr_bvec()
453 sge->length = bv.bv_len; in rdma_rw_init_map_wrs_bvec()
456 total_len += bv.bv_len; in rdma_rw_init_map_wrs_bvec()
459 bvec_iter_advance_single(bvecs, iter, bv.bv_len); in rdma_rw_init_map_wrs_bvec()
514 mapped_len, bv.bv_len, dir, 0); in rdma_rw_init_iova_wrs_bvec()
518 mapped_len += bv.bv_len; in rdma_rw_init_iova_wrs_bvec()
519 bvec_iter_advance(bvec, iter, bv.bv_len); in rdma_rw_init_iova_wrs_bvec()
/linux/drivers/xen/
H A Dbiomerge.c15 return bfn1 + PFN_DOWN(vec1->bv_offset + vec1->bv_len) == bfn2; in xen_biovec_phys_mergeable()
/linux/fs/netfs/
H A Diterator.c123 len = bvecs[ix].bv_len; in netfs_limit_bvec()
132 len = min3(n, bvecs[ix].bv_len - skip, max_size); in netfs_limit_bvec()
/linux/include/linux/
H A Dbio.h162 bio_advance_iter_single((bio), &(iter), (bvl).bv_len))
171 bio_advance_iter_single((bio), &(iter), (bvl).bv_len))
185 #define bio_iter_last(bvec, iter) ((iter).bi_size == (bvec).bv_len)
289 fi->_seg_count = bvec->bv_len; in bio_first_folio()
/linux/fs/erofs/
H A Dzdata.c1502 bvec->bv_len = PAGE_SIZE; in z_erofs_fill_bio_vec()
1528 bvec->bv_len = round_up(zbv.end, bs) - bvec->bv_offset; in z_erofs_fill_bio_vec()
1734 if (cur + bvec.bv_len > end) in z_erofs_submit_queue()
1735 bvec.bv_len = end - cur; in z_erofs_submit_queue()
1736 DBG_BUGON(bvec.bv_len < sb->s_blocksize); in z_erofs_submit_queue()
1762 if (!bio_add_page(bio, bvec.bv_page, bvec.bv_len, in z_erofs_submit_queue()
1765 last_pa = cur + bvec.bv_len; in z_erofs_submit_queue()
1767 } while ((cur += bvec.bv_len) < end); in z_erofs_submit_queue()

1234