/linux/include/linux/ |
H A D | bvec.h | 44 static inline void bvec_set_page(struct bio_vec *bv, struct page *page, in bvec_set_page() argument 47 bv->bv_page = page; in bvec_set_page() 48 bv->bv_len = len; in bvec_set_page() 49 bv->bv_offset = offset; in bvec_set_page() 59 static inline void bvec_set_folio(struct bio_vec *bv, struct folio *folio, in bvec_set_folio() argument 62 bvec_set_page(bv, &folio->page, len, offset); in bvec_set_folio() 71 static inline void bvec_set_virt(struct bio_vec *bv, void *vaddr, in bvec_set_virt() argument 74 bvec_set_page(bv, virt_to_page(vaddr), len, offset_in_page(vaddr)); in bvec_set_virt() 89 struct bio_vec bv; member 140 static inline bool bvec_iter_advance(const struct bio_vec *bv, in bvec_iter_advance() argument [all …]
|
/linux/drivers/md/bcache/ |
H A D | util.c | 234 struct bio_vec *bv = bio->bi_io_vec; in bch_bio_map() local 239 bv->bv_offset = base ? offset_in_page(base) : 0; in bch_bio_map() 242 for (; size; bio->bi_vcnt++, bv++) { in bch_bio_map() 243 bv->bv_offset = 0; in bch_bio_map() 244 start: bv->bv_len = min_t(size_t, PAGE_SIZE - bv->bv_offset, in bch_bio_map() 247 bv->bv_page = is_vmalloc_addr(base) in bch_bio_map() 251 base += bv->bv_len; in bch_bio_map() 254 size -= bv->bv_len; in bch_bio_map() 271 struct bio_vec *bv; in bch_bio_alloc_pages() local 277 for (i = 0, bv = bio->bi_io_vec; i < bio->bi_vcnt; bv++, i++) { in bch_bio_alloc_pages() [all …]
|
H A D | debug.c | 112 struct bio_vec bv, cbv; in bch_data_verify() local 130 bio_for_each_segment(bv, bio, iter) { in bch_data_verify() 131 void *p1 = bvec_kmap_local(&bv); in bch_data_verify() 137 cache_set_err_on(memcmp(p1, p2, bv.bv_len), in bch_data_verify() 145 bio_advance_iter(check, &citer, bv.bv_len); in bch_data_verify()
|
/linux/arch/parisc/kernel/ |
H A D | entry.S | 161 bv,n 0(%r3) 772 bv %r0(%r2) 781 bv %r0(%r2) 940 bv %r0(%r20) 969 bv %r0(%r1) 1699 bv %r0(%r2) 1895 bv %r0(%r19) /* jumps to schedule() */ 1926 bv %r0(%rp) 1994 bv,n (%r1) 2090 bv,n (%r1) [all …]
|
H A D | real2.S | 79 bv 0(%r31) 101 bv 0(%rp) 121 bv 0(%r2) 136 bv 0(%r2) 177 bv 0(%r2) 214 bv 0(%r2) 260 bv 0(%r31) 273 bv 0(%rp) 288 bv %r0(%r2)
|
H A D | hpmc.S | 154 bv (r3) /* call pdce_proc */ 169 bv (%r3) /* call pdce_proc */ 199 bv (%r3) /* call pdce_proc */ 221 bv (%r5) 269 bv (%r3) /* call pdce_proc */
|
H A D | head.S | 101 bv,n (%r1) 208 bv,n (%rp) 250 bv (%rp) 263 bv (%r3) 277 bv (%r3) 396 bv (%rp)
|
H A D | pacache.S | 175 2: bv %r0(%r2) 184 bv,n %r0(%r2) 245 bv %r0(%r2) 306 bv %r0(%r2) 369 bv %r0(%r2) 480 bv %r0(%r2) 660 bv %r0(%r2) 728 bv %r0(%r2) 777 bv %r0(%r2) 826 bv %r0(%r2) [all …]
|
/linux/drivers/net/ethernet/netronome/nfp/bpf/ |
H A D | main.c | 66 struct nfp_bpf_vnic *bv; in nfp_bpf_vnic_alloc() local 79 bv = kzalloc(sizeof(*bv), GFP_KERNEL); in nfp_bpf_vnic_alloc() 80 if (!bv) in nfp_bpf_vnic_alloc() 82 nn->app_priv = bv; in nfp_bpf_vnic_alloc() 88 bv->start_off = nn_readw(nn, NFP_NET_CFG_BPF_START); in nfp_bpf_vnic_alloc() 89 bv->tgt_done = nn_readw(nn, NFP_NET_CFG_BPF_DONE); in nfp_bpf_vnic_alloc() 99 struct nfp_bpf_vnic *bv = nn->app_priv; in nfp_bpf_vnic_free() local 101 WARN_ON(bv->tc_prog); in nfp_bpf_vnic_free() 102 kfree(bv); in nfp_bpf_vnic_free() 111 struct nfp_bpf_vnic *bv; in nfp_bpf_setup_tc_block_cb() local [all …]
|
/linux/block/ |
H A D | blk-merge.c | 21 static inline void bio_get_first_bvec(struct bio *bio, struct bio_vec *bv) in bio_get_first_bvec() argument 23 *bv = mp_bvec_iter_bvec(bio->bi_io_vec, bio->bi_iter); in bio_get_first_bvec() 26 static inline void bio_get_last_bvec(struct bio *bio, struct bio_vec *bv) in bio_get_last_bvec() argument 31 bio_get_first_bvec(bio, bv); in bio_get_last_bvec() 32 if (bv->bv_len == bio->bi_iter.bi_size) in bio_get_last_bvec() 42 *bv = bio->bi_io_vec[idx]; in bio_get_last_bvec() 49 bv->bv_len = iter.bi_bvec_done; in bio_get_last_bvec() 270 const struct bio_vec *bv, unsigned *nsegs, unsigned *bytes, in bvec_split_segs() argument 274 unsigned len = min(bv->bv_len, max_len); in bvec_split_segs() 279 seg_size = get_max_segment_size(lim, bvec_phys(bv) + total_len, len); in bvec_split_segs() [all …]
|
H A D | bio.c | 162 void bvec_free(mempool_t *pool, struct bio_vec *bv, unsigned short nr_vecs) in bvec_free() argument 167 mempool_free(bv, pool); in bvec_free() 169 kmem_cache_free(biovec_slab(nr_vecs)->slab, bv); in bvec_free() 622 struct bio_vec bv; in zero_fill_bio_iter() local 625 __bio_for_each_segment(bv, bio, iter, start) in zero_fill_bio_iter() 626 memzero_bvec(&bv); in zero_fill_bio_iter() 642 struct bio_vec bv; in bio_truncate() local 653 bio_for_each_segment(bv, bio, iter) { in bio_truncate() 654 if (done + bv.bv_len > new_size) { in bio_truncate() 661 zero_user(bv.bv_page, bv.bv_offset + offset, in bio_truncate() [all …]
|
H A D | t10-pi.c | 381 struct bio_vec bv; in blk_integrity_generate() local 387 bio_for_each_segment(bv, bio, bviter) { in blk_integrity_generate() 388 void *kaddr = bvec_kmap_local(&bv); in blk_integrity_generate() 391 iter.data_size = bv.bv_len; in blk_integrity_generate() 413 struct bio_vec bv; in blk_integrity_verify() local 423 __bio_for_each_segment(bv, bio, bviter, bip->bio_iter) { in blk_integrity_verify() 424 void *kaddr = bvec_kmap_local(&bv); in blk_integrity_verify() 428 iter.data_size = bv.bv_len; in blk_integrity_verify()
|
H A D | bio-integrity.c | 107 static void bio_integrity_unpin_bvec(struct bio_vec *bv, int nr_vecs, in bio_integrity_unpin_bvec() argument 113 if (dirty && !PageCompound(bv[i].bv_page)) in bio_integrity_unpin_bvec() 114 set_page_dirty_lock(bv[i].bv_page); in bio_integrity_unpin_bvec() 115 unpin_user_page(bv[i].bv_page); in bio_integrity_unpin_bvec() 171 struct bio_vec *bv = &bip->bip_vec[bip->bip_vcnt - 1]; in bio_integrity_add_page() local 174 if (bvec_try_merge_hw_page(q, bv, page, len, offset, in bio_integrity_add_page() 188 if (bvec_gap_to_prev(&q->limits, bv, offset)) in bio_integrity_add_page()
|
H A D | blk-crypto.c | 218 struct bio_vec bv; in bio_crypt_check_alignment() local 220 bio_for_each_segment(bv, bio, iter) { in bio_crypt_check_alignment() 221 if (!IS_ALIGNED(bv.bv_len | bv.bv_offset, data_unit_size)) in bio_crypt_check_alignment()
|
/linux/fs/bcachefs/ |
H A D | checksum.c | 262 struct bio_vec bv; in __bch2_checksum_bio() local 278 __bio_for_each_segment(bv, bio, *iter, *iter) { in __bch2_checksum_bio() 279 void *p = kmap_local_page(bv.bv_page) + bv.bv_offset; in __bch2_checksum_bio() 281 bch2_checksum_update(&state, p, bv.bv_len); in __bch2_checksum_bio() 285 __bio_for_each_bvec(bv, bio, *iter, *iter) in __bch2_checksum_bio() 286 bch2_checksum_update(&state, page_address(bv.bv_page) + bv.bv_offset, in __bch2_checksum_bio() 287 bv.bv_len); in __bch2_checksum_bio() 301 __bio_for_each_segment(bv, bio, *iter, *iter) { in __bch2_checksum_bio() 302 void *p = kmap_local_page(bv.bv_page) + bv.bv_offset; in __bch2_checksum_bio() 304 crypto_shash_update(desc, p, bv.bv_len); in __bch2_checksum_bio() [all …]
|
H A D | compress.c | 43 struct bio_vec bv; in bio_phys_contig() local 47 __bio_for_each_bvec(bv, bio, iter, start) { in bio_phys_contig() 49 expected_start != page_address(bv.bv_page) + bv.bv_offset) in bio_phys_contig() 52 expected_start = page_address(bv.bv_page) + in bio_phys_contig() 53 bv.bv_offset + bv.bv_len; in bio_phys_contig() 63 struct bio_vec bv; in __bio_map_or_bounce() local 81 __bio_for_each_segment(bv, bio, iter, start) { in __bio_map_or_bounce() 83 bv.bv_offset) in __bio_map_or_bounce() 86 if (bv.bv_len < iter.bi_size && in __bio_map_or_bounce() 87 bv.bv_offset + bv.bv_len < PAGE_SIZE) in __bio_map_or_bounce() [all …]
|
H A D | fs-io.h | 20 static inline struct folio_vec biovec_to_foliovec(struct bio_vec bv) in biovec_to_foliovec() argument 23 struct folio *folio = page_folio(bv.bv_page); in biovec_to_foliovec() 24 size_t offset = (folio_page_idx(folio, bv.bv_page) << PAGE_SHIFT) + in biovec_to_foliovec() 25 bv.bv_offset; in biovec_to_foliovec() 26 size_t len = min_t(size_t, folio_size(folio) - offset, bv.bv_len); in biovec_to_foliovec()
|
H A D | util.c | 673 struct bio_vec bv; in memcpy_to_bio() local 676 __bio_for_each_segment(bv, dst, iter, dst_iter) { in memcpy_to_bio() 677 void *dstp = kmap_local_page(bv.bv_page); in memcpy_to_bio() 679 memcpy(dstp + bv.bv_offset, src, bv.bv_len); in memcpy_to_bio() 682 src += bv.bv_len; in memcpy_to_bio() 688 struct bio_vec bv; in memcpy_from_bio() local 691 __bio_for_each_segment(bv, src, iter, src_iter) { in memcpy_from_bio() 692 void *srcp = kmap_local_page(bv.bv_page); in memcpy_from_bio() 694 memcpy(dst, srcp + bv.bv_offset, bv.bv_len); in memcpy_from_bio() 697 dst += bv.bv_len; in memcpy_from_bio()
|
/linux/drivers/block/ |
H A D | n64cart.c | 62 static bool n64cart_do_bvec(struct device *dev, struct bio_vec *bv, u32 pos) in n64cart_do_bvec() argument 68 WARN_ON_ONCE((bv->bv_offset & (MIN_ALIGNMENT - 1)) || in n64cart_do_bvec() 69 (bv->bv_len & (MIN_ALIGNMENT - 1))); in n64cart_do_bvec() 71 dma_addr = dma_map_bvec(dev, bv, DMA_FROM_DEVICE, 0); in n64cart_do_bvec() 79 n64cart_write_reg(PI_WRITE_REG, bv->bv_len - 1); in n64cart_do_bvec() 83 dma_unmap_page(dev, dma_addr, bv->bv_len, DMA_FROM_DEVICE); in n64cart_do_bvec()
|
/linux/fs/netfs/ |
H A D | iterator.c | 40 struct bio_vec *bv = NULL; in netfs_extract_user_iter() local 54 bv_size = array_size(max_pages, sizeof(*bv)); in netfs_extract_user_iter() 55 bv = kvmalloc(bv_size, GFP_KERNEL); in netfs_extract_user_iter() 56 if (!bv) in netfs_extract_user_iter() 64 pages = (void *)bv + bv_size - pg_size; in netfs_extract_user_iter() 92 bvec_set_page(bv + npages + i, *pages++, len - offset, offset); in netfs_extract_user_iter() 100 iov_iter_bvec(new, orig->data_source, bv, npages, orig_len - count); in netfs_extract_user_iter()
|
/linux/drivers/s390/block/ |
H A D | dasd_fba.c | 436 struct bio_vec bv; in dasd_fba_build_cp_regular() local 457 rq_for_each_segment(bv, req, iter) { in dasd_fba_build_cp_regular() 458 if (bv.bv_len & (blksize - 1)) in dasd_fba_build_cp_regular() 461 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_fba_build_cp_regular() 462 if (idal_is_needed (page_address(bv.bv_page), bv.bv_len)) in dasd_fba_build_cp_regular() 463 cidaw += bv.bv_len / blksize; in dasd_fba_build_cp_regular() 499 rq_for_each_segment(bv, req, iter) { in dasd_fba_build_cp_regular() 500 dst = bvec_virt(&bv); in dasd_fba_build_cp_regular() 505 memcpy(copy + bv.bv_offset, dst, bv.bv_len); in dasd_fba_build_cp_regular() 507 dst = copy + bv.bv_offset; in dasd_fba_build_cp_regular() [all …]
|
/linux/net/ceph/ |
H A D | messenger_v2.c | 146 struct bio_vec bv; in do_try_sendpage() local 154 bvec_set_page(&bv, it->bvec->bv_page, in do_try_sendpage() 168 if (sendpage_ok(bv.bv_page)) in do_try_sendpage() 173 iov_iter_bvec(&msg.msg_iter, ITER_SOURCE, &bv, 1, bv.bv_len); in do_try_sendpage() 232 static void set_in_bvec(struct ceph_connection *con, const struct bio_vec *bv) in set_in_bvec() argument 236 con->v2.in_bvec = *bv; in set_in_bvec() 237 iov_iter_bvec(&con->v2.in_iter, ITER_DEST, &con->v2.in_bvec, 1, bv->bv_len); in set_in_bvec() 273 static void set_out_bvec(struct ceph_connection *con, const struct bio_vec *bv, in set_out_bvec() argument 279 con->v2.out_bvec = *bv; in set_out_bvec() 860 struct bio_vec *bv) in get_bvec_at() argument [all …]
|
/linux/drivers/md/ |
H A D | dm-io-rewind.c | 12 static inline bool dm_bvec_iter_rewind(const struct bio_vec *bv, in dm_bvec_iter_rewind() argument 27 while (idx >= 0 && bytes && bytes > bv[idx].bv_len) { in dm_bvec_iter_rewind() 28 bytes -= bv[idx].bv_len; in dm_bvec_iter_rewind() 41 iter->bi_bvec_done = bv[idx].bv_len - bytes; in dm_bvec_iter_rewind()
|
H A D | dm-ebs-target.c | 65 static int __ebs_rw_bvec(struct ebs_c *ec, enum req_op op, struct bio_vec *bv, in __ebs_rw_bvec() argument 71 unsigned int bv_len = bv->bv_len; in __ebs_rw_bvec() 76 if (unlikely(!bv->bv_page || !bv_len)) in __ebs_rw_bvec() 79 pa = bvec_virt(bv); in __ebs_rw_bvec() 102 flush_dcache_page(bv->bv_page); in __ebs_rw_bvec() 104 flush_dcache_page(bv->bv_page); in __ebs_rw_bvec() 125 struct bio_vec bv; in __ebs_rw_bio() local 128 bio_for_each_bvec(bv, bio, iter) { in __ebs_rw_bio() 129 rr = __ebs_rw_bvec(ec, op, &bv, &iter); in __ebs_rw_bio()
|
/linux/fs/btrfs/ |
H A D | bio.c | 173 struct bio_vec *bv = bio_first_bvec_all(&repair_bbio->bio); in btrfs_end_repair_bio() local 180 ASSERT(folio_order(page_folio(bv->bv_page)) == 0); in btrfs_end_repair_bio() 183 !btrfs_data_csum_ok(repair_bbio, dev, 0, bv)) { in btrfs_end_repair_bio() 203 page_folio(bv->bv_page), bv->bv_offset, mirror); in btrfs_end_repair_bio() 220 struct bio_vec *bv, in repair_one_sector() argument 254 __bio_add_page(repair_bio, bv->bv_page, bv->bv_len, bv->bv_offset); in repair_one_sector() 293 struct bio_vec bv = bio_iter_iovec(&bbio->bio, *iter); in btrfs_check_read_bio() local 295 bv.bv_len = min(bv.bv_len, sectorsize); in btrfs_check_read_bio() 296 if (status || !btrfs_data_csum_ok(bbio, dev, offset, &bv)) in btrfs_check_read_bio() 297 fbio = repair_one_sector(bbio, offset, &bv, fbio); in btrfs_check_read_bio()
|