| /linux/lib/ |
| H A D | iov_iter.c | 106 n -= fault_in_readable(i->ubuf + i->iov_offset, n); in fault_in_iov_iter_readable() 114 for (p = iter_iov(i), skip = i->iov_offset; count; p++, skip = 0) { in fault_in_iov_iter_readable() 149 n -= fault_in_safe_writeable(i->ubuf + i->iov_offset, n); in fault_in_iov_iter_writeable() 157 for (p = iter_iov(i), skip = i->iov_offset; count; p++, skip = 0) { in fault_in_iov_iter_writeable() 185 .iov_offset = 0, in iov_iter_init() 510 size += i->iov_offset; in iov_iter_bvec_advance() 517 i->iov_offset = size; in iov_iter_bvec_advance() 530 size += i->iov_offset; // from beginning of current segment in iov_iter_iovec_advance() 536 i->iov_offset = size; in iov_iter_iovec_advance() 555 size += i->iov_offset; /* From beginning of current segment. */ in iov_iter_folioq_advance() [all …]
|
| H A D | scatterlist.c | 1175 unsigned long start = iter->iov_offset; in extract_bvec_to_sg() 1221 unsigned long start = iter->iov_offset; in extract_kvec_to_sg() 1284 size_t offset = iter->iov_offset; in extract_folioq_to_sg() 1326 iter->iov_offset = offset; in extract_folioq_to_sg() 1344 loff_t start = iter->xarray_start + iter->iov_offset; in extract_xarray_to_sg()
|
| /linux/include/linux/ |
| H A D | iov_iter.h | 30 remain = step(base + iter->iov_offset, 0, len, priv, priv2); in iterate_ubuf() 32 iter->iov_offset += progress; in iterate_ubuf() 45 size_t progress = 0, skip = iter->iov_offset; in iterate_iovec() 66 iter->iov_offset = skip; in iterate_iovec() 79 size_t progress = 0, skip = iter->iov_offset; in iterate_kvec() 100 iter->iov_offset = skip; in iterate_kvec() 113 size_t progress = 0, skip = iter->iov_offset; in iterate_bvec() 139 iter->iov_offset = skip; in iterate_bvec() 153 size_t progress = 0, skip = iter->iov_offset; in iterate_folioq() 195 iter->iov_offset = skip; in iterate_folioq() [all …]
|
| /linux/drivers/net/ethernet/google/gve/ |
| H A D | gve_tx.c | 105 iov[0].iov_offset = fifo->head; in gve_tx_alloc_fifo() 116 iov[1].iov_offset = 0; /* Start of fifo*/ in gve_tx_alloc_fifo() 551 u64 iov_offset, u64 iov_len) in gve_dma_sync_for_device() argument 553 u64 last_page = (iov_offset + iov_len - 1) / PAGE_SIZE; in gve_dma_sync_for_device() 554 u64 first_page = iov_offset / PAGE_SIZE; in gve_dma_sync_for_device() 598 info->iov[hdr_nfrags - 1].iov_offset, skb->len); in gve_tx_add_skb_copy() 601 tx->tx_fifo.base + info->iov[hdr_nfrags - 1].iov_offset, in gve_tx_add_skb_copy() 604 info->iov[hdr_nfrags - 1].iov_offset, in gve_tx_add_skb_copy() 621 info->iov[i].iov_offset); in gve_tx_add_skb_copy() 624 tx->tx_fifo.base + info->iov[i].iov_offset, in gve_tx_add_skb_copy() [all …]
|
| H A D | gve.h | 357 u32 iov_offset; /* offset into this segment */ member
|
| /linux/fs/netfs/ |
| H A D | iterator.c | 115 size_t skip = iter->iov_offset + start_offset; in netfs_limit_bvec() 156 loff_t pos = iter->xarray_start + iter->iov_offset; in netfs_limit_xarray() 215 start_offset += iter->iov_offset; in netfs_limit_folioq()
|
| H A D | read_pgpriv2.c | 73 creq->buffer.iter.iov_offset = cache->submit_off; in netfs_pgpriv2_copy_folio() 86 creq->buffer.iter.iov_offset = 0; in netfs_pgpriv2_copy_folio()
|
| H A D | write_issue.c | 803 wreq->buffer.iter.iov_offset = 0; in netfs_write_folio_single()
|
| /linux/lib/tests/ |
| H A D | kunit_iov_iter.c | 451 KUNIT_EXPECT_EQ(test, iter.iov_offset, pr->to % PAGE_SIZE); in iov_kunit_copy_to_folioq() 513 KUNIT_EXPECT_EQ(test, iter.iov_offset, pr->to % PAGE_SIZE); in iov_kunit_copy_from_folioq() 612 KUNIT_EXPECT_EQ(test, iter.iov_offset, size); in iov_kunit_copy_to_xarray() 670 KUNIT_EXPECT_EQ(test, iter.iov_offset, size); in iov_kunit_copy_from_xarray() 1005 KUNIT_EXPECT_EQ(test, iter.iov_offset, pr->to - pr->from); in iov_kunit_extract_pages_xarray()
|
| /linux/net/9p/ |
| H A D | trans_virtio.c | 350 p = data->kvec->iov_base + data->iov_offset; in p9_get_mapped_pages()
|
| /linux/fs/smb/client/ |
| H A D | smbdirect.c | 3154 unsigned long start = iter->iov_offset; in smb_extract_bvec_to_rdma() 3195 unsigned long start = iter->iov_offset; in smb_extract_kvec_to_rdma() 3254 size_t offset = iter->iov_offset; in smb_extract_folioq_to_rdma() 3296 iter->iov_offset = offset; in smb_extract_folioq_to_rdma()
|
| /linux/include/trace/events/ |
| H A D | afs.h | 930 __entry->offset = msg->msg_iter.xarray_start + msg->msg_iter.iov_offset; 954 __entry->offset = msg->msg_iter.xarray_start + msg->msg_iter.iov_offset;
|
| /linux/io_uring/ |
| H A D | net.c | 1417 bi.bi_bvec_done = from->iov_offset; in io_sg_from_iter() 1436 from->iov_offset = bi.bi_bvec_done; in io_sg_from_iter()
|
| H A D | rw.c | 717 addr = iter->ubuf + iter->iov_offset; in loop_rw_iter()
|
| H A D | rsrc.c | 1093 iter->iov_offset = offset; in io_import_fixed()
|
| /linux/fs/nfs/ |
| H A D | localio.c | 387 size_t skip = i->iov_offset; in nfs_iov_iter_aligned_bvec()
|
| /linux/drivers/block/ |
| H A D | zloop.c | 522 iter.iov_offset = rq->bio->bi_iter.bi_bvec_done; in zloop_rw()
|
| H A D | loop.c | 387 iter.iov_offset = offset; in lo_rw_aio()
|
| /linux/drivers/nvme/host/ |
| H A D | tcp.c | 318 return req->iter.bvec->bv_offset + req->iter.iov_offset; in nvme_tcp_req_cur_offset() 368 req->iter.iov_offset = offset; in nvme_tcp_init_iter()
|
| /linux/net/vmw_vsock/ |
| H A D | virtio_transport_common.c | 55 if (iov_iter->iov_offset) in virtio_transport_can_zcopy()
|
| /linux/block/ |
| H A D | bio.c | 1204 bio->bi_iter.bi_bvec_done = iter->iov_offset; in bio_iov_bvec_set()
|
| /linux/net/ceph/ |
| H A D | messenger_v2.c | 156 it->bvec->bv_len - it->iov_offset), in do_try_sendpage() 157 it->bvec->bv_offset + it->iov_offset); in do_try_sendpage()
|
| /linux/fs/nfsd/ |
| H A D | vfs.c | 1286 return (unsigned long)(iter->bvec->bv_offset + iter->iov_offset); in iov_iter_bvec_offset()
|
| /linux/fs/fuse/ |
| H A D | file.c | 1534 return (unsigned long)iter_iov(ii)->iov_base + ii->iov_offset; in fuse_get_user_addr()
|