Home
last modified time | relevance | path

Searched refs:bi_opf (Results 1 – 25 of 58) sorted by relevance

123

/linux/block/
H A Dblk-core.c338 if (bio->bi_opf & REQ_NOWAIT) { in __bio_queue_enter()
526 if (op_is_flush(bio->bi_opf) && !bio_sectors(bio)) in bio_check_ro()
568 current->comm, bio->bi_bdev, bio->bi_opf, in bio_check_eod()
622 bio->bi_opf |= REQ_NOMERGE; in blk_check_zone_append()
639 if ((bio->bi_opf & REQ_POLLED) && in __submit_bio()
792 if ((bio->bi_opf & REQ_NOWAIT) && !bdev_nowait(bdev)) in submit_bio_noacct()
817 if (op_is_flush(bio->bi_opf)) { in submit_bio_noacct()
822 bio->bi_opf &= ~(REQ_PREFLUSH | REQ_FUA); in submit_bio_noacct()
834 if (bio->bi_opf & REQ_ATOMIC) { in submit_bio_noacct()
1281 sizeof_field(struct bio, bi_opf)); in blk_dev_init()
H A Dbio.c222 bio->bi_opf = opf; in bio_init()
285 bio->bi_opf = opf; in bio_reset()
841 if (bio->bi_opf & REQ_ALLOC_CACHE) in bio_put()
889 bio = bio_alloc_bioset(bdev, 0, bio_src->bi_opf, gfp, bs); in bio_alloc_clone()
918 bio_init(bio, bdev, bio_src->bi_io_vec, 0, bio_src->bi_opf); in bio_init_clone()
1005 bio->bi_opf |= REQ_NOMERGE; in __bio_add_page()
1278 bio->bi_opf |= REQ_NOMERGE; in bio_iov_iter_get_pages()
1492 bio->bi_opf |= REQ_SYNC; in bio_await()
1828 if (bio->bi_opf & REQ_ATOMIC) in bio_split()
1861 if (WARN_ON_ONCE(bio->bi_opf & REQ_ATOMIC && size)) in bio_trim()
H A Dfops.c85 bio.bi_opf |= REQ_ATOMIC; in __blkdev_direct_IO_simple()
96 bio.bi_opf |= REQ_NOWAIT; in __blkdev_direct_IO_simple()
238 bio->bi_opf |= REQ_NOWAIT; in __blkdev_direct_IO()
376 bio->bi_opf |= REQ_ATOMIC; in __blkdev_direct_IO_async()
379 bio->bi_opf |= REQ_NOWAIT; in __blkdev_direct_IO_async()
382 bio->bi_opf |= REQ_POLLED; in __blkdev_direct_IO_async()
H A Dbio-integrity.c130 bio->bi_opf &= ~REQ_INTEGRITY; in bio_integrity_free()
142 bio->bi_opf |= REQ_INTEGRITY; in bio_integrity_init()
413 bio->bi_opf |= REQ_NOMERGE; in bio_integrity_map_user()
H A Dbio-integrity-auto.c31 bid->bio->bi_opf &= ~REQ_INTEGRITY; in bio_integrity_finish()
H A Dblk-mq.c1048 req->cmd_flags |= req->bio->bi_opf & REQ_FAILFAST_MASK; in blk_update_request()
1391 if (rq->bio && rq->bio->bi_opf & REQ_POLLED) in blk_mq_start_request()
2690 if (bio->bi_opf & REQ_RAHEAD) in blk_mq_bio_to_request()
3054 .cmd_flags = bio->bi_opf, in blk_mq_get_new_requests()
3113 rq->cmd_flags = bio->bi_opf; in blk_mq_use_cached_rq()
3145 const int is_sync = op_is_sync(bio->bi_opf); in blk_mq_submit_bio()
3155 rq = blk_mq_peek_cached_request(plug, q, bio->bi_opf); in blk_mq_submit_bio()
3189 if ((bio->bi_opf & REQ_POLLED) && !blk_mq_can_poll(q)) { in blk_mq_submit_bio()
3218 if (bio->bi_opf & REQ_NOWAIT) in blk_mq_submit_bio()
3241 if (op_is_flush(bio->bi_opf) && blk_insert_flush(rq)) in blk_mq_submit_bio()
[all …]
/linux/drivers/md/dm-vdo/
H A Dvio.h124 blk_opf_t bi_opf, physical_block_number_t pbn);
127 blk_opf_t bi_opf, physical_block_number_t pbn);
129 blk_opf_t bi_opf, physical_block_number_t pbn);
/linux/drivers/md/bcache/
H A Dwriteback.h123 return (op_is_sync(bio->bi_opf) || in should_writeback()
124 bio->bi_opf & (REQ_META|REQ_PRIO) || in should_writeback()
H A Drequest.c202 bio->bi_opf &= ~(REQ_PREFLUSH|REQ_FUA); in CLOSURE_CALLBACK()
247 n->bi_opf = REQ_OP_WRITE; in CLOSURE_CALLBACK()
405 if ((bio->bi_opf & (REQ_RAHEAD|REQ_BACKGROUND))) { in check_should_bypass()
406 if (!(bio->bi_opf & (REQ_META|REQ_PRIO)) && in check_should_bypass()
667 bio->bi_opf & REQ_PREFLUSH)) { in backing_request_endio()
758 s->iop.flush_journal = op_is_flush(bio->bi_opf); in search_alloc()
1033 if (bio->bi_opf & REQ_PREFLUSH) { in cached_dev_write()
/linux/include/trace/events/
H A Dbcache.h31 blk_fill_rwbs(__entry->rwbs, bio->bi_opf);
105 blk_fill_rwbs(__entry->rwbs, bio->bi_opf);
140 blk_fill_rwbs(__entry->rwbs, bio->bi_opf);
171 blk_fill_rwbs(__entry->rwbs, bio->bi_opf);
241 blk_fill_rwbs(__entry->rwbs, bio->bi_opf);
H A Dblock.h326 blk_fill_rwbs(__entry->rwbs, bio->bi_opf);
353 blk_fill_rwbs(__entry->rwbs, bio->bi_opf);
506 blk_fill_rwbs(__entry->rwbs, bio->bi_opf);
547 blk_fill_rwbs(__entry->rwbs, bio->bi_opf);
628 blk_fill_rwbs(__entry->rwbs, bio->bi_opf);
/linux/drivers/md/
H A Ddm-io.c483 if ((io_req->bi_opf & REQ_OP_MASK) == REQ_OP_READ) { in dp_init()
508 if (num_regions > 1 && !op_is_write(io_req->bi_opf)) { in dm_io()
519 io_req->bi_opf, &dp, sync_error_bits, ioprio); in dm_io()
521 async_io(io_req->client, num_regions, where, io_req->bi_opf, &dp, in dm_io()
H A Draid1.c492 (bio->bi_opf & MD_FAILFAST) && in raid1_end_write_request()
1250 behind_bio = bio_alloc_bioset(NULL, vcnt, bio->bi_opf, GFP_NOIO, in alloc_behind_master_bio()
1357 bio->bi_opf & REQ_NOWAIT)) { in raid1_read_request()
1420 read_bio->bi_opf &= ~REQ_NOWAIT; in raid1_read_request()
1428 read_bio->bi_opf |= MD_FAILFAST; in raid1_read_request()
1459 if (bio->bi_opf & REQ_NOWAIT) in wait_blocked_rdev()
1513 if (bio->bi_opf & REQ_NOWAIT) { in raid1_write_request()
1529 bio->bi_opf & REQ_NOWAIT)) { in raid1_write_request()
1599 if (bio->bi_opf & REQ_ATOMIC) in raid1_write_request()
1666 mbio->bi_opf &= ~REQ_NOWAIT; in raid1_write_request()
[all …]
H A Draid10.c492 (bio->bi_opf & MD_FAILFAST)) { in raid10_end_write_request()
1126 if (!wait_barrier(conf, bio->bi_opf & REQ_NOWAIT)) { in regular_request_wait()
1134 if (bio->bi_opf & REQ_NOWAIT) { in regular_request_wait()
1226 read_bio->bi_opf &= ~REQ_NOWAIT; in raid10_read_request()
1236 read_bio->bi_opf |= MD_FAILFAST; in raid10_read_request()
1260 mbio->bi_opf &= ~REQ_NOWAIT; in raid10_write_one_disk()
1272 mbio->bi_opf |= MD_FAILFAST; in raid10_write_one_disk()
1358 if (bio->bi_opf & REQ_NOWAIT) { in raid10_write_request()
1390 if (bio->bi_opf & REQ_NOWAIT) { in raid10_write_request()
1464 if (bio->bi_opf & REQ_ATOMIC) in raid10_write_request()
[all …]
H A Ddm-flakey.c22 (((bio)->bi_opf & (fc)->corrupt_bio_flags) == (fc)->corrupt_bio_flags)
379 (bio_data_dir(bio) == WRITE) ? 'w' : 'r', bio->bi_opf, in corrupt_bio_common()
444 bio_init_inline(clone, fc->dev->bdev, nr_iovecs, bio->bi_opf); in clone_bio()
H A Ddm-stripe.c278 if (bio->bi_opf & REQ_PREFLUSH) { in stripe_map()
414 if (bio->bi_opf & REQ_RAHEAD) in stripe_end_io()
H A Draid1-10.c303 return !(bio->bi_opf & (REQ_RAHEAD | REQ_NOWAIT)) && in raid1_should_handle_error()
H A Ddm-log-writes.c653 bool flush_bio = (bio->bi_opf & REQ_PREFLUSH); in log_writes_map()
654 bool fua_bio = (bio->bi_opf & REQ_FUA); in log_writes_map()
656 bool meta_bio = (bio->bi_opf & REQ_META); in log_writes_map()
H A Ddm-region-hash.c406 if (bio->bi_opf & REQ_PREFLUSH) { in dm_rh_mark_nosync()
534 if (bio->bi_opf & REQ_PREFLUSH || bio_op(bio) == REQ_OP_DISCARD) in dm_rh_inc_pending()
H A Ddm-clone-target.c271 return op_is_flush(bio->bi_opf) && in bio_triggers_commit()
424 if (!(bio->bi_opf & REQ_FUA)) { in complete_overwrite_bio()
1262 if ((bio->bi_opf & REQ_PREFLUSH) && dest_dev_flushed) { in process_deferred_flush_bios()
1331 if (bio->bi_opf & REQ_PREFLUSH) { in clone_map()
/linux/include/linux/
H A Ddm-io.h62 blk_opf_t bi_opf; /* Request type and flags */ member
/linux/drivers/block/
H A Dbrd.c143 blk_opf_t opf = bio->bi_opf; in brd_rw_bvec()
206 if (unlikely(op_is_discard(bio->bi_opf))) { in brd_submit_bio()
/linux/fs/xfs/
H A Dxfs_zone_gc.c779 chunk->bio.bi_opf &= ~REQ_OP_WRITE; in xfs_zone_gc_submit_write()
780 chunk->bio.bi_opf |= REQ_OP_ZONE_APPEND; in xfs_zone_gc_submit_write()
967 bio->bi_opf &= ~REQ_OP_ZONE_RESET; in xfs_submit_zone_reset_bio()
968 bio->bi_opf |= REQ_OP_DISCARD; in xfs_submit_zone_reset_bio()
/linux/fs/iomap/
H A Dbio.c128 bio->bi_opf |= REQ_RAHEAD; in iomap_read_alloc_bio()
/linux/Documentation/admin-guide/device-mapper/
H A Ddm-flakey.rst67 Perform the replacement only if bio->bi_opf has all the

123