| /linux/block/ |
| H A D | blk-core.c | 338 if (bio->bi_opf & REQ_NOWAIT) { in __bio_queue_enter() 526 if (op_is_flush(bio->bi_opf) && !bio_sectors(bio)) in bio_check_ro() 568 current->comm, bio->bi_bdev, bio->bi_opf, in bio_check_eod() 622 bio->bi_opf |= REQ_NOMERGE; in blk_check_zone_append() 639 if ((bio->bi_opf & REQ_POLLED) && in __submit_bio() 792 if ((bio->bi_opf & REQ_NOWAIT) && !bdev_nowait(bdev)) in submit_bio_noacct() 817 if (op_is_flush(bio->bi_opf)) { in submit_bio_noacct() 822 bio->bi_opf &= ~(REQ_PREFLUSH | REQ_FUA); in submit_bio_noacct() 834 if (bio->bi_opf & REQ_ATOMIC) { in submit_bio_noacct() 1281 sizeof_field(struct bio, bi_opf)); in blk_dev_init()
|
| H A D | bio.c | 222 bio->bi_opf = opf; in bio_init() 285 bio->bi_opf = opf; in bio_reset() 841 if (bio->bi_opf & REQ_ALLOC_CACHE) in bio_put() 889 bio = bio_alloc_bioset(bdev, 0, bio_src->bi_opf, gfp, bs); in bio_alloc_clone() 918 bio_init(bio, bdev, bio_src->bi_io_vec, 0, bio_src->bi_opf); in bio_init_clone() 1005 bio->bi_opf |= REQ_NOMERGE; in __bio_add_page() 1278 bio->bi_opf |= REQ_NOMERGE; in bio_iov_iter_get_pages() 1492 bio->bi_opf |= REQ_SYNC; in bio_await() 1828 if (bio->bi_opf & REQ_ATOMIC) in bio_split() 1861 if (WARN_ON_ONCE(bio->bi_opf & REQ_ATOMIC && size)) in bio_trim()
|
| H A D | fops.c | 85 bio.bi_opf |= REQ_ATOMIC; in __blkdev_direct_IO_simple() 96 bio.bi_opf |= REQ_NOWAIT; in __blkdev_direct_IO_simple() 238 bio->bi_opf |= REQ_NOWAIT; in __blkdev_direct_IO() 376 bio->bi_opf |= REQ_ATOMIC; in __blkdev_direct_IO_async() 379 bio->bi_opf |= REQ_NOWAIT; in __blkdev_direct_IO_async() 382 bio->bi_opf |= REQ_POLLED; in __blkdev_direct_IO_async()
|
| H A D | bio-integrity.c | 130 bio->bi_opf &= ~REQ_INTEGRITY; in bio_integrity_free() 142 bio->bi_opf |= REQ_INTEGRITY; in bio_integrity_init() 413 bio->bi_opf |= REQ_NOMERGE; in bio_integrity_map_user()
|
| H A D | bio-integrity-auto.c | 31 bid->bio->bi_opf &= ~REQ_INTEGRITY; in bio_integrity_finish()
|
| H A D | blk-mq.c | 1048 req->cmd_flags |= req->bio->bi_opf & REQ_FAILFAST_MASK; in blk_update_request() 1391 if (rq->bio && rq->bio->bi_opf & REQ_POLLED) in blk_mq_start_request() 2690 if (bio->bi_opf & REQ_RAHEAD) in blk_mq_bio_to_request() 3054 .cmd_flags = bio->bi_opf, in blk_mq_get_new_requests() 3113 rq->cmd_flags = bio->bi_opf; in blk_mq_use_cached_rq() 3145 const int is_sync = op_is_sync(bio->bi_opf); in blk_mq_submit_bio() 3155 rq = blk_mq_peek_cached_request(plug, q, bio->bi_opf); in blk_mq_submit_bio() 3189 if ((bio->bi_opf & REQ_POLLED) && !blk_mq_can_poll(q)) { in blk_mq_submit_bio() 3218 if (bio->bi_opf & REQ_NOWAIT) in blk_mq_submit_bio() 3241 if (op_is_flush(bio->bi_opf) && blk_insert_flush(rq)) in blk_mq_submit_bio() [all …]
|
| /linux/drivers/md/dm-vdo/ |
| H A D | vio.h | 124 blk_opf_t bi_opf, physical_block_number_t pbn); 127 blk_opf_t bi_opf, physical_block_number_t pbn); 129 blk_opf_t bi_opf, physical_block_number_t pbn);
|
| /linux/drivers/md/bcache/ |
| H A D | writeback.h | 123 return (op_is_sync(bio->bi_opf) || in should_writeback() 124 bio->bi_opf & (REQ_META|REQ_PRIO) || in should_writeback()
|
| H A D | request.c | 202 bio->bi_opf &= ~(REQ_PREFLUSH|REQ_FUA); in CLOSURE_CALLBACK() 247 n->bi_opf = REQ_OP_WRITE; in CLOSURE_CALLBACK() 405 if ((bio->bi_opf & (REQ_RAHEAD|REQ_BACKGROUND))) { in check_should_bypass() 406 if (!(bio->bi_opf & (REQ_META|REQ_PRIO)) && in check_should_bypass() 667 bio->bi_opf & REQ_PREFLUSH)) { in backing_request_endio() 758 s->iop.flush_journal = op_is_flush(bio->bi_opf); in search_alloc() 1033 if (bio->bi_opf & REQ_PREFLUSH) { in cached_dev_write()
|
| /linux/include/trace/events/ |
| H A D | bcache.h | 31 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 105 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 140 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 171 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 241 blk_fill_rwbs(__entry->rwbs, bio->bi_opf);
|
| H A D | block.h | 326 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 353 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 506 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 547 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 628 blk_fill_rwbs(__entry->rwbs, bio->bi_opf);
|
| /linux/drivers/md/ |
| H A D | dm-io.c | 483 if ((io_req->bi_opf & REQ_OP_MASK) == REQ_OP_READ) { in dp_init() 508 if (num_regions > 1 && !op_is_write(io_req->bi_opf)) { in dm_io() 519 io_req->bi_opf, &dp, sync_error_bits, ioprio); in dm_io() 521 async_io(io_req->client, num_regions, where, io_req->bi_opf, &dp, in dm_io()
|
| H A D | raid1.c | 492 (bio->bi_opf & MD_FAILFAST) && in raid1_end_write_request() 1250 behind_bio = bio_alloc_bioset(NULL, vcnt, bio->bi_opf, GFP_NOIO, in alloc_behind_master_bio() 1357 bio->bi_opf & REQ_NOWAIT)) { in raid1_read_request() 1420 read_bio->bi_opf &= ~REQ_NOWAIT; in raid1_read_request() 1428 read_bio->bi_opf |= MD_FAILFAST; in raid1_read_request() 1459 if (bio->bi_opf & REQ_NOWAIT) in wait_blocked_rdev() 1513 if (bio->bi_opf & REQ_NOWAIT) { in raid1_write_request() 1529 bio->bi_opf & REQ_NOWAIT)) { in raid1_write_request() 1599 if (bio->bi_opf & REQ_ATOMIC) in raid1_write_request() 1666 mbio->bi_opf &= ~REQ_NOWAIT; in raid1_write_request() [all …]
|
| H A D | raid10.c | 492 (bio->bi_opf & MD_FAILFAST)) { in raid10_end_write_request() 1126 if (!wait_barrier(conf, bio->bi_opf & REQ_NOWAIT)) { in regular_request_wait() 1134 if (bio->bi_opf & REQ_NOWAIT) { in regular_request_wait() 1226 read_bio->bi_opf &= ~REQ_NOWAIT; in raid10_read_request() 1236 read_bio->bi_opf |= MD_FAILFAST; in raid10_read_request() 1260 mbio->bi_opf &= ~REQ_NOWAIT; in raid10_write_one_disk() 1272 mbio->bi_opf |= MD_FAILFAST; in raid10_write_one_disk() 1358 if (bio->bi_opf & REQ_NOWAIT) { in raid10_write_request() 1390 if (bio->bi_opf & REQ_NOWAIT) { in raid10_write_request() 1464 if (bio->bi_opf & REQ_ATOMIC) in raid10_write_request() [all …]
|
| H A D | dm-flakey.c | 22 (((bio)->bi_opf & (fc)->corrupt_bio_flags) == (fc)->corrupt_bio_flags) 379 (bio_data_dir(bio) == WRITE) ? 'w' : 'r', bio->bi_opf, in corrupt_bio_common() 444 bio_init_inline(clone, fc->dev->bdev, nr_iovecs, bio->bi_opf); in clone_bio()
|
| H A D | dm-stripe.c | 278 if (bio->bi_opf & REQ_PREFLUSH) { in stripe_map() 414 if (bio->bi_opf & REQ_RAHEAD) in stripe_end_io()
|
| H A D | raid1-10.c | 303 return !(bio->bi_opf & (REQ_RAHEAD | REQ_NOWAIT)) && in raid1_should_handle_error()
|
| H A D | dm-log-writes.c | 653 bool flush_bio = (bio->bi_opf & REQ_PREFLUSH); in log_writes_map() 654 bool fua_bio = (bio->bi_opf & REQ_FUA); in log_writes_map() 656 bool meta_bio = (bio->bi_opf & REQ_META); in log_writes_map()
|
| H A D | dm-region-hash.c | 406 if (bio->bi_opf & REQ_PREFLUSH) { in dm_rh_mark_nosync() 534 if (bio->bi_opf & REQ_PREFLUSH || bio_op(bio) == REQ_OP_DISCARD) in dm_rh_inc_pending()
|
| H A D | dm-clone-target.c | 271 return op_is_flush(bio->bi_opf) && in bio_triggers_commit() 424 if (!(bio->bi_opf & REQ_FUA)) { in complete_overwrite_bio() 1262 if ((bio->bi_opf & REQ_PREFLUSH) && dest_dev_flushed) { in process_deferred_flush_bios() 1331 if (bio->bi_opf & REQ_PREFLUSH) { in clone_map()
|
| /linux/include/linux/ |
| H A D | dm-io.h | 62 blk_opf_t bi_opf; /* Request type and flags */ member
|
| /linux/drivers/block/ |
| H A D | brd.c | 143 blk_opf_t opf = bio->bi_opf; in brd_rw_bvec() 206 if (unlikely(op_is_discard(bio->bi_opf))) { in brd_submit_bio()
|
| /linux/fs/xfs/ |
| H A D | xfs_zone_gc.c | 779 chunk->bio.bi_opf &= ~REQ_OP_WRITE; in xfs_zone_gc_submit_write() 780 chunk->bio.bi_opf |= REQ_OP_ZONE_APPEND; in xfs_zone_gc_submit_write() 967 bio->bi_opf &= ~REQ_OP_ZONE_RESET; in xfs_submit_zone_reset_bio() 968 bio->bi_opf |= REQ_OP_DISCARD; in xfs_submit_zone_reset_bio()
|
| /linux/fs/iomap/ |
| H A D | bio.c | 128 bio->bi_opf |= REQ_RAHEAD; in iomap_read_alloc_bio()
|
| /linux/Documentation/admin-guide/device-mapper/ |
| H A D | dm-flakey.rst | 67 Perform the replacement only if bio->bi_opf has all the
|