Home
last modified time | relevance | path

Searched refs:opf (Results 1 – 25 of 44) sorted by relevance

12

/linux/arch/sparc/kernel/
H A Dvisemul.c294 static void edge(struct pt_regs *regs, unsigned int insn, unsigned int opf) in edge() argument
307 switch (opf) { in edge()
352 switch (opf) { in edge()
372 static void array(struct pt_regs *regs, unsigned int insn, unsigned int opf) in array() argument
394 switch (opf) { in array()
477 static void pformat(struct pt_regs *regs, unsigned int insn, unsigned int opf) in pformat() argument
483 scale = (gsr >> 3) & (opf == FPACK16_OPF ? 0xf : 0x1f); in pformat()
484 switch (opf) { in pformat()
589 static void pmul(struct pt_regs *regs, unsigned int insn, unsigned int opf) in pmul() argument
594 switch (opf) { in pmul()
[all …]
/linux/drivers/md/
H A Ddm-io.c306 static void do_region(const blk_opf_t opf, unsigned int region, in do_region() argument
319 const enum req_op op = opf & REQ_OP_MASK; in do_region()
353 bio = bio_alloc_bioset(where->bdev, num_bvecs, opf, GFP_NOIO, in do_region()
382 WARN_ON_ONCE(opf & REQ_ATOMIC && remaining); in do_region()
386 static void dispatch_io(blk_opf_t opf, unsigned int num_regions, in dispatch_io() argument
401 if (where[i].count || (opf & REQ_PREFLUSH)) in dispatch_io()
402 do_region(opf, i, where + i, dp, io, ioprio); in dispatch_io()
413 struct dm_io_region *where, blk_opf_t opf, in async_io() argument
429 dispatch_io(opf, num_regions, where, dp, io, ioprio); in async_io()
446 struct dm_io_region *where, blk_opf_t opf, struct dpages *dp, in sync_io() argument
[all …]
H A Ddm-snap-persistent.c232 static int chunk_io(struct pstore *ps, void *area, chunk_t chunk, blk_opf_t opf, in chunk_io() argument
241 .bi_opf = opf, in chunk_io()
288 static int area_io(struct pstore *ps, blk_opf_t opf) in area_io() argument
292 return chunk_io(ps, ps->area, chunk, opf, 0); in area_io()
/linux/drivers/block/
H A Dbrd.c86 blk_opf_t opf) in brd_insert_page() argument
88 gfp_t gfp = (opf & REQ_NOWAIT) ? GFP_NOWAIT : GFP_NOIO; in brd_insert_page()
143 blk_opf_t opf = bio->bi_opf; in brd_rw_bvec() local
150 if (!page && op_is_write(opf)) { in brd_rw_bvec()
151 page = brd_insert_page(brd, sector, opf); in brd_rw_bvec()
157 if (op_is_write(opf)) { in brd_rw_bvec()
173 if (PTR_ERR(page) == -ENOMEM && (opf & REQ_NOWAIT)) in brd_rw_bvec()
/linux/drivers/scsi/device_handler/
H A Dscsi_dh_hp_sw.c83 blk_opf_t opf = REQ_OP_DRV_IN | REQ_FAILFAST_DEV | in hp_sw_tur() local
103 res = scsi_execute_cmd(sdev, cmd, opf, NULL, 0, HP_SW_TIMEOUT, in hp_sw_tur()
132 blk_opf_t opf = REQ_OP_DRV_IN | REQ_FAILFAST_DEV | in hp_sw_start_stop() local
157 res = scsi_execute_cmd(sdev, cmd, opf, NULL, 0, HP_SW_TIMEOUT, in hp_sw_start_stop()
H A Dscsi_dh_emc.c242 blk_opf_t opf = REQ_OP_DRV_OUT | REQ_FAILFAST_DEV | in send_trespass_cmd() local
269 err = scsi_execute_cmd(sdev, cdb, opf, csdev->buffer, len, in send_trespass_cmd()
H A Dscsi_dh_alua.c130 blk_opf_t opf = REQ_OP_DRV_IN | REQ_FAILFAST_DEV | in submit_rtpg() local
145 return scsi_execute_cmd(sdev, cdb, opf, buff, bufflen, in submit_rtpg()
163 blk_opf_t opf = REQ_OP_DRV_OUT | REQ_FAILFAST_DEV | in submit_stpg() local
180 return scsi_execute_cmd(sdev, cdb, opf, stpg_data, in submit_stpg()
/linux/include/linux/
H A Dbio.h353 blk_opf_t opf, gfp_t gfp_mask,
366 unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp_mask) in bio_alloc() argument
368 return bio_alloc_bioset(bdev, nr_vecs, opf, gfp_mask, &fs_bio_set); in bio_alloc()
426 unsigned short max_vecs, blk_opf_t opf);
428 unsigned short max_vecs, blk_opf_t opf) in bio_init_inline() argument
430 bio_init(bio, bdev, bio_inline_vecs(bio), max_vecs, opf); in bio_init_inline()
433 void bio_reset(struct bio *bio, struct block_device *bdev, blk_opf_t opf);
434 void bio_reuse(struct bio *bio, blk_opf_t opf);
741 unsigned int nr_pages, blk_opf_t opf, gfp_t gfp);
/linux/block/
H A Dbio.c246 unsigned short max_vecs, blk_opf_t opf) in bio_init() argument
250 bio->bi_opf = opf; in bio_init()
302 void bio_reset(struct bio *bio, struct block_device *bdev, blk_opf_t opf) in bio_reset() argument
313 bio->bi_opf = opf; in bio_reset()
332 void bio_reuse(struct bio *bio, blk_opf_t opf) in bio_reuse() argument
342 bio_reset(bio, bio->bi_bdev, opf); in bio_reuse()
410 unsigned int nr_pages, blk_opf_t opf, gfp_t gfp) in blk_next_bio() argument
412 return bio_chain_and_submit(bio, bio_alloc(bdev, nr_pages, opf, gfp)); in blk_next_bio()
487 unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp, in bio_alloc_percpu_cache() argument
508 bio_init_inline(bio, bdev, nr_vecs, opf); in bio_alloc_percpu_cache()
[all …]
H A Dblk-mq.h90 static inline enum hctx_type blk_mq_get_hctx_type(blk_opf_t opf) in blk_mq_get_hctx_type() argument
97 if (opf & REQ_POLLED) in blk_mq_get_hctx_type()
99 else if ((opf & REQ_OP_MASK) == REQ_OP_READ) in blk_mq_get_hctx_type()
109 static inline struct blk_mq_hw_ctx *blk_mq_map_queue(blk_opf_t opf, in blk_mq_map_queue() argument
112 return ctx->hctxs[blk_mq_get_hctx_type(opf)]; in blk_mq_map_queue()
H A Dblk-wbt.c537 static inline unsigned int get_limit(struct rq_wb *rwb, blk_opf_t opf) in get_limit() argument
541 if ((opf & REQ_OP_MASK) == REQ_OP_DISCARD) in get_limit()
552 if ((opf & REQ_HIPRIO) || wb_recent_wait(rwb)) in get_limit()
554 else if ((opf & REQ_BACKGROUND) || close_io(rwb)) { in get_limit()
569 blk_opf_t opf; member
575 return rq_wait_inc_below(rqw, get_limit(data->rwb, data->opf)); in wbt_inflight_cb()
589 blk_opf_t opf) in __wbt_wait() argument
595 .opf = opf, in __wbt_wait()
H A Dbfq-cgroup.c223 blk_opf_t opf) in bfqg_stats_update_io_add() argument
225 blkg_rwstat_add(&bfqg->stats.queued, opf, 1); in bfqg_stats_update_io_add()
231 void bfqg_stats_update_io_remove(struct bfq_group *bfqg, blk_opf_t opf) in bfqg_stats_update_io_remove() argument
233 blkg_rwstat_add(&bfqg->stats.queued, opf, -1); in bfqg_stats_update_io_remove()
236 void bfqg_stats_update_io_merged(struct bfq_group *bfqg, blk_opf_t opf) in bfqg_stats_update_io_merged() argument
238 blkg_rwstat_add(&bfqg->stats.merged, opf, 1); in bfqg_stats_update_io_merged()
242 u64 io_start_time_ns, blk_opf_t opf) in bfqg_stats_update_completion() argument
248 blkg_rwstat_add(&stats->service_time, opf, in bfqg_stats_update_completion()
251 blkg_rwstat_add(&stats->wait_time, opf, in bfqg_stats_update_completion()
257 void bfqg_stats_update_io_remove(struct bfq_group *bfqg, blk_opf_t opf) { } in bfqg_stats_update_io_remove() argument
[all …]
H A Dfops.c31 blk_opf_t opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; in dio_bio_write_op() local
35 opf |= REQ_FUA; in dio_bio_write_op()
36 return opf; in dio_bio_write_op()
182 blk_opf_t opf = is_read ? REQ_OP_READ : dio_bio_write_op(iocb); in __blkdev_direct_IO() local
186 bio = bio_alloc_bioset(bdev, nr_pages, opf, GFP_KERNEL, in __blkdev_direct_IO()
262 bio = bio_alloc(bdev, nr_pages, opf, GFP_KERNEL); in __blkdev_direct_IO()
327 blk_opf_t opf = is_read ? REQ_OP_READ : dio_bio_write_op(iocb); in __blkdev_direct_IO_async() local
333 bio = bio_alloc_bioset(bdev, nr_pages, opf, GFP_KERNEL, in __blkdev_direct_IO_async()
H A Dbfq-iosched.h1067 void bfqg_stats_update_io_remove(struct bfq_group *bfqg, blk_opf_t opf);
1068 void bfqg_stats_update_io_merged(struct bfq_group *bfqg, blk_opf_t opf);
1070 u64 io_start_time_ns, blk_opf_t opf);
1078 blk_opf_t opf);
H A Dblk-mq.c599 blk_opf_t opf, in blk_mq_rq_cache_fill() argument
606 .cmd_flags = opf, in blk_mq_rq_cache_fill()
627 blk_opf_t opf, in blk_mq_alloc_cached_request() argument
639 rq = blk_mq_rq_cache_fill(q, plug, opf, flags); in blk_mq_alloc_cached_request()
647 if (blk_mq_get_hctx_type(opf) != rq->mq_hctx->type) in blk_mq_alloc_cached_request()
649 if (op_is_flush(rq->cmd_flags) != op_is_flush(opf)) in blk_mq_alloc_cached_request()
656 rq->cmd_flags = opf; in blk_mq_alloc_cached_request()
661 struct request *blk_mq_alloc_request(struct request_queue *q, blk_opf_t opf, in blk_mq_alloc_request() argument
666 rq = blk_mq_alloc_cached_request(q, opf, flags); in blk_mq_alloc_request()
672 .cmd_flags = opf, in blk_mq_alloc_request()
[all …]
/linux/drivers/target/
H A Dtarget_core_iblock.c355 blk_opf_t opf) in iblock_get_bio() argument
364 bio = bio_alloc_bioset(ib_dev->ibd_bd, bio_max_segs(sg_num), opf, in iblock_get_bio()
752 blk_opf_t opf; in iblock_execute_rw() local
765 opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; in iblock_execute_rw()
773 opf |= REQ_FUA; in iblock_execute_rw()
775 opf |= REQ_FUA; in iblock_execute_rw()
779 opf |= REQ_ATOMIC; in iblock_execute_rw()
781 opf = REQ_OP_READ; in iblock_execute_rw()
796 bio = iblock_get_bio(cmd, block_lba, sgl_nents, opf); in iblock_execute_rw()
829 bio = iblock_get_bio(cmd, block_lba, sg_num, opf); in iblock_execute_rw()
/linux/fs/gfs2/
H A Dlops.c263 bio_end_io_t *end_io, blk_opf_t opf) in gfs2_log_alloc_bio() argument
266 struct bio *bio = bio_alloc(sb->s_bdev, BIO_MAX_VECS, opf, GFP_NOIO); in gfs2_log_alloc_bio()
293 struct bio **biop, blk_opf_t opf, in gfs2_log_get_bio() argument
308 *biop = gfs2_log_alloc_bio(sdp, blkno, end_io, opf); in gfs2_log_get_bio()
329 u64 blkno, blk_opf_t opf) in gfs2_log_write() argument
334 bio = gfs2_log_get_bio(sdp, blkno, &jd->jd_log_bio, opf, in gfs2_log_write()
339 opf, gfs2_end_log_write, true); in gfs2_log_write()
482 sector_t sector, blk_opf_t opf) in gfs2_chain_bio() argument
486 new = bio_alloc(prev->bi_bdev, nr_iovecs, opf, GFP_NOIO); in gfs2_chain_bio()
/linux/kernel/trace/
H A Dblktrace.c311 const blk_opf_t opf, u64 what, int error, in __blk_add_trace() argument
324 const enum req_op op = opf & REQ_OP_MASK; in __blk_add_trace()
331 what |= MASK_TC_BIT(opf, SYNC); in __blk_add_trace()
332 what |= MASK_TC_BIT(opf, RAHEAD); in __blk_add_trace()
333 what |= MASK_TC_BIT(opf, META); in __blk_add_trace()
334 what |= MASK_TC_BIT(opf, PREFLUSH); in __blk_add_trace()
335 what |= MASK_TC_BIT(opf, FUA); in __blk_add_trace()
2152 void blk_fill_rwbs(char *rwbs, blk_opf_t opf) in blk_fill_rwbs() argument
2156 if (opf & REQ_PREFLUSH) in blk_fill_rwbs()
2159 switch (opf & REQ_OP_MASK) { in blk_fill_rwbs()
[all …]
/linux/fs/nilfs2/
H A Dbtnode.c88 sector_t pblocknr, blk_opf_t opf, in nilfs_btnode_submit_block() argument
121 if (opf & REQ_RAHEAD) { in nilfs_btnode_submit_block()
139 submit_bh(opf, bh); in nilfs_btnode_submit_block()
H A Dmdt.c116 nilfs_mdt_submit_block(struct inode *inode, unsigned long blkoff, blk_opf_t opf, in nilfs_mdt_submit_block() argument
131 if (opf & REQ_RAHEAD) { in nilfs_mdt_submit_block()
153 submit_bh(opf, bh); in nilfs_mdt_submit_block()
157 opf & REQ_OP_MASK); in nilfs_mdt_submit_block()
/linux/fs/
H A Dmpage.c168 blk_opf_t opf = REQ_OP_READ; in do_mpage_readpage() local
174 opf |= REQ_RAHEAD; in do_mpage_readpage()
287 args->bio = bio_alloc(bdev, bio_max_segs(args->nr_pages), opf, in do_mpage_readpage()
H A Dbuffer.c58 static void submit_bh_wbc(blk_opf_t opf, struct buffer_head *bh,
2777 static void submit_bh_wbc(blk_opf_t opf, struct buffer_head *bh, in submit_bh_wbc() argument
2781 const enum req_op op = opf & REQ_OP_MASK; in submit_bh_wbc()
2797 opf |= REQ_META; in submit_bh_wbc()
2799 opf |= REQ_PRIO; in submit_bh_wbc()
2801 bio = bio_alloc(bh->b_bdev, 1, opf, GFP_NOIO); in submit_bh_wbc()
2824 void submit_bh(blk_opf_t opf, struct buffer_head *bh) in submit_bh() argument
2826 submit_bh_wbc(opf, bh, WRITE_LIFE_NOT_SET, NULL); in submit_bh()
/linux/drivers/nvme/target/
H A Dzns.c536 const blk_opf_t opf = REQ_OP_ZONE_APPEND | REQ_SYNC | REQ_IDLE; in nvmet_bdev_execute_zone_append() local
575 ARRAY_SIZE(req->inline_bvec), opf); in nvmet_bdev_execute_zone_append()
577 bio = bio_alloc(req->ns->bdev, req->sg_cnt, opf, GFP_KERNEL); in nvmet_bdev_execute_zone_append()
/linux/fs/iomap/
H A Ddirect-io.c54 struct iomap_dio *dio, unsigned short nr_vecs, blk_opf_t opf) in iomap_dio_alloc_bio() argument
57 return bio_alloc_bioset(iter->iomap.bdev, nr_vecs, opf, in iomap_dio_alloc_bio()
59 return bio_alloc(iter->iomap.bdev, nr_vecs, opf, GFP_KERNEL); in iomap_dio_alloc_bio()
/linux/include/scsi/
H A Dscsi_cmnd.h397 struct request *scsi_alloc_request(struct request_queue *q, blk_opf_t opf,

12