/linux/drivers/infiniband/ulp/rtrs/ |
H A D | rtrs-srv-stats.c | 44 sum.dir[WRITE].cnt += r->dir[WRITE].cnt; in rtrs_srv_stats_rdma_to_str() 45 sum.dir[WRITE].size_total += r->dir[WRITE].size_total; in rtrs_srv_stats_rdma_to_str() 50 sum.dir[WRITE].cnt, sum.dir[WRITE].size_total); in rtrs_srv_stats_rdma_to_str()
|
H A D | rtrs-clt-stats.c | 94 sum.dir[WRITE].cnt += r->dir[WRITE].cnt; in rtrs_clt_stats_rdma_to_str() 95 sum.dir[WRITE].size_total += r->dir[WRITE].size_total; in rtrs_clt_stats_rdma_to_str() 101 sum.dir[WRITE].cnt, sum.dir[WRITE].size_total, in rtrs_clt_stats_rdma_to_str()
|
H A D | rtrs-srv-trace.h | 72 { WRITE, "WRITE" }),
|
/linux/drivers/md/ |
H A D | dm-stats.c | 191 atomic_read(&shared->in_flight[WRITE]); in dm_stat_in_flight() 234 atomic_read(&shared->in_flight[WRITE])); in dm_stats_cleanup() 347 atomic_set(&s->stat_shared[ni].in_flight[WRITE], 0); in dm_stats_create() 550 in_flight_write = (unsigned int)atomic_read(&shared->in_flight[WRITE]); in dm_stat_round() 554 p->io_ticks[WRITE] += difference; in dm_stat_round() 696 ((bi_rw == WRITE) == in dm_stats_account_io() 697 (READ_ONCE(last->last_rw) == WRITE)) in dm_stats_account_io() 732 shared->tmp.sectors[WRITE] = 0; in __dm_stat_init_temporary_percpu_totals() 734 shared->tmp.ios[WRITE] = 0; in __dm_stat_init_temporary_percpu_totals() 736 shared->tmp.merges[WRITE] = 0; in __dm_stat_init_temporary_percpu_totals() [all …]
|
H A D | dm-flakey.c | 149 fc->corrupt_bio_rw = WRITE; in parse_features() 218 ((fc->corrupt_bio_byte && fc->corrupt_bio_rw == WRITE) || in parse_features() 224 ((fc->corrupt_bio_byte && fc->corrupt_bio_rw == WRITE) || in parse_features() 379 (bio_data_dir(bio) == WRITE) ? 'w' : 'r', bio->bi_opf, in corrupt_bio_common() 542 if (fc->corrupt_bio_byte && fc->corrupt_bio_rw == WRITE) { in flakey_map() 638 (fc->corrupt_bio_rw == WRITE) ? 'w' : 'r', in flakey_status()
|
/linux/block/ |
H A D | blk-throttle.c | 263 INIT_LIST_HEAD(&sq->queued[WRITE]); in throtl_service_queue_init() 286 for (rw = READ; rw <= WRITE; rw++) { in throtl_pd_alloc() 293 tg->bps[WRITE] = U64_MAX; in throtl_pd_alloc() 295 tg->iops[WRITE] = UINT_MAX; in throtl_pd_alloc() 343 for (rw = READ; rw <= WRITE; rw++) { in tg_update_has_rules() 759 __tg_update_carryover(tg, WRITE, &bytes[WRITE], &ios[WRITE]); in tg_update_carryover() 763 bytes[READ], bytes[WRITE], ios[READ], ios[WRITE]); in tg_update_carryover() 976 bio = throtl_peek_queued(&sq->queued[WRITE]); in tg_update_disptime() 1065 while ((bio = throtl_peek_queued(&sq->queued[WRITE])) && in throtl_dispatch_tg() 1068 tg_dispatch_one_bio(tg, WRITE); in throtl_dispatch_tg() [all …]
|
/linux/drivers/s390/net/ |
H A D | ctcm_sysfs.c | 96 priv->channel[WRITE]->prof.maxmulti); in ctcm_print_statistics() 98 priv->channel[WRITE]->prof.maxcqueue); in ctcm_print_statistics() 100 priv->channel[WRITE]->prof.doios_single); in ctcm_print_statistics() 102 priv->channel[WRITE]->prof.doios_multi); in ctcm_print_statistics() 104 priv->channel[WRITE]->prof.txlen); in ctcm_print_statistics() 106 jiffies_to_usecs(priv->channel[WRITE]->prof.tx_time)); in ctcm_print_statistics() 133 memset(&priv->channel[WRITE]->prof, 0, in stats_write()
|
/linux/Documentation/admin-guide/device-mapper/ |
H A D | log-writes.rst | 8 There is a log_write_entry written for every WRITE request and the target is 10 that is in the WRITE requests is copied into the log to make the replay happen 17 cache. This means that normal WRITE requests are not actually logged until the 22 This works by attaching all WRITE requests to a list once the write completes. 42 Any REQ_OP_DISCARD requests are treated like WRITE requests. Otherwise we would 43 have all the DISCARD requests, and then the WRITE requests and then the FLUSH 46 WRITE block 1, DISCARD block 1, FLUSH 50 DISCARD 1, WRITE 1, FLUSH
|
/linux/drivers/s390/block/ |
H A D | dasd_fba.c | 84 if (rw == WRITE) in define_extent() 104 if (rw == WRITE) in locate_record() 354 define_extent(ccw++, cqr->data, WRITE, blksize, first_rec, count); in dasd_fba_build_cp_discard() 366 locate_record(ccw++, LO_data++, WRITE, cur_pos, wz_count); in dasd_fba_build_cp_discard() 386 locate_record(ccw++, LO_data++, WRITE, cur_pos, d_count); in dasd_fba_build_cp_discard() 404 locate_record(ccw++, LO_data++, WRITE, cur_pos, wz_count); in dasd_fba_build_cp_discard() 445 } else if (rq_data_dir(req) == WRITE) { in dasd_fba_build_cp_regular() 504 if (copy && rq_data_dir(req) == WRITE) in dasd_fba_build_cp_regular()
|
/linux/fs/f2fs/ |
H A D | file.c | 590 f2fs_down_write(&F2FS_I(inode)->i_gc_rwsem[WRITE]); in finish_preallocate_blocks() 597 f2fs_up_write(&F2FS_I(inode)->i_gc_rwsem[WRITE]); in finish_preallocate_blocks() 944 if (f2fs_sb_has_blkzoned(sbi) && (rw == WRITE) && in f2fs_force_buffered_io() 981 if (!f2fs_force_buffered_io(inode, WRITE)) { in f2fs_getattr() 1140 f2fs_down_write(&fi->i_gc_rwsem[WRITE]); in f2fs_setattr() 1154 f2fs_up_write(&fi->i_gc_rwsem[WRITE]); in f2fs_setattr() 1300 f2fs_down_write(&F2FS_I(inode)->i_gc_rwsem[WRITE]); in f2fs_punch_hole() 1310 f2fs_up_write(&F2FS_I(inode)->i_gc_rwsem[WRITE]); in f2fs_punch_hole() 1547 f2fs_down_write(&F2FS_I(inode)->i_gc_rwsem[WRITE]); in f2fs_do_collapse() 1559 f2fs_up_write(&F2FS_I(inode)->i_gc_rwsem[WRITE]); in f2fs_do_collapse() [all …]
|
H A D | verity.c | 206 f2fs_down_write(&F2FS_I(inode)->i_gc_rwsem[WRITE]); in f2fs_end_enable_verity() 214 f2fs_up_write(&F2FS_I(inode)->i_gc_rwsem[WRITE]); in f2fs_end_enable_verity()
|
/linux/fs/bcachefs/ |
H A D | compress.c | 276 data = __bounce_alloc(c, dst_len, WRITE); in bch2_bio_uncompress_inplace() 320 ? __bio_map_or_bounce(c, dst, dst_iter, WRITE) in bch2_bio_uncompress() 321 : __bounce_alloc(c, dst_len, WRITE); in bch2_bio_uncompress() 458 dst_data = bio_map_or_bounce(c, dst, WRITE); in __bio_compress() 621 mempool_exit(&c->compression_bounce[WRITE]); in bch2_fs_compress_exit() 661 if (!mempool_initialized(&c->compression_bounce[WRITE]) && in __bch2_fs_compress_init() 662 mempool_init_kvmalloc_pool(&c->compression_bounce[WRITE], in __bch2_fs_compress_init()
|
H A D | super.c | 316 bch2_dev_io_ref_stop(ca, WRITE); in __bch2_fs_read_only() 502 enumerated_ref_start(&ca->io_ref[WRITE]); in __bch2_fs_read_write() 612 bch2_io_clock_exit(&c->io_clock[WRITE]); in __bch2_fs_free() 1003 bch2_io_clock_init(&c->io_clock[WRITE]) ?: in bch2_fs_alloc() 1373 WARN_ON(!enumerated_ref_is_zero(&ca->io_ref[WRITE])); in bch2_dev_free() 1394 bch2_time_stats_quantiles_exit(&ca->io_latency[WRITE]); in bch2_dev_free() 1397 enumerated_ref_exit(&ca->io_ref[WRITE]); in bch2_dev_free() 1497 bch2_time_stats_quantiles_init(&ca->io_latency[WRITE]); in __bch2_dev_alloc() 1522 enumerated_ref_init(&ca->io_ref[WRITE], BCH_DEV_WRITE_REF_NR, NULL) || in __bch2_dev_alloc() 1587 BUG_ON(!enumerated_ref_is_zero(&ca->io_ref[WRITE])); in __bch2_dev_attach_bdev() [all …]
|
H A D | movinggc.c | 311 atomic64_read(&c->io_clock[WRITE].now)) << 9); in bch2_copygc_wait_to_text() 316 atomic64_read(&c->io_clock[WRITE].now) - in bch2_copygc_wait_to_text() 349 struct io_clock *clock = &c->io_clock[WRITE]; in bch2_copygc_thread()
|
H A D | alloc_background.c | 360 prt_printf(out, "io_time[WRITE] %llu\n", a->io_time[WRITE]); in __bch2_alloc_v4_to_text() 411 .io_time[WRITE] = u.write_time, in __bch2_alloc_to_v4() 894 new_a->io_time[WRITE]= bch2_current_io_time(c, WRITE); in bch2_trigger_alloc() 1960 enumerated_ref_put(&ca->io_ref[WRITE], BCH_DEV_WRITE_REF_dev_do_discards); in bch2_do_discards_work() 1971 if (!bch2_dev_get_ioref(c, ca->dev_idx, WRITE, BCH_DEV_WRITE_REF_dev_do_discards)) in bch2_dev_do_discards() 1977 enumerated_ref_put(&ca->io_ref[WRITE], BCH_DEV_WRITE_REF_dev_do_discards); in bch2_dev_do_discards() 2055 enumerated_ref_put(&ca->io_ref[WRITE], BCH_DEV_WRITE_REF_discard_one_bucket_fast); in bch2_do_discards_fast_work() 2069 if (!bch2_dev_get_ioref(c, ca->dev_idx, WRITE, BCH_DEV_WRITE_REF_discard_one_bucket_fast)) in bch2_discard_one_bucket_fast() 2075 enumerated_ref_put(&ca->io_ref[WRITE], BCH_DEV_WRITE_REF_discard_one_bucket_fast); in bch2_discard_one_bucket_fast() 2273 enumerated_ref_put(&ca->io_ref[WRITE], BCH_DEV_WRITE_REF_do_invalidates); in bch2_do_invalidates_work() [all …]
|
/linux/tools/testing/selftests/powerpc/benchmarks/ |
H A D | context_switch.c | 171 #define WRITE 1 macro 191 assert(write(pipe_fd2[WRITE], &c, 1) == 1); in pipe_thread1() 203 assert(write(pipe_fd1[WRITE], &c, 1) == 1); in pipe_thread2()
|
/linux/drivers/block/drbd/ |
H A D | drbd_req.c | 33 req->rq_state = (bio_data_dir(bio_src) == WRITE ? RQ_WRITE : 0) in drbd_req_new() 819 if (bio_data_dir(req->master_bio) == WRITE) in __req_mod() 1235 if (rw == WRITE && req->private_bio && req->i.size in drbd_request_prepare() 1327 if (rw == WRITE) { in drbd_send_and_submit() 1354 if (rw != WRITE) { in drbd_send_and_submit() 1365 if (rw == WRITE) in drbd_send_and_submit() 1371 if (rw == WRITE) { in drbd_send_and_submit() 1400 &device->pending_master_completion[rw == WRITE]); in drbd_send_and_submit() 1405 &device->pending_completion[rw == WRITE]); in drbd_send_and_submit() 1451 if (rw == WRITE /* rw != WRITE should not even end up here! */ in submit_fast_path()
|
/linux/Documentation/PCI/endpoint/ |
H A D | pci-test-function.rst | 78 the COPY/WRITE command. 83 for the READ/WRITE/COPY and raise IRQ (Legacy/MSI) commands.
|
/linux/fs/ |
H A D | read_write.c | 655 ret = rw_verify_area(WRITE, file, pos, count); in kernel_write() 677 ret = rw_verify_area(WRITE, file, pos, count); in vfs_write() 944 ret = rw_verify_area(WRITE, file, &iocb->ki_pos, tot_len); in vfs_iocb_iter_write() 976 ret = rw_verify_area(WRITE, file, ppos, tot_len); in vfs_iter_write() 981 ret = do_iter_readv_writev(file, iter, ppos, WRITE, flags); in vfs_iter_write() 1051 ret = rw_verify_area(WRITE, file, pos, tot_len); in vfs_writev() 1057 ret = do_iter_readv_writev(file, &iter, pos, WRITE, flags); in vfs_writev() 1059 ret = do_loop_readv_writev(file, &iter, pos, WRITE, flags); in vfs_writev() 1367 retval = rw_verify_area(WRITE, fd_file(out), &out_pos, count); in do_sendfile() 1572 ret = rw_verify_area(WRITE, file_out, &pos_out, len); in vfs_copy_file_range()
|
/linux/rust/kernel/mm/ |
H A D | virt.rs | 339 (self.flags() & flags::WRITE) != 0 in writable() 401 pub const WRITE: vm_flags_t = bindings::VM_WRITE as vm_flags_t; constant
|
/linux/tools/perf/trace/beauty/ |
H A D | flock.c | 44 P_CMD(WRITE); in syscall_arg__scnprintf_flock()
|
/linux/drivers/block/ |
H A D | swim3.c | 332 if (rq_data_dir(req) == WRITE) { in swim3_queue_rq() 443 if (rq_data_dir(req) == WRITE) in setup_transfer() 460 if (rq_data_dir(req) == WRITE) { in setup_transfer() 477 if (rq_data_dir(req) == WRITE) in setup_transfer() 648 (rq_data_dir(fs->cur_req)==WRITE? "writ": "read"), in xfer_timeout() 733 if (rq_data_dir(req) == WRITE) in swim3_interrupt() 766 rq_data_dir(req) == WRITE? "writ": "read", in swim3_interrupt()
|
/linux/net/rxrpc/ |
H A D | output.c | 274 iov_iter_kvec(&msg.msg_iter, WRITE, kv, nr_kv, len); in rxrpc_send_ack_packet() 412 iov_iter_kvec(&msg.msg_iter, WRITE, iov, 1, sizeof(pkt)); in rxrpc_send_abort_packet() 667 iov_iter_kvec(&msg.msg_iter, WRITE, call->local->kvec, 1 + req->n, len); in rxrpc_send_data_packet() 789 iov_iter_kvec(&msg.msg_iter, WRITE, iov, 2, len); in rxrpc_send_conn_abort() 862 iov_iter_kvec(&msg.msg_iter, WRITE, iov, ioc, size); in rxrpc_reject_packet() 911 iov_iter_kvec(&msg.msg_iter, WRITE, iov, 2, len); in rxrpc_send_keepalive() 952 iov_iter_bvec(&msg.msg_iter, WRITE, bvec, nr_sg, len); in rxrpc_send_response()
|
/linux/Documentation/ABI/testing/ |
H A D | sysfs-devices-platform-_UDC_-gadget | 17 the SCSI WRITE(10,12) commands when a gadget in USB Mass
|
/linux/include/trace/misc/ |
H A D | fs.h | 46 __fmode_flag(WRITE), \
|