| /linux/io_uring/ |
| H A D | net.h | 28 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags); 32 int io_sendmsg(struct io_kiocb *req, unsigned int issue_flags); 34 int io_send(struct io_kiocb *req, unsigned int issue_flags); 37 int io_recvmsg(struct io_kiocb *req, unsigned int issue_flags); 38 int io_recv(struct io_kiocb *req, unsigned int issue_flags); 43 int io_accept(struct io_kiocb *req, unsigned int issue_flags); 46 int io_socket(struct io_kiocb *req, unsigned int issue_flags); 49 int io_connect(struct io_kiocb *req, unsigned int issue_flags); 51 int io_send_zc(struct io_kiocb *req, unsigned int issue_flags); 52 int io_sendmsg_zc(struct io_kiocb *req, unsigned int issue_flags); [all...] |
| H A D | cmd_net.c | 11 unsigned int issue_flags) in io_uring_cmd_getsockopt() argument 14 bool compat = !!(issue_flags & IO_URING_F_COMPAT); in io_uring_cmd_getsockopt() 38 unsigned int issue_flags) in io_uring_cmd_setsockopt() argument 41 bool compat = !!(issue_flags & IO_URING_F_COMPAT); in io_uring_cmd_setsockopt() 57 struct sk_buff *skb, unsigned issue_flags) in io_process_timestamp_skb() argument 85 return io_uring_cmd_post_mshot_cqe32(cmd, issue_flags, cqe); in io_process_timestamp_skb() 90 unsigned int issue_flags) in io_uring_cmd_timestamp() argument 98 if (!(issue_flags & IO_URING_F_CQE32)) in io_uring_cmd_timestamp() 100 ret = io_cmd_poll_multishot(cmd, issue_flags, EPOLLERR); in io_uring_cmd_timestamp() 122 if (!io_process_timestamp_skb(cmd, sk, skb, issue_flags)) in io_uring_cmd_timestamp() 135 io_uring_cmd_sock(struct io_uring_cmd * cmd,unsigned int issue_flags) io_uring_cmd_sock() argument [all...] |
| H A D | openclose.c | 117 int io_openat2(struct io_kiocb *req, unsigned int issue_flags) in io_openat2() argument 131 if (issue_flags & IO_URING_F_NONBLOCK) { in io_openat2() 156 (!resolve_nonblock && (issue_flags & IO_URING_F_NONBLOCK))) in io_openat2() 161 if ((issue_flags & IO_URING_F_NONBLOCK) && !nonblock_set) in io_openat2() 167 ret = io_fixed_fd_install(req, issue_flags, file, in io_openat2() 178 int io_openat(struct io_kiocb *req, unsigned int issue_flags) in io_openat() argument 180 return io_openat2(req, issue_flags); in io_openat() 191 int __io_close_fixed(struct io_ring_ctx *ctx, unsigned int issue_flags, in __io_close_fixed() argument 196 io_ring_submit_lock(ctx, issue_flags); in __io_close_fixed() 198 io_ring_submit_unlock(ctx, issue_flags); in __io_close_fixed() [all …]
|
| H A D | openclose.h | 3 int __io_close_fixed(struct io_ring_ctx *ctx, unsigned int issue_flags, 7 int io_openat(struct io_kiocb *req, unsigned int issue_flags); 11 int io_openat2(struct io_kiocb *req, unsigned int issue_flags); 14 int io_close(struct io_kiocb *req, unsigned int issue_flags); 17 int io_pipe(struct io_kiocb *req, unsigned int issue_flags); 20 int io_install_fixed_fd(struct io_kiocb *req, unsigned int issue_flags);
|
| H A D | net.c | 136 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags) in io_shutdown() argument 142 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_shutdown() 166 static void io_netmsg_recycle(struct io_kiocb *req, unsigned int issue_flags) in io_netmsg_recycle() argument 171 if (unlikely(issue_flags & IO_URING_F_UNLOCKED)) { in io_netmsg_recycle() 450 unsigned int issue_flags) in io_req_msg_cleanup() argument 452 io_netmsg_recycle(req, issue_flags); in io_req_msg_cleanup() 537 int io_sendmsg(struct io_kiocb *req, unsigned int issue_flags) in io_sendmsg() argument 555 if (issue_flags & IO_URING_F_NONBLOCK) in io_sendmsg() 565 if (ret == -EAGAIN && (issue_flags & IO_URING_F_NONBLOCK)) in io_sendmsg() 577 io_req_msg_cleanup(req, issue_flags); in io_sendmsg() [all …]
|
| H A D | fs.h | 4 int io_renameat(struct io_kiocb *req, unsigned int issue_flags); 8 int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags); 12 int io_mkdirat(struct io_kiocb *req, unsigned int issue_flags); 16 int io_symlinkat(struct io_kiocb *req, unsigned int issue_flags); 19 int io_linkat(struct io_kiocb *req, unsigned int issue_flags);
|
| H A D | poll.h | 27 * Must only be called inside issue_flags & IO_URING_F_MULTISHOT, or 36 int io_poll_add(struct io_kiocb *req, unsigned int issue_flags); 39 int io_poll_remove(struct io_kiocb *req, unsigned int issue_flags); 43 unsigned issue_flags); 44 int io_arm_apoll(struct io_kiocb *req, unsigned issue_flags, __poll_t mask); 45 int io_arm_poll_handler(struct io_kiocb *req, unsigned issue_flags);
|
| H A D | rw.c | 112 unsigned int issue_flags) in __io_import_rw_buffer() argument 123 *sel = io_buffer_select(req, &sqe_len, io->buf_group, issue_flags); in __io_import_rw_buffer() 135 unsigned int issue_flags) in io_import_rw_buffer() argument 139 ret = __io_import_rw_buffer(rw, req, io, sel, issue_flags); in io_import_rw_buffer() 147 static void io_rw_recycle(struct io_kiocb *req, unsigned int issue_flags) in io_rw_recycle() argument 151 if (unlikely(issue_flags & IO_URING_F_UNLOCKED)) in io_rw_recycle() 162 static void io_req_rw_cleanup(struct io_kiocb *req, unsigned int issue_flags) in io_req_rw_cleanup() argument 193 io_rw_recycle(req, issue_flags); in io_req_rw_cleanup() 366 static int io_init_rw_fixed(struct io_kiocb *req, unsigned int issue_flags, in io_init_rw_fixed() argument 377 issue_flags); in io_init_rw_fixed() [all …]
|
| H A D | poll.c | 528 static void io_poll_add_hash(struct io_kiocb *req, unsigned int issue_flags) in io_poll_add_hash() argument 532 io_ring_submit_lock(ctx, issue_flags); in io_poll_add_hash() 534 io_ring_submit_unlock(ctx, issue_flags); in io_poll_add_hash() 546 unsigned issue_flags) in __io_arm_poll_handler() argument 568 ipt->owning = issue_flags & IO_URING_F_UNLOCKED; in __io_arm_poll_handler() 599 io_poll_add_hash(req, issue_flags); in __io_arm_poll_handler() 608 io_poll_add_hash(req, issue_flags); in __io_arm_poll_handler() 646 unsigned issue_flags) in io_req_alloc_apoll() argument 655 if (!(issue_flags & IO_URING_F_UNLOCKED)) in io_req_alloc_apoll() 670 int io_arm_apoll(struct io_kiocb *req, unsigned issue_flags, __poll_t mask) in io_arm_apoll() argument [all …]
|
| H A D | rw.h | 43 int io_read(struct io_kiocb *req, unsigned int issue_flags); 44 int io_write(struct io_kiocb *req, unsigned int issue_flags); 45 int io_read_fixed(struct io_kiocb *req, unsigned int issue_flags); 46 int io_write_fixed(struct io_kiocb *req, unsigned int issue_flags); 51 int io_read_mshot(struct io_kiocb *req, unsigned int issue_flags);
|
| H A D | xattr.h | 6 int io_fsetxattr(struct io_kiocb *req, unsigned int issue_flags); 9 int io_setxattr(struct io_kiocb *req, unsigned int issue_flags); 12 int io_fgetxattr(struct io_kiocb *req, unsigned int issue_flags); 15 int io_getxattr(struct io_kiocb *req, unsigned int issue_flags);
|
| H A D | sync.c | 40 int io_sync_file_range(struct io_kiocb *req, unsigned int issue_flags) in io_sync_file_range() argument 46 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_sync_file_range() 70 int io_fsync(struct io_kiocb *req, unsigned int issue_flags) in io_fsync() argument 77 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_fsync() 99 int io_fallocate(struct io_kiocb *req, unsigned int issue_flags) in io_fallocate() argument 105 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_fallocate()
|
| H A D | fs.c | 81 int io_renameat(struct io_kiocb *req, unsigned int issue_flags) in io_renameat() argument 86 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_renameat() 130 int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags) in io_unlinkat() argument 135 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_unlinkat() 177 int io_mkdirat(struct io_kiocb *req, unsigned int issue_flags) in io_mkdirat() argument 182 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_mkdirat() 227 int io_symlinkat(struct io_kiocb *req, unsigned int issue_flags) in io_symlinkat() argument 232 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_symlinkat() 272 int io_linkat(struct io_kiocb *req, unsigned int issue_flags) in io_linkat() argument 277 WARN_ON_ONCE(issue_flags in io_linkat() [all...] |
| H A D | kbuf.c | 104 bool io_kbuf_recycle_legacy(struct io_kiocb *req, unsigned issue_flags) in io_kbuf_recycle_legacy() argument 110 io_ring_submit_lock(ctx, issue_flags); in io_kbuf_recycle_legacy() 118 io_ring_submit_unlock(ctx, issue_flags); in io_kbuf_recycle_legacy() 158 static bool io_should_commit(struct io_kiocb *req, unsigned int issue_flags) in io_should_commit() argument 170 if (issue_flags & IO_URING_F_UNLOCKED) in io_should_commit() 181 unsigned int issue_flags) in io_ring_buffer_select() argument 205 if (io_should_commit(req, issue_flags)) { in io_ring_buffer_select() 213 unsigned buf_group, unsigned int issue_flags) in io_buffer_select() argument 219 io_ring_submit_lock(req->ctx, issue_flags); in io_buffer_select() 224 sel = io_ring_buffer_select(req, len, bl, issue_flags); in io_buffer_select() [all …]
|
| H A D | cancel.c | 107 unsigned issue_flags) in io_try_cancel() argument 122 ret = io_poll_cancel(ctx, cd, issue_flags); in io_try_cancel() 126 ret = io_waitid_cancel(ctx, cd, issue_flags); in io_try_cancel() 130 ret = io_futex_cancel(ctx, cd, issue_flags); in io_try_cancel() 170 unsigned int issue_flags) in __io_async_cancel() argument 178 ret = io_try_cancel(tctx, cd, issue_flags); in __io_async_cancel() 188 io_ring_submit_lock(ctx, issue_flags); in __io_async_cancel() 200 io_ring_submit_unlock(ctx, issue_flags); in __io_async_cancel() 204 int io_async_cancel(struct io_kiocb *req, unsigned int issue_flags) in io_async_cancel() argument 222 issue_flags); in io_async_cancel() [all …]
|
| H A D | sync.h | 4 int io_sync_file_range(struct io_kiocb *req, unsigned int issue_flags); 7 int io_fsync(struct io_kiocb *req, unsigned int issue_flags); 9 int io_fallocate(struct io_kiocb *req, unsigned int issue_flags);
|
| H A D | uring_cmd.h | 11 int io_uring_cmd(struct io_kiocb *req, unsigned int issue_flags); 20 unsigned int issue_flags, 26 unsigned int issue_flags, __poll_t mask);
|
| H A D | advise.c | 51 int io_madvise(struct io_kiocb *req, unsigned int issue_flags) in io_madvise() argument 57 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_madvise() 96 int io_fadvise(struct io_kiocb *req, unsigned int issue_flags) in io_fadvise() argument 101 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK && io_fadvise_force_async(fa)); in io_fadvise()
|
| H A D | cancel.h | 19 int io_async_cancel(struct io_kiocb *req, unsigned int issue_flags); 22 unsigned int issue_flags); 33 unsigned int issue_flags, struct hlist_head *list,
|
| H A D | waitid.c | 168 unsigned int issue_flags) in io_waitid_cancel() argument 170 return io_cancel_remove(ctx, cd, issue_flags, &ctx->waitid_list, __io_waitid_cancel); in io_waitid_cancel() 281 int io_waitid(struct io_kiocb *req, unsigned int issue_flags) in io_waitid() argument 306 io_ring_submit_lock(ctx, issue_flags); in io_waitid() 326 io_ring_submit_unlock(ctx, issue_flags); in io_waitid() 334 io_ring_submit_unlock(ctx, issue_flags); in io_waitid() 342 io_ring_submit_unlock(ctx, issue_flags); in io_waitid()
|
| H A D | rsrc.h | 63 unsigned issue_flags); 66 unsigned issue_flags); 69 unsigned nr_iovs, unsigned issue_flags); 119 int io_files_update(struct io_kiocb *req, unsigned int issue_flags);
|
| H A D | zcrx.h | 76 unsigned issue_flags, unsigned int *len); 90 unsigned issue_flags, unsigned int *len) in io_zcrx_recv() argument 106 int io_recvzc(struct io_kiocb *req, unsigned int issue_flags);
|
| H A D | rsrc.c | 447 unsigned int issue_flags) in io_files_update_with_index_alloc() argument 469 ret = io_fixed_fd_install(req, issue_flags, file, in io_files_update_with_index_alloc() 474 __io_close_fixed(req->ctx, issue_flags, ret); in io_files_update_with_index_alloc() 485 int io_files_update(struct io_kiocb *req, unsigned int issue_flags) in io_files_update() argument 500 ret = io_files_update_with_index_alloc(req, issue_flags); in io_files_update() 502 io_ring_submit_lock(ctx, issue_flags); in io_files_update() 505 io_ring_submit_unlock(ctx, issue_flags); in io_files_update() 941 unsigned int issue_flags) in io_buffer_register_bvec() argument 952 io_ring_submit_lock(ctx, issue_flags); in io_buffer_register_bvec() 998 io_ring_submit_unlock(ctx, issue_flags); in io_buffer_register_bvec() 1004 io_buffer_unregister_bvec(struct io_uring_cmd * cmd,unsigned int index,unsigned int issue_flags) io_buffer_unregister_bvec() argument 1120 io_find_buf_node(struct io_kiocb * req,unsigned issue_flags) io_find_buf_node() argument 1144 io_import_reg_buf(struct io_kiocb * req,struct iov_iter * iter,u64 buf_addr,size_t len,int ddir,unsigned issue_flags) io_import_reg_buf() argument 1491 io_import_reg_vec(int ddir,struct iov_iter * iter,struct io_kiocb * req,struct iou_vec * vec,unsigned nr_iovs,unsigned issue_flags) io_import_reg_vec() argument [all...] |
| /linux/fs/fuse/ |
| H A D | dev_uring.c | 501 unsigned int issue_flags) in fuse_uring_cancel() argument 523 io_uring_cmd_done(cmd, -ENOTCONN, issue_flags); in fuse_uring_cancel() 527 static void fuse_uring_prepare_cancel(struct io_uring_cmd *cmd, int issue_flags, in fuse_uring_prepare_cancel() argument 531 io_uring_cmd_mark_cancelable(cmd, issue_flags); in fuse_uring_prepare_cancel() 724 unsigned int issue_flags) in fuse_uring_send_next_to_ring() argument 741 io_uring_cmd_done(cmd, 0, issue_flags); in fuse_uring_send_next_to_ring() 816 unsigned int issue_flags) in fuse_uring_commit() argument 845 unsigned int issue_flags) in fuse_uring_next_fuse_req() argument 857 err = fuse_uring_send_next_to_ring(ent, req, issue_flags); in fuse_uring_next_fuse_req() 879 static int fuse_uring_commit_fetch(struct io_uring_cmd *cmd, int issue_flags, in fuse_uring_commit_fetch() argument [all …]
|
| /linux/drivers/nvme/host/ |
| H A D | ioctl.c | 440 struct io_uring_cmd *ioucmd, unsigned int issue_flags, bool vec) in nvme_uring_cmd_io() argument 490 ddir, &iter, issue_flags); in nvme_uring_cmd_io() 493 ddir, &iter, ioucmd, issue_flags); in nvme_uring_cmd_io() 500 if (issue_flags & IO_URING_F_NONBLOCK) { in nvme_uring_cmd_io() 504 if (issue_flags & IO_URING_F_IOPOLL) in nvme_uring_cmd_io() 631 static int nvme_uring_cmd_checks(unsigned int issue_flags) in nvme_uring_cmd_checks() argument 635 if ((issue_flags & (IO_URING_F_SQE128|IO_URING_F_CQE32)) != in nvme_uring_cmd_checks() 642 unsigned int issue_flags) in nvme_ns_uring_cmd() argument 647 ret = nvme_uring_cmd_checks(issue_flags); in nvme_ns_uring_cmd() 653 ret = nvme_uring_cmd_io(ctrl, ns, ioucmd, issue_flags, false); in nvme_ns_uring_cmd() [all …]
|