| /linux/include/linux/io_uring/ |
| H A D | cmd.h | 48 unsigned int issue_flags); 53 unsigned issue_flags); 63 unsigned issue_flags, bool is_cqe32); 74 unsigned int issue_flags); 85 unsigned int issue_flags); 92 struct io_br_sel *sel, unsigned int issue_flags); 98 unsigned int issue_flags) in io_uring_cmd_import_fixed() argument 106 unsigned issue_flags) in io_uring_cmd_import_fixed_vec() argument 111 u64 ret2, unsigned issue_flags, bool is_cqe32) in __io_uring_cmd_done() argument 119 unsigned int issue_flags) in io_uring_cmd_mark_cancelable() argument [all …]
|
| /linux/io_uring/ |
| H A D | net.h | 29 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags); 33 int io_sendmsg(struct io_kiocb *req, unsigned int issue_flags); 35 int io_send(struct io_kiocb *req, unsigned int issue_flags); 38 int io_recvmsg(struct io_kiocb *req, unsigned int issue_flags); 39 int io_recv(struct io_kiocb *req, unsigned int issue_flags); 44 int io_accept(struct io_kiocb *req, unsigned int issue_flags); 47 int io_socket(struct io_kiocb *req, unsigned int issue_flags); 51 int io_connect(struct io_kiocb *req, unsigned int issue_flags); 53 int io_send_zc(struct io_kiocb *req, unsigned int issue_flags); 54 int io_sendmsg_zc(struct io_kiocb *req, unsigned int issue_flags); [all...] |
| H A D | cmd_net.c | 12 unsigned int issue_flags) in io_uring_cmd_getsockopt() 15 bool compat = !!(issue_flags & IO_URING_F_COMPAT); in io_uring_cmd_getsockopt() 39 unsigned int issue_flags) in io_uring_cmd_setsockopt() 42 bool compat = !!(issue_flags & IO_URING_F_COMPAT); in io_uring_cmd_setsockopt() 58 struct sk_buff *skb, unsigned issue_flags) in io_process_timestamp_skb() 86 return io_uring_cmd_post_mshot_cqe32(cmd, issue_flags, cqe); in io_process_timestamp_skb() 91 unsigned int issue_flags) in io_uring_cmd_timestamp() 99 if (!(issue_flags & IO_URING_F_CQE32)) in io_uring_cmd_timestamp() 101 ret = io_cmd_poll_multishot(cmd, issue_flags, EPOLLERR); in io_uring_cmd_timestamp() 123 if (!io_process_timestamp_skb(cmd, sk, skb, issue_flags)) in io_uring_cmd_timestamp() 11 io_uring_cmd_getsockopt(struct socket * sock,struct io_uring_cmd * cmd,unsigned int issue_flags) io_uring_cmd_getsockopt() argument 38 io_uring_cmd_setsockopt(struct socket * sock,struct io_uring_cmd * cmd,unsigned int issue_flags) io_uring_cmd_setsockopt() argument 57 io_process_timestamp_skb(struct io_uring_cmd * cmd,struct sock * sk,struct sk_buff * skb,unsigned issue_flags) io_process_timestamp_skb() argument 90 io_uring_cmd_timestamp(struct socket * sock,struct io_uring_cmd * cmd,unsigned int issue_flags) io_uring_cmd_timestamp() argument 135 io_uring_cmd_sock(struct io_uring_cmd * cmd,unsigned int issue_flags) io_uring_cmd_sock() argument [all...] |
| H A D | openclose.c | 123 int io_openat2(struct io_kiocb *req, unsigned int issue_flags) in io_openat2() argument 138 if (issue_flags & IO_URING_F_NONBLOCK) { in io_openat2() 163 (issue_flags & IO_URING_F_NONBLOCK)) { in io_openat2() 172 if ((issue_flags & IO_URING_F_NONBLOCK) && !nonblock_set) in io_openat2() 178 ret = io_fixed_fd_install(req, issue_flags, file, in io_openat2() 188 int io_openat(struct io_kiocb *req, unsigned int issue_flags) in io_openat() argument 190 return io_openat2(req, issue_flags); in io_openat() 200 int __io_close_fixed(struct io_ring_ctx *ctx, unsigned int issue_flags, in __io_close_fixed() argument 205 io_ring_submit_lock(ctx, issue_flags); in __io_close_fixed() 207 io_ring_submit_unlock(ctx, issue_flags); in __io_close_fixed() [all …]
|
| H A D | openclose.h | 5 int __io_close_fixed(struct io_ring_ctx *ctx, unsigned int issue_flags, 9 int io_openat(struct io_kiocb *req, unsigned int issue_flags); 14 int io_openat2(struct io_kiocb *req, unsigned int issue_flags); 17 int io_close(struct io_kiocb *req, unsigned int issue_flags); 20 int io_pipe(struct io_kiocb *req, unsigned int issue_flags); 23 int io_install_fixed_fd(struct io_kiocb *req, unsigned int issue_flags);
|
| H A D | net.c | 136 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags) in io_shutdown() argument 142 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_shutdown() 166 static void io_netmsg_recycle(struct io_kiocb *req, unsigned int issue_flags) in io_netmsg_recycle() argument 171 if (unlikely(issue_flags & IO_URING_F_UNLOCKED)) { in io_netmsg_recycle() 450 unsigned int issue_flags) in io_req_msg_cleanup() argument 452 io_netmsg_recycle(req, issue_flags); in io_req_msg_cleanup() 541 int io_sendmsg(struct io_kiocb *req, unsigned int issue_flags) in io_sendmsg() 559 if (issue_flags & IO_URING_F_NONBLOCK) in io_sendmsg() 569 if (ret == -EAGAIN && (issue_flags & IO_URING_F_NONBLOCK)) in io_sendmsg() 581 io_req_msg_cleanup(req, issue_flags); in io_sendmsg() 537 io_sendmsg(struct io_kiocb * req,unsigned int issue_flags) io_sendmsg() argument 586 io_send_select_buffer(struct io_kiocb * req,unsigned int issue_flags,struct io_br_sel * sel,struct io_async_msghdr * kmsg) io_send_select_buffer() argument 634 io_send(struct io_kiocb * req,unsigned int issue_flags) io_send() argument 841 io_recv_finish(struct io_kiocb * req,struct io_async_msghdr * kmsg,struct io_br_sel * sel,bool mshot_finished,unsigned issue_flags) io_recv_finish() argument 1005 io_recvmsg(struct io_kiocb * req,unsigned int issue_flags) io_recvmsg() argument 1093 io_recv_buf_select(struct io_kiocb * req,struct io_async_msghdr * kmsg,struct io_br_sel * sel,unsigned int issue_flags) io_recv_buf_select() argument 1164 io_recv(struct io_kiocb * req,unsigned int issue_flags) io_recv() argument 1268 io_recvzc(struct io_kiocb * req,unsigned int issue_flags) io_recvzc() argument 1438 io_send_zc_import(struct io_kiocb * req,unsigned int issue_flags) io_send_zc_import() argument 1451 io_send_zc(struct io_kiocb * req,unsigned int issue_flags) io_send_zc() argument 1520 io_sendmsg_zc(struct io_kiocb * req,unsigned int issue_flags) io_sendmsg_zc() argument 1641 io_accept(struct io_kiocb * req,unsigned int issue_flags) io_accept() argument 1723 io_socket(struct io_kiocb * req,unsigned int issue_flags) io_socket() argument 1775 io_connect(struct io_kiocb * req,unsigned int issue_flags) io_connect() argument 1845 io_bind(struct io_kiocb * req,unsigned int issue_flags) io_bind() argument 1874 io_listen(struct io_kiocb * req,unsigned int issue_flags) io_listen() argument [all...] |
| H A D | fs.h | 4 int io_renameat(struct io_kiocb *req, unsigned int issue_flags); 8 int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags); 12 int io_mkdirat(struct io_kiocb *req, unsigned int issue_flags); 16 int io_symlinkat(struct io_kiocb *req, unsigned int issue_flags); 19 int io_linkat(struct io_kiocb *req, unsigned int issue_flags);
|
| H A D | rw.c | 112 unsigned int issue_flags) in __io_import_rw_buffer() argument 123 *sel = io_buffer_select(req, &sqe_len, io->buf_group, issue_flags); in __io_import_rw_buffer() 135 unsigned int issue_flags) in io_import_rw_buffer() argument 139 ret = __io_import_rw_buffer(rw, req, io, sel, issue_flags); in io_import_rw_buffer() 147 static bool io_rw_recycle(struct io_kiocb *req, unsigned int issue_flags) in io_rw_recycle() argument 151 if (unlikely(issue_flags & IO_URING_F_UNLOCKED)) in io_rw_recycle() 165 static void io_req_rw_cleanup(struct io_kiocb *req, unsigned int issue_flags) in io_req_rw_cleanup() argument 196 if (!io_rw_recycle(req, issue_flags)) { in io_req_rw_cleanup() 373 static int io_init_rw_fixed(struct io_kiocb *req, unsigned int issue_flags, in io_init_rw_fixed() argument 384 issue_flags); in io_init_rw_fixed() [all …]
|
| H A D | poll.h | 27 * Must only be called inside issue_flags & IO_URING_F_MULTISHOT, or 36 int io_poll_add(struct io_kiocb *req, unsigned int issue_flags); 39 int io_poll_remove(struct io_kiocb *req, unsigned int issue_flags); 43 unsigned issue_flags); 44 int io_arm_apoll(struct io_kiocb *req, unsigned issue_flags, __poll_t mask); 45 int io_arm_poll_handler(struct io_kiocb *req, unsigned issue_flags);
|
| H A D | poll.c | 528 static void io_poll_add_hash(struct io_kiocb *req, unsigned int issue_flags) in io_poll_add_hash() argument 532 io_ring_submit_lock(ctx, issue_flags); in io_poll_add_hash() 534 io_ring_submit_unlock(ctx, issue_flags); in io_poll_add_hash() 546 unsigned issue_flags) in __io_arm_poll_handler() argument 568 ipt->owning = issue_flags & IO_URING_F_UNLOCKED; in __io_arm_poll_handler() 599 io_poll_add_hash(req, issue_flags); in __io_arm_poll_handler() 608 io_poll_add_hash(req, issue_flags); in __io_arm_poll_handler() 646 unsigned issue_flags) in io_req_alloc_apoll() argument 655 if (!(issue_flags & IO_URING_F_UNLOCKED)) in io_req_alloc_apoll() 670 int io_arm_apoll(struct io_kiocb *req, unsigned issue_flags, __poll_t mask) in io_arm_apoll() argument [all …]
|
| H A D | rw.h | 43 int io_read(struct io_kiocb *req, unsigned int issue_flags); 44 int io_write(struct io_kiocb *req, unsigned int issue_flags); 45 int io_read_fixed(struct io_kiocb *req, unsigned int issue_flags); 46 int io_write_fixed(struct io_kiocb *req, unsigned int issue_flags); 51 int io_read_mshot(struct io_kiocb *req, unsigned int issue_flags);
|
| H A D | futex.c | 116 unsigned int issue_flags) in io_futex_cancel() argument 118 return io_cancel_remove(ctx, cd, issue_flags, &ctx->futex_list, __io_futex_cancel); in io_futex_cancel() 221 int io_futexv_wait(struct io_kiocb *req, unsigned int issue_flags) in io_futexv_wait() argument 228 io_ring_submit_lock(ctx, issue_flags); in io_futexv_wait() 236 io_ring_submit_unlock(ctx, issue_flags); in io_futexv_wait() 270 io_ring_submit_unlock(ctx, issue_flags); in io_futexv_wait() 274 int io_futex_wait(struct io_kiocb *req, unsigned int issue_flags) in io_futex_wait() argument 286 io_ring_submit_lock(ctx, issue_flags); in io_futex_wait() 304 io_ring_submit_unlock(ctx, issue_flags); in io_futex_wait() 310 io_ring_submit_unlock(ctx, issue_flags); in io_futex_wait() [all …]
|
| H A D | xattr.c | 97 int io_fgetxattr(struct io_kiocb *req, unsigned int issue_flags) in io_fgetxattr() argument 102 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_fgetxattr() 109 int io_getxattr(struct io_kiocb *req, unsigned int issue_flags) in io_getxattr() argument 115 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_getxattr() 174 int io_fsetxattr(struct io_kiocb *req, unsigned int issue_flags) in io_fsetxattr() argument 179 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_fsetxattr() 186 int io_setxattr(struct io_kiocb *req, unsigned int issue_flags) in io_setxattr() argument 192 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_setxattr()
|
| H A D | xattr.h | 6 int io_fsetxattr(struct io_kiocb *req, unsigned int issue_flags); 9 int io_setxattr(struct io_kiocb *req, unsigned int issue_flags); 12 int io_fgetxattr(struct io_kiocb *req, unsigned int issue_flags); 15 int io_getxattr(struct io_kiocb *req, unsigned int issue_flags);
|
| H A D | sync.c | 40 int io_sync_file_range(struct io_kiocb *req, unsigned int issue_flags) in io_sync_file_range() argument 46 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_sync_file_range() 72 int io_fsync(struct io_kiocb *req, unsigned int issue_flags) in io_fsync() 79 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_fsync() 101 int io_fallocate(struct io_kiocb *req, unsigned int issue_flags) in io_fallocate() 107 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_fallocate() 70 io_fsync(struct io_kiocb * req,unsigned int issue_flags) io_fsync() argument 99 io_fallocate(struct io_kiocb * req,unsigned int issue_flags) io_fallocate() argument
|
| H A D | fs.c | 82 int io_renameat(struct io_kiocb *req, unsigned int issue_flags) in io_renameat() 89 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_renameat() 134 int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags) in io_unlinkat() 140 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_unlinkat() 183 int io_mkdirat(struct io_kiocb *req, unsigned int issue_flags) in io_mkdirat() 189 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_mkdirat() 235 int io_symlinkat(struct io_kiocb *req, unsigned int issue_flags) in io_symlinkat() 242 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_linkat_prep() 283 int io_linkat(struct io_kiocb *req, unsigned int issue_flags) in io_linkat() 290 WARN_ON_ONCE(issue_flags in io_link_cleanup() 81 io_renameat(struct io_kiocb * req,unsigned int issue_flags) io_renameat() argument 130 io_unlinkat(struct io_kiocb * req,unsigned int issue_flags) io_unlinkat() argument 177 io_mkdirat(struct io_kiocb * req,unsigned int issue_flags) io_mkdirat() argument 227 io_symlinkat(struct io_kiocb * req,unsigned int issue_flags) io_symlinkat() argument 272 io_linkat(struct io_kiocb * req,unsigned int issue_flags) io_linkat() argument [all...] |
| H A D | kbuf.c | 104 bool io_kbuf_recycle_legacy(struct io_kiocb *req, unsigned issue_flags) in io_kbuf_recycle_legacy() argument 110 io_ring_submit_lock(ctx, issue_flags); in io_kbuf_recycle_legacy() 118 io_ring_submit_unlock(ctx, issue_flags); in io_kbuf_recycle_legacy() 158 static bool io_should_commit(struct io_kiocb *req, unsigned int issue_flags) in io_should_commit() argument 170 if (issue_flags & IO_URING_F_UNLOCKED) in io_should_commit() 181 unsigned int issue_flags) in io_ring_buffer_select() argument 205 if (io_should_commit(req, issue_flags)) { in io_ring_buffer_select() 213 unsigned buf_group, unsigned int issue_flags) in io_buffer_select() argument 219 io_ring_submit_lock(req->ctx, issue_flags); in io_buffer_select() 224 sel = io_ring_buffer_select(req, len, bl, issue_flags); in io_buffer_select() [all …]
|
| H A D | cancel.c | 106 unsigned issue_flags) in io_try_cancel() argument 121 ret = io_poll_cancel(ctx, cd, issue_flags); in io_try_cancel() 125 ret = io_waitid_cancel(ctx, cd, issue_flags); in io_try_cancel() 129 ret = io_futex_cancel(ctx, cd, issue_flags); in io_try_cancel() 169 unsigned int issue_flags) in __io_async_cancel() argument 177 ret = io_try_cancel(tctx, cd, issue_flags); in __io_async_cancel() 187 io_ring_submit_lock(ctx, issue_flags); in __io_async_cancel() 199 io_ring_submit_unlock(ctx, issue_flags); in __io_async_cancel() 203 int io_async_cancel(struct io_kiocb *req, unsigned int issue_flags) in io_async_cancel() argument 221 issue_flags); in io_async_cancel() [all …]
|
| H A D | sync.h | 4 int io_sync_file_range(struct io_kiocb *req, unsigned int issue_flags); 7 int io_fsync(struct io_kiocb *req, unsigned int issue_flags); 9 int io_fallocate(struct io_kiocb *req, unsigned int issue_flags);
|
| H A D | uring_cmd.h | 11 int io_uring_cmd(struct io_kiocb *req, unsigned int issue_flags); 20 unsigned int issue_flags, 26 unsigned int issue_flags, __poll_t mask);
|
| H A D | advise.c | 51 int io_madvise(struct io_kiocb *req, unsigned int issue_flags) in io_madvise() argument 57 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_madvise() 96 int io_fadvise(struct io_kiocb *req, unsigned int issue_flags) in io_fadvise() argument 101 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK && io_fadvise_force_async(fa)); in io_fadvise()
|
| H A D | cancel.h | 17 int io_async_cancel(struct io_kiocb *req, unsigned int issue_flags); 20 unsigned int issue_flags); 31 unsigned int issue_flags, struct hlist_head *list,
|
| /linux/fs/fuse/ |
| H A D | dev_uring.c | 501 unsigned int issue_flags) in fuse_uring_cancel() argument 523 io_uring_cmd_done(cmd, -ENOTCONN, issue_flags); in fuse_uring_cancel() 527 static void fuse_uring_prepare_cancel(struct io_uring_cmd *cmd, int issue_flags, in fuse_uring_prepare_cancel() argument 531 io_uring_cmd_mark_cancelable(cmd, issue_flags); in fuse_uring_prepare_cancel() 724 unsigned int issue_flags) in fuse_uring_send_next_to_ring() argument 741 io_uring_cmd_done(cmd, 0, issue_flags); in fuse_uring_send_next_to_ring() 816 unsigned int issue_flags) in fuse_uring_commit() argument 845 unsigned int issue_flags) in fuse_uring_next_fuse_req() argument 857 err = fuse_uring_send_next_to_ring(ent, req, issue_flags); in fuse_uring_next_fuse_req() 879 static int fuse_uring_commit_fetch(struct io_uring_cmd *cmd, int issue_flags, in fuse_uring_commit_fetch() argument [all …]
|
| /linux/drivers/nvme/host/ |
| H A D | ioctl.c | 447 struct io_uring_cmd *ioucmd, unsigned int issue_flags, bool vec) in nvme_uring_cmd_io() argument 498 ddir, &iter, issue_flags); in nvme_uring_cmd_io() 501 ddir, &iter, ioucmd, issue_flags); in nvme_uring_cmd_io() 508 if (issue_flags & IO_URING_F_NONBLOCK) { in nvme_uring_cmd_io() 512 if (issue_flags & IO_URING_F_IOPOLL) in nvme_uring_cmd_io() 639 static int nvme_uring_cmd_checks(unsigned int issue_flags) in nvme_uring_cmd_checks() argument 643 if ((issue_flags & (IO_URING_F_SQE128|IO_URING_F_CQE32)) != in nvme_uring_cmd_checks() 650 unsigned int issue_flags) in nvme_ns_uring_cmd() argument 655 ret = nvme_uring_cmd_checks(issue_flags); in nvme_ns_uring_cmd() 661 ret = nvme_uring_cmd_io(ctrl, ns, ioucmd, issue_flags, false); in nvme_ns_uring_cmd() [all …]
|
| /linux/drivers/block/ |
| H A D | ublk_drv.c | 139 unsigned int issue_flags; member 747 io_uring_cmd_done(fcmd->cmd, res, data->issue_flags); in ublk_batch_deinit_fetch_buf() 753 unsigned int issue_flags) in ublk_batch_fetch_post_cqe() argument 755 if (io_uring_mshot_cmd_post_cqe(fcmd->cmd, sel, issue_flags)) in ublk_batch_fetch_post_cqe() 1568 int res, unsigned issue_flags) in ublk_complete_io_cmd() argument 1573 io_uring_cmd_done(cmd, res, issue_flags); in ublk_complete_io_cmd() 1626 unsigned int issue_flags) in ublk_auto_buf_register() argument 1631 io->buf.auto_reg.index, issue_flags); in ublk_auto_buf_register() 1652 unsigned int issue_flags) in ublk_auto_buf_dispatch() argument 1655 issue_flags); in ublk_auto_buf_dispatch() [all …]
|