Home
last modified time | relevance | path

Searched refs:issue_flags (Results 1 – 25 of 47) sorted by relevance

12

/linux/include/linux/io_uring/
H A Dcmd.h48 unsigned int issue_flags);
53 unsigned issue_flags);
63 unsigned issue_flags, bool is_cqe32);
74 unsigned int issue_flags);
85 unsigned int issue_flags);
92 struct io_br_sel *sel, unsigned int issue_flags);
98 unsigned int issue_flags) in io_uring_cmd_import_fixed() argument
106 unsigned issue_flags) in io_uring_cmd_import_fixed_vec() argument
111 u64 ret2, unsigned issue_flags, bool is_cqe32) in __io_uring_cmd_done() argument
119 unsigned int issue_flags) in io_uring_cmd_mark_cancelable() argument
[all …]
/linux/io_uring/
H A Dnet.h29 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags);
33 int io_sendmsg(struct io_kiocb *req, unsigned int issue_flags);
35 int io_send(struct io_kiocb *req, unsigned int issue_flags);
38 int io_recvmsg(struct io_kiocb *req, unsigned int issue_flags);
39 int io_recv(struct io_kiocb *req, unsigned int issue_flags);
44 int io_accept(struct io_kiocb *req, unsigned int issue_flags);
47 int io_socket(struct io_kiocb *req, unsigned int issue_flags);
51 int io_connect(struct io_kiocb *req, unsigned int issue_flags);
53 int io_send_zc(struct io_kiocb *req, unsigned int issue_flags);
54 int io_sendmsg_zc(struct io_kiocb *req, unsigned int issue_flags);
[all...]
H A Dcmd_net.c12 unsigned int issue_flags) in io_uring_cmd_getsockopt()
15 bool compat = !!(issue_flags & IO_URING_F_COMPAT); in io_uring_cmd_getsockopt()
39 unsigned int issue_flags) in io_uring_cmd_setsockopt()
42 bool compat = !!(issue_flags & IO_URING_F_COMPAT); in io_uring_cmd_setsockopt()
58 struct sk_buff *skb, unsigned issue_flags) in io_process_timestamp_skb()
86 return io_uring_cmd_post_mshot_cqe32(cmd, issue_flags, cqe); in io_process_timestamp_skb()
91 unsigned int issue_flags) in io_uring_cmd_timestamp()
99 if (!(issue_flags & IO_URING_F_CQE32)) in io_uring_cmd_timestamp()
101 ret = io_cmd_poll_multishot(cmd, issue_flags, EPOLLERR); in io_uring_cmd_timestamp()
123 if (!io_process_timestamp_skb(cmd, sk, skb, issue_flags)) in io_uring_cmd_timestamp()
11 io_uring_cmd_getsockopt(struct socket * sock,struct io_uring_cmd * cmd,unsigned int issue_flags) io_uring_cmd_getsockopt() argument
38 io_uring_cmd_setsockopt(struct socket * sock,struct io_uring_cmd * cmd,unsigned int issue_flags) io_uring_cmd_setsockopt() argument
57 io_process_timestamp_skb(struct io_uring_cmd * cmd,struct sock * sk,struct sk_buff * skb,unsigned issue_flags) io_process_timestamp_skb() argument
90 io_uring_cmd_timestamp(struct socket * sock,struct io_uring_cmd * cmd,unsigned int issue_flags) io_uring_cmd_timestamp() argument
135 io_uring_cmd_sock(struct io_uring_cmd * cmd,unsigned int issue_flags) io_uring_cmd_sock() argument
[all...]
H A Dopenclose.c123 int io_openat2(struct io_kiocb *req, unsigned int issue_flags) in io_openat2() argument
138 if (issue_flags & IO_URING_F_NONBLOCK) { in io_openat2()
163 (issue_flags & IO_URING_F_NONBLOCK)) { in io_openat2()
172 if ((issue_flags & IO_URING_F_NONBLOCK) && !nonblock_set) in io_openat2()
178 ret = io_fixed_fd_install(req, issue_flags, file, in io_openat2()
188 int io_openat(struct io_kiocb *req, unsigned int issue_flags) in io_openat() argument
190 return io_openat2(req, issue_flags); in io_openat()
200 int __io_close_fixed(struct io_ring_ctx *ctx, unsigned int issue_flags, in __io_close_fixed() argument
205 io_ring_submit_lock(ctx, issue_flags); in __io_close_fixed()
207 io_ring_submit_unlock(ctx, issue_flags); in __io_close_fixed()
[all …]
H A Dopenclose.h5 int __io_close_fixed(struct io_ring_ctx *ctx, unsigned int issue_flags,
9 int io_openat(struct io_kiocb *req, unsigned int issue_flags);
14 int io_openat2(struct io_kiocb *req, unsigned int issue_flags);
17 int io_close(struct io_kiocb *req, unsigned int issue_flags);
20 int io_pipe(struct io_kiocb *req, unsigned int issue_flags);
23 int io_install_fixed_fd(struct io_kiocb *req, unsigned int issue_flags);
H A Dnet.c136 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags) in io_shutdown() argument
142 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_shutdown()
166 static void io_netmsg_recycle(struct io_kiocb *req, unsigned int issue_flags) in io_netmsg_recycle() argument
171 if (unlikely(issue_flags & IO_URING_F_UNLOCKED)) { in io_netmsg_recycle()
450 unsigned int issue_flags) in io_req_msg_cleanup() argument
452 io_netmsg_recycle(req, issue_flags); in io_req_msg_cleanup()
541 int io_sendmsg(struct io_kiocb *req, unsigned int issue_flags) in io_sendmsg()
559 if (issue_flags & IO_URING_F_NONBLOCK) in io_sendmsg()
569 if (ret == -EAGAIN && (issue_flags & IO_URING_F_NONBLOCK)) in io_sendmsg()
581 io_req_msg_cleanup(req, issue_flags); in io_sendmsg()
537 io_sendmsg(struct io_kiocb * req,unsigned int issue_flags) io_sendmsg() argument
586 io_send_select_buffer(struct io_kiocb * req,unsigned int issue_flags,struct io_br_sel * sel,struct io_async_msghdr * kmsg) io_send_select_buffer() argument
634 io_send(struct io_kiocb * req,unsigned int issue_flags) io_send() argument
841 io_recv_finish(struct io_kiocb * req,struct io_async_msghdr * kmsg,struct io_br_sel * sel,bool mshot_finished,unsigned issue_flags) io_recv_finish() argument
1005 io_recvmsg(struct io_kiocb * req,unsigned int issue_flags) io_recvmsg() argument
1093 io_recv_buf_select(struct io_kiocb * req,struct io_async_msghdr * kmsg,struct io_br_sel * sel,unsigned int issue_flags) io_recv_buf_select() argument
1164 io_recv(struct io_kiocb * req,unsigned int issue_flags) io_recv() argument
1268 io_recvzc(struct io_kiocb * req,unsigned int issue_flags) io_recvzc() argument
1438 io_send_zc_import(struct io_kiocb * req,unsigned int issue_flags) io_send_zc_import() argument
1451 io_send_zc(struct io_kiocb * req,unsigned int issue_flags) io_send_zc() argument
1520 io_sendmsg_zc(struct io_kiocb * req,unsigned int issue_flags) io_sendmsg_zc() argument
1641 io_accept(struct io_kiocb * req,unsigned int issue_flags) io_accept() argument
1723 io_socket(struct io_kiocb * req,unsigned int issue_flags) io_socket() argument
1775 io_connect(struct io_kiocb * req,unsigned int issue_flags) io_connect() argument
1845 io_bind(struct io_kiocb * req,unsigned int issue_flags) io_bind() argument
1874 io_listen(struct io_kiocb * req,unsigned int issue_flags) io_listen() argument
[all...]
H A Dfs.h4 int io_renameat(struct io_kiocb *req, unsigned int issue_flags);
8 int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags);
12 int io_mkdirat(struct io_kiocb *req, unsigned int issue_flags);
16 int io_symlinkat(struct io_kiocb *req, unsigned int issue_flags);
19 int io_linkat(struct io_kiocb *req, unsigned int issue_flags);
H A Drw.c112 unsigned int issue_flags) in __io_import_rw_buffer() argument
123 *sel = io_buffer_select(req, &sqe_len, io->buf_group, issue_flags); in __io_import_rw_buffer()
135 unsigned int issue_flags) in io_import_rw_buffer() argument
139 ret = __io_import_rw_buffer(rw, req, io, sel, issue_flags); in io_import_rw_buffer()
147 static bool io_rw_recycle(struct io_kiocb *req, unsigned int issue_flags) in io_rw_recycle() argument
151 if (unlikely(issue_flags & IO_URING_F_UNLOCKED)) in io_rw_recycle()
165 static void io_req_rw_cleanup(struct io_kiocb *req, unsigned int issue_flags) in io_req_rw_cleanup() argument
196 if (!io_rw_recycle(req, issue_flags)) { in io_req_rw_cleanup()
373 static int io_init_rw_fixed(struct io_kiocb *req, unsigned int issue_flags, in io_init_rw_fixed() argument
384 issue_flags); in io_init_rw_fixed()
[all …]
H A Dpoll.h27 * Must only be called inside issue_flags & IO_URING_F_MULTISHOT, or
36 int io_poll_add(struct io_kiocb *req, unsigned int issue_flags);
39 int io_poll_remove(struct io_kiocb *req, unsigned int issue_flags);
43 unsigned issue_flags);
44 int io_arm_apoll(struct io_kiocb *req, unsigned issue_flags, __poll_t mask);
45 int io_arm_poll_handler(struct io_kiocb *req, unsigned issue_flags);
H A Dpoll.c528 static void io_poll_add_hash(struct io_kiocb *req, unsigned int issue_flags) in io_poll_add_hash() argument
532 io_ring_submit_lock(ctx, issue_flags); in io_poll_add_hash()
534 io_ring_submit_unlock(ctx, issue_flags); in io_poll_add_hash()
546 unsigned issue_flags) in __io_arm_poll_handler() argument
568 ipt->owning = issue_flags & IO_URING_F_UNLOCKED; in __io_arm_poll_handler()
599 io_poll_add_hash(req, issue_flags); in __io_arm_poll_handler()
608 io_poll_add_hash(req, issue_flags); in __io_arm_poll_handler()
646 unsigned issue_flags) in io_req_alloc_apoll() argument
655 if (!(issue_flags & IO_URING_F_UNLOCKED)) in io_req_alloc_apoll()
670 int io_arm_apoll(struct io_kiocb *req, unsigned issue_flags, __poll_t mask) in io_arm_apoll() argument
[all …]
H A Drw.h43 int io_read(struct io_kiocb *req, unsigned int issue_flags);
44 int io_write(struct io_kiocb *req, unsigned int issue_flags);
45 int io_read_fixed(struct io_kiocb *req, unsigned int issue_flags);
46 int io_write_fixed(struct io_kiocb *req, unsigned int issue_flags);
51 int io_read_mshot(struct io_kiocb *req, unsigned int issue_flags);
H A Dfutex.c116 unsigned int issue_flags) in io_futex_cancel() argument
118 return io_cancel_remove(ctx, cd, issue_flags, &ctx->futex_list, __io_futex_cancel); in io_futex_cancel()
221 int io_futexv_wait(struct io_kiocb *req, unsigned int issue_flags) in io_futexv_wait() argument
228 io_ring_submit_lock(ctx, issue_flags); in io_futexv_wait()
236 io_ring_submit_unlock(ctx, issue_flags); in io_futexv_wait()
270 io_ring_submit_unlock(ctx, issue_flags); in io_futexv_wait()
274 int io_futex_wait(struct io_kiocb *req, unsigned int issue_flags) in io_futex_wait() argument
286 io_ring_submit_lock(ctx, issue_flags); in io_futex_wait()
304 io_ring_submit_unlock(ctx, issue_flags); in io_futex_wait()
310 io_ring_submit_unlock(ctx, issue_flags); in io_futex_wait()
[all …]
H A Dxattr.c97 int io_fgetxattr(struct io_kiocb *req, unsigned int issue_flags) in io_fgetxattr() argument
102 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_fgetxattr()
109 int io_getxattr(struct io_kiocb *req, unsigned int issue_flags) in io_getxattr() argument
115 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_getxattr()
174 int io_fsetxattr(struct io_kiocb *req, unsigned int issue_flags) in io_fsetxattr() argument
179 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_fsetxattr()
186 int io_setxattr(struct io_kiocb *req, unsigned int issue_flags) in io_setxattr() argument
192 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_setxattr()
H A Dxattr.h6 int io_fsetxattr(struct io_kiocb *req, unsigned int issue_flags);
9 int io_setxattr(struct io_kiocb *req, unsigned int issue_flags);
12 int io_fgetxattr(struct io_kiocb *req, unsigned int issue_flags);
15 int io_getxattr(struct io_kiocb *req, unsigned int issue_flags);
H A Dsync.c40 int io_sync_file_range(struct io_kiocb *req, unsigned int issue_flags) in io_sync_file_range() argument
46 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_sync_file_range()
72 int io_fsync(struct io_kiocb *req, unsigned int issue_flags) in io_fsync()
79 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_fsync()
101 int io_fallocate(struct io_kiocb *req, unsigned int issue_flags) in io_fallocate()
107 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_fallocate()
70 io_fsync(struct io_kiocb * req,unsigned int issue_flags) io_fsync() argument
99 io_fallocate(struct io_kiocb * req,unsigned int issue_flags) io_fallocate() argument
H A Dfs.c82 int io_renameat(struct io_kiocb *req, unsigned int issue_flags) in io_renameat()
89 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_renameat()
134 int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags) in io_unlinkat()
140 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_unlinkat()
183 int io_mkdirat(struct io_kiocb *req, unsigned int issue_flags) in io_mkdirat()
189 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_mkdirat()
235 int io_symlinkat(struct io_kiocb *req, unsigned int issue_flags) in io_symlinkat()
242 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_linkat_prep()
283 int io_linkat(struct io_kiocb *req, unsigned int issue_flags) in io_linkat()
290 WARN_ON_ONCE(issue_flags in io_link_cleanup()
81 io_renameat(struct io_kiocb * req,unsigned int issue_flags) io_renameat() argument
130 io_unlinkat(struct io_kiocb * req,unsigned int issue_flags) io_unlinkat() argument
177 io_mkdirat(struct io_kiocb * req,unsigned int issue_flags) io_mkdirat() argument
227 io_symlinkat(struct io_kiocb * req,unsigned int issue_flags) io_symlinkat() argument
272 io_linkat(struct io_kiocb * req,unsigned int issue_flags) io_linkat() argument
[all...]
H A Dkbuf.c104 bool io_kbuf_recycle_legacy(struct io_kiocb *req, unsigned issue_flags) in io_kbuf_recycle_legacy() argument
110 io_ring_submit_lock(ctx, issue_flags); in io_kbuf_recycle_legacy()
118 io_ring_submit_unlock(ctx, issue_flags); in io_kbuf_recycle_legacy()
158 static bool io_should_commit(struct io_kiocb *req, unsigned int issue_flags) in io_should_commit() argument
170 if (issue_flags & IO_URING_F_UNLOCKED) in io_should_commit()
181 unsigned int issue_flags) in io_ring_buffer_select() argument
205 if (io_should_commit(req, issue_flags)) { in io_ring_buffer_select()
213 unsigned buf_group, unsigned int issue_flags) in io_buffer_select() argument
219 io_ring_submit_lock(req->ctx, issue_flags); in io_buffer_select()
224 sel = io_ring_buffer_select(req, len, bl, issue_flags); in io_buffer_select()
[all …]
H A Dcancel.c106 unsigned issue_flags) in io_try_cancel() argument
121 ret = io_poll_cancel(ctx, cd, issue_flags); in io_try_cancel()
125 ret = io_waitid_cancel(ctx, cd, issue_flags); in io_try_cancel()
129 ret = io_futex_cancel(ctx, cd, issue_flags); in io_try_cancel()
169 unsigned int issue_flags) in __io_async_cancel() argument
177 ret = io_try_cancel(tctx, cd, issue_flags); in __io_async_cancel()
187 io_ring_submit_lock(ctx, issue_flags); in __io_async_cancel()
199 io_ring_submit_unlock(ctx, issue_flags); in __io_async_cancel()
203 int io_async_cancel(struct io_kiocb *req, unsigned int issue_flags) in io_async_cancel() argument
221 issue_flags); in io_async_cancel()
[all …]
H A Dsync.h4 int io_sync_file_range(struct io_kiocb *req, unsigned int issue_flags);
7 int io_fsync(struct io_kiocb *req, unsigned int issue_flags);
9 int io_fallocate(struct io_kiocb *req, unsigned int issue_flags);
H A During_cmd.h11 int io_uring_cmd(struct io_kiocb *req, unsigned int issue_flags);
20 unsigned int issue_flags,
26 unsigned int issue_flags, __poll_t mask);
H A Dadvise.c51 int io_madvise(struct io_kiocb *req, unsigned int issue_flags) in io_madvise() argument
57 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_madvise()
96 int io_fadvise(struct io_kiocb *req, unsigned int issue_flags) in io_fadvise() argument
101 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK && io_fadvise_force_async(fa)); in io_fadvise()
H A Dcancel.h17 int io_async_cancel(struct io_kiocb *req, unsigned int issue_flags);
20 unsigned int issue_flags);
31 unsigned int issue_flags, struct hlist_head *list,
/linux/fs/fuse/
H A Ddev_uring.c501 unsigned int issue_flags) in fuse_uring_cancel() argument
523 io_uring_cmd_done(cmd, -ENOTCONN, issue_flags); in fuse_uring_cancel()
527 static void fuse_uring_prepare_cancel(struct io_uring_cmd *cmd, int issue_flags, in fuse_uring_prepare_cancel() argument
531 io_uring_cmd_mark_cancelable(cmd, issue_flags); in fuse_uring_prepare_cancel()
724 unsigned int issue_flags) in fuse_uring_send_next_to_ring() argument
741 io_uring_cmd_done(cmd, 0, issue_flags); in fuse_uring_send_next_to_ring()
816 unsigned int issue_flags) in fuse_uring_commit() argument
845 unsigned int issue_flags) in fuse_uring_next_fuse_req() argument
857 err = fuse_uring_send_next_to_ring(ent, req, issue_flags); in fuse_uring_next_fuse_req()
879 static int fuse_uring_commit_fetch(struct io_uring_cmd *cmd, int issue_flags, in fuse_uring_commit_fetch() argument
[all …]
/linux/drivers/nvme/host/
H A Dioctl.c447 struct io_uring_cmd *ioucmd, unsigned int issue_flags, bool vec) in nvme_uring_cmd_io() argument
498 ddir, &iter, issue_flags); in nvme_uring_cmd_io()
501 ddir, &iter, ioucmd, issue_flags); in nvme_uring_cmd_io()
508 if (issue_flags & IO_URING_F_NONBLOCK) { in nvme_uring_cmd_io()
512 if (issue_flags & IO_URING_F_IOPOLL) in nvme_uring_cmd_io()
639 static int nvme_uring_cmd_checks(unsigned int issue_flags) in nvme_uring_cmd_checks() argument
643 if ((issue_flags & (IO_URING_F_SQE128|IO_URING_F_CQE32)) != in nvme_uring_cmd_checks()
650 unsigned int issue_flags) in nvme_ns_uring_cmd() argument
655 ret = nvme_uring_cmd_checks(issue_flags); in nvme_ns_uring_cmd()
661 ret = nvme_uring_cmd_io(ctrl, ns, ioucmd, issue_flags, false); in nvme_ns_uring_cmd()
[all …]
/linux/drivers/block/
H A Dublk_drv.c139 unsigned int issue_flags; member
747 io_uring_cmd_done(fcmd->cmd, res, data->issue_flags); in ublk_batch_deinit_fetch_buf()
753 unsigned int issue_flags) in ublk_batch_fetch_post_cqe() argument
755 if (io_uring_mshot_cmd_post_cqe(fcmd->cmd, sel, issue_flags)) in ublk_batch_fetch_post_cqe()
1568 int res, unsigned issue_flags) in ublk_complete_io_cmd() argument
1573 io_uring_cmd_done(cmd, res, issue_flags); in ublk_complete_io_cmd()
1626 unsigned int issue_flags) in ublk_auto_buf_register() argument
1631 io->buf.auto_reg.index, issue_flags); in ublk_auto_buf_register()
1652 unsigned int issue_flags) in ublk_auto_buf_dispatch() argument
1655 issue_flags); in ublk_auto_buf_dispatch()
[all …]

12