Home
last modified time | relevance | path

Searched refs:issue_flags (Results 1 – 25 of 41) sorted by relevance

12

/linux/io_uring/
H A Dnet.h28 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags);
32 int io_sendmsg(struct io_kiocb *req, unsigned int issue_flags);
34 int io_send(struct io_kiocb *req, unsigned int issue_flags);
37 int io_recvmsg(struct io_kiocb *req, unsigned int issue_flags);
38 int io_recv(struct io_kiocb *req, unsigned int issue_flags);
43 int io_accept(struct io_kiocb *req, unsigned int issue_flags);
46 int io_socket(struct io_kiocb *req, unsigned int issue_flags);
49 int io_connect(struct io_kiocb *req, unsigned int issue_flags);
51 int io_send_zc(struct io_kiocb *req, unsigned int issue_flags);
52 int io_sendmsg_zc(struct io_kiocb *req, unsigned int issue_flags);
[all...]
H A Dcmd_net.c11 unsigned int issue_flags) in io_uring_cmd_getsockopt() argument
14 bool compat = !!(issue_flags & IO_URING_F_COMPAT); in io_uring_cmd_getsockopt()
38 unsigned int issue_flags) in io_uring_cmd_setsockopt() argument
41 bool compat = !!(issue_flags & IO_URING_F_COMPAT); in io_uring_cmd_setsockopt()
57 struct sk_buff *skb, unsigned issue_flags) in io_process_timestamp_skb() argument
85 return io_uring_cmd_post_mshot_cqe32(cmd, issue_flags, cqe); in io_process_timestamp_skb()
90 unsigned int issue_flags) in io_uring_cmd_timestamp() argument
98 if (!(issue_flags & IO_URING_F_CQE32)) in io_uring_cmd_timestamp()
100 ret = io_cmd_poll_multishot(cmd, issue_flags, EPOLLERR); in io_uring_cmd_timestamp()
122 if (!io_process_timestamp_skb(cmd, sk, skb, issue_flags)) in io_uring_cmd_timestamp()
135 io_uring_cmd_sock(struct io_uring_cmd * cmd,unsigned int issue_flags) io_uring_cmd_sock() argument
[all...]
H A Dopenclose.c117 int io_openat2(struct io_kiocb *req, unsigned int issue_flags) in io_openat2() argument
131 if (issue_flags & IO_URING_F_NONBLOCK) { in io_openat2()
156 (!resolve_nonblock && (issue_flags & IO_URING_F_NONBLOCK))) in io_openat2()
161 if ((issue_flags & IO_URING_F_NONBLOCK) && !nonblock_set) in io_openat2()
167 ret = io_fixed_fd_install(req, issue_flags, file, in io_openat2()
178 int io_openat(struct io_kiocb *req, unsigned int issue_flags) in io_openat() argument
180 return io_openat2(req, issue_flags); in io_openat()
191 int __io_close_fixed(struct io_ring_ctx *ctx, unsigned int issue_flags, in __io_close_fixed() argument
196 io_ring_submit_lock(ctx, issue_flags); in __io_close_fixed()
198 io_ring_submit_unlock(ctx, issue_flags); in __io_close_fixed()
[all …]
H A Dopenclose.h3 int __io_close_fixed(struct io_ring_ctx *ctx, unsigned int issue_flags,
7 int io_openat(struct io_kiocb *req, unsigned int issue_flags);
11 int io_openat2(struct io_kiocb *req, unsigned int issue_flags);
14 int io_close(struct io_kiocb *req, unsigned int issue_flags);
17 int io_pipe(struct io_kiocb *req, unsigned int issue_flags);
20 int io_install_fixed_fd(struct io_kiocb *req, unsigned int issue_flags);
H A Dnet.c136 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags) in io_shutdown() argument
142 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_shutdown()
166 static void io_netmsg_recycle(struct io_kiocb *req, unsigned int issue_flags) in io_netmsg_recycle() argument
171 if (unlikely(issue_flags & IO_URING_F_UNLOCKED)) { in io_netmsg_recycle()
450 unsigned int issue_flags) in io_req_msg_cleanup() argument
452 io_netmsg_recycle(req, issue_flags); in io_req_msg_cleanup()
537 int io_sendmsg(struct io_kiocb *req, unsigned int issue_flags) in io_sendmsg() argument
555 if (issue_flags & IO_URING_F_NONBLOCK) in io_sendmsg()
565 if (ret == -EAGAIN && (issue_flags & IO_URING_F_NONBLOCK)) in io_sendmsg()
577 io_req_msg_cleanup(req, issue_flags); in io_sendmsg()
[all …]
H A Dfs.h4 int io_renameat(struct io_kiocb *req, unsigned int issue_flags);
8 int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags);
12 int io_mkdirat(struct io_kiocb *req, unsigned int issue_flags);
16 int io_symlinkat(struct io_kiocb *req, unsigned int issue_flags);
19 int io_linkat(struct io_kiocb *req, unsigned int issue_flags);
H A Dpoll.h27 * Must only be called inside issue_flags & IO_URING_F_MULTISHOT, or
36 int io_poll_add(struct io_kiocb *req, unsigned int issue_flags);
39 int io_poll_remove(struct io_kiocb *req, unsigned int issue_flags);
43 unsigned issue_flags);
44 int io_arm_apoll(struct io_kiocb *req, unsigned issue_flags, __poll_t mask);
45 int io_arm_poll_handler(struct io_kiocb *req, unsigned issue_flags);
H A Drw.c112 unsigned int issue_flags) in __io_import_rw_buffer() argument
123 *sel = io_buffer_select(req, &sqe_len, io->buf_group, issue_flags); in __io_import_rw_buffer()
135 unsigned int issue_flags) in io_import_rw_buffer() argument
139 ret = __io_import_rw_buffer(rw, req, io, sel, issue_flags); in io_import_rw_buffer()
147 static void io_rw_recycle(struct io_kiocb *req, unsigned int issue_flags) in io_rw_recycle() argument
151 if (unlikely(issue_flags & IO_URING_F_UNLOCKED)) in io_rw_recycle()
162 static void io_req_rw_cleanup(struct io_kiocb *req, unsigned int issue_flags) in io_req_rw_cleanup() argument
193 io_rw_recycle(req, issue_flags); in io_req_rw_cleanup()
366 static int io_init_rw_fixed(struct io_kiocb *req, unsigned int issue_flags, in io_init_rw_fixed() argument
377 issue_flags); in io_init_rw_fixed()
[all …]
H A Dpoll.c528 static void io_poll_add_hash(struct io_kiocb *req, unsigned int issue_flags) in io_poll_add_hash() argument
532 io_ring_submit_lock(ctx, issue_flags); in io_poll_add_hash()
534 io_ring_submit_unlock(ctx, issue_flags); in io_poll_add_hash()
546 unsigned issue_flags) in __io_arm_poll_handler() argument
568 ipt->owning = issue_flags & IO_URING_F_UNLOCKED; in __io_arm_poll_handler()
599 io_poll_add_hash(req, issue_flags); in __io_arm_poll_handler()
608 io_poll_add_hash(req, issue_flags); in __io_arm_poll_handler()
646 unsigned issue_flags) in io_req_alloc_apoll() argument
655 if (!(issue_flags & IO_URING_F_UNLOCKED)) in io_req_alloc_apoll()
670 int io_arm_apoll(struct io_kiocb *req, unsigned issue_flags, __poll_t mask) in io_arm_apoll() argument
[all …]
H A Drw.h43 int io_read(struct io_kiocb *req, unsigned int issue_flags);
44 int io_write(struct io_kiocb *req, unsigned int issue_flags);
45 int io_read_fixed(struct io_kiocb *req, unsigned int issue_flags);
46 int io_write_fixed(struct io_kiocb *req, unsigned int issue_flags);
51 int io_read_mshot(struct io_kiocb *req, unsigned int issue_flags);
H A Dxattr.h6 int io_fsetxattr(struct io_kiocb *req, unsigned int issue_flags);
9 int io_setxattr(struct io_kiocb *req, unsigned int issue_flags);
12 int io_fgetxattr(struct io_kiocb *req, unsigned int issue_flags);
15 int io_getxattr(struct io_kiocb *req, unsigned int issue_flags);
H A Dsync.c40 int io_sync_file_range(struct io_kiocb *req, unsigned int issue_flags) in io_sync_file_range() argument
46 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_sync_file_range()
70 int io_fsync(struct io_kiocb *req, unsigned int issue_flags) in io_fsync() argument
77 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_fsync()
99 int io_fallocate(struct io_kiocb *req, unsigned int issue_flags) in io_fallocate() argument
105 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_fallocate()
H A Dfs.c81 int io_renameat(struct io_kiocb *req, unsigned int issue_flags) in io_renameat() argument
86 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_renameat()
130 int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags) in io_unlinkat() argument
135 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_unlinkat()
177 int io_mkdirat(struct io_kiocb *req, unsigned int issue_flags) in io_mkdirat() argument
182 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_mkdirat()
227 int io_symlinkat(struct io_kiocb *req, unsigned int issue_flags) in io_symlinkat() argument
232 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_symlinkat()
272 int io_linkat(struct io_kiocb *req, unsigned int issue_flags) in io_linkat() argument
277 WARN_ON_ONCE(issue_flags in io_linkat()
[all...]
H A Dkbuf.c104 bool io_kbuf_recycle_legacy(struct io_kiocb *req, unsigned issue_flags) in io_kbuf_recycle_legacy() argument
110 io_ring_submit_lock(ctx, issue_flags); in io_kbuf_recycle_legacy()
118 io_ring_submit_unlock(ctx, issue_flags); in io_kbuf_recycle_legacy()
158 static bool io_should_commit(struct io_kiocb *req, unsigned int issue_flags) in io_should_commit() argument
170 if (issue_flags & IO_URING_F_UNLOCKED) in io_should_commit()
181 unsigned int issue_flags) in io_ring_buffer_select() argument
205 if (io_should_commit(req, issue_flags)) { in io_ring_buffer_select()
213 unsigned buf_group, unsigned int issue_flags) in io_buffer_select() argument
219 io_ring_submit_lock(req->ctx, issue_flags); in io_buffer_select()
224 sel = io_ring_buffer_select(req, len, bl, issue_flags); in io_buffer_select()
[all …]
H A Dcancel.c107 unsigned issue_flags) in io_try_cancel() argument
122 ret = io_poll_cancel(ctx, cd, issue_flags); in io_try_cancel()
126 ret = io_waitid_cancel(ctx, cd, issue_flags); in io_try_cancel()
130 ret = io_futex_cancel(ctx, cd, issue_flags); in io_try_cancel()
170 unsigned int issue_flags) in __io_async_cancel() argument
178 ret = io_try_cancel(tctx, cd, issue_flags); in __io_async_cancel()
188 io_ring_submit_lock(ctx, issue_flags); in __io_async_cancel()
200 io_ring_submit_unlock(ctx, issue_flags); in __io_async_cancel()
204 int io_async_cancel(struct io_kiocb *req, unsigned int issue_flags) in io_async_cancel() argument
222 issue_flags); in io_async_cancel()
[all …]
H A Dsync.h4 int io_sync_file_range(struct io_kiocb *req, unsigned int issue_flags);
7 int io_fsync(struct io_kiocb *req, unsigned int issue_flags);
9 int io_fallocate(struct io_kiocb *req, unsigned int issue_flags);
H A During_cmd.h11 int io_uring_cmd(struct io_kiocb *req, unsigned int issue_flags);
20 unsigned int issue_flags,
26 unsigned int issue_flags, __poll_t mask);
H A Dadvise.c51 int io_madvise(struct io_kiocb *req, unsigned int issue_flags) in io_madvise() argument
57 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK); in io_madvise()
96 int io_fadvise(struct io_kiocb *req, unsigned int issue_flags) in io_fadvise() argument
101 WARN_ON_ONCE(issue_flags & IO_URING_F_NONBLOCK && io_fadvise_force_async(fa)); in io_fadvise()
H A Dcancel.h19 int io_async_cancel(struct io_kiocb *req, unsigned int issue_flags);
22 unsigned int issue_flags);
33 unsigned int issue_flags, struct hlist_head *list,
H A Dwaitid.c168 unsigned int issue_flags) in io_waitid_cancel() argument
170 return io_cancel_remove(ctx, cd, issue_flags, &ctx->waitid_list, __io_waitid_cancel); in io_waitid_cancel()
281 int io_waitid(struct io_kiocb *req, unsigned int issue_flags) in io_waitid() argument
306 io_ring_submit_lock(ctx, issue_flags); in io_waitid()
326 io_ring_submit_unlock(ctx, issue_flags); in io_waitid()
334 io_ring_submit_unlock(ctx, issue_flags); in io_waitid()
342 io_ring_submit_unlock(ctx, issue_flags); in io_waitid()
H A Drsrc.h63 unsigned issue_flags);
66 unsigned issue_flags);
69 unsigned nr_iovs, unsigned issue_flags);
119 int io_files_update(struct io_kiocb *req, unsigned int issue_flags);
H A Dzcrx.h76 unsigned issue_flags, unsigned int *len);
90 unsigned issue_flags, unsigned int *len) in io_zcrx_recv() argument
106 int io_recvzc(struct io_kiocb *req, unsigned int issue_flags);
H A Drsrc.c447 unsigned int issue_flags) in io_files_update_with_index_alloc() argument
469 ret = io_fixed_fd_install(req, issue_flags, file, in io_files_update_with_index_alloc()
474 __io_close_fixed(req->ctx, issue_flags, ret); in io_files_update_with_index_alloc()
485 int io_files_update(struct io_kiocb *req, unsigned int issue_flags) in io_files_update() argument
500 ret = io_files_update_with_index_alloc(req, issue_flags); in io_files_update()
502 io_ring_submit_lock(ctx, issue_flags); in io_files_update()
505 io_ring_submit_unlock(ctx, issue_flags); in io_files_update()
941 unsigned int issue_flags) in io_buffer_register_bvec() argument
952 io_ring_submit_lock(ctx, issue_flags); in io_buffer_register_bvec()
998 io_ring_submit_unlock(ctx, issue_flags); in io_buffer_register_bvec()
1004 io_buffer_unregister_bvec(struct io_uring_cmd * cmd,unsigned int index,unsigned int issue_flags) io_buffer_unregister_bvec() argument
1120 io_find_buf_node(struct io_kiocb * req,unsigned issue_flags) io_find_buf_node() argument
1144 io_import_reg_buf(struct io_kiocb * req,struct iov_iter * iter,u64 buf_addr,size_t len,int ddir,unsigned issue_flags) io_import_reg_buf() argument
1491 io_import_reg_vec(int ddir,struct iov_iter * iter,struct io_kiocb * req,struct iou_vec * vec,unsigned nr_iovs,unsigned issue_flags) io_import_reg_vec() argument
[all...]
/linux/fs/fuse/
H A Ddev_uring.c501 unsigned int issue_flags) in fuse_uring_cancel() argument
523 io_uring_cmd_done(cmd, -ENOTCONN, issue_flags); in fuse_uring_cancel()
527 static void fuse_uring_prepare_cancel(struct io_uring_cmd *cmd, int issue_flags, in fuse_uring_prepare_cancel() argument
531 io_uring_cmd_mark_cancelable(cmd, issue_flags); in fuse_uring_prepare_cancel()
724 unsigned int issue_flags) in fuse_uring_send_next_to_ring() argument
741 io_uring_cmd_done(cmd, 0, issue_flags); in fuse_uring_send_next_to_ring()
816 unsigned int issue_flags) in fuse_uring_commit() argument
845 unsigned int issue_flags) in fuse_uring_next_fuse_req() argument
857 err = fuse_uring_send_next_to_ring(ent, req, issue_flags); in fuse_uring_next_fuse_req()
879 static int fuse_uring_commit_fetch(struct io_uring_cmd *cmd, int issue_flags, in fuse_uring_commit_fetch() argument
[all …]
/linux/drivers/nvme/host/
H A Dioctl.c440 struct io_uring_cmd *ioucmd, unsigned int issue_flags, bool vec) in nvme_uring_cmd_io() argument
490 ddir, &iter, issue_flags); in nvme_uring_cmd_io()
493 ddir, &iter, ioucmd, issue_flags); in nvme_uring_cmd_io()
500 if (issue_flags & IO_URING_F_NONBLOCK) { in nvme_uring_cmd_io()
504 if (issue_flags & IO_URING_F_IOPOLL) in nvme_uring_cmd_io()
631 static int nvme_uring_cmd_checks(unsigned int issue_flags) in nvme_uring_cmd_checks() argument
635 if ((issue_flags & (IO_URING_F_SQE128|IO_URING_F_CQE32)) != in nvme_uring_cmd_checks()
642 unsigned int issue_flags) in nvme_ns_uring_cmd() argument
647 ret = nvme_uring_cmd_checks(issue_flags); in nvme_ns_uring_cmd()
653 ret = nvme_uring_cmd_io(ctrl, ns, ioucmd, issue_flags, false); in nvme_ns_uring_cmd()
[all …]

12