Home
last modified time | relevance | path

Searched refs:io_kiocb (Results 1 – 25 of 33) sorted by relevance

12

/linux/io_uring/
H A Dfs.h3 int io_renameat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
4 int io_renameat(struct io_kiocb *req, unsigned int issue_flags);
5 void io_renameat_cleanup(struct io_kiocb *req);
7 int io_unlinkat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
8 int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags);
9 void io_unlinkat_cleanup(struct io_kiocb *req);
11 int io_mkdirat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
12 int io_mkdirat(struct io_kiocb *req, unsigned int issue_flags);
13 void io_mkdirat_cleanup(struct io_kiocb *req);
15 int io_symlinkat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
[all …]
H A Dkbuf.h70 struct io_br_sel io_buffer_select(struct io_kiocb *req, size_t *len,
72 int io_buffers_select(struct io_kiocb *req, struct buf_sel_arg *arg,
74 int io_buffers_peek(struct io_kiocb *req, struct buf_sel_arg *arg,
78 int io_remove_buffers_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
79 int io_provide_buffers_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
80 int io_manage_buffers_legacy(struct io_kiocb *req, unsigned int issue_flags);
86 bool io_kbuf_recycle_legacy(struct io_kiocb *req, unsigned issue_flags);
87 void io_kbuf_drop_legacy(struct io_kiocb *req);
89 unsigned int __io_put_kbufs(struct io_kiocb *req, struct io_buffer_list *bl,
91 bool io_kbuf_commit(struct io_kiocb *req,
[all …]
H A Dxattr.h3 void io_xattr_cleanup(struct io_kiocb *req);
5 int io_fsetxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
6 int io_fsetxattr(struct io_kiocb *req, unsigned int issue_flags);
8 int io_setxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
9 int io_setxattr(struct io_kiocb *req, unsigned int issue_flags);
11 int io_fgetxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
12 int io_fgetxattr(struct io_kiocb *req, unsigned int issue_flags);
14 int io_getxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
15 int io_getxattr(struct io_kiocb *req, unsigned int issue_flags);
H A Dtimeout.c24 struct io_kiocb *head;
26 struct io_kiocb *prev;
77 static struct io_kiocb *__io_disarm_linked_timeout(struct io_kiocb *req,
78 struct io_kiocb *link);
80 static inline bool io_is_timeout_noseq(struct io_kiocb *req) in io_is_timeout_noseq()
88 static inline void io_put_req(struct io_kiocb *req) in io_put_req()
112 struct io_kiocb *req = tw_req.req; in io_timeout_complete()
138 struct io_kiocb *req; in io_flush_killed_timeouts()
151 static void io_kill_timeout(struct io_kiocb *req, struct list_head *list) in io_kill_timeout()
175 struct io_kiocb *req = cmd_to_io_kiocb(timeout); in io_flush_timeouts()
[all …]
H A Dio_uring.h125 struct io_kiocb *req;
176 void io_req_defer_failed(struct io_kiocb *req, s32 res);
179 bool io_req_post_cqe(struct io_kiocb *req, s32 res, u32 cflags);
180 bool io_req_post_cqe32(struct io_kiocb *req, struct io_uring_cqe src_cqe[2]);
183 unsigned io_linked_nr(struct io_kiocb *req);
184 void io_req_track_inflight(struct io_kiocb *req);
185 struct file *io_file_get_normal(struct io_kiocb *req, int fd);
186 struct file *io_file_get_fixed(struct io_kiocb *req, int fd,
190 void io_req_task_queue(struct io_kiocb *req);
192 void io_req_task_queue_fail(struct io_kiocb *req, int ret);
[all …]
H A Dpoll.c36 struct io_kiocb *req;
59 static inline struct io_kiocb *wqe_to_req(struct wait_queue_entry *wqe) in wqe_to_req()
63 return (struct io_kiocb *)(priv & ~IO_WQE_F_DOUBLE); in wqe_to_req()
73 static bool io_poll_get_ownership_slowpath(struct io_kiocb *req) in io_poll_get_ownership_slowpath()
94 static inline bool io_poll_get_ownership(struct io_kiocb *req) in io_poll_get_ownership()
101 static void io_poll_mark_cancelled(struct io_kiocb *req) in io_poll_mark_cancelled()
106 static struct io_poll *io_poll_get_double(struct io_kiocb *req) in io_poll_get_double()
114 static struct io_poll *io_poll_get_single(struct io_kiocb *req) in io_poll_get_single()
121 static void io_poll_req_insert(struct io_kiocb *req) in io_poll_req_insert()
171 static void io_poll_remove_entries(struct io_kiocb *req) in io_poll_remove_entries()
[all …]
H A Drw.c38 static bool io_file_supports_nowait(struct io_kiocb *req, __poll_t mask) in io_file_supports_nowait()
64 static int io_iov_buffer_select_prep(struct io_kiocb *req) in io_iov_buffer_select_prep()
83 static int io_import_vec(int ddir, struct io_kiocb *req, in io_import_vec()
110 static int __io_import_rw_buffer(int ddir, struct io_kiocb *req, in __io_import_rw_buffer()
132 static inline int io_import_rw_buffer(int rw, struct io_kiocb *req, in io_import_rw_buffer()
147 static bool io_rw_recycle(struct io_kiocb *req, unsigned int issue_flags) in io_rw_recycle()
165 static void io_req_rw_cleanup(struct io_kiocb *req, unsigned int issue_flags) in io_req_rw_cleanup()
204 static int io_rw_alloc_async(struct io_kiocb *req) in io_rw_alloc_async()
232 static int io_prep_rw_pi(struct io_kiocb *req, struct io_rw *rw, int ddir, in io_prep_rw_pi()
259 static int __io_prep_rw(struct io_kiocb *req, const struct io_uring_sqe *sqe, in __io_prep_rw()
[all …]
H A Dsync.h3 int io_sfr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
4 int io_sync_file_range(struct io_kiocb *req, unsigned int issue_flags);
6 int io_fsync_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
7 int io_fsync(struct io_kiocb *req, unsigned int issue_flags);
9 int io_fallocate(struct io_kiocb *req, unsigned int issue_flags);
10 int io_fallocate_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
H A Dxattr.c25 void io_xattr_cleanup(struct io_kiocb *req) in io_xattr_cleanup()
34 static void io_xattr_finish(struct io_kiocb *req, int ret) in io_xattr_finish()
42 static int __io_getxattr_prep(struct io_kiocb *req, in __io_getxattr_prep()
74 int io_fgetxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_fgetxattr_prep()
79 int io_getxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_getxattr_prep()
97 int io_fgetxattr(struct io_kiocb *req, unsigned int issue_flags) in io_fgetxattr()
109 int io_getxattr(struct io_kiocb *req, unsigned int issue_flags) in io_getxattr()
122 static int __io_setxattr_prep(struct io_kiocb *req, in __io_setxattr_prep()
151 int io_setxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_setxattr_prep()
169 int io_fsetxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_fsetxattr_prep()
[all …]
H A Dopenclose.c53 static int __io_openat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in __io_openat_prep()
85 void io_openat_bpf_populate(struct io_uring_bpf_ctx *bctx, struct io_kiocb *req) in io_openat_bpf_populate()
94 int io_openat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_openat_prep()
104 int io_openat2_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_openat2_prep()
123 int io_openat2(struct io_kiocb *req, unsigned int issue_flags) in io_openat2()
188 int io_openat(struct io_kiocb *req, unsigned int issue_flags) in io_openat()
193 void io_open_cleanup(struct io_kiocb *req) in io_open_cleanup()
212 static inline int io_close_fixed(struct io_kiocb *req, unsigned int issue_flags) in io_close_fixed()
219 int io_close_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_close_prep()
236 int io_close(struct io_kiocb *req, unsigned int issue_flags) in io_close()
[all …]
H A Dtw.c23 struct io_kiocb *req, *tmp; in io_fallback_req_func()
62 struct io_kiocb *req = container_of(node, struct io_kiocb, in io_handle_tw_list()
91 struct io_kiocb *req; in __io_fallback_tw()
94 req = container_of(node, struct io_kiocb, io_task_work.node); in __io_fallback_tw()
170 void io_req_local_work_add(struct io_kiocb *req, unsigned flags) in io_req_local_work_add()
192 struct io_kiocb *first_req = container_of(head, in io_req_local_work_add()
193 struct io_kiocb, in io_req_local_work_add()
239 void io_req_normal_work_add(struct io_kiocb *req) in io_req_normal_work_add()
267 void io_req_task_work_add_remote(struct io_kiocb *req, unsigned flags) in io_req_task_work_add_remote()
313 struct io_kiocb *req = container_of(*node, struct io_kiocb, in __io_run_local_work_loop()
H A Dadvise.h3 int io_madvise_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
4 int io_madvise(struct io_kiocb *req, unsigned int issue_flags);
6 int io_fadvise_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
7 int io_fadvise(struct io_kiocb *req, unsigned int issue_flags);
H A Dwaitid.c35 static void io_waitid_free(struct io_kiocb *req) in io_waitid_free()
68 static bool io_waitid_copy_si(struct io_kiocb *req, int signo) in io_waitid_copy_si()
97 static int io_waitid_finish(struct io_kiocb *req, int ret) in io_waitid_finish()
112 static void io_waitid_remove_wq(struct io_kiocb *req) in io_waitid_remove_wq()
128 static void io_waitid_complete(struct io_kiocb *req, int ret) in io_waitid_complete()
146 static bool __io_waitid_cancel(struct io_kiocb *req) in __io_waitid_cancel()
179 static inline bool io_waitid_drop_issue_ref(struct io_kiocb *req) in io_waitid_drop_issue_ref()
199 struct io_kiocb *req = tw_req.req; in io_waitid_cb()
241 struct io_kiocb *req = iwa->req; in io_waitid_wait()
260 int io_waitid_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_waitid_prep()
[all …]
H A Dfutex.c27 struct io_kiocb *req;
56 struct io_kiocb *req = tw_req.req; in io_futex_complete()
67 struct io_kiocb *req = tw_req.req; in io_futexv_complete()
92 static bool __io_futex_cancel(struct io_kiocb *req) in __io_futex_cancel()
127 int io_futex_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_futex_prep()
159 struct io_kiocb *req = q->wake_data; in io_futex_wakev_fn()
174 int io_futexv_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_futexv_prep()
213 struct io_kiocb *req = ifd->req; in io_futex_wake_fn()
223 int io_futexv_wait(struct io_kiocb *req, unsigned int issue_flags) in io_futexv_wait()
276 int io_futex_wait(struct io_kiocb *req, unsigned int issue_flags) in io_futex_wait()
[all …]
H A Dio_uring.c121 static void io_queue_sqe(struct io_kiocb *req, unsigned int extra_flags);
153 static void io_poison_cached_req(struct io_kiocb *req) in io_poison_cached_req()
163 static void io_poison_req(struct io_kiocb *req) in io_poison_req()
173 static inline void req_fail_link_node(struct io_kiocb *req, int res) in req_fail_link_node()
179 static inline void io_req_add_to_cache(struct io_kiocb *req, struct io_ring_ctx *ctx) in io_req_add_to_cache()
309 static void io_clean_op(struct io_kiocb *req) in io_clean_op()
336 inline void io_req_track_inflight(struct io_kiocb *req) in io_req_track_inflight()
344 static struct io_kiocb *__io_prep_linked_timeout(struct io_kiocb *req) in __io_prep_linked_timeout()
358 static void io_prep_async_work(struct io_kiocb *req) in io_prep_async_work()
390 static void io_prep_async_link(struct io_kiocb *req) in io_prep_async_link()
[all …]
H A During_cmd.h11 int io_uring_cmd(struct io_kiocb *req, unsigned int issue_flags);
12 int io_uring_cmd_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
13 void io_uring_cmd_sqe_copy(struct io_kiocb *req);
14 void io_uring_cmd_cleanup(struct io_kiocb *req);
H A Dnet.c85 struct io_kiocb *notif;
123 int io_shutdown_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_shutdown_prep()
136 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags) in io_shutdown()
166 static void io_netmsg_recycle(struct io_kiocb *req, unsigned int issue_flags) in io_netmsg_recycle()
185 static struct io_async_msghdr *io_msg_alloc_async(struct io_kiocb *req) in io_msg_alloc_async()
200 static inline void io_mshot_prep_retry(struct io_kiocb *req, in io_mshot_prep_retry()
211 static int io_net_import_vec(struct io_kiocb *req, struct io_async_msghdr *iomsg, in io_net_import_vec()
238 static int io_compat_msg_copy_hdr(struct io_kiocb *req, in io_compat_msg_copy_hdr()
289 static int io_msg_copy_hdr(struct io_kiocb *req, struct io_async_msghdr *iomsg, in io_msg_copy_hdr()
342 void io_sendmsg_recvmsg_cleanup(struct io_kiocb *req) in io_sendmsg_recvmsg_cleanup()
[all …]
H A Dnotif.h26 struct io_kiocb *io_alloc_notif(struct io_ring_ctx *ctx);
30 static inline struct io_notif_data *io_notif_to_data(struct io_kiocb *notif) in io_notif_to_data()
35 static inline void io_notif_flush(struct io_kiocb *notif) in io_notif_flush()
43 static inline int io_notif_account_mem(struct io_kiocb *notif, unsigned len) in io_notif_account_mem()
H A Drsrc.h66 struct io_rsrc_node *io_find_buf_node(struct io_kiocb *req,
68 int io_import_reg_buf(struct io_kiocb *req, struct iov_iter *iter,
72 struct io_kiocb *req, struct iou_vec *vec,
74 int io_prep_reg_iovec(struct io_kiocb *req, struct iou_vec *iv,
128 int io_files_update(struct io_kiocb *req, unsigned int issue_flags);
129 int io_files_update_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
H A Dstatx.h3 int io_statx_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
4 int io_statx(struct io_kiocb *req, unsigned int issue_flags);
5 void io_statx_cleanup(struct io_kiocb *req);
H A Dkbuf.c62 bool io_kbuf_commit(struct io_kiocb *req, in io_kbuf_commit()
99 void io_kbuf_drop_legacy(struct io_kiocb *req) in io_kbuf_drop_legacy()
108 bool io_kbuf_recycle_legacy(struct io_kiocb *req, unsigned issue_flags) in io_kbuf_recycle_legacy()
135 static void __user *io_provided_buffer_select(struct io_kiocb *req, size_t *len, in io_provided_buffer_select()
156 static int io_provided_buffers_select(struct io_kiocb *req, size_t *len, in io_provided_buffers_select()
171 static bool io_should_commit(struct io_kiocb *req, unsigned int issue_flags) in io_should_commit()
192 static struct io_br_sel io_ring_buffer_select(struct io_kiocb *req, size_t *len, in io_ring_buffer_select()
226 struct io_br_sel io_buffer_select(struct io_kiocb *req, size_t *len, in io_buffer_select()
249 static int io_ring_buffers_peek(struct io_kiocb *req, struct buf_sel_arg *arg, in io_ring_buffers_peek()
331 int io_buffers_select(struct io_kiocb *req, struct buf_sel_arg *arg, in io_buffers_select()
[all …]
H A Dadvise.c31 int io_madvise_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_madvise_prep()
51 int io_madvise(struct io_kiocb *req, unsigned int issue_flags) in io_madvise()
79 int io_fadvise_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_fadvise_prep()
96 int io_fadvise(struct io_kiocb *req, unsigned int issue_flags) in io_fadvise()
H A Dnop.h3 int io_nop_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
4 int io_nop(struct io_kiocb *req, unsigned int issue_flags);
H A Dtruncate.h3 int io_ftruncate_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
4 int io_ftruncate(struct io_kiocb *req, unsigned int issue_flags);
/linux/include/trace/events/
H A Dio_uring.h110 TP_PROTO(struct io_kiocb *req, int fd),
142 TP_PROTO(struct io_kiocb *req, bool hashed),
186 TP_PROTO(struct io_kiocb *req),
225 TP_PROTO(struct io_kiocb *req, struct io_kiocb *target_req),
285 TP_PROTO(struct io_kiocb *req, struct io_kiocb *link),
366 TP_PROTO(struct io_kiocb *req),
410 TP_PROTO(struct io_kiocb *req, int mask, int events),
451 TP_PROTO(struct io_kiocb *req, int mask),
492 TP_PROTO(const struct io_uring_sqe *sqe, struct io_kiocb *req, int error),

12