Lines Matching refs:io_ring_ctx
122 static void __io_req_caches_free(struct io_ring_ctx *ctx);
179 static inline void io_req_add_to_cache(struct io_kiocb *req, struct io_ring_ctx *ctx) in io_req_add_to_cache()
188 struct io_ring_ctx *ctx = container_of(ref, struct io_ring_ctx, refs); in io_ring_ctx_ref_free()
215 static void io_free_alloc_caches(struct io_ring_ctx *ctx) in io_free_alloc_caches()
225 static __cold struct io_ring_ctx *io_ring_ctx_alloc(struct io_uring_params *p) in io_ring_ctx_alloc()
227 struct io_ring_ctx *ctx; in io_ring_ctx_alloc()
395 struct io_ring_ctx *ctx = req->ctx; in io_prep_async_link()
456 static __cold noinline void io_queue_deferred(struct io_ring_ctx *ctx) in io_queue_deferred()
479 void __io_commit_cqring_flush(struct io_ring_ctx *ctx) in __io_commit_cqring_flush()
489 static inline void __io_cq_lock(struct io_ring_ctx *ctx) in __io_cq_lock()
495 static inline void io_cq_lock(struct io_ring_ctx *ctx) in io_cq_lock()
501 static inline void __io_cq_unlock_post(struct io_ring_ctx *ctx) in __io_cq_unlock_post()
514 static void io_cq_unlock_post(struct io_ring_ctx *ctx) in io_cq_unlock_post()
523 static void __io_cqring_overflow_flush(struct io_ring_ctx *ctx, bool dying) in __io_cqring_overflow_flush()
580 static void io_cqring_overflow_kill(struct io_ring_ctx *ctx) in io_cqring_overflow_kill()
586 void io_cqring_do_overflow_flush(struct io_ring_ctx *ctx) in io_cqring_do_overflow_flush()
593 void io_cqring_overflow_flush_locked(struct io_ring_ctx *ctx) in io_cqring_overflow_flush_locked()
634 static __cold bool io_cqring_add_overflow(struct io_ring_ctx *ctx, in io_cqring_add_overflow()
660 static struct io_overflow_cqe *io_alloc_ocqe(struct io_ring_ctx *ctx, in io_alloc_ocqe()
693 static bool io_fill_nop_cqe(struct io_ring_ctx *ctx, unsigned int off) in io_fill_nop_cqe()
712 bool io_cqe_cache_refill(struct io_ring_ctx *ctx, bool overflow, bool cqe32) in io_cqe_cache_refill()
754 static bool io_fill_cqe_aux32(struct io_ring_ctx *ctx, in io_fill_cqe_aux32()
769 static bool io_fill_cqe_aux(struct io_ring_ctx *ctx, u64 user_data, s32 res, in io_fill_cqe_aux()
796 static __cold void io_cqe_overflow(struct io_ring_ctx *ctx, struct io_cqe *cqe, in io_cqe_overflow()
807 static __cold bool io_cqe_overflow_locked(struct io_ring_ctx *ctx, in io_cqe_overflow_locked()
817 bool io_post_aux_cqe(struct io_ring_ctx *ctx, u64 user_data, s32 res, u32 cflags) in io_post_aux_cqe()
836 void io_add_aux_cqe(struct io_ring_ctx *ctx, u64 user_data, s32 res, u32 cflags) in io_add_aux_cqe()
855 struct io_ring_ctx *ctx = req->ctx; in io_req_post_cqe()
887 struct io_ring_ctx *ctx = req->ctx; in io_req_post_cqe32()
908 struct io_ring_ctx *ctx = req->ctx; in io_req_complete_post()
964 __cold bool __io_alloc_req_refill(struct io_ring_ctx *ctx) in __io_alloc_req_refill()
1007 struct io_ring_ctx *ctx = req->ctx; in __io_req_find_next_prep()
1042 struct io_ring_ctx *ctx = req->ctx; in io_req_task_submit()
1076 struct io_ring_ctx *ctx = req->ctx; in io_req_put_rsrc_nodes()
1086 static void io_free_batch_list(struct io_ring_ctx *ctx, in io_free_batch_list()
1128 void __io_submit_flush_completions(struct io_ring_ctx *ctx) in __io_submit_flush_completions()
1169 __cold void io_iopoll_try_reap_events(struct io_ring_ctx *ctx) in io_iopoll_try_reap_events()
1196 static int io_iopoll_check(struct io_ring_ctx *ctx, unsigned int min_events) in io_iopoll_check()
1277 struct io_ring_ctx *ctx = req->ctx; in io_iopoll_req_issued()
1332 struct io_ring_ctx *ctx = req->ctx; in io_drain_req()
1561 struct io_ring_ctx *ctx = req->ctx; in io_file_get_fixed()
1672 static inline bool io_check_restriction(struct io_ring_ctx *ctx, in io_check_restriction()
1692 static void io_init_drain(struct io_ring_ctx *ctx) in io_init_drain()
1717 static int io_init_req(struct io_ring_ctx *ctx, struct io_kiocb *req, in io_init_req()
1836 struct io_ring_ctx *ctx = req->ctx; in io_submit_fail_init()
1870 static inline int io_submit_sqe(struct io_ring_ctx *ctx, struct io_kiocb *req, in io_submit_sqe()
1929 static void io_submit_state_end(struct io_ring_ctx *ctx) in io_submit_state_end()
1954 static void io_commit_sqring(struct io_ring_ctx *ctx) in io_commit_sqring()
1978 static bool io_get_sqe(struct io_ring_ctx *ctx, const struct io_uring_sqe **sqe) in io_get_sqe()
2010 int io_submit_sqes(struct io_ring_ctx *ctx, unsigned int nr) in io_submit_sqes()
2066 static void io_rings_free(struct io_ring_ctx *ctx) in io_rings_free()
2128 static __cold void __io_req_caches_free(struct io_ring_ctx *ctx) in __io_req_caches_free()
2145 static __cold void io_req_caches_free(struct io_ring_ctx *ctx) in io_req_caches_free()
2151 static __cold void io_ring_ctx_free(struct io_ring_ctx *ctx) in io_ring_ctx_free()
2206 struct io_ring_ctx *ctx = container_of(cb, struct io_ring_ctx, in io_activate_pollwq_cb()
2221 __cold void io_activate_pollwq(struct io_ring_ctx *ctx) in io_activate_pollwq()
2245 struct io_ring_ctx *ctx = file->private_data; in io_uring_poll()
2285 struct io_ring_ctx *ctx;
2307 struct io_ring_ctx *ctx = container_of(work, struct io_ring_ctx, exit_work); in io_ring_exit_work()
2409 static __cold void io_ring_ctx_wait_and_kill(struct io_ring_ctx *ctx) in io_ring_ctx_wait_and_kill()
2434 struct io_ring_ctx *ctx = file->private_data; in io_uring_release()
2441 static struct io_uring_reg_wait *io_get_ext_arg_reg(struct io_ring_ctx *ctx, in io_get_ext_arg_reg()
2460 static int io_validate_ext_arg(struct io_ring_ctx *ctx, unsigned flags, in io_validate_ext_arg()
2476 static int io_get_ext_arg(struct io_ring_ctx *ctx, unsigned flags, in io_get_ext_arg()
2588 struct io_ring_ctx *ctx; in SYSCALL_DEFINE6()
2715 static __cold int io_allocate_scq_urings(struct io_ring_ctx *ctx, in io_allocate_scq_urings()
2779 static struct file *io_uring_get_file(struct io_ring_ctx *ctx) in io_uring_get_file()
2948 static void io_ctx_restriction_clone(struct io_ring_ctx *ctx, in io_ctx_restriction_clone()
2965 struct io_ring_ctx *ctx; in io_uring_create()