| /linux/tools/include/io_uring/ |
| H A D | mini_liburing.h | 25 struct io_uring_cqe *cqes; 50 struct io_uring_cqe *cqes; 79 sq->ring_sz += p->cq_entries * sizeof(struct io_uring_cqe); in io_uring_mmap() 108 cq->ring_sz = p->cq_off.cqes + p->cq_entries * sizeof(struct io_uring_cqe); in io_uring_mmap() 183 struct io_uring_cqe **cqe_ptr) in io_uring_wait_cqe()
|
| /linux/io_uring/ |
| H A D | io_uring.h | 180 bool io_req_post_cqe32(struct io_kiocb *req, struct io_uring_cqe src_cqe[2]); 258 struct io_uring_cqe **ret, in io_get_cqe_overflow() 280 static inline bool io_get_cqe(struct io_ring_ctx *ctx, struct io_uring_cqe **ret, in io_get_cqe() 287 struct io_uring_cqe **cqe_ret) in io_defer_get_uncommited_cqe() 299 struct io_uring_cqe *cqe; in io_fill_cqe_req()
|
| H A D | uring_cmd.h | 21 struct io_uring_cqe cqe[2]);
|
| H A D | cmd_net.c | 76 struct io_uring_cqe cqe[2]; in io_process_timestamp_skb() 82 BUILD_BUG_ON(sizeof(struct io_uring_cqe) != sizeof(struct io_timespec)); in io_process_timestamp_skb()
|
| H A D | fdinfo.c | 156 struct io_uring_cqe *cqe; in __io_uring_show_fdinfo() 244 struct io_uring_cqe *cqe = &ocqe->cqe; in __io_uring_show_fdinfo()
|
| H A D | io_uring.c | 533 size_t cqe_size = sizeof(struct io_uring_cqe); in __io_cqring_overflow_flush() 534 struct io_uring_cqe *cqe; in __io_cqring_overflow_flush() 670 ocq_size += sizeof(struct io_uring_cqe); in io_alloc_ocqe() 696 struct io_uring_cqe *cqe = &ctx->rings->cqes[off]; in io_fill_nop_cqe() 755 struct io_uring_cqe src_cqe[2]) in io_fill_cqe_aux32() 757 struct io_uring_cqe *cqe; in io_fill_cqe_aux32() 773 struct io_uring_cqe *cqe; in io_fill_cqe_aux() 885 bool io_req_post_cqe32(struct io_kiocb *req, struct io_uring_cqe cqe[2]) in io_req_post_cqe32()
|
| H A D | register.c | 636 cq_size = sizeof(struct io_uring_cqe); in io_register_resize_rings()
|
| H A D | zcrx.c | 1294 struct io_uring_cqe *cqe; in io_zcrx_queue_cqe()
|
| /linux/include/linux/ |
| H A D | io_uring_types.h | 224 struct io_uring_cqe cqes[] ____cacheline_aligned_in_smp; 389 struct io_uring_cqe *cqe_cached; 390 struct io_uring_cqe *cqe_sentinel; 793 struct io_uring_cqe cqe;
|
| /linux/tools/testing/vsock/ |
| H A D | vsock_uring_test.c | 63 struct io_uring_cqe *cqe; in vsock_io_uring_client() 144 struct io_uring_cqe *cqe; in vsock_io_uring_server()
|
| /linux/tools/testing/selftests/ublk/ |
| H A D | fault_inject.c | 103 const struct io_uring_cqe *cqe)
|
| H A D | kublk.h | 149 const struct io_uring_cqe *); 562 const struct io_uring_cqe *cqe);
|
| H A D | null.c | 95 const struct io_uring_cqe *cqe) in ublk_null_io_done()
|
| H A D | kublk.c | 76 struct io_uring_cqe *cqe; in __ublk_ctrl_cmd() 832 struct io_uring_cqe *cqe) in ublksrv_handle_tgt_cqe() 846 const struct io_uring_cqe *cqe) in ublk_handle_uring_cmd() 885 struct io_uring_cqe *cqe, void *data) in ublk_handle_cqe() 917 struct io_uring_cqe *cqe; in ublk_reap_events_uring()
|
| H A D | file_backed.c | 169 const struct io_uring_cqe *cqe) in ublk_loop_memset_file()
|
| H A D | stripe.c | 235 const struct io_uring_cqe *cqe) in ublk_stripe_io_done()
|
| /linux/tools/testing/selftests/x86/ |
| H A D | lam.c | 105 struct io_uring_cqe *cqes; 490 cring->ring_sz = p.cq_off.cqes + p.cq_entries * sizeof(struct io_uring_cqe); in mmap_io_uring() 571 struct io_uring_cqe *cqe; in handle_uring_cq()
|
| /linux/tools/testing/selftests/net/ |
| H A D | io_uring_zerocopy_tx.c | 97 struct io_uring_cqe *cqe; in do_tx()
|
| /linux/tools/include/uapi/linux/ |
| H A D | io_uring.h | 392 struct io_uring_cqe { struct
|
| /linux/include/uapi/linux/ |
| H A D | io_uring.h | 500 struct io_uring_cqe { struct
|
| /linux/include/trace/events/ |
| H A D | io_uring.h | 323 TP_PROTO(struct io_ring_ctx *ctx, void *req, struct io_uring_cqe *cqe),
|
| /linux/tools/testing/selftests/bpf/prog_tests/ |
| H A D | sockopt.c | 995 struct io_uring_cqe *cqe; in uring_sockopt()
|
| /linux/tools/testing/selftests/mm/ |
| H A D | cow.c | 414 struct io_uring_cqe *cqe; in do_test_iouring()
|