Searched refs:sq_entries (Results 1 – 12 of 12) sorted by relevance
| /linux/io_uring/ |
| H A D | fdinfo.c | 64 unsigned int sq_mask = ctx->sq_entries - 1, cq_mask = ctx->cq_entries - 1; in __io_uring_show_fdinfo() 70 unsigned int cq_entries, sq_entries; in __io_uring_show_fdinfo() local 93 sq_entries = min(sq_tail - sq_head, ctx->sq_entries); in __io_uring_show_fdinfo() 94 for (i = 0; i < sq_entries; i++) { in __io_uring_show_fdinfo()
|
| H A D | io_uring.c | 1748 (unsigned)(sqe - ctx->sq_sqes) >= ctx->sq_entries - 1) in io_init_req() 1978 unsigned mask = ctx->sq_entries - 1; in io_get_sqe() 1984 if (unlikely(head >= ctx->sq_entries)) { in io_get_sqe() 1989 head = array_index_nospec(head, ctx->sq_entries); in io_get_sqe() 2016 entries = ctx->sq_entries; in io_submit_sqes() 2073 static int rings_size(unsigned int flags, unsigned int sq_entries, in rings_size() argument 2085 if (sq_entries < 2) in rings_size() 2095 rl->sq_size = array_size(sqe_size, sq_entries); in rings_size() 2116 sq_array_size = array_size(sizeof(u32), sq_entries); in rings_size() 2694 ctx->sq_entries = p->sq_entries; in io_allocate_scq_urings() [all …]
|
| H A D | register.c | 537 WRITE_ONCE(n.rings->sq_ring_mask, p->sq_entries - 1); in io_register_resize_rings() 539 WRITE_ONCE(n.rings->sq_ring_entries, p->sq_entries); in io_register_resize_rings() 591 if (tail - old_head > p->sq_entries) in io_register_resize_rings() 594 unsigned src_head = i & (ctx->sq_entries - 1); in io_register_resize_rings() 595 unsigned dst_head = i & (p->sq_entries - 1); in io_register_resize_rings() 633 ctx->sq_entries = p->sq_entries; in io_register_resize_rings()
|
| H A D | io_uring.h | 445 return READ_ONCE(r->sq.tail) - READ_ONCE(r->sq.head) == ctx->sq_entries; in io_sqring_full() 455 return min(entries, ctx->sq_entries); in io_sqring_entries()
|
| H A D | tctx.c | 41 concurrency = min(ctx->sq_entries, 4 * num_online_cpus()); in io_init_wq_offload()
|
| /linux/include/trace/events/ |
| H A D | io_uring.h | 29 TP_PROTO(int fd, void *ctx, u32 sq_entries, u32 cq_entries, u32 flags), 31 TP_ARGS(fd, ctx, sq_entries, cq_entries, flags), 36 __field( u32, sq_entries ) 44 __entry->sq_entries = sq_entries; 50 __entry->ctx, __entry->fd, __entry->sq_entries,
|
| /linux/tools/include/io_uring/ |
| H A D | mini_liburing.h | 82 sq->ring_sz += p->sq_entries * sizeof(unsigned int); in io_uring_mmap() 98 size = p->sq_entries * sizeof(struct io_uring_sqe); in io_uring_mmap() 113 munmap(sq->sqes, p->sq_entries * sizeof(struct io_uring_sqe)); in io_uring_mmap()
|
| /linux/drivers/net/ethernet/ibm/ehea/ |
| H A D | ehea_main.c | 53 static int sq_entries = EHEA_DEF_ENTRIES_SQ; variable 61 module_param(sq_entries, int, 0); 77 MODULE_PARM_DESC(sq_entries, " Number of entries for the Send Queue " 2255 pr_cfg.max_entries_scq = sq_entries * 2; in ehea_port_res_setup() 2256 pr_cfg.max_entries_sq = sq_entries; in ehea_port_res_setup() 2262 pr_cfg_small_rx.max_entries_scq = sq_entries; in ehea_port_res_setup() 2263 pr_cfg_small_rx.max_entries_sq = sq_entries; in ehea_port_res_setup() 2963 port->sig_comp_iv = sq_entries / 10; in ehea_setup_single_port() 3514 if ((sq_entries < EHEA_MIN_ENTRIES_QP) || in check_module_parm() 3515 (sq_entries > EHEA_MAX_ENTRIES_SQ)) { in check_module_parm()
|
| /linux/tools/include/uapi/linux/ |
| H A D | io_uring.h | 486 __u32 sq_entries; member
|
| /linux/include/linux/ |
| H A D | io_uring_types.h | 321 unsigned sq_entries; member
|
| /linux/tools/testing/selftests/x86/ |
| H A D | lam.c | 489 sring->ring_sz = p.sq_off.array + p.sq_entries * sizeof(unsigned int); in mmap_io_uring() 529 s->sq_ring.queue.sqes = mmap(0, p.sq_entries * sizeof(struct io_uring_sqe), in mmap_io_uring()
|
| /linux/include/uapi/linux/ |
| H A D | io_uring.h | 608 __u32 sq_entries; member
|