Searched refs:sqes (Results 1 – 13 of 13) sorted by relevance
| /linux/tools/include/io_uring/ |
| H A D | mini_liburing.h | 36 struct io_uring_sqe *sqes; member 99 sq->sqes = mmap(0, size, PROT_READ | PROT_WRITE, in io_uring_mmap() 101 if (sq->sqes == MAP_FAILED) { in io_uring_mmap() 113 munmap(sq->sqes, p->sq_entries * sizeof(struct io_uring_sqe)); in io_uring_mmap() 179 return &sq->sqes[sq->sqe_tail++ & *sq->kring_mask]; in io_uring_get_sqe() 253 munmap(sq->sqes, *sq->kring_entries * sizeof(struct io_uring_sqe)); in io_uring_queue_exit()
|
| /linux/io_uring/ |
| H A D | uring_cmd.h | 8 struct io_uring_sqe sqes[2]; member
|
| /linux/tools/testing/selftests/ublk/ |
| H A D | kublk.h | 375 struct io_uring_sqe *sqes[], int nr_sqes) in ublk_io_alloc_sqes() argument 385 sqes[i] = io_uring_get_sqe(ring); in ublk_io_alloc_sqes() 386 if (!sqes[i]) in ublk_io_alloc_sqes()
|
| /linux/tools/testing/selftests/x86/ |
| H A D | lam.c | 106 struct io_uring_sqe *sqes; member 529 s->sq_ring.queue.sqes = mmap(0, p.sq_entries * sizeof(struct io_uring_sqe), in mmap_io_uring() 532 if (s->sq_ring.queue.sqes == MAP_FAILED) { in mmap_io_uring() 640 sqe = &ring->sq_ring.queue.sqes[index]; in handle_uring_sq()
|
| /linux/drivers/nvme/host/ |
| H A D | apple.c | 135 struct nvme_command *sqes; member 301 memcpy(&q->sqes[q->sq_tail], cmd, sizeof(*cmd)); in apple_nvme_submit_cmd_t8015() 303 memcpy((void *)q->sqes + (q->sq_tail << APPLE_NVME_IOSQES), in apple_nvme_submit_cmd_t8015() 332 memcpy(&q->sqes[tag], cmd, sizeof(*cmd)); in apple_nvme_submit_cmd_t8103() 1370 q->sqes = dmam_alloc_coherent(anv->dev, iosq_size, in apple_nvme_queue_alloc() 1372 if (!q->sqes) in apple_nvme_queue_alloc()
|
| H A D | pci.c | 34 #define SQ_SIZE(q) ((q)->q_depth << (q)->sqes) 383 u8 sqes; member 733 memcpy(nvmeq->sq_cmds + (nvmeq->sq_tail << nvmeq->sqes), in nvme_sq_copy_cmd() 2110 nvmeq->sqes = qid ? dev->io_sqes : NVME_ADM_SQES; in nvme_alloc_queue()
|
| /linux/drivers/nvme/target/ |
| H A D | passthru.c | 133 id->sqes = min_t(__u8, ((0x6 << 4) | 0x6), id->sqes); in nvmet_passthru_override_id_ctrl()
|
| H A D | admin-cmd.c | 726 id->sqes = (0x6 << 4) | 0x6; in nvmet_execute_identify_ctrl()
|
| /linux/drivers/net/ethernet/intel/idpf/ |
| H A D | xsk.c | 283 .sqes = xdpsq->tx_buf, in idpf_xsk_tx_prep() 301 .sqes = xdpsq->tx_buf, in idpf_xsk_xmit_prep()
|
| H A D | xdp.c | 345 .sqes = xdpsq->tx_buf, in idpf_xdp_tx_prep()
|
| /linux/include/net/libeth/ |
| H A D | xdp.h | 422 struct libeth_sqe *sqes; member 746 sqe = &sq->sqes[i]; in libeth_xdp_tx_fill_buf() 1003 sqe = &sq->sqes[i]; in libeth_xdp_xmit_fill_buf()
|
| H A D | xsk.h | 124 sqe = &sq->sqes[i]; in libeth_xsk_tx_fill_buf()
|
| /linux/include/linux/ |
| H A D | nvme.h | 366 __u8 sqes; member
|