| /linux/io_uring/ |
| H A D | fs.c | 50 int io_renameat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_renameat_prep() argument 56 if (sqe->buf_index || sqe->splice_fd_in) in io_renameat_prep() 61 ren->old_dfd = READ_ONCE(sqe->fd); in io_renameat_prep() 62 oldf = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_renameat_prep() 63 newf = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in io_renameat_prep() 64 ren->new_dfd = READ_ONCE(sqe->len); in io_renameat_prep() 65 ren->flags = READ_ONCE(sqe->rename_flags); in io_renameat_prep() 107 int io_unlinkat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_unlinkat_prep() 113 if (sqe in io_unlinkat_prep() 104 io_unlinkat_prep(struct io_kiocb * req,const struct io_uring_sqe * sqe) io_unlinkat_prep() argument 154 io_mkdirat_prep(struct io_kiocb * req,const struct io_uring_sqe * sqe) io_mkdirat_prep() argument 198 io_symlinkat_prep(struct io_kiocb * req,const struct io_uring_sqe * sqe) io_symlinkat_prep() argument 241 io_linkat_prep(struct io_kiocb * req,const struct io_uring_sqe * sqe) io_linkat_prep() argument [all...] |
| H A D | sync.c | 25 int io_sfr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_sfr_prep() argument 29 if (unlikely(sqe->addr || sqe->buf_index || sqe->splice_fd_in)) in io_sfr_prep() 32 sync->off = READ_ONCE(sqe->off); in io_sfr_prep() 33 sync->len = READ_ONCE(sqe->len); in io_sfr_prep() 34 sync->flags = READ_ONCE(sqe->sync_range_flags); in io_sfr_prep() 53 int io_fsync_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_fsync_prep() argument 57 if (unlikely(sqe->addr || sqe in io_fsync_prep() 85 io_fallocate_prep(struct io_kiocb * req,const struct io_uring_sqe * sqe) io_fallocate_prep() argument [all...] |
| H A D | openclose.c | 53 static int __io_openat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in __io_openat_prep() argument 59 if (unlikely(sqe->buf_index)) in __io_openat_prep() 68 open->dfd = READ_ONCE(sqe->fd); in __io_openat_prep() 69 fname = u64_to_user_ptr(READ_ONCE(sqe->addr)); in __io_openat_prep() 75 open->file_slot = READ_ONCE(sqe->file_index); in __io_openat_prep() 94 int io_openat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_openat_prep() argument 97 u64 mode = READ_ONCE(sqe->len); in io_openat_prep() 98 u64 flags = READ_ONCE(sqe->open_flags); in io_openat_prep() 101 return __io_openat_prep(req, sqe); in io_openat_prep() 104 int io_openat2_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_openat2_prep() argument [all …]
|
| H A D | xattr.c | 43 const struct io_uring_sqe *sqe) in __io_getxattr_prep() argument 51 name = u64_to_user_ptr(READ_ONCE(sqe->addr)); in __io_getxattr_prep() 52 ix->ctx.value = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in __io_getxattr_prep() 53 ix->ctx.size = READ_ONCE(sqe->len); in __io_getxattr_prep() 54 ix->ctx.flags = READ_ONCE(sqe->xattr_flags); in __io_getxattr_prep() 74 int io_fgetxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_fgetxattr_prep() argument 76 return __io_getxattr_prep(req, sqe); in io_fgetxattr_prep() 79 int io_getxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_getxattr_prep() argument 88 ret = __io_getxattr_prep(req, sqe); in io_getxattr_prep() 92 path = u64_to_user_ptr(READ_ONCE(sqe->addr3)); in io_getxattr_prep() [all …]
|
| H A D | advise.c | 31 int io_madvise_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_madvise_prep() argument 36 if (sqe->buf_index || sqe->splice_fd_in) in io_madvise_prep() 39 ma->addr = READ_ONCE(sqe->addr); in io_madvise_prep() 40 ma->len = READ_ONCE(sqe->off); in io_madvise_prep() 42 ma->len = READ_ONCE(sqe->len); in io_madvise_prep() 43 ma->advice = READ_ONCE(sqe->fadvise_advice); in io_madvise_prep() 79 int io_fadvise_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_fadvise_prep() argument 83 if (sqe->buf_index || sqe in io_fadvise_prep() [all...] |
| H A D | mock_file.c | 67 const struct io_uring_sqe *sqe = cmd->sqe; in io_cmd_copy_regbuf() local 74 ubuf = u64_to_user_ptr(READ_ONCE(sqe->addr3)); in io_cmd_copy_regbuf() 75 iovec = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_cmd_copy_regbuf() 76 iovec_len = READ_ONCE(sqe->len); in io_cmd_copy_regbuf() 77 flags = READ_ONCE(sqe->file_index); in io_cmd_copy_regbuf() 79 if (unlikely(sqe->ioprio || sqe->__pad1)) in io_cmd_copy_regbuf() 213 const struct io_uring_sqe *sqe = cmd->sqe; in io_create_mock_file() local 225 uarg = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_create_mock_file() 226 uarg_size = READ_ONCE(sqe->len); in io_create_mock_file() 228 if (sqe->ioprio || sqe->__pad1 || sqe->addr3 || sqe->file_index) in io_create_mock_file() [all …]
|
| H A D | truncate.c | 23 int io_ftruncate_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_ftruncate_prep() argument 27 if (sqe->rw_flags || sqe->addr || sqe->len || sqe->buf_index || in io_ftruncate_prep() 28 sqe->splice_fd_in || sqe->addr3) in io_ftruncate_prep() 31 ft->len = READ_ONCE(sqe->off); in io_ftruncate_prep()
|
| H A D | net.h | 28 int io_shutdown_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 32 int io_sendmsg_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 37 int io_recvmsg_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 43 int io_accept_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 46 int io_socket_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 50 int io_connect_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 55 int io_send_zc_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 58 int io_bind_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 61 int io_listen_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
| H A D | net.c | 89 * The UAPI flags are the lower 8 bits, as that's all sqe->ioprio will hold 123 int io_shutdown_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_shutdown_prep() argument 127 if (unlikely(sqe->off || sqe->addr || sqe->rw_flags || in io_shutdown_prep() 128 sqe->buf_index || sqe->splice_fd_in)) in io_shutdown_prep() 131 shutdown->how = READ_ONCE(sqe->len); in io_shutdown_prep() 349 static int io_send_setup(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_send_setup() argument 357 sr->buf = u64_to_user_ptr(READ_ONCE(sqe in io_send_setup() 390 io_sendmsg_setup(struct io_kiocb * req,const struct io_uring_sqe * sqe) io_sendmsg_setup() argument 416 io_sendmsg_prep(struct io_kiocb * req,const struct io_uring_sqe * sqe) io_sendmsg_prep() argument 777 io_recvmsg_prep(struct io_kiocb * req,const struct io_uring_sqe * sqe) io_recvmsg_prep() argument 1240 io_recvzc_prep(struct io_kiocb * req,const struct io_uring_sqe * sqe) io_recvzc_prep() argument 1320 io_send_zc_prep(struct io_kiocb * req,const struct io_uring_sqe * sqe) io_send_zc_prep() argument 1607 io_accept_prep(struct io_kiocb * req,const struct io_uring_sqe * sqe) io_accept_prep() argument 1702 io_socket_prep(struct io_kiocb * req,const struct io_uring_sqe * sqe) io_socket_prep() argument 1756 io_connect_prep(struct io_kiocb * req,const struct io_uring_sqe * sqe) io_connect_prep() argument 1827 io_bind_prep(struct io_kiocb * req,const struct io_uring_sqe * sqe) io_bind_prep() argument 1863 io_listen_prep(struct io_kiocb * req,const struct io_uring_sqe * sqe) io_listen_prep() argument [all...] |
| H A D | futex.c | 127 int io_futex_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_futex_prep() argument 132 if (unlikely(sqe->len || sqe->futex_flags || sqe->buf_index || in io_futex_prep() 133 sqe->file_index)) in io_futex_prep() 136 iof->uaddr = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_futex_prep() 137 iof->futex_val = READ_ONCE(sqe->addr2); in io_futex_prep() 138 iof->futex_mask = READ_ONCE(sqe->addr3); in io_futex_prep() 139 flags = READ_ONCE(sqe->fd); in io_futex_prep() 172 int io_futexv_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_futexv_prep() argument 179 if (unlikely(sqe->fd || sqe->buf_index || sqe->file_index || in io_futexv_prep() 180 sqe->addr2 || sqe->futex_flags || sqe->addr3)) in io_futexv_prep() [all …]
|
| H A D | cmd_net.c | 14 const struct io_uring_sqe *sqe = cmd->sqe; in io_uring_cmd_getsockopt() 19 level = READ_ONCE(sqe->level); in io_uring_cmd_getsockopt() 23 optval = u64_to_user_ptr(READ_ONCE(sqe->optval)); in io_uring_cmd_getsockopt() 24 optname = READ_ONCE(sqe->optname); in io_uring_cmd_getsockopt() 25 optlen = READ_ONCE(sqe->optlen); in io_uring_cmd_getsockopt() 41 const struct io_uring_sqe *sqe = cmd->sqe; in io_uring_cmd_setsockopt() 47 optval = u64_to_user_ptr(READ_ONCE(sqe->optval)); in io_uring_cmd_setsockopt() 48 optname = READ_ONCE(sqe in io_uring_cmd_setsockopt() 13 const struct io_uring_sqe *sqe = cmd->sqe; io_uring_cmd_getsockopt() local 40 const struct io_uring_sqe *sqe = cmd->sqe; io_uring_cmd_setsockopt() local [all...] |
| H A D | epoll.c | 29 int io_epoll_ctl_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_epoll_ctl_prep() 33 if (sqe->buf_index || sqe->splice_fd_in) in io_epoll_ctl_prep() 36 epoll->epfd = READ_ONCE(sqe->fd); in io_epoll_ctl_prep() 37 epoll->op = READ_ONCE(sqe->len); in io_epoll_ctl_prep() 38 epoll->fd = READ_ONCE(sqe->off); in io_epoll_ctl_prep() 43 ev = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_epoll_ctl_prep() 67 int io_epoll_wait_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) 71 if (sqe->off || sqe 24 io_epoll_ctl_prep(struct io_kiocb * req,const struct io_uring_sqe * sqe) io_epoll_ctl_prep() argument [all...] |
| H A D | rw.h | 35 int io_prep_read_fixed(struct io_kiocb *req, const struct io_uring_sqe *sqe); 36 int io_prep_write_fixed(struct io_kiocb *req, const struct io_uring_sqe *sqe); 37 int io_prep_readv_fixed(struct io_kiocb *req, const struct io_uring_sqe *sqe); 38 int io_prep_writev_fixed(struct io_kiocb *req, const struct io_uring_sqe *sqe); 39 int io_prep_readv(struct io_kiocb *req, const struct io_uring_sqe *sqe); 40 int io_prep_writev(struct io_kiocb *req, const struct io_uring_sqe *sqe); 41 int io_prep_read(struct io_kiocb *req, const struct io_uring_sqe *sqe); 42 int io_prep_write(struct io_kiocb *req, const struct io_uring_sqe *sqe); 50 int io_read_mshot_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
| H A D | fs.h | 3 int io_renameat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 7 int io_unlinkat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 11 int io_mkdirat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 15 int io_symlinkat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 18 int io_linkat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
| /linux/tools/testing/selftests/ublk/ |
| H A D | null.c | 47 struct io_uring_sqe *sqe, int q_id, unsigned buf_idx) in __setup_nop_io() 51 io_uring_prep_nop(sqe); in __setup_nop_io() 52 sqe->buf_index = buf_idx; in __setup_nop_io() 53 sqe->flags |= IOSQE_FIXED_FILE; in __setup_nop_io() 54 sqe->rw_flags = IORING_NOP_FIXED_BUFFER | IORING_NOP_INJECT_RESULT; in __setup_nop_io() 55 sqe->len = iod->nr_sectors << 9; /* injected result */ in __setup_nop_io() 56 sqe->user_data = build_user_data(tag, ublk_op, 0, q_id, 1); in __setup_nop_io() 63 struct io_uring_sqe *sqe[3]; in null_queue_zc_io() 66 ublk_io_alloc_sqes(t, sqe, 3); in null_queue_zc_io() 68 io_uring_prep_buf_register(sqe[ in null_queue_zc_io() 46 __setup_nop_io(int tag,const struct ublksrv_io_desc * iod,struct io_uring_sqe * sqe,int q_id) __setup_nop_io() argument 62 struct io_uring_sqe *sqe[3]; null_queue_zc_io() local 85 struct io_uring_sqe *sqe[1]; null_queue_auto_zc_io() local [all...] |
| H A D | file_backed.c | 20 struct io_uring_sqe *sqe[1]; in loop_queue_flush_io() local 22 ublk_io_alloc_sqes(t, sqe, 1); in loop_queue_flush_io() 23 io_uring_prep_fsync(sqe[0], ublk_get_registered_fd(q, 1) /*fds[1]*/, IORING_FSYNC_DATASYNC); in loop_queue_flush_io() 24 io_uring_sqe_set_flags(sqe[0], IOSQE_FIXED_FILE); in loop_queue_flush_io() 26 sqe[0]->user_data = build_user_data(tag, ublk_op, 0, q->q_id, 1); in loop_queue_flush_io() 40 struct io_uring_sqe *sqe[3]; in loop_queue_tgt_rw_io() local 45 ublk_io_alloc_sqes(t, sqe, 1); in loop_queue_tgt_rw_io() 47 io_uring_prep_rw(op, sqe[0], ublk_get_registered_fd(q, 2), in loop_queue_tgt_rw_io() 51 sqe[0]->flags = IOSQE_FIXED_FILE; in loop_queue_tgt_rw_io() 53 sqe[0]->user_data = build_user_data(tag, ublk_op, 1, q->q_id, 1); in loop_queue_tgt_rw_io() [all …]
|
| H A D | kublk.h | 405 static inline void __io_uring_prep_buf_reg_unreg(struct io_uring_sqe *sqe, in __io_uring_prep_buf_reg_unreg() argument 408 struct ublksrv_io_cmd *cmd = (struct ublksrv_io_cmd *)sqe->cmd; in __io_uring_prep_buf_reg_unreg() 411 io_uring_prep_read(sqe, dev_fd, 0, 0, 0); in __io_uring_prep_buf_reg_unreg() 412 sqe->opcode = IORING_OP_URING_CMD; in __io_uring_prep_buf_reg_unreg() 414 sqe->flags &= ~IOSQE_FIXED_FILE; in __io_uring_prep_buf_reg_unreg() 416 sqe->flags |= IOSQE_FIXED_FILE; in __io_uring_prep_buf_reg_unreg() 423 static inline void io_uring_prep_buf_register(struct io_uring_sqe *sqe, in io_uring_prep_buf_register() argument 426 __io_uring_prep_buf_reg_unreg(sqe, q, tag, q_id, index); in io_uring_prep_buf_register() 427 sqe->cmd_op = UBLK_U_IO_REGISTER_IO_BUF; in io_uring_prep_buf_register() 430 static inline void io_uring_prep_buf_unregister(struct io_uring_sqe *sqe, in io_uring_prep_buf_unregister() argument [all …]
|
| H A D | stripe.c | 133 struct io_uring_sqe *sqe[NR_STRIPE]; in stripe_queue_tgt_rw_io() local 143 ublk_io_alloc_sqes(t, sqe, s->nr + extra); in stripe_queue_tgt_rw_io() 146 io_uring_prep_buf_register(sqe[0], q, tag, q->q_id, buf_idx); in stripe_queue_tgt_rw_io() 147 sqe[0]->flags |= IOSQE_CQE_SKIP_SUCCESS | IOSQE_IO_HARDLINK; in stripe_queue_tgt_rw_io() 148 sqe[0]->user_data = build_user_data(tag, in stripe_queue_tgt_rw_io() 149 ublk_cmd_op_nr(sqe[0]->cmd_op), 0, q->q_id, 1); in stripe_queue_tgt_rw_io() 155 io_uring_prep_rw(op, sqe[i], in stripe_queue_tgt_rw_io() 160 io_uring_sqe_set_flags(sqe[i], IOSQE_FIXED_FILE); in stripe_queue_tgt_rw_io() 162 sqe[i]->buf_index = buf_idx; in stripe_queue_tgt_rw_io() 164 sqe[i]->flags |= IOSQE_IO_HARDLINK; in stripe_queue_tgt_rw_io() [all …]
|
| /linux/drivers/crypto/hisilicon/zip/ |
| H A D | zip_crypto.c | 46 #define GET_REQ_FROM_SQE(sqe) ((u64)(sqe)->dw26 | (u64)(sqe)->dw27 << 32) argument 79 void (*fill_addr)(struct hisi_zip_sqe *sqe, struct hisi_zip_req *req); 80 void (*fill_buf_size)(struct hisi_zip_sqe *sqe, struct hisi_zip_req *req); 81 void (*fill_buf_type)(struct hisi_zip_sqe *sqe, u8 buf_type); 82 void (*fill_req_type)(struct hisi_zip_sqe *sqe, u8 req_type); 83 void (*fill_win_size)(struct hisi_zip_sqe *sqe, u8 win_size); 84 void (*fill_tag)(struct hisi_zip_sqe *sqe, struct hisi_zip_req *req); 85 void (*fill_sqe_type)(struct hisi_zip_sqe *sqe, u8 sqe_type); 86 u32 (*get_status)(struct hisi_zip_sqe *sqe); 87 u32 (*get_dstlen)(struct hisi_zip_sqe *sqe); [all …]
|
| /linux/drivers/infiniband/sw/siw/ |
| H A D | siw_qp.c | 251 wqe->sqe.flags = 0; in siw_qp_mpa_rts() 252 wqe->sqe.num_sge = 1; in siw_qp_mpa_rts() 253 wqe->sqe.sge[0].length = 0; in siw_qp_mpa_rts() 254 wqe->sqe.sge[0].laddr = 0; in siw_qp_mpa_rts() 255 wqe->sqe.sge[0].lkey = 0; in siw_qp_mpa_rts() 260 wqe->sqe.rkey = 1; in siw_qp_mpa_rts() 261 wqe->sqe.raddr = 0; in siw_qp_mpa_rts() 265 wqe->sqe.opcode = SIW_OP_WRITE; in siw_qp_mpa_rts() 269 wqe->sqe.opcode = SIW_OP_READ; in siw_qp_mpa_rts() 276 siw_read_to_orq(rreq, &wqe->sqe); in siw_qp_mpa_rts() [all …]
|
| H A D | siw_qp_tx.c | 52 struct siw_sge *sge = &wqe->sqe.sge[0]; in siw_try_1seg() 55 if (bytes > MAX_HDR_INLINE || wqe->sqe.num_sge != 1) in siw_try_1seg() 62 memcpy(paddr, &wqe->sqe.sge[1], bytes); in siw_try_1seg() 135 c_tx->pkt.rreq.sink_stag = htonl(wqe->sqe.sge[0].lkey); in siw_qp_prepare_tx() 137 cpu_to_be64(wqe->sqe.sge[0].laddr); in siw_qp_prepare_tx() 138 c_tx->pkt.rreq.source_stag = htonl(wqe->sqe.rkey); in siw_qp_prepare_tx() 139 c_tx->pkt.rreq.source_to = cpu_to_be64(wqe->sqe.raddr); in siw_qp_prepare_tx() 140 c_tx->pkt.rreq.read_size = htonl(wqe->sqe.sge[0].length); in siw_qp_prepare_tx() 183 c_tx->pkt.send_inv.inval_stag = cpu_to_be32(wqe->sqe.rkey); in siw_qp_prepare_tx() 195 c_tx->pkt.rwrite.sink_stag = htonl(wqe->sqe.rkey); in siw_qp_prepare_tx() [all …]
|
| H A D | siw_verbs.c | 652 struct siw_sqe *sqe) in siw_copy_inline_sgl() argument 655 void *kbuf = &sqe->sge[1]; in siw_copy_inline_sgl() 658 sqe->sge[0].laddr = (uintptr_t)kbuf; in siw_copy_inline_sgl() 659 sqe->sge[0].lkey = 0; in siw_copy_inline_sgl() 677 sqe->sge[0].length = max(bytes, 0); in siw_copy_inline_sgl() 678 sqe->num_sge = bytes > 0 ? 1 : 0; in siw_copy_inline_sgl() 690 struct siw_sqe sqe = {}; in siw_sq_flush_wr() local 694 sqe.opcode = SIW_OP_WRITE; in siw_sq_flush_wr() 697 sqe.opcode = SIW_OP_READ; in siw_sq_flush_wr() 700 sqe.opcode = SIW_OP_READ_LOCAL_INV; in siw_sq_flush_wr() [all …]
|
| /linux/include/net/libeth/ |
| H A D | tx.h | 124 static inline void libeth_tx_complete(struct libeth_sqe *sqe, in libeth_tx_complete() argument 127 switch (sqe->type) { in libeth_tx_complete() 133 dma_unmap_page(cp->dev, dma_unmap_addr(sqe, dma), in libeth_tx_complete() 134 dma_unmap_len(sqe, len), DMA_TO_DEVICE); in libeth_tx_complete() 140 switch (sqe->type) { in libeth_tx_complete() 142 cp->ss->packets += sqe->packets; in libeth_tx_complete() 143 cp->ss->bytes += sqe->bytes; in libeth_tx_complete() 145 napi_consume_skb(sqe->skb, cp->napi); in libeth_tx_complete() 148 kfree(sqe->raw); in libeth_tx_complete() 154 sqe->type = LIBETH_SQE_EMPTY; in libeth_tx_complete() [all …]
|
| /linux/tools/include/io_uring/ |
| H A D | mini_liburing.h | 259 static inline void io_uring_prep_cmd(struct io_uring_sqe *sqe, int op, in io_uring_prep_cmd() argument 265 memset(sqe, 0, sizeof(*sqe)); in io_uring_prep_cmd() 266 sqe->opcode = (__u8)IORING_OP_URING_CMD; in io_uring_prep_cmd() 267 sqe->fd = sockfd; in io_uring_prep_cmd() 268 sqe->cmd_op = op; in io_uring_prep_cmd() 270 sqe->level = level; in io_uring_prep_cmd() 271 sqe->optname = optname; in io_uring_prep_cmd() 272 sqe->optval = (unsigned long long)optval; in io_uring_prep_cmd() 273 sqe->optlen = optlen; in io_uring_prep_cmd() 287 static inline void io_uring_prep_send(struct io_uring_sqe *sqe, int sockfd, in io_uring_prep_send() argument [all …]
|
| /linux/drivers/net/ethernet/qlogic/qed/ |
| H A D | qed_nvmetcp_fw_funcs.c | 68 if (!task_params->sqe) in init_sqe() 71 memset(task_params->sqe, 0, sizeof(*task_params->sqe)); in init_sqe() 72 task_params->sqe->task_id = cpu_to_le16(task_params->itid); in init_sqe() 79 SET_FIELD(task_params->sqe->contlen_cdbsize, in init_sqe() 81 SET_FIELD(task_params->sqe->flags, NVMETCP_WQE_WQE_TYPE, in init_sqe() 94 SET_FIELD(task_params->sqe->flags, NVMETCP_WQE_NUM_SGES, num_sges); in init_sqe() 95 SET_FIELD(task_params->sqe->contlen_cdbsize, NVMETCP_WQE_CONT_LEN, buf_size); in init_sqe() 99 SET_FIELD(task_params->sqe->flags, NVMETCP_WQE_WQE_TYPE, in init_sqe() 101 SET_FIELD(task_params->sqe->contlen_cdbsize, in init_sqe() 106 SET_FIELD(task_params->sqe->flags, NVMETCP_WQE_WQE_TYPE, in init_sqe() [all …]
|