Home
last modified time | relevance | path

Searched full:req (Results 1 – 25 of 1908) sorted by relevance

12345678910>>...77

/linux/drivers/staging/greybus/
H A Daudio_apbridgea.c16 struct audio_apbridgea_set_config_request req; in gb_audio_apbridgea_set_config() local
18 req.hdr.type = AUDIO_APBRIDGEA_TYPE_SET_CONFIG; in gb_audio_apbridgea_set_config()
19 req.hdr.i2s_port = cpu_to_le16(i2s_port); in gb_audio_apbridgea_set_config()
20 req.format = cpu_to_le32(format); in gb_audio_apbridgea_set_config()
21 req.rate = cpu_to_le32(rate); in gb_audio_apbridgea_set_config()
22 req.mclk_freq = cpu_to_le32(mclk_freq); in gb_audio_apbridgea_set_config()
24 return gb_hd_output(connection->hd, &req, sizeof(req), in gb_audio_apbridgea_set_config()
33 struct audio_apbridgea_register_cport_request req; in gb_audio_apbridgea_register_cport() local
36 req.hdr.type = AUDIO_APBRIDGEA_TYPE_REGISTER_CPORT; in gb_audio_apbridgea_register_cport()
37 req.hdr.i2s_port = cpu_to_le16(i2s_port); in gb_audio_apbridgea_register_cport()
[all …]
H A Daudio_gb.c50 struct gb_audio_get_control_request req; in gb_audio_gb_get_control() local
54 req.control_id = control_id; in gb_audio_gb_get_control()
55 req.index = index; in gb_audio_gb_get_control()
58 &req, sizeof(req), &resp, sizeof(resp)); in gb_audio_gb_get_control()
72 struct gb_audio_set_control_request req; in gb_audio_gb_set_control() local
74 req.control_id = control_id; in gb_audio_gb_set_control()
75 req.index = index; in gb_audio_gb_set_control()
76 memcpy(&req.value, value, sizeof(req.value)); in gb_audio_gb_set_control()
79 &req, sizeof(req), NULL, 0); in gb_audio_gb_set_control()
86 struct gb_audio_enable_widget_request req; in gb_audio_gb_enable_widget() local
[all …]
/linux/drivers/nvme/target/
H A Dadmin-cmd.c15 static void nvmet_execute_delete_sq(struct nvmet_req *req) in nvmet_execute_delete_sq() argument
17 struct nvmet_ctrl *ctrl = req->sq->ctrl; in nvmet_execute_delete_sq()
18 u16 sqid = le16_to_cpu(req->cmd->delete_queue.qid); in nvmet_execute_delete_sq()
22 status = nvmet_report_invalid_opcode(req); in nvmet_execute_delete_sq()
38 nvmet_req_complete(req, status); in nvmet_execute_delete_sq()
41 static void nvmet_execute_create_sq(struct nvmet_req *req) in nvmet_execute_create_sq() argument
43 struct nvmet_ctrl *ctrl = req->sq->ctrl; in nvmet_execute_create_sq()
44 struct nvme_command *cmd = req->cmd; in nvmet_execute_create_sq()
53 status = nvmet_report_invalid_opcode(req); in nvmet_execute_create_sq()
80 nvmet_req_complete(req, status); in nvmet_execute_create_sq()
[all …]
H A Dio-cmd-file.c76 static ssize_t nvmet_file_submit_bvec(struct nvmet_req *req, loff_t pos, in nvmet_file_submit_bvec() argument
79 struct kiocb *iocb = &req->f.iocb; in nvmet_file_submit_bvec()
84 if (req->cmd->rw.opcode == nvme_cmd_write) { in nvmet_file_submit_bvec()
85 if (req->cmd->rw.control & cpu_to_le16(NVME_RW_FUA)) in nvmet_file_submit_bvec()
87 call_iter = req->ns->file->f_op->write_iter; in nvmet_file_submit_bvec()
90 call_iter = req->ns->file->f_op->read_iter; in nvmet_file_submit_bvec()
94 iov_iter_bvec(&iter, rw, req->f.bvec, nr_segs, count); in nvmet_file_submit_bvec()
97 iocb->ki_filp = req->ns->file; in nvmet_file_submit_bvec()
105 struct nvmet_req *req = container_of(iocb, struct nvmet_req, f.iocb); in nvmet_file_io_done() local
108 if (req->f.bvec != req->inline_bvec) { in nvmet_file_io_done()
[all …]
H A Dfabrics-cmd-auth.c34 static u8 nvmet_auth_negotiate(struct nvmet_req *req, void *d) in nvmet_auth_negotiate() argument
36 struct nvmet_ctrl *ctrl = req->sq->ctrl; in nvmet_auth_negotiate()
41 __func__, ctrl->cntlid, req->sq->qid, in nvmet_auth_negotiate()
45 req->sq->dhchap_tid = le16_to_cpu(data->t_id); in nvmet_auth_negotiate()
46 req->sq->sc_c = data->sc_c; in nvmet_auth_negotiate()
51 if (req->sq->qid) in nvmet_auth_negotiate()
55 if (nvmet_queue_tls_keyid(req->sq)) in nvmet_auth_negotiate()
59 if (!nvmet_queue_tls_keyid(req->sq)) in nvmet_auth_negotiate()
89 __func__, ctrl->cntlid, req->sq->qid); in nvmet_auth_negotiate()
93 __func__, ctrl->cntlid, req->sq->qid, in nvmet_auth_negotiate()
[all …]
H A Dzns.c69 void nvmet_execute_identify_ctrl_zns(struct nvmet_req *req) in nvmet_execute_identify_ctrl_zns() argument
71 u8 zasl = req->sq->ctrl->subsys->zasl; in nvmet_execute_identify_ctrl_zns()
72 struct nvmet_ctrl *ctrl = req->sq->ctrl; in nvmet_execute_identify_ctrl_zns()
87 status = nvmet_copy_to_sgl(req, 0, id, sizeof(*id)); in nvmet_execute_identify_ctrl_zns()
91 nvmet_req_complete(req, status); in nvmet_execute_identify_ctrl_zns()
94 void nvmet_execute_identify_ns_zns(struct nvmet_req *req) in nvmet_execute_identify_ns_zns() argument
101 if (le32_to_cpu(req->cmd->identify.nsid) == NVME_NSID_ALL) { in nvmet_execute_identify_ns_zns()
102 req->error_loc = offsetof(struct nvme_identify, nsid); in nvmet_execute_identify_ns_zns()
113 status = nvmet_req_find_ns(req); in nvmet_execute_identify_ns_zns()
117 if (nvmet_ns_revalidate(req->ns)) { in nvmet_execute_identify_ns_zns()
[all …]
H A Dio-cmd-bdev.c130 u16 blk_to_nvme_status(struct nvmet_req *req, blk_status_t blk_sts) in blk_to_nvme_status() argument
145 req->error_loc = offsetof(struct nvme_rw_command, length); in blk_to_nvme_status()
149 req->error_loc = offsetof(struct nvme_rw_command, slba); in blk_to_nvme_status()
153 req->error_loc = offsetof(struct nvme_common_command, opcode); in blk_to_nvme_status()
157 req->error_loc = offsetof(struct nvme_rw_command, nsid); in blk_to_nvme_status()
162 req->error_loc = offsetof(struct nvme_common_command, opcode); in blk_to_nvme_status()
165 switch (req->cmd->common.opcode) { in blk_to_nvme_status()
168 req->error_slba = le64_to_cpu(req->cmd->rw.slba); in blk_to_nvme_status()
171 req->error_slba = in blk_to_nvme_status()
172 le64_to_cpu(req->cmd->write_zeroes.slba); in blk_to_nvme_status()
[all …]
/linux/drivers/media/mc/
H A Dmc-request.c39 static void media_request_clean(struct media_request *req) in media_request_clean() argument
44 WARN_ON(req->state != MEDIA_REQUEST_STATE_CLEANING); in media_request_clean()
45 WARN_ON(req->updating_count); in media_request_clean()
46 WARN_ON(req->access_count); in media_request_clean()
48 list_for_each_entry_safe(obj, obj_safe, &req->objects, list) { in media_request_clean()
53 req->updating_count = 0; in media_request_clean()
54 req->access_count = 0; in media_request_clean()
55 WARN_ON(req->num_incomplete_objects); in media_request_clean()
56 req->num_incomplete_objects = 0; in media_request_clean()
57 wake_up_interruptible_all(&req->poll_wait); in media_request_clean()
[all …]
/linux/io_uring/
H A Drw.c38 static bool io_file_supports_nowait(struct io_kiocb *req, __poll_t mask) in io_file_supports_nowait() argument
41 if (req->flags & REQ_F_SUPPORT_NOWAIT) in io_file_supports_nowait()
44 if (io_file_can_poll(req)) { in io_file_supports_nowait()
47 return vfs_poll(req->file, &pt) & mask; in io_file_supports_nowait()
64 static int io_iov_buffer_select_prep(struct io_kiocb *req) in io_iov_buffer_select_prep() argument
68 struct io_rw *rw = io_kiocb_to_cmd(req, struct io_rw); in io_iov_buffer_select_prep()
73 if (io_is_compat(req->ctx)) in io_iov_buffer_select_prep()
83 static int io_import_vec(int ddir, struct io_kiocb *req, in io_import_vec() argument
100 io_is_compat(req->ctx)); in io_import_vec()
104 req->flags |= REQ_F_NEED_CLEANUP; in io_import_vec()
[all …]
H A Dpoll.c36 struct io_kiocb *req; member
73 static bool io_poll_get_ownership_slowpath(struct io_kiocb *req) in io_poll_get_ownership_slowpath() argument
82 v = atomic_fetch_or(IO_POLL_RETRY_FLAG, &req->poll_refs); in io_poll_get_ownership_slowpath()
85 return !(atomic_fetch_inc(&req->poll_refs) & IO_POLL_REF_MASK); in io_poll_get_ownership_slowpath()
94 static inline bool io_poll_get_ownership(struct io_kiocb *req) in io_poll_get_ownership() argument
96 if (unlikely(atomic_read(&req->poll_refs) >= IO_POLL_REF_BIAS)) in io_poll_get_ownership()
97 return io_poll_get_ownership_slowpath(req); in io_poll_get_ownership()
98 return !(atomic_fetch_inc(&req->poll_refs) & IO_POLL_REF_MASK); in io_poll_get_ownership()
101 static void io_poll_mark_cancelled(struct io_kiocb *req) in io_poll_mark_cancelled() argument
103 atomic_or(IO_POLL_CANCEL_FLAG, &req->poll_refs); in io_poll_mark_cancelled()
[all …]
H A Dnet.c123 int io_shutdown_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_shutdown_prep() argument
125 struct io_shutdown *shutdown = io_kiocb_to_cmd(req, struct io_shutdown); in io_shutdown_prep()
132 req->flags |= REQ_F_FORCE_ASYNC; in io_shutdown_prep()
136 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags) in io_shutdown() argument
138 struct io_shutdown *shutdown = io_kiocb_to_cmd(req, struct io_shutdown); in io_shutdown()
144 sock = sock_from_file(req->file); in io_shutdown()
149 io_req_set_res(req, ret, 0); in io_shutdown()
166 static void io_netmsg_recycle(struct io_kiocb *req, unsigned int issue_flags) in io_netmsg_recycle() argument
168 struct io_async_msghdr *hdr = req->async_data; in io_netmsg_recycle()
181 if (io_alloc_cache_put(&req->ctx->netmsg_cache, hdr)) in io_netmsg_recycle()
[all …]
/linux/tools/testing/selftests/net/tcp_ao/lib/
H A Dnetlink.c73 test_print("req buf is too small: %zu < %zu", req_sz, nl_size); in rtattr_pack()
142 } req; in __add_veth() local
146 memset(&req, 0, sizeof(req)); in __add_veth()
147 req.nh.nlmsg_len = NLMSG_LENGTH(sizeof(req.info)); in __add_veth()
148 req.nh.nlmsg_type = RTM_NEWLINK; in __add_veth()
149 req.nh.nlmsg_flags = flags; in __add_veth()
150 req.nh.nlmsg_seq = seq; in __add_veth()
151 req.info.ifi_family = AF_UNSPEC; in __add_veth()
152 req.info.ifi_change = 0xFFFFFFFF; in __add_veth()
154 if (rtattr_pack(&req.nh, sizeof(req), IFLA_IFNAME, name, strlen(name))) in __add_veth()
[all …]
/linux/drivers/peci/
H A Drequest.c88 static u8 peci_request_data_cc(struct peci_request *req) in peci_request_data_cc() argument
90 return req->rx.buf[0]; in peci_request_data_cc()
95 * @req: the PECI request that contains response data with completion code
102 int peci_request_status(struct peci_request *req) in peci_request_status() argument
104 u8 cc = peci_request_data_cc(req); in peci_request_status()
107 dev_dbg(&req->device->dev, "ret: %#02x\n", cc); in peci_request_status()
133 static int peci_request_xfer(struct peci_request *req) in peci_request_xfer() argument
135 struct peci_device *device = req->device; in peci_request_xfer()
140 ret = controller->ops->xfer(controller, device->addr, req); in peci_request_xfer()
146 static int peci_request_xfer_retry(struct peci_request *req) in peci_request_xfer_retry() argument
[all …]
/linux/crypto/
H A Dahash.c32 static int ahash_def_finup(struct ahash_request *req);
102 int crypto_hash_walk_first(struct ahash_request *req, in crypto_hash_walk_first() argument
105 walk->total = req->nbytes; in crypto_hash_walk_first()
111 walk->flags = req->base.flags; in crypto_hash_walk_first()
113 if (ahash_request_isvirt(req)) { in crypto_hash_walk_first()
114 walk->data = req->svirt; in crypto_hash_walk_first()
116 return req->nbytes; in crypto_hash_walk_first()
119 walk->sg = req->src; in crypto_hash_walk_first()
162 static inline struct shash_desc *prepare_shash_desc(struct ahash_request *req, in prepare_shash_desc() argument
165 struct shash_desc *desc = ahash_request_ctx(req); in prepare_shash_desc()
[all …]
H A Dseqiv.c21 static void seqiv_aead_encrypt_complete2(struct aead_request *req, int err) in seqiv_aead_encrypt_complete2() argument
23 struct aead_request *subreq = aead_request_ctx(req); in seqiv_aead_encrypt_complete2()
32 geniv = crypto_aead_reqtfm(req); in seqiv_aead_encrypt_complete2()
33 memcpy(req->iv, subreq->iv, crypto_aead_ivsize(geniv)); in seqiv_aead_encrypt_complete2()
41 struct aead_request *req = data; in seqiv_aead_encrypt_complete() local
43 seqiv_aead_encrypt_complete2(req, err); in seqiv_aead_encrypt_complete()
44 aead_request_complete(req, err); in seqiv_aead_encrypt_complete()
47 static int seqiv_aead_encrypt(struct aead_request *req) in seqiv_aead_encrypt() argument
49 struct crypto_aead *geniv = crypto_aead_reqtfm(req); in seqiv_aead_encrypt()
51 struct aead_request *subreq = aead_request_ctx(req); in seqiv_aead_encrypt()
[all …]
/linux/drivers/infiniband/hw/bng_re/
H A Dbng_tlv.h26 static inline u8 __get_cmdq_base_opcode(struct cmdq_base *req, u32 size) in __get_cmdq_base_opcode() argument
28 if (HAS_TLV_HEADER(req) && size > TLV_BYTES) in __get_cmdq_base_opcode()
29 return ((struct cmdq_base *)GET_TLV_DATA(req))->opcode; in __get_cmdq_base_opcode()
31 return req->opcode; in __get_cmdq_base_opcode()
34 static inline void __set_cmdq_base_opcode(struct cmdq_base *req, in __set_cmdq_base_opcode() argument
37 if (HAS_TLV_HEADER(req) && size > TLV_BYTES) in __set_cmdq_base_opcode()
38 ((struct cmdq_base *)GET_TLV_DATA(req))->opcode = val; in __set_cmdq_base_opcode()
40 req->opcode = val; in __set_cmdq_base_opcode()
43 static inline __le16 __get_cmdq_base_cookie(struct cmdq_base *req, u32 size) in __get_cmdq_base_cookie() argument
45 if (HAS_TLV_HEADER(req) && size > TLV_BYTES) in __get_cmdq_base_cookie()
[all …]
/linux/net/handshake/
H A Drequest.c66 static bool handshake_req_hash_add(struct handshake_req *req) in handshake_req_hash_add() argument
71 &req->hr_rhash, in handshake_req_hash_add()
76 static void handshake_req_destroy(struct handshake_req *req) in handshake_req_destroy() argument
78 if (req->hr_proto->hp_destroy) in handshake_req_destroy()
79 req->hr_proto->hp_destroy(req); in handshake_req_destroy()
80 rhashtable_remove_fast(&handshake_rhashtbl, &req->hr_rhash, in handshake_req_destroy()
82 kfree(req); in handshake_req_destroy()
88 struct handshake_req *req; in handshake_sk_destruct() local
90 req = handshake_req_hash_lookup(sk); in handshake_sk_destruct()
91 if (!req) in handshake_sk_destruct()
[all …]
/linux/drivers/block/drbd/
H A Ddrbd_req.c26 struct drbd_request *req; in drbd_req_new() local
28 req = mempool_alloc(&drbd_request_mempool, GFP_NOIO); in drbd_req_new()
29 if (!req) in drbd_req_new()
31 memset(req, 0, sizeof(*req)); in drbd_req_new()
33 req->rq_state = (bio_data_dir(bio_src) == WRITE ? RQ_WRITE : 0) in drbd_req_new()
36 req->device = device; in drbd_req_new()
37 req->master_bio = bio_src; in drbd_req_new()
38 req->epoch = 0; in drbd_req_new()
40 drbd_clear_interval(&req->i); in drbd_req_new()
41 req->i.sector = bio_src->bi_iter.bi_sector; in drbd_req_new()
[all …]
/linux/drivers/accel/ivpu/
H A Divpu_jsm_msg.c98 struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_REGISTER_DB }; in ivpu_jsm_register_db() local
102 req.payload.register_db.db_idx = db_id; in ivpu_jsm_register_db()
103 req.payload.register_db.jobq_base = jobq_base; in ivpu_jsm_register_db()
104 req.payload.register_db.jobq_size = jobq_size; in ivpu_jsm_register_db()
105 req.payload.register_db.host_ssid = ctx_id; in ivpu_jsm_register_db()
107 ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_REGISTER_DB_DONE, &resp, in ivpu_jsm_register_db()
117 struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_UNREGISTER_DB }; in ivpu_jsm_unregister_db() local
121 req.payload.unregister_db.db_idx = db_id; in ivpu_jsm_unregister_db()
123 ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_UNREGISTER_DB_DONE, &resp, in ivpu_jsm_unregister_db()
133 struct vpu_jsm_msg req = { .type = VPU_JSM_MSG_QUERY_ENGINE_HB }; in ivpu_jsm_get_heartbeat() local
[all …]
/linux/drivers/infiniband/hw/bnxt_re/
H A Dqplib_tlv.h48 static inline u8 __get_cmdq_base_opcode(struct cmdq_base *req, u32 size) in __get_cmdq_base_opcode() argument
50 if (HAS_TLV_HEADER(req) && size > TLV_BYTES) in __get_cmdq_base_opcode()
51 return ((struct cmdq_base *)GET_TLV_DATA(req))->opcode; in __get_cmdq_base_opcode()
53 return req->opcode; in __get_cmdq_base_opcode()
56 static inline void __set_cmdq_base_opcode(struct cmdq_base *req, in __set_cmdq_base_opcode() argument
59 if (HAS_TLV_HEADER(req) && size > TLV_BYTES) in __set_cmdq_base_opcode()
60 ((struct cmdq_base *)GET_TLV_DATA(req))->opcode = val; in __set_cmdq_base_opcode()
62 req->opcode = val; in __set_cmdq_base_opcode()
65 static inline __le16 __get_cmdq_base_cookie(struct cmdq_base *req, u32 size) in __get_cmdq_base_cookie() argument
67 if (HAS_TLV_HEADER(req) && size > TLV_BYTES) in __get_cmdq_base_cookie()
[all …]
/linux/fs/nfs/
H A Dpagelist.c37 const struct nfs_page *req; member
42 const struct nfs_page *req) in nfs_page_iter_page_init() argument
44 i->req = req; in nfs_page_iter_page_init()
50 const struct nfs_page *req = i->req; in nfs_page_iter_page_advance() local
53 i->count = (tmp < req->wb_bytes) ? tmp : req->wb_bytes; in nfs_page_iter_page_advance()
58 const struct nfs_page *req = i->req; in nfs_page_iter_page_get() local
61 if (i->count != req->wb_bytes) { in nfs_page_iter_page_get()
62 size_t base = i->count + req->wb_pgbase; in nfs_page_iter_page_get()
65 page = nfs_page_to_page(req, base); in nfs_page_iter_page_get()
102 hdr->req = nfs_list_entry(mirror->pg_list.next); in nfs_pgheader_init()
[all …]
/linux/drivers/net/ethernet/marvell/prestera/
H A Dprestera_hw.c781 /* structure that are part of req/resp fw messages */ in prestera_hw_build_tests()
1026 struct prestera_msg_port_info_req req = { in prestera_hw_port_info_get() local
1033 &req.cmd, sizeof(req), &resp.ret, sizeof(resp)); in prestera_hw_port_info_get()
1046 struct prestera_msg_switch_attr_req req = { in prestera_hw_switch_mac_set() local
1050 ether_addr_copy(req.param.mac, mac); in prestera_hw_switch_mac_set()
1053 &req.cmd, sizeof(req)); in prestera_hw_switch_mac_set()
1059 struct prestera_msg_common_req req; in prestera_hw_switch_init() local
1067 &req.cmd, sizeof(req), in prestera_hw_switch_init()
1094 struct prestera_msg_switch_attr_req req = { in prestera_hw_switch_ageing_set() local
1102 &req.cmd, sizeof(req)); in prestera_hw_switch_ageing_set()
[all …]
/linux/fs/smb/server/
H A Dtransport_ipc.c310 static int ipc_server_config_on_startup(struct ksmbd_startup_request *req) in ipc_server_config_on_startup() argument
314 ksmbd_set_fd_limit(req->file_max); in ipc_server_config_on_startup()
315 server_conf.flags = req->flags; in ipc_server_config_on_startup()
316 server_conf.signing = req->signing; in ipc_server_config_on_startup()
317 server_conf.tcp_port = req->tcp_port; in ipc_server_config_on_startup()
318 server_conf.ipc_timeout = req->ipc_timeout * HZ; in ipc_server_config_on_startup()
319 if (check_mul_overflow(req->deadtime, SMB_ECHO_INTERVAL, in ipc_server_config_on_startup()
324 server_conf.share_fake_fscaps = req->share_fake_fscaps; in ipc_server_config_on_startup()
325 ksmbd_init_domain(req->sub_auth); in ipc_server_config_on_startup()
327 if (req->smb2_max_read) in ipc_server_config_on_startup()
[all …]
/linux/include/trace/events/
H A Dio_uring.h101 * @req: pointer to a submitted request
110 TP_PROTO(struct io_kiocb *req, int fd),
112 TP_ARGS(req, fd),
116 __field( void *, req )
122 __entry->ctx = req->ctx;
123 __entry->req = req;
124 __entry->user_data = req->cqe.user_data;
128 TP_printk("ring %p, req %p, user_data 0x%llx, fd %d",
129 __entry->ctx, __entry->req, __entry->user_data, __entry->fd)
135 * @req: pointer to a submitted request
[all …]
/linux/net/sunrpc/
H A Dbackchannel_rqst.c56 static void xprt_free_allocation(struct rpc_rqst *req) in xprt_free_allocation() argument
60 dprintk("RPC: free allocations for req= %p\n", req); in xprt_free_allocation()
61 WARN_ON_ONCE(test_bit(RPC_BC_PA_IN_USE, &req->rq_bc_pa_state)); in xprt_free_allocation()
62 xbufp = &req->rq_rcv_buf; in xprt_free_allocation()
64 xbufp = &req->rq_snd_buf; in xprt_free_allocation()
66 kfree(req); in xprt_free_allocation()
94 struct rpc_rqst *req; in xprt_alloc_bc_req() local
97 req = kzalloc(sizeof(*req), gfp_flags); in xprt_alloc_bc_req()
98 if (req == NULL) in xprt_alloc_bc_req()
101 req->rq_xprt = xprt; in xprt_alloc_bc_req()
[all …]

12345678910>>...77