| /linux/include/linux/surface_aggregator/ |
| H A D | controller.h | 146 struct ssam_request_sync **rqst, 149 void ssam_request_sync_free(struct ssam_request_sync *rqst); 151 int ssam_request_sync_init(struct ssam_request_sync *rqst, 163 static inline void ssam_request_sync_set_data(struct ssam_request_sync *rqst, in ssam_request_sync_set_data() argument 166 ssh_request_set_data(&rqst->base, ptr, len); in ssam_request_sync_set_data() 178 static inline void ssam_request_sync_set_resp(struct ssam_request_sync *rqst, in ssam_request_sync_set_resp() argument 181 rqst->resp = resp; in ssam_request_sync_set_resp() 185 struct ssam_request_sync *rqst); 204 static inline int ssam_request_sync_wait(struct ssam_request_sync *rqst) in ssam_request_sync_wait() argument 206 wait_for_completion(&rqst->comp); in ssam_request_sync_wait() [all …]
|
| /linux/drivers/platform/surface/aggregator/ |
| H A D | ssh_request_layer.c | 86 static u16 ssh_request_get_rqid(struct ssh_request *rqst) in ssh_request_get_rqid() argument 88 return get_unaligned_le16(rqst->packet.data.ptr in ssh_request_get_rqid() 92 static u32 ssh_request_get_rqid_safe(struct ssh_request *rqst) in ssh_request_get_rqid_safe() argument 94 if (!rqst->packet.data.ptr) in ssh_request_get_rqid_safe() 97 return ssh_request_get_rqid(rqst); in ssh_request_get_rqid_safe() 100 static void ssh_rtl_queue_remove(struct ssh_request *rqst) in ssh_rtl_queue_remove() argument 102 struct ssh_rtl *rtl = ssh_request_rtl(rqst); in ssh_rtl_queue_remove() 106 if (!test_and_clear_bit(SSH_REQUEST_SF_QUEUED_BIT, &rqst->state)) { in ssh_rtl_queue_remove() 111 list_del(&rqst->node); in ssh_rtl_queue_remove() 114 ssh_request_put(rqst); in ssh_rtl_queue_remove() [all …]
|
| H A D | ssh_msgb.h | 173 const struct ssam_request *rqst) in msgb_push_cmd() argument 182 msgb_push_frame(msgb, type, sizeof(struct ssh_command) + rqst->length, seq); in msgb_push_cmd() 191 __msgb_push_u8(msgb, rqst->target_category); /* Target category. */ in msgb_push_cmd() 192 __msgb_push_u8(msgb, rqst->target_id); /* Target ID. */ in msgb_push_cmd() 194 __msgb_push_u8(msgb, rqst->instance_id); /* Instance ID. */ in msgb_push_cmd() 196 __msgb_push_u8(msgb, rqst->command_id); /* Command ID. */ in msgb_push_cmd() 199 msgb_push_buf(msgb, rqst->payload, rqst->length); in msgb_push_cmd()
|
| H A D | controller.c | 1530 static void ssam_request_sync_complete(struct ssh_request *rqst, in ssam_request_sync_complete() argument 1534 struct ssh_rtl *rtl = ssh_request_rtl(rqst); in ssam_request_sync_complete() 1537 r = container_of(rqst, struct ssam_request_sync, base); in ssam_request_sync_complete() 1569 static void ssam_request_sync_release(struct ssh_request *rqst) in ssam_request_sync_release() argument 1571 complete_all(&container_of(rqst, struct ssam_request_sync, base)->comp); in ssam_request_sync_release() 1600 struct ssam_request_sync **rqst, in ssam_request_sync_alloc() argument 1605 *rqst = kzalloc(sizeof(**rqst) + msglen, flags); in ssam_request_sync_alloc() 1606 if (!*rqst) in ssam_request_sync_alloc() 1609 buffer->ptr = (u8 *)(*rqst + 1); in ssam_request_sync_alloc() 1631 void ssam_request_sync_free(struct ssam_request_sync *rqst) in ssam_request_sync_free() argument [all …]
|
| H A D | ssh_request_layer.h | 121 static inline struct ssh_rtl *ssh_request_rtl(struct ssh_request *rqst) in ssh_request_rtl() argument 125 ptl = READ_ONCE(rqst->packet.ptl); in ssh_request_rtl() 129 int ssh_rtl_submit(struct ssh_rtl *rtl, struct ssh_request *rqst); 130 bool ssh_rtl_cancel(struct ssh_request *rqst, bool pending); 140 int ssh_request_init(struct ssh_request *rqst, enum ssam_request_flags flags,
|
| /linux/net/sunrpc/xprtrdma/ |
| H A D | backchannel.c | 57 static int rpcrdma_bc_marshal_reply(struct rpc_rqst *rqst) in rpcrdma_bc_marshal_reply() argument 59 struct rpcrdma_xprt *r_xprt = rpcx_to_rdmax(rqst->rq_xprt); in rpcrdma_bc_marshal_reply() 60 struct rpcrdma_req *req = rpcr_to_rdmar(rqst); in rpcrdma_bc_marshal_reply() 65 rdmab_data(req->rl_rdmabuf), rqst); in rpcrdma_bc_marshal_reply() 70 *p++ = rqst->rq_xid; in rpcrdma_bc_marshal_reply() 79 &rqst->rq_snd_buf, rpcrdma_noch_pullup)) in rpcrdma_bc_marshal_reply() 82 trace_xprtrdma_cb_reply(r_xprt, rqst); in rpcrdma_bc_marshal_reply() 98 int xprt_rdma_bc_send_reply(struct rpc_rqst *rqst) in xprt_rdma_bc_send_reply() argument 100 struct rpc_xprt *xprt = rqst->rq_xprt; in xprt_rdma_bc_send_reply() 102 struct rpcrdma_req *req = rpcr_to_rdmar(rqst); in xprt_rdma_bc_send_reply() [all …]
|
| H A D | rpc_rdma.c | 127 struct rpc_rqst *rqst) in rpcrdma_args_inline() argument 129 struct xdr_buf *xdr = &rqst->rq_snd_buf; in rpcrdma_args_inline() 159 struct rpc_rqst *rqst) in rpcrdma_results_inline() argument 161 return rqst->rq_rcv_buf.buflen <= r_xprt->rx_ep->re_max_inline_recv; in rpcrdma_results_inline() 170 const struct rpc_rqst *rqst) in rpcrdma_nonpayload_inline() argument 172 const struct xdr_buf *buf = &rqst->rq_rcv_buf; in rpcrdma_nonpayload_inline() 297 struct rpc_rqst *rqst, in rpcrdma_encode_read_list() argument 309 pos = rqst->rq_snd_buf.head[0].iov_len; in rpcrdma_encode_read_list() 312 rpcrdma_xdr_cursor_init(&cur, &rqst->rq_snd_buf, pos, rtype); in rpcrdma_encode_read_list() 322 trace_xprtrdma_chunk_read(rqst->rq_task, pos, mr, in rpcrdma_encode_read_list() [all …]
|
| H A D | svc_rdma_backchannel.c | 76 struct rpc_rqst *rqst, in svc_rdma_bc_sendto() argument 84 &rqst->rq_snd_buf); in svc_rdma_bc_sendto() 91 get_page(virt_to_page(rqst->rq_buffer)); in svc_rdma_bc_sendto() 103 struct rpc_rqst *rqst = task->tk_rqstp; in xprt_rdma_bc_allocate() local 104 size_t size = rqst->rq_callsize; in xprt_rdma_bc_allocate() 116 rqst->rq_buffer = page_address(page); in xprt_rdma_bc_allocate() 118 rqst->rq_rbuffer = kmalloc(rqst->rq_rcvsize, GFP_NOIO | __GFP_NOWARN); in xprt_rdma_bc_allocate() 119 if (!rqst->rq_rbuffer) { in xprt_rdma_bc_allocate() 129 struct rpc_rqst *rqst = task->tk_rqstp; in xprt_rdma_bc_free() local 131 put_page(virt_to_page(rqst->rq_buffer)); in xprt_rdma_bc_free() [all …]
|
| H A D | transport.c | 521 struct rpc_rqst *rqst = &req->rl_slot; in xprt_rdma_alloc_slot() local 523 if (!xprt_wake_up_backlog(xprt, rqst)) { in xprt_rdma_alloc_slot() 524 memset(rqst, 0, sizeof(*rqst)); in xprt_rdma_alloc_slot() 537 xprt_rdma_free_slot(struct rpc_xprt *xprt, struct rpc_rqst *rqst) in xprt_rdma_free_slot() argument 542 rpcrdma_reply_put(&r_xprt->rx_buf, rpcr_to_rdmar(rqst)); in xprt_rdma_free_slot() 543 if (!xprt_wake_up_backlog(xprt, rqst)) { in xprt_rdma_free_slot() 544 memset(rqst, 0, sizeof(*rqst)); in xprt_rdma_free_slot() 545 rpcrdma_buffer_put(&r_xprt->rx_buf, rpcr_to_rdmar(rqst)); in xprt_rdma_free_slot() 573 struct rpc_rqst *rqst = task->tk_rqstp; in xprt_rdma_allocate() local 574 struct rpcrdma_xprt *r_xprt = rpcx_to_rdmax(rqst->rq_xprt); in xprt_rdma_allocate() [all …]
|
| /linux/drivers/hid/surface-hid/ |
| H A D | surface_hid.c | 46 struct ssam_request rqst; in ssam_hid_get_descriptor() local 59 rqst.target_category = shid->uid.category; in ssam_hid_get_descriptor() 60 rqst.target_id = shid->uid.target; in ssam_hid_get_descriptor() 61 rqst.command_id = SURFACE_HID_CID_GET_DESCRIPTOR; in ssam_hid_get_descriptor() 62 rqst.instance_id = shid->uid.instance; in ssam_hid_get_descriptor() 63 rqst.flags = SSAM_REQUEST_HAS_RESPONSE; in ssam_hid_get_descriptor() 64 rqst.length = sizeof(struct surface_hid_buffer_slice); in ssam_hid_get_descriptor() 65 rqst.payload = buffer; in ssam_hid_get_descriptor() 83 status = ssam_retry(ssam_request_do_sync_onstack, shid->ctrl, &rqst, &rsp, in ssam_hid_get_descriptor() 116 struct ssam_request rqst; in ssam_hid_set_raw_report() local [all …]
|
| H A D | surface_kbd.c | 36 struct ssam_request rqst; in ssam_kbd_get_descriptor() local 40 rqst.target_category = shid->uid.category; in ssam_kbd_get_descriptor() 41 rqst.target_id = shid->uid.target; in ssam_kbd_get_descriptor() 42 rqst.command_id = SURFACE_KBD_CID_GET_DESCRIPTOR; in ssam_kbd_get_descriptor() 43 rqst.instance_id = shid->uid.instance; in ssam_kbd_get_descriptor() 44 rqst.flags = SSAM_REQUEST_HAS_RESPONSE; in ssam_kbd_get_descriptor() 45 rqst.length = sizeof(entry); in ssam_kbd_get_descriptor() 46 rqst.payload = &entry; in ssam_kbd_get_descriptor() 52 status = ssam_retry(ssam_request_do_sync_onstack, shid->ctrl, &rqst, &rsp, sizeof(entry)); in ssam_kbd_get_descriptor() 67 struct ssam_request rqst; in ssam_kbd_set_caps_led() local [all …]
|
| /linux/fs/smb/client/ |
| H A D | smb2inode.c | 178 struct smb_rqst *rqst, int *num_rqst) in set_next_compound() argument 183 smb2_set_next_command(tcon, &rqst[*num_rqst]); in set_next_compound() 185 smb2_set_related(&rqst[*num_rqst]); in set_next_compound() 221 struct smb_rqst *rqst; in smb2_compound_op() local 242 rqst = &vars->rqst[0]; in smb2_compound_op() 285 rqst[num_rqst].rq_iov = &vars->open_iov[0]; in smb2_compound_op() 286 rqst[num_rqst].rq_nvec = SMB2_CREATE_IOV_SIZE; in smb2_compound_op() 288 &rqst[num_rqst], &oplock, &vars->oparms, in smb2_compound_op() 294 smb2_set_next_command(tcon, &rqst[num_rqst]); in smb2_compound_op() 309 rqst[num_rqst].rq_iov = &vars->qi_iov; in smb2_compound_op() [all …]
|
| H A D | smb1encrypt.c | 26 static int cifs_calc_signature(struct smb_rqst *rqst, in cifs_calc_signature() argument 31 if (!rqst->rq_iov || !signature || !server) in cifs_calc_signature() 43 rqst, server, signature, in cifs_calc_signature() 48 int cifs_sign_rqst(struct smb_rqst *rqst, struct TCP_Server_Info *server, in cifs_sign_rqst() argument 53 struct smb_hdr *cifs_pdu = (struct smb_hdr *)rqst->rq_iov[0].iov_base; in cifs_sign_rqst() 78 rc = cifs_calc_signature(rqst, server, smb_signature); in cifs_sign_rqst() 87 int cifs_verify_signature(struct smb_rqst *rqst, in cifs_verify_signature() argument 94 struct smb_hdr *cifs_pdu = (struct smb_hdr *)rqst->rq_iov[0].iov_base; in cifs_verify_signature() 126 rc = cifs_calc_signature(rqst, server, what_we_think_sig_should_be); in cifs_verify_signature()
|
| H A D | smb2pdu.c | 1080 struct smb_rqst rqst; in SMB2_negotiate() local 1158 memset(&rqst, 0, sizeof(struct smb_rqst)); in SMB2_negotiate() 1159 rqst.rq_iov = iov; in SMB2_negotiate() 1160 rqst.rq_nvec = 1; in SMB2_negotiate() 1163 &rqst, &resp_buftype, flags, &rsp_iov); in SMB2_negotiate() 1614 struct smb_rqst rqst; in SMB2_sess_sendreceive() local 1623 memset(&rqst, 0, sizeof(struct smb_rqst)); in SMB2_sess_sendreceive() 1624 rqst.rq_iov = sess_data->iov; in SMB2_sess_sendreceive() 1625 rqst.rq_nvec = 2; in SMB2_sess_sendreceive() 1630 &rqst, in SMB2_sess_sendreceive() [all …]
|
| H A D | smb2proto.h | 39 int smb2_verify_signature(struct smb_rqst *rqst, 45 struct smb_rqst *rqst); 47 struct smb_rqst *rqst); 118 struct smb_rqst *rqst); 119 void smb2_set_next_command(struct cifs_tcon *tcon, struct smb_rqst *rqst); 120 void smb2_set_related(struct smb_rqst *rqst); 121 void smb2_set_replay(struct TCP_Server_Info *server, struct smb_rqst *rqst); 143 struct smb_rqst *rqst, __u8 *oplock, 145 void SMB2_open_free(struct smb_rqst *rqst); 151 struct smb_rqst *rqst, u64 persistent_fid, [all …]
|
| H A D | smbdirect.c | 418 struct smb_rqst *rqst; in smbd_send() local 451 rqst = &rqst_array[rqst_idx]; in smbd_send() 454 rqst_idx, smb_rqst_len(server, rqst)); in smbd_send() 455 for (i = 0; i < rqst->rq_nvec; i++) in smbd_send() 456 dump_smb(rqst->rq_iov[i].iov_base, rqst->rq_iov[i].iov_len); in smbd_send() 459 rqst_idx, rqst->rq_nvec, remaining_data_length, in smbd_send() 460 iov_iter_count(&rqst->rq_iter), smb_rqst_len(server, rqst)); in smbd_send() 464 for (i = 0; i < rqst->rq_nvec; i++) in smbd_send() 465 klen += rqst->rq_iov[i].iov_len; in smbd_send() 466 iov_iter_kvec(&iter, ITER_SOURCE, rqst->rq_iov, rqst->rq_nvec, klen); in smbd_send() [all …]
|
| H A D | transport.c | 214 smb_rqst_len(struct TCP_Server_Info *server, struct smb_rqst *rqst) in smb_rqst_len() argument 221 if (!is_smb1(server) && rqst->rq_nvec >= 2 && in smb_rqst_len() 222 rqst->rq_iov[0].iov_len == 4) { in smb_rqst_len() 223 iov = &rqst->rq_iov[1]; in smb_rqst_len() 224 nvec = rqst->rq_nvec - 1; in smb_rqst_len() 226 iov = rqst->rq_iov; in smb_rqst_len() 227 nvec = rqst->rq_nvec; in smb_rqst_len() 234 buflen += iov_iter_count(&rqst->rq_iter); in smb_rqst_len() 239 struct smb_rqst *rqst) in __smb_send_rqst() argument 257 rc = smbd_send(server, num_rqst, rqst); in __smb_send_rqst() [all …]
|
| H A D | smb2transport.c | 207 smb2_calc_signature(struct smb_rqst *rqst, struct TCP_Server_Info *server) in smb2_calc_signature() argument 211 struct kvec *iov = rqst->rq_iov; in smb2_calc_signature() 237 drqst = *rqst; in smb2_calc_signature() 459 smb3_calc_signature(struct smb_rqst *rqst, struct TCP_Server_Info *server) in smb3_calc_signature() argument 463 struct kvec *iov = rqst->rq_iov; in smb3_calc_signature() 471 return smb2_calc_signature(rqst, server); in smb3_calc_signature() 497 drqst = *rqst; in smb3_calc_signature() 514 smb2_sign_rqst(struct smb_rqst *rqst, struct TCP_Server_Info *server) in smb2_sign_rqst() argument 521 shdr = (struct smb2_hdr *)rqst->rq_iov[0].iov_base; in smb2_sign_rqst() 542 return smb3_calc_signature(rqst, server); in smb2_sign_rqst() [all …]
|
| H A D | smb2ops.c | 1189 struct smb_rqst *rqst; in smb2_set_ea() local 1230 rqst = vars->rqst; in smb2_set_ea() 1276 rqst[0].rq_iov = vars->open_iov; in smb2_set_ea() 1277 rqst[0].rq_nvec = SMB2_CREATE_IOV_SIZE; in smb2_set_ea() 1290 &rqst[0], &oplock, &oparms, utf16_path); in smb2_set_ea() 1293 smb2_set_next_command(tcon, &rqst[0]); in smb2_set_ea() 1297 rqst[1].rq_iov = vars->si_iov; in smb2_set_ea() 1298 rqst[1].rq_nvec = 1; in smb2_set_ea() 1316 &rqst[1], COMPOUND_FID, in smb2_set_ea() 1322 smb2_set_next_command(tcon, &rqst[1]); in smb2_set_ea() [all …]
|
| H A D | smb1transport.c | 105 cifs_setup_async_request(struct TCP_Server_Info *server, struct smb_rqst *rqst) in cifs_setup_async_request() argument 108 struct smb_hdr *hdr = (struct smb_hdr *)rqst->rq_iov[0].iov_base; in cifs_setup_async_request() 119 rc = cifs_sign_rqst(rqst, server, &mid->sequence_number); in cifs_setup_async_request() 167 struct smb_rqst rqst = { .rq_iov = iov, in cifs_check_receive() local 173 rc = cifs_verify_signature(&rqst, server, in cifs_check_receive() 192 struct smb_rqst *rqst) in cifs_setup_request() argument 195 struct smb_hdr *hdr = (struct smb_hdr *)rqst->rq_iov[0].iov_base; in cifs_setup_request() 201 rc = cifs_sign_rqst(rqst, server, &mid->sequence_number); in cifs_setup_request() 214 struct smb_rqst rqst = { in SendReceive2() local 220 &rqst, resp_buf_type, flags, resp_iov); in SendReceive2() [all …]
|
| H A D | cached_dir.c | 145 struct smb_rqst rqst[2]; in open_cached_dir() local 264 memset(rqst, 0, sizeof(rqst)); in open_cached_dir() 270 rqst[0].rq_iov = open_iov; in open_cached_dir() 271 rqst[0].rq_nvec = SMB2_CREATE_IOV_SIZE; in open_cached_dir() 286 &rqst[0], &oplock, &oparms, utf16_path); in open_cached_dir() 297 smb2_set_next_command(tcon, &rqst[0]); in open_cached_dir() 300 rqst[1].rq_iov = qi_iov; in open_cached_dir() 301 rqst[1].rq_nvec = 1; in open_cached_dir() 304 &rqst[1], COMPOUND_FID, in open_cached_dir() 312 smb2_set_related(&rqst[1]); in open_cached_dir() [all …]
|
| H A D | cifsproto.h | 95 int cifs_call_async(struct TCP_Server_Info *server, struct smb_rqst *rqst, 101 struct TCP_Server_Info *server, struct smb_rqst *rqst, 105 const int num_rqst, struct smb_rqst *rqst, 110 struct smb_rqst *rqst); 118 struct smb_rqst *rqst, struct mid_q_entry *mid, in send_cancel() argument 122 server->ops->send_cancel(ses, server, rqst, mid, xid) : 0; in send_cancel() 347 int __cifs_calc_signature(struct smb_rqst *rqst, 510 static inline int cifs_get_num_sgs(const struct smb_rqst *rqst, in cifs_get_num_sgs() argument 533 data_size = iov_iter_count(&rqst[i].rq_iter); in cifs_get_num_sgs() 540 WARN_ON_ONCE(user_backed_iter(&rqst[i].rq_iter))) in cifs_get_num_sgs() [all …]
|
| /linux/drivers/nvme/host/ |
| H A D | fc.h | 42 acc->rqst.desc_tag = cpu_to_be32(FCNVME_LSDESC_RQST); in nvme_fc_format_rsp_hdr() 43 acc->rqst.desc_len = in nvme_fc_format_rsp_hdr() 45 acc->rqst.w0.ls_cmd = rqst_ls_cmd; in nvme_fc_format_rsp_hdr() 194 struct fcnvme_ls_disconnect_assoc_rqst *rqst) in nvmefc_vldt_lsreq_discon_assoc() argument 200 else if (rqst->desc_list_len != in nvmefc_vldt_lsreq_discon_assoc() 204 else if (rqst->associd.desc_tag != cpu_to_be32(FCNVME_LSDESC_ASSOC_ID)) in nvmefc_vldt_lsreq_discon_assoc() 206 else if (rqst->associd.desc_len != in nvmefc_vldt_lsreq_discon_assoc() 210 else if (rqst->discon_cmd.desc_tag != in nvmefc_vldt_lsreq_discon_assoc() 213 else if (rqst->discon_cmd.desc_len != in nvmefc_vldt_lsreq_discon_assoc() 221 else if (rqst->discon_cmd.rsvd8[0]) in nvmefc_vldt_lsreq_discon_assoc()
|
| /linux/include/trace/events/ |
| H A D | rpcrdma.h | 598 const struct rpc_rqst *rqst 601 TP_ARGS(r_xprt, rqst), 610 __entry->xid = be32_to_cpu(rqst->rq_xid); 625 const struct rpc_rqst *rqst \ 627 TP_ARGS(r_xprt, rqst)) 797 const struct rpc_rqst *rqst = &req->rl_slot; 799 __entry->task_id = rqst->rq_task->tk_pid; 800 __entry->client_id = rqst->rq_task->tk_client->cl_clid; 856 const struct rpc_rqst *rqst = &req->rl_slot; 858 __entry->task_id = rqst->rq_task->tk_pid; [all …]
|
| /linux/drivers/platform/surface/ |
| H A D | surface_acpi_notify.c | 532 struct ssam_request *rqst, in san_rqst_fixup_suspended() argument 535 if (rqst->target_category == SSAM_SSH_TC_BAS && rqst->command_id == 0x0D) { in san_rqst_fixup_suspended() 567 struct ssam_request rqst; in san_rqst() local 575 rqst.target_category = gsb_rqst->tc; in san_rqst() 576 rqst.target_id = gsb_rqst->tid; in san_rqst() 577 rqst.command_id = gsb_rqst->cid; in san_rqst() 578 rqst.instance_id = gsb_rqst->iid; in san_rqst() 579 rqst.flags = gsb_rqst->snc ? SSAM_REQUEST_HAS_RESPONSE : 0; in san_rqst() 580 rqst.length = get_unaligned(&gsb_rqst->cdl); in san_rqst() 581 rqst.payload = &gsb_rqst->pld[0]; in san_rqst() [all …]
|