/linux/net/rds/ |
H A D | ib_send.c | 151 struct rds_ib_send_work *send, in rds_ib_send_unmap_op() argument 157 switch (send->s_wr.opcode) { in rds_ib_send_unmap_op() 159 if (send->s_op) { in rds_ib_send_unmap_op() 160 rm = container_of(send->s_op, struct rds_message, data); in rds_ib_send_unmap_op() 161 rds_ib_send_unmap_data(ic, send->s_op, wc_status); in rds_ib_send_unmap_op() 166 if (send->s_op) { in rds_ib_send_unmap_op() 167 rm = container_of(send->s_op, struct rds_message, rdma); in rds_ib_send_unmap_op() 168 rds_ib_send_unmap_rdma(ic, send->s_op, wc_status); in rds_ib_send_unmap_op() 173 if (send->s_op) { in rds_ib_send_unmap_op() 174 rm = container_of(send->s_op, struct rds_message, atomic); in rds_ib_send_unmap_op() [all …]
|
/linux/tools/testing/selftests/net/ |
H A D | udpgso.c | 65 int tlen; /* send() buffer size, may exceed mss */ 66 bool tfail; /* send() call is expected to fail */ 70 bool v6_ext_hdr; /* send() dgrams with IPv6 extension headers */ 85 /* no GSO: send a single byte */ 90 /* no GSO: send a single MSS */ 95 /* no GSO: send a single MSS + 1B: fail */ 100 /* send a single MSS: will fall back to no GSO */ 106 /* send a single MSS + 1B */ 113 /* send exactly 2 MSS */ 119 /* send 2 MSS + 1B */ [all …]
|
H A D | tls.c | 230 /* Send some data through with ULP but no keys */ 239 EXPECT_EQ(send(self->fd, test_str, send_len, 0), send_len); in TEST_F() 284 EXPECT_EQ(send(self->fd, test_str, send_len, 0), -1); 443 EXPECT_EQ(send(self->fd, test_str, to_send, 0), to_send); in chunked_sendfile() 517 EXPECT_GE(send(self->fd, buf, send_len, 0), 0); in TEST_F() 529 EXPECT_EQ(send(self->fd, test_str, send_len, 0), send_len); in TEST_F() 540 EXPECT_EQ(send(self->fd, test_str, send_len, MSG_MORE), send_len); in TEST_F() 542 EXPECT_EQ(send(self->fd, test_str, send_len, 0), send_len); in TEST_F() 554 EXPECT_EQ(send(self->fd, test_str, send_len, MSG_MORE), send_len); in TEST_F() 564 EXPECT_EQ(send(sel in TEST_F() [all...] |
/linux/drivers/isdn/hardware/mISDN/ |
H A D | netjet.c | 76 struct tiger_dma send; member 182 bc->bch.nr, fill, cnt, idx, card->send.idx); in fill_mem() 189 val = card->send.start[idx]; in fill_mem() 192 card->send.start[idx++] = val; in fill_mem() 193 if (idx >= card->send.size) in fill_mem() 209 fill_mem(bc, 0, card->send.size, 0xff); in mode_tiger() 228 bc->free = card->send.size / 2; in mode_tiger() 242 bc->free = card->send.size / 2; in mode_tiger() 259 card->send.dmacur = inl(card->base + NJ_DMA_READ_ADR); in mode_tiger() 261 card->send.idx = (card->send.dmacur - card->send.dmastart) >> 2; in mode_tiger() [all …]
|
/linux/tools/testing/selftests/bpf/ |
H A D | test_flow_dissector.sh | 113 # Send 10 IPv4/UDP packets from port 8. Filter should not drop any. 115 # Send 10 IPv4/UDP packets from port 9. Filter should drop all. 117 # Send 10 IPv4/UDP packets from port 10. Filter should not drop any. 121 # Send 10 IPv4/UDP packets from port 8. Filter should not drop any. 123 # Send 10 IPv4/UDP packets from port 9. Filter should drop all. 125 # Send 10 IPv4/UDP packets from port 10. Filter should not drop any. 129 # Send 10 IPv4/IPv4/UDP packets from port 8. Filter should not drop any. 132 # Send 10 IPv4/IPv4/UDP packets from port 9. Filter should drop all. 135 # Send 10 IPv4/IPv4/UDP packets from port 10. Filter should not drop any. 140 # Send 10 IPv4/GRE/IPv4/UDP packets from port 8. Filter should not drop any. [all …]
|
/linux/net/llc/ |
H A D | llc_s_ac.c | 46 * @skb: the event to send 69 * llc_sap_action_send_xid_c - send XID PDU as response to XID REQ 71 * @skb: the event to send 73 * Send a XID command PDU to MAC layer in response to a XID REQUEST 94 * llc_sap_action_send_xid_r - send XID PDU resp to MAC for received XID 96 * @skb: the event to send 98 * Send XID response PDU to MAC in response to an earlier received XID 125 * llc_sap_action_send_test_c - send TEST PDU to MAC in resp to TEST REQ 127 * @skb: the event to send 129 * Send a TEST command PDU to the MAC layer in response to a TEST REQUEST [all …]
|
/linux/include/rdma/ |
H A D | rdmavt_qp.h | 60 * RVT_S_SIGNAL_REQ_WR - set if QP send WRs contain completion signaled 61 * RVT_S_BUSY - send tasklet is processing the QP 70 * RVT_S_WAIT_DMA - waiting for send DMA queue to drain before generating 71 * next send completion entry not via send DMA 72 * RVT_S_WAIT_PIO - waiting for a send buffer to be available 76 * RVT_S_WAIT_PSN - waiting for a packet to exit the send DMA queue 78 * RVT_S_SEND_ONE - send one packet, request ACK, then wait for ACK 79 * RVT_S_ECN - a BECN was queued to the send engine 116 * Wait flags that would prevent send wor [all...] |
H A D | ib_mad.h | 451 * @seg_count: The number of RMPP segments allocated for this send. 532 * @mad_send_wc: Send work completion information on the sent MAD. 540 * @send_buf: Send buffer if found, else NULL 543 * MADs received in response to a send request operation will be handed to 544 * the user before the send operation completes. All data buffers given 586 * ib_mad_send_wc - MAD send completion information. 587 * @send_buf: Send MAD data buffer associated with the send MAD request. 624 * for the corresponding send request. 655 * ib_register_mad_agent - Register to send/receive MADs. 663 * @rmpp_version: If set, indicates that the client will send [all …]
|
/linux/tools/testing/selftests/net/netfilter/ |
H A D | xt_string.sh | 51 send() { # (offset) function 67 send 0 68 send $((1000 - patlen)) 76 send 1000 77 send $((1400 - patlen)) 85 send $((1500 - patlen)) 93 send 1495 101 send 1500 109 send $((1600 - patlen)) 117 send $((1600 - patlen + 1)) [all …]
|
/linux/drivers/media/pci/intel/ipu6/ |
H A D | ipu6-fw-isys.h | 32 /* Single dedicated send queue for proxy interface */ 36 /* Send queues layout */ 388 * @send_irq_sof_discarded: send irq on discarded frame sof response 390 * and send the response 392 * whether to send the response 393 * @send_irq_eof_discarded: send irq on discarded frame eof response 395 * and send the response 397 * whether to send the response 398 * @send_resp_sof_discarded: send response for discarded frame sof detected, 400 * @send_resp_eof_discarded: send response for discarded frame eof detected, [all …]
|
/linux/drivers/misc/vmw_vmci/ |
H A D | vmci_route.c | 37 * them again when we do the actual send to ensure that we do in vmci_route() 50 * cannot send it to the hypervisor. It must come in vmci_route() 57 * We must be acting as a guest in order to send to in vmci_route() 63 /* And we cannot send if the source is the host context. */ in vmci_route() 79 /* Send from local client down to the hypervisor. */ in vmci_route() 88 * guest, then we need to send it down to the host. in vmci_route() 98 * may send vmci event datagrams to the host in vmci_route() 99 * itself, but it will never send datagrams to in vmci_route() 116 /* Send it from local client down to the host. */ in vmci_route() 186 * send it down, since we have no guest in vmci_route() [all …]
|
/linux/Documentation/crypto/ |
H A D | userspace-if.rst | 62 send()/write() system call family. The result of the cipher operation is 79 particular cipher instance. When invoking send/write or recv/read 80 system calls to send data to the kernel or obtain data from the 88 the input buffer used for the send/write system call and the output 120 Using the send() system call, the application provides the data that 121 should be processed with the message digest. The send system call allows 124 - MSG_MORE: If this flag is set, the send system call acts like a 126 calculated. If the flag is not set, the send system call calculates 154 Before data can be sent to the kernel using the write/send system call 165 send/recv system call family. That cmsghdr data structure holds the [all …]
|
/linux/Documentation/networking/ |
H A D | rds.rst | 55 Sockets must be bound before you can send or receive data. 79 setsockopt(SOL_SOCKET): send and receive buffer size 80 RDS honors the send and receive buffer size socket options. 90 mark the port as "congested" and send a congestion update to 104 An attempt to send a message that exceeds SO_SNDSIZE will 107 An attempt to send a message that would take the total number 111 An attempt to send a message to a destination that is marked 135 POLLOUT is a little harder. Since you can essentially send 137 there's room on the send queue (ie the number of bytes queued 152 it detects a timeout. For instance, if it tried to send a message, [all …]
|
H A D | tls.rst | 69 For example, we can send an encrypted hello world record as follows: 74 send(sock, msg, strlen(msg)); 76 send() data is directly encrypted from the userspace buffer provided 77 to the encrypted kernel send buffer if possible. 79 The sendfile system call will send the file's data over TLS records of maximum 88 TLS records are created and sent after each send() call, unless 93 This buffer is allocated at the time send() is called, such that 94 either the entire send() call will return -ENOMEM (or block waiting 95 for memory), or the encryption will always succeed. If send() returns 123 Send TLS control messages [all …]
|
/linux/drivers/infiniband/hw/hfi1/ |
H A D | pio.c | 13 * Send Context functions 32 /* global control of PIO send */ 89 /* number of send context memory pools */ 92 /* Send Context Size (SCS) wildcards */ 96 /* Send Context Count (SCC) wildcards */ 101 /* Send Context Size (SCS) constants */ 109 /* default send context sizes */ 122 /* send context memory pool configuration */ 177 * Read the send context memory pool configuration and send context 179 * counts and sizes for the send context types. [all …]
|
H A D | pio.h | 8 /* send context types */ 13 #define SC_MAX 4 /* count of send context types */ 15 /* invalid send context index */ 39 struct send_context *sc;/* back pointer to owning send context */ 55 /* per-NUMA send context */ 97 /* send context flags */ 123 /* send context configuration sizes (one per type) */ 132 * Since the mapping now allows for non-uniform send contexts per vl, the 133 * number of send contexts for a vl is either the vl_scontexts[vl] or 144 * For the case where n > nactual, the send contexts are assigned [all …]
|
/linux/drivers/gpu/drm/amd/display/dc/dio/dcn30/ |
H A D | dcn30_dio_stream_encoder.c | 55 uint32_t cont, send, line; in enc3_update_hdmi_info_packet() local 67 /* send packet(s) every frame */ in enc3_update_hdmi_info_packet() 68 send = 1; in enc3_update_hdmi_info_packet() 69 /* select line number to send packets on */ in enc3_update_hdmi_info_packet() 73 send = 0; in enc3_update_hdmi_info_packet() 84 HDMI_GENERIC0_SEND, send); in enc3_update_hdmi_info_packet() 91 HDMI_GENERIC1_SEND, send); in enc3_update_hdmi_info_packet() 98 HDMI_GENERIC2_SEND, send); in enc3_update_hdmi_info_packet() 105 HDMI_GENERIC3_SEND, send); in enc3_update_hdmi_info_packet() 112 HDMI_GENERIC4_SEND, send); in enc3_update_hdmi_info_packet() [all …]
|
/linux/net/sunrpc/xprtrdma/ |
H A D | svc_rdma_sendto.c | 53 * a Send WR conveying the transport header and the RPC message itself to 68 * The logic here depends on Send Queue and completion ordering. Since 69 * the Send WR is always posted last, it will always complete last. Thus 78 * When the Send WR is constructed, it also gets its own svc_rdma_send_ctxt. 80 * ctxt, the Send WR is posted, and sendto returns. 82 * The svc_rdma_send_ctxt is presented when the Send WR completes. The 83 * Send completion handler finally releases the Reply's pages. 85 * This mechanism also assumes that completions on the transport's Send 87 * and Send completion running at the same time could release pages that 92 * - If the Send WR is posted successfully, it will either complete [all …]
|
/linux/net/atm/ |
H A D | raw.c | 59 return vcc->dev->ops->send(vcc, skb); in atm_send_aal0() 67 vcc->send = atm_send_aal0; in atm_init_aal0() 77 vcc->send = vcc->dev->ops->send_bh; in atm_init_aal34() 79 vcc->send = vcc->dev->ops->send; in atm_init_aal34() 89 vcc->send = vcc->dev->ops->send_bh; in atm_init_aal5() 91 vcc->send = vcc->dev->ops->send; in atm_init_aal5()
|
/linux/Documentation/devicetree/bindings/mailbox/ |
H A D | apple,mailbox.yaml | 17 One of the two FIFOs is used to send data to a co-processor while the other 52 - description: send fifo is empty interrupt 53 - description: send fifo is not empty interrupt 59 - const: send-empty 60 - const: send-not-empty 85 interrupt-names = "send-empty", "send-not-empty",
|
/linux/Documentation/infiniband/ |
H A D | tag_matching.rst | 6 source send operations to destination receives. The following parameters must 14 The ordering rules require that when more than one pair of send and receive 15 message envelopes may match, the pair that includes the earliest posted-send 31 1. The Eager protocol- the complete message is sent when the send is 32 processed by the sender. A completion send is received in the send_cq 46 to the MPI receive routines in the posted receive list and posts send messages 47 using the MPI send routines. The head of the posted receive list may be 50 When send is initiated and arrives at the receive side, if there is no
|
/linux/tools/testing/selftests/net/packetdrill/ |
H A D | tcp_zerocopy_closed.pkt | 2 // send with MSG_ZEROCOPY on a non-established socket 4 // verify that a send in state TCP_CLOSE correctly aborts the zerocopy 7 // First send on a closed socket and wait for (absent) notification. 8 // Then connect and send and verify that notification nr. is zero. 15 +0 send(4, ..., 4000, MSG_ZEROCOPY) = -1 EPIPE (Broken pipe) 28 +0 send(4, ..., 4000, MSG_ZEROCOPY) = 4000
|
/linux/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_guc_ct.c | 64 * | 0x0000 | H2G `CTB Descriptor`_ (send) | | 68 * | 0x1000 | H2G `CT Buffer`_ (send) | n*4K | 80 * of the receive buffer (relative to the send) to ensure a G2H response 142 spin_lock_init(&ct->ctbs.send.lock); in intel_guc_ct_init_early() 218 static int ct_register_buffer(struct intel_guc_ct *ct, bool send, in ct_register_buffer() argument 223 err = intel_guc_self_cfg64(ct_to_guc(ct), send ? in ct_register_buffer() 230 err = intel_guc_self_cfg64(ct_to_guc(ct), send ? in ct_register_buffer() 237 err = intel_guc_self_cfg32(ct_to_guc(ct), send ? in ct_register_buffer() 244 send ? "SEND" : "RECV", ERR_PTR(err)); in ct_register_buffer() 284 /* store pointers to desc and cmds for send ctb */ in intel_guc_ct_init() [all …]
|
/linux/Documentation/hid/ |
H A D | uhid.rst | 48 The first thing you should do is send a UHID_CREATE2 event. This will 62 If you want to send data on the interrupt channel to the HID subsystem, you send 63 a HID_INPUT2 event with your raw data payload. If the kernel wants to send data 74 If your device disconnects, you should send a UHID_DESTROY event. This will 75 unregister the device. You can now send UHID_CREATE2 again to register a new 91 This creates the internal HID device. No I/O is possible until you send this 103 You must send UHID_CREATE2 before sending input to the kernel! This event 159 send UHID_INPUT2 events to the kernel. 166 This is sent if the HID device driver wants to send raw data to the I/O 180 GET_REPORT (or if it fails), you must send a UHID_GET_REPORT_REPLY to the [all …]
|
/linux/drivers/gpu/drm/amd/pm/powerplay/smumgr/ |
H A D | smu9_smumgr.c | 88 * Send a message to the SMC, and do not wait for its response. 90 * @param msg the message to send. 108 * Send a message to the SMC, and wait for its response. 110 * @param msg the message to send. 129 dev_err(adev->dev, "Failed to send message: 0x%x, ret value: 0x%x\n", msg, ret); in smu9_send_msg_to_smc() 135 * Send a message to the SMC with parameter 137 * @param msg: the message to send. 138 * @param parameter: the parameter to send
|