/freebsd/sys/ofed/drivers/infiniband/ulp/sdp/ |
H A D | sdp_rx.c | 43 sdp_handle_disconn(struct sdp_sock *ssk) in sdp_handle_disconn() argument 46 sdp_dbg(ssk->socket, "%s\n", __func__); in sdp_handle_disconn() 48 SDP_WLOCK_ASSERT(ssk); in sdp_handle_disconn() 49 if (TCPS_HAVERCVDFIN(ssk->state) == 0) in sdp_handle_disconn() 50 socantrcvmore(ssk->socket); in sdp_handle_disconn() 52 switch (ssk->state) { in sdp_handle_disconn() 55 ssk->state = TCPS_CLOSE_WAIT; in sdp_handle_disconn() 60 sdp_dbg(ssk->socket, in sdp_handle_disconn() 64 sdp_cancel_dreq_wait_timeout(ssk); in sdp_handle_disconn() 65 ssk->qp_active = 0; in sdp_handle_disconn() [all …]
|
H A D | sdp_main.c | 113 sdp_destroy_cma(struct sdp_sock *ssk) in sdp_destroy_cma() argument 116 if (ssk->id == NULL) in sdp_destroy_cma() 118 rdma_destroy_id(ssk->id); in sdp_destroy_cma() 119 ssk->id = NULL; in sdp_destroy_cma() 123 sdp_pcbbind(struct sdp_sock *ssk, struct sockaddr *nam, struct ucred *cred) in sdp_pcbbind() argument 129 SDP_WLOCK_ASSERT(ssk); in sdp_pcbbind() 131 if (ssk->lport != 0 || ssk->laddr != INADDR_ANY) in sdp_pcbbind() 134 SDP_WUNLOCK(ssk); in sdp_pcbbind() 135 if (ssk->id == NULL) in sdp_pcbbind() 136 ssk->id = rdma_create_id(&init_net, sdp_cma_handler, ssk, RDMA_PS_SDP, IB_QPT_RC); in sdp_pcbbind() [all …]
|
H A D | sdp_tx.c | 41 static int sdp_process_tx_cq(struct sdp_sock *ssk); 45 sdp_xmit_poll(struct sdp_sock *ssk, int force) in sdp_xmit_poll() argument 49 SDP_WLOCK_ASSERT(ssk); in sdp_xmit_poll() 50 sdp_prf(ssk->socket, NULL, "%s", __func__); in sdp_xmit_poll() 54 if (!callout_pending(&ssk->tx_ring.timer)) in sdp_xmit_poll() 55 callout_reset(&ssk->tx_ring.timer, SDP_TX_POLL_TIMEOUT, in sdp_xmit_poll() 56 sdp_poll_tx_timeout, ssk); in sdp_xmit_poll() 59 if (force || (++ssk->tx_ring.poll_cnt & (SDP_TX_POLL_MODER - 1)) == 0) in sdp_xmit_poll() 60 wc_processed = sdp_process_tx_cq(ssk); in sdp_xmit_poll() 66 sdp_post_send(struct sdp_sock *ssk, struct mbuf *mb) in sdp_post_send() argument [all …]
|
H A D | sdp_cma.c | 85 struct sdp_sock *ssk; in sdp_init_qp() local 90 ssk = sdp_sk(sk); in sdp_init_qp() 91 ssk->max_sge = sdp_get_max_dev_sge(device); in sdp_init_qp() 92 sdp_dbg(sk, "Max sges: %d\n", ssk->max_sge); in sdp_init_qp() 94 qp_init_attr.cap.max_send_sge = MIN(ssk->max_sge, SDP_MAX_SEND_SGES); in sdp_init_qp() 98 qp_init_attr.cap.max_recv_sge = MIN(ssk->max_sge, SDP_MAX_RECV_SGES); in sdp_init_qp() 102 ssk->sdp_dev = ib_get_client_data(device, &sdp_client); in sdp_init_qp() 103 if (!ssk->sdp_dev) { in sdp_init_qp() 109 rc = sdp_rx_ring_create(ssk, device); in sdp_init_qp() 113 rc = sdp_tx_ring_create(ssk, device); in sdp_init_qp() [all …]
|
H A D | sdp_bcopy.c | 107 sdp_nagle_off(struct sdp_sock *ssk, struct mbuf *mb) in sdp_nagle_off() argument 118 (ssk->flags & SDP_NODELAY) || in sdp_nagle_off() 119 !ssk->nagle_last_unacked || in sdp_nagle_off() 120 mb->m_pkthdr.len >= ssk->xmit_size_goal / 4 || in sdp_nagle_off() 124 unsigned long mseq = ring_head(ssk->tx_ring); in sdp_nagle_off() 125 ssk->nagle_last_unacked = mseq; in sdp_nagle_off() 127 if (!callout_pending(&ssk->nagle_timer)) { in sdp_nagle_off() 128 callout_reset(&ssk->nagle_timer, SDP_NAGLE_TIMEOUT, in sdp_nagle_off() 129 sdp_nagle_timeout, ssk); in sdp_nagle_off() 130 sdp_dbg_data(ssk->socket, "Starting nagle timer\n"); in sdp_nagle_off() [all …]
|
H A D | sdp_zcopy.c | 53 struct sdp_sock *ssk = sdp_sk(sk); in sdp_post_srcavail() local 60 WARN_ON(ssk->tx_sa); in sdp_post_srcavail() 83 ssk->tx_sa = tx_sa; in sdp_post_srcavail() 87 payload_len = MIN(payload_len, ssk->xmit_size_goal - sizeof(struct sdp_srcah)); in sdp_post_srcavail() 103 mb_entail(sk, ssk, mb); in sdp_post_srcavail() 105 ssk->write_seq += payload_len; in sdp_post_srcavail() 118 struct sdp_sock *ssk = sdp_sk(sk); in sdp_post_srcavail_cancel() local 121 sdp_dbg_data(ssk->socket, "Posting srcavail cancel\n"); in sdp_post_srcavail_cancel() 124 mb_entail(sk, ssk, mb); in sdp_post_srcavail_cancel() 126 sdp_post_sends(ssk, 0); in sdp_post_srcavail_cancel() [all …]
|
H A D | sdp.h | 148 #define rx_ring_posted(ssk) ring_posted(ssk->rx_ring) argument 150 #define tx_ring_posted(ssk) (ring_posted(ssk->tx_ring) + \ argument 151 (ssk->tx_ring.rdma_inflight ? ssk->tx_ring.rdma_inflight->busy : 0)) 153 #define tx_ring_posted(ssk) ring_posted(ssk->tx_ring) argument 220 struct sdp_sock *ssk; member 280 #define tx_credits(ssk) (atomic_read(&ssk->tx_ring.credits)) argument 367 #define rcv_nxt(ssk) atomic_read(&(ssk->rcv_nxt)) argument 372 #define mseq_ack(ssk) (atomic_read(&ssk->mseq_ack)) argument 379 #define remote_credits(ssk) (atomic_read(&ssk->remote_credits)) argument 406 #define SDP_RLOCK(ssk) rw_rlock(&(ssk)->lock) argument [all …]
|
H A D | sdp_proc.c | 61 struct sdp_sock *ssk; in sdp_get_idx() local 64 list_for_each_entry(ssk, &sock_list, sock_list) { in sdp_get_idx() 66 return ssk; in sdp_get_idx()
|
/freebsd/contrib/unbound/dnscrypt/ |
H A D | dnscrypt.c | 1035 struct shared_secret_cache_key* ssk = (struct shared_secret_cache_key*)k; in dnsc_shared_secrets_sizefunc() local 1037 + lock_get_mem(&ssk->entry.lock); in dnsc_shared_secrets_sizefunc() 1039 (void)ssk; /* otherwise ssk is unused if no threading, or fixed locksize */ in dnsc_shared_secrets_sizefunc() 1052 struct shared_secret_cache_key* ssk = (struct shared_secret_cache_key*)k; in dnsc_shared_secrets_delkeyfunc() local 1053 lock_rw_destroy(&ssk->entry.lock); in dnsc_shared_secrets_delkeyfunc() 1054 free(ssk); in dnsc_shared_secrets_delkeyfunc()
|