/freebsd/sys/dev/cxgbe/ |
H A D | t4_smt.c | 112 struct sge_wrq *wrq; in t4_write_sme() local 121 MPASS(e->wrq != NULL); in t4_write_sme() 122 wrq = e->wrq; in t4_write_sme() 123 sc = wrq->adapter; in t4_write_sme() 124 MPASS(wrq->adapter != NULL); in t4_write_sme() 132 req = start_wrq_wr(wrq, howmany(sizeof(*req), 16), &cookie); in t4_write_sme() 161 t6req = start_wrq_wr(wrq, howmany(sizeof(*t6req), 16), &cookie); in t4_write_sme() 178 commit_wrq_wr(wrq, req, &cookie); in t4_write_sme() 211 e->wrq = &sc->sge.ctrlq[0]; in t4_smt_set_switching()
|
H A D | t4_l2t.c | 172 struct sge_wrq *wrq; in t4_write_l2e() local 178 MPASS(e->wrq != NULL); in t4_write_l2e() 180 wrq = e->wrq; in t4_write_l2e() 181 sc = wrq->adapter; in t4_write_l2e() 183 req = start_wrq_wr(wrq, howmany(sizeof(*req), 16), &cookie); in t4_write_l2e() 189 commit_wrq_wr(wrq, req, &cookie); in t4_write_l2e() 231 e->wrq == (struct sge_wrq *)txq && in t4_l2t_alloc_tls() 268 e->wrq = (struct sge_wrq *)txq; in t4_l2t_alloc_tls() 302 e->wrq = &sc->sge.ctrlq[0]; in t4_l2t_alloc_switching()
|
H A D | t4_sge.c | 2139 struct sge_wrq *wrq = arg; in wrq_tx_drain() local 2140 struct sge_eq *eq = &wrq->eq; in wrq_tx_drain() 2143 if (TAILQ_EMPTY(&wrq->incomplete_wrs) && !STAILQ_EMPTY(&wrq->wr_list)) in wrq_tx_drain() 2144 drain_wrq_wr_list(wrq->adapter, wrq); in wrq_tx_drain() 2149 drain_wrq_wr_list(struct adapter *sc, struct sge_wrq *wrq) in drain_wrq_wr_list() argument 2151 struct sge_eq *eq = &wrq->eq; in drain_wrq_wr_list() 2158 MPASS(TAILQ_EMPTY(&wrq->incomplete_wrs)); in drain_wrq_wr_list() 2159 wr = STAILQ_FIRST(&wrq->wr_list); in drain_wrq_wr_list() 2171 MPASS(wr->wrq == wrq); in drain_wrq_wr_list() 2191 wrq->tx_wrs_copied++; in drain_wrq_wr_list() [all …]
|
H A D | adapter.h | 709 struct sge_wrq *wrq; member 758 struct sge_wrq wrq; member 1536 alloc_wrqe(int wr_len, struct sge_wrq *wrq) in alloc_wrqe() argument 1545 wr->wrq = wrq; in alloc_wrqe() 1564 struct sge_wrq *wrq = wr->wrq; in t4_wrq_tx() local 1566 TXQ_LOCK(wrq); in t4_wrq_tx() 1567 if (__predict_true(wrq->eq.flags & EQ_HW_ALLOCATED)) in t4_wrq_tx() 1568 t4_wrq_tx_locked(sc, wrq, wr); in t4_wrq_tx() 1571 TXQ_UNLOCK(wrq); in t4_wrq_tx()
|
H A D | t4_smt.h | 50 struct sge_wrq *wrq; /* queue to use for write_sme */ member
|
H A D | t4_l2t.h | 64 struct sge_wrq *wrq; /* queue to use for write_l2e */ member
|
H A D | t4_main.c | 2016 struct sge_wrq *wrq; in stop_lld() local 2071 TXQ_LOCK(&ofld_txq->wrq); in stop_lld() 2072 ofld_txq->wrq.eq.flags &= ~EQ_HW_ALLOCATED; in stop_lld() 2073 TXQ_UNLOCK(&ofld_txq->wrq); in stop_lld() 2090 wrq = &sc->sge.ctrlq[i]; in stop_lld() 2091 TXQ_LOCK(wrq); in stop_lld() 2092 wrq->eq.flags &= ~EQ_HW_ALLOCATED; in stop_lld() 2093 TXQ_UNLOCK(wrq); in stop_lld() 2094 quiesce_wrq(wrq); in stop_lld() 7187 quiesce_wrq(struct sge_wrq *wrq) in quiesce_wrq() argument [all …]
|
/freebsd/sys/dev/cxgbe/tom/ |
H A D | t4_cpl_io.c | 107 wr = alloc_wrqe(roundup2(flowclen, 16), &toep->ofld_txq->wrq); in send_flowc_wr() 198 &toep->ofld_txq->wrq)) == NULL) { in update_tx_rate_limit() 262 wr = alloc_wrqe(sizeof(*req), &toep->ofld_txq->wrq); in send_reset() 466 wr = alloc_wrqe(sizeof(*req), &toep->ofld_txq->wrq); in t4_close_conn() 805 &toep->ofld_txq->wrq); in t4_push_frames() 825 &toep->ofld_txq->wrq); in t4_push_frames() 986 wr = alloc_wrqe(roundup2(plen, 16), &toep->ofld_txq->wrq); in write_iscsi_mbuf_wr() 1068 wr = alloc_wrqe(roundup2(wr_len, 16), &toep->ofld_txq->wrq); in write_iscsi_mbuf_wr() 1499 wr = alloc_wrqe(sizeof(*cpl), &ofld_txq->wrq); in send_abort_rpl() 1944 t4_set_tcb_field(struct adapter *sc, struct sge_wrq *wrq, struct toepcb *toep, in t4_set_tcb_field() argument [all …]
|
H A D | t4_ddp.c | 87 struct sge_wrq *wrq, int tid, struct ddp_rcv_buffer *drb); 1620 t4_write_page_pods_for_ps(struct adapter *sc, struct sge_wrq *wrq, int tid, in t4_write_page_pods_for_ps() argument 1651 wr = alloc_wrqe(len, wrq); in t4_write_page_pods_for_ps() 1699 t4_write_page_pods_for_rcvbuf(struct adapter *sc, struct sge_wrq *wrq, int tid, in t4_write_page_pods_for_rcvbuf() argument 1733 wr = alloc_wrqe(len, wrq); in t4_write_page_pods_for_rcvbuf() 1809 struct ppod_reservation *prsv, struct bio *bp, struct mbufq *wrq) in t4_write_page_pods_for_bio() argument 1878 mbufq_enqueue(wrq, m); in t4_write_page_pods_for_bio() 1887 struct mbufq *wrq) in t4_write_page_pods_for_buf() argument 1966 mbufq_enqueue(wrq, m); in t4_write_page_pods_for_buf() 1977 int xferlen, struct mbufq *wrq) in t4_write_page_pods_for_sgl() argument [all …]
|
H A D | t4_tls.c | 68 t4_set_tcb_field(sc, &toep->ofld_txq->wrq, toep, word, mask, val, 0, 0); in t4_set_tls_tcb_field() 93 t4_set_tcb_field(sc, &toep->ofld_txq->wrq, toep, W_TCB_T_FLAGS, in t4_set_rx_quiesce() 176 wr = alloc_wrqe(TLS_KEY_WR_SZ, &toep->ofld_txq->wrq); in tls_program_key_id() 641 wr = alloc_wrqe(roundup2(wr_len, 16), &toep->ofld_txq->wrq); in t4_push_ktls()
|
H A D | t4_tom_l2t.c | 407 e->wrq = &sc->sge.ctrlq[pi->port_id]; in t4_l2t_get()
|
H A D | t4_listen.c | 498 wr = alloc_wrqe(roundup2(flowclen, 16), &ofld_txq->wrq); in send_flowc_wr_synqe() 548 &sc->sge.ofld_txq[synqe->params.txq_idx].wrq); in send_abort_rpl_synqe()
|
H A D | t4_tom.c | 872 flowc = start_wrq_wr(&toep->ofld_txq->wrq, flowclen16, &cookie); in send_mss_flowc_wr() 891 commit_wrq_wr(&toep->ofld_txq->wrq, flowc, &cookie); in send_mss_flowc_wr()
|
/freebsd/sys/dev/cxgbe/iw_cxgbe/ |
H A D | qp.c | 1133 wr = alloc_wrqe(sizeof(*wqe), &toep->ofld_txq->wrq); in post_terminate() 1265 wr = alloc_wrqe(sizeof(*wqe), &toep->ofld_txq->wrq); in rdma_fini() 1359 wr = alloc_wrqe(sizeof(*wqe), &toep->ofld_txq->wrq); in rdma_init()
|
/freebsd/sys/dev/cxgbe/cxgbei/ |
H A D | icl_cxgbei.c | 1040 wr = alloc_wrqe(roundup2(flowclen, 16), &toep->ofld_txq->wrq); in send_iscsi_flowc_wr()
|