Lines Matching refs:txq

1133 static void txq_stop(struct sge_eth_txq *txq)  in txq_stop()  argument
1135 netif_tx_stop_queue(txq->txq); in txq_stop()
1136 txq->q.stops++; in txq_stop()
1164 struct sge_eth_txq *txq; in t4vf_eth_xmit() local
1195 txq = &adapter->sge.ethtxq[pi->first_qset + qidx]; in t4vf_eth_xmit()
1205 reclaim_completed_tx(adapter, &txq->q, true); in t4vf_eth_xmit()
1214 credits = txq_avail(&txq->q) - ndesc; in t4vf_eth_xmit()
1223 txq_stop(txq); in t4vf_eth_xmit()
1237 txq->mapping_err++; in t4vf_eth_xmit()
1252 txq_stop(txq); in t4vf_eth_xmit()
1263 wr = (void *)&txq->q.desc[txq->q.pidx]; in t4vf_eth_xmit()
1319 txq->tso++; in t4vf_eth_xmit()
1320 txq->tx_cso += ssi->gso_segs; in t4vf_eth_xmit()
1337 txq->tx_cso++; in t4vf_eth_xmit()
1347 txq->vlan_ins++; in t4vf_eth_xmit()
1362 T4_TRACE5(adapter->tb[txq->q.cntxt_id & 7], in t4vf_eth_xmit()
1364 ndesc, credits, txq->q.pidx, skb->len, ssi->nr_frags); in t4vf_eth_xmit()
1376 inline_tx_skb(skb, &txq->q, cpl + 1); in t4vf_eth_xmit()
1417 struct sge_txq *tq = &txq->q; in t4vf_eth_xmit()
1446 txq_advance(&txq->q, ndesc); in t4vf_eth_xmit()
1448 ring_tx_db(adapter, &txq->q, ndesc); in t4vf_eth_xmit()
2131 struct sge_eth_txq *txq = &s->ethtxq[i]; in sge_tx_timer_cb() local
2133 if (reclaimable(&txq->q) && __netif_tx_trylock(txq->txq)) { in sge_tx_timer_cb()
2134 int avail = reclaimable(&txq->q); in sge_tx_timer_cb()
2139 free_tx_desc(adapter, &txq->q, avail, true); in sge_tx_timer_cb()
2140 txq->q.in_use -= avail; in sge_tx_timer_cb()
2141 __netif_tx_unlock(txq->txq); in sge_tx_timer_cb()
2410 int t4vf_sge_alloc_eth_txq(struct adapter *adapter, struct sge_eth_txq *txq, in t4vf_sge_alloc_eth_txq() argument
2424 nentries = txq->q.size + s->stat_len / sizeof(struct tx_desc); in t4vf_sge_alloc_eth_txq()
2430 txq->q.desc = alloc_ring(adapter->pdev_dev, txq->q.size, in t4vf_sge_alloc_eth_txq()
2433 &txq->q.phys_addr, &txq->q.sdesc, s->stat_len); in t4vf_sge_alloc_eth_txq()
2434 if (!txq->q.desc) in t4vf_sge_alloc_eth_txq()
2466 cmd.eqaddr = cpu_to_be64(txq->q.phys_addr); in t4vf_sge_alloc_eth_txq()
2478 kfree(txq->q.sdesc); in t4vf_sge_alloc_eth_txq()
2479 txq->q.sdesc = NULL; in t4vf_sge_alloc_eth_txq()
2482 txq->q.desc, txq->q.phys_addr); in t4vf_sge_alloc_eth_txq()
2483 txq->q.desc = NULL; in t4vf_sge_alloc_eth_txq()
2487 txq->q.in_use = 0; in t4vf_sge_alloc_eth_txq()
2488 txq->q.cidx = 0; in t4vf_sge_alloc_eth_txq()
2489 txq->q.pidx = 0; in t4vf_sge_alloc_eth_txq()
2490 txq->q.stat = (void *)&txq->q.desc[txq->q.size]; in t4vf_sge_alloc_eth_txq()
2491 txq->q.cntxt_id = FW_EQ_ETH_CMD_EQID_G(be32_to_cpu(rpl.eqid_pkd)); in t4vf_sge_alloc_eth_txq()
2492 txq->q.bar2_addr = bar2_address(adapter, in t4vf_sge_alloc_eth_txq()
2493 txq->q.cntxt_id, in t4vf_sge_alloc_eth_txq()
2495 &txq->q.bar2_qid); in t4vf_sge_alloc_eth_txq()
2496 txq->q.abs_id = in t4vf_sge_alloc_eth_txq()
2498 txq->txq = devq; in t4vf_sge_alloc_eth_txq()
2499 txq->tso = 0; in t4vf_sge_alloc_eth_txq()
2500 txq->tx_cso = 0; in t4vf_sge_alloc_eth_txq()
2501 txq->vlan_ins = 0; in t4vf_sge_alloc_eth_txq()
2502 txq->q.stops = 0; in t4vf_sge_alloc_eth_txq()
2503 txq->q.restarts = 0; in t4vf_sge_alloc_eth_txq()
2504 txq->mapping_err = 0; in t4vf_sge_alloc_eth_txq()
2565 struct sge_eth_txq *txq = s->ethtxq; in t4vf_free_sge_resources() local
2570 for (qs = 0; qs < adapter->sge.ethqsets; qs++, rxq++, txq++) { in t4vf_free_sge_resources()
2573 if (txq->q.desc) { in t4vf_free_sge_resources()
2574 t4vf_eth_eq_free(adapter, txq->q.cntxt_id); in t4vf_free_sge_resources()
2575 free_tx_desc(adapter, &txq->q, txq->q.in_use, true); in t4vf_free_sge_resources()
2576 kfree(txq->q.sdesc); in t4vf_free_sge_resources()
2577 free_txq(adapter, &txq->q); in t4vf_free_sge_resources()