Home
last modified time | relevance | path

Searched refs:tx_q (Results 1 – 25 of 25) sorted by relevance

/linux/drivers/net/ethernet/qualcomm/emac/
H A Demac-mac.c305 writel(upper_32_bits(adpt->tx_q.tpd.dma_addr), in emac_mac_dma_rings_config()
308 writel(lower_32_bits(adpt->tx_q.tpd.dma_addr), in emac_mac_dma_rings_config()
311 writel(adpt->tx_q.tpd.count & TPD_RING_SIZE_BMSK, in emac_mac_dma_rings_config()
587 struct emac_tx_queue *tx_q = &adpt->tx_q; in emac_tx_q_descs_free() local
592 if (!tx_q->tpd.tpbuff) in emac_tx_q_descs_free()
595 for (i = 0; i < tx_q->tpd.count; i++) { in emac_tx_q_descs_free()
596 struct emac_buffer *tpbuf = GET_TPD_BUFFER(tx_q, i); in emac_tx_q_descs_free()
610 size = sizeof(struct emac_buffer) * tx_q->tpd.count; in emac_tx_q_descs_free()
611 memset(tx_q->tpd.tpbuff, 0, size); in emac_tx_q_descs_free()
614 memset(tx_q->tpd.v_addr, 0, tx_q->tpd.size); in emac_tx_q_descs_free()
[all …]
H A Demac-mac.h231 struct emac_tx_queue *tx_q,
233 void emac_mac_tx_process(struct emac_adapter *adpt, struct emac_tx_queue *tx_q);
H A Demac.h341 struct emac_tx_queue tx_q; member
/linux/drivers/net/ethernet/stmicro/stmmac/
H A Dstmmac_main.c341 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; in stmmac_tx_avail() local
344 if (tx_q->dirty_tx > tx_q->cur_tx) in stmmac_tx_avail()
345 avail = tx_q->dirty_tx - tx_q->cur_tx - 1; in stmmac_tx_avail()
347 avail = priv->dma_conf.dma_tx_size - tx_q->cur_tx + tx_q->dirty_tx - 1; in stmmac_tx_avail()
377 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; in stmmac_eee_tx_busy() local
379 if (tx_q->dirty_tx != tx_q->cur_tx) in stmmac_eee_tx_busy()
1400 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; in stmmac_display_tx_rings() local
1405 head_tx = (void *)tx_q->dma_etx; in stmmac_display_tx_rings()
1407 } else if (tx_q->tbs & STMMAC_TBS_AVAIL) { in stmmac_display_tx_rings()
1408 head_tx = (void *)tx_q->dma_entx; in stmmac_display_tx_rings()
[all …]
/linux/drivers/net/ethernet/intel/idpf/
H A Didpf_txrx.c177 static int idpf_tx_buf_alloc_all(struct idpf_tx_queue *tx_q) in idpf_tx_buf_alloc_all() argument
182 if (idpf_queue_has(FLOW_SCH_EN, tx_q)) in idpf_tx_buf_alloc_all()
183 tx_q->buf_pool_size = U16_MAX; in idpf_tx_buf_alloc_all()
185 tx_q->buf_pool_size = tx_q->desc_count; in idpf_tx_buf_alloc_all()
186 tx_q->tx_buf = kcalloc(tx_q->buf_pool_size, sizeof(*tx_q->tx_buf), in idpf_tx_buf_alloc_all()
188 if (!tx_q->tx_buf) in idpf_tx_buf_alloc_all()
202 struct idpf_tx_queue *tx_q) in idpf_tx_desc_alloc() argument
204 struct device *dev = tx_q->dev; in idpf_tx_desc_alloc()
208 err = idpf_tx_buf_alloc_all(tx_q); in idpf_tx_desc_alloc()
212 tx_q->size = tx_q->desc_count * sizeof(*tx_q->base_tx); in idpf_tx_desc_alloc()
[all …]
H A Didpf_ptp.c645 int idpf_ptp_request_ts(struct idpf_tx_queue *tx_q, struct sk_buff *skb, in idpf_ptp_request_ts() argument
652 spin_lock(&tx_q->cached_tstamp_caps->latches_lock); in idpf_ptp_request_ts()
654 head = &tx_q->cached_tstamp_caps->latches_free; in idpf_ptp_request_ts()
656 spin_unlock(&tx_q->cached_tstamp_caps->latches_lock); in idpf_ptp_request_ts()
669 &tx_q->cached_tstamp_caps->latches_in_use); in idpf_ptp_request_ts()
670 spin_unlock(&tx_q->cached_tstamp_caps->latches_lock); in idpf_ptp_request_ts()
H A Didpf_ptp.h286 int idpf_ptp_request_ts(struct idpf_tx_queue *tx_q, struct sk_buff *skb,
371 static inline int idpf_ptp_request_ts(struct idpf_tx_queue *tx_q, in idpf_ptp_request_ts() argument
H A Didpf_txrx.h1101 void idpf_tx_buf_hw_update(struct idpf_tx_queue *tx_q, u32 val,
1104 netdev_tx_t idpf_tx_drop_skb(struct idpf_tx_queue *tx_q, struct sk_buff *skb);
1109 struct idpf_tx_queue *tx_q);
/linux/drivers/net/wwan/t7xx/
H A Dt7xx_hif_dpmaif.c211 struct dpmaif_tx_queue *tx_q; in t7xx_dpmaif_rxtx_sw_allocs() local
236 tx_q = &dpmaif_ctrl->txq[tx_idx]; in t7xx_dpmaif_rxtx_sw_allocs()
237 tx_q->index = tx_idx; in t7xx_dpmaif_rxtx_sw_allocs()
238 tx_q->dpmaif_ctrl = dpmaif_ctrl; in t7xx_dpmaif_rxtx_sw_allocs()
239 ret = t7xx_dpmaif_txq_init(tx_q); in t7xx_dpmaif_rxtx_sw_allocs()
261 tx_q = &dpmaif_ctrl->txq[i]; in t7xx_dpmaif_rxtx_sw_allocs()
262 t7xx_dpmaif_txq_free(tx_q); in t7xx_dpmaif_rxtx_sw_allocs()
282 struct dpmaif_tx_queue *tx_q; in t7xx_dpmaif_sw_release() local
289 tx_q = &dpmaif_ctrl->txq[i]; in t7xx_dpmaif_sw_release()
290 t7xx_dpmaif_txq_free(tx_q); in t7xx_dpmaif_sw_release()
/linux/drivers/net/ethernet/intel/ice/
H A Dice_dcb_lib.h47 static inline bool ice_find_q_in_range(u16 low, u16 high, unsigned int tx_q) in ice_find_q_in_range() argument
49 return (tx_q >= low) && (tx_q < high); in ice_find_q_in_range()
/linux/net/nfc/nci/
H A Duart.c44 skb = skb_dequeue(&nu->tx_q); in nci_uart_dequeue()
56 return skb_queue_empty(&nu->tx_q); in nci_uart_queue_empty()
122 skb_queue_head_init(&nu->tx_q); in nci_uart_set_driver()
186 skb_queue_purge(&nu->tx_q); in nci_uart_tty_close()
363 skb_queue_tail(&nu->tx_q, skb); in nci_uart_send()
H A Ddata.c146 spin_lock_irqsave(&ndev->tx_q.lock, flags); in nci_queue_tx_data_frags()
149 __skb_queue_tail(&ndev->tx_q, skb_frag); in nci_queue_tx_data_frags()
151 spin_unlock_irqrestore(&ndev->tx_q.lock, flags); in nci_queue_tx_data_frags()
185 skb_queue_tail(&ndev->tx_q, skb); in nci_send_data()
H A Dcore.c543 skb_queue_purge(&ndev->tx_q); in nci_open_device()
576 skb_queue_purge(&ndev->tx_q); in nci_close_device()
1269 skb_queue_head_init(&ndev->tx_q); in nci_register_device()
1506 skb = skb_dequeue(&ndev->tx_q); in nci_tx_work()
/linux/drivers/net/wireless/mediatek/mt7601u/
H A Ddma.c312 struct mt7601u_tx_queue *q = &dev->tx_q[ep]; in mt7601u_dma_submit_tx()
477 if (!dev->tx_q) in mt7601u_free_tx()
481 mt7601u_free_tx_queue(&dev->tx_q[i]); in mt7601u_free_tx()
505 dev->tx_q = devm_kcalloc(dev->dev, __MT_EP_OUT_MAX, in mt7601u_alloc_tx()
506 sizeof(*dev->tx_q), GFP_KERNEL); in mt7601u_alloc_tx()
507 if (!dev->tx_q) in mt7601u_alloc_tx()
511 if (mt7601u_alloc_tx_queue(dev, &dev->tx_q[i])) in mt7601u_alloc_tx()
H A Dmt7601u.h201 struct mt7601u_tx_queue *tx_q; member
/linux/drivers/net/fddi/skfp/h/
H A Dhwmtm.h242 #define HWM_GET_TX_USED(smc,queue) (int) (smc)->hw.fp.tx_q[queue].tx_used
261 (smc)->hw.fp.tx_q[queue].tx_curr_put
H A Dfplustm.h191 struct s_smt_tx_queue tx_q[USED_QUEUES] ; member
/linux/net/bluetooth/
H A Dl2cap_core.c449 skb_queue_head_init(&chan->tx_q); in l2cap_chan_create()
573 skb_queue_head_init(&chan->tx_q); in l2cap_le_flowctl_init()
684 skb_queue_purge(&chan->tx_q); in l2cap_chan_del()
699 skb_queue_purge(&chan->tx_q); in l2cap_chan_del()
1933 skb_queue_splice_tail_init(skbs, &chan->tx_q); in l2cap_streaming_send()
1935 while (!skb_queue_empty(&chan->tx_q)) { in l2cap_streaming_send()
1937 skb = skb_dequeue(&chan->tx_q); in l2cap_streaming_send()
2013 if (skb_queue_is_last(&chan->tx_q, skb)) in l2cap_ertm_send()
2016 chan->tx_send_head = skb_queue_next(&chan->tx_q, skb); in l2cap_ertm_send()
2023 chan->unacked_frames, skb_queue_len(&chan->tx_q)); in l2cap_ertm_send()
[all …]
H A Dhci_conn.c1044 skb_queue_head_init(&conn->tx_q.queue); in __hci_conn_add()
1227 skb_queue_purge(&conn->tx_q.queue); in hci_conn_del()
3044 struct tx_queue *comp = &conn->tx_q; in hci_conn_tx_queue()
3107 struct tx_queue *comp = &conn->tx_q; in hci_conn_tx_dequeue()
/linux/drivers/net/ethernet/chelsio/cxgb3/
H A Dadapter.h208 struct netdev_queue *tx_q; /* associated netdev TX queue */ member
H A Dsge.c1992 netif_tx_wake_queue(qs->tx_q); in restart_tx()
2915 if (__netif_tx_trylock(qs->tx_q)) { in sge_timer_tx()
2918 __netif_tx_unlock(qs->tx_q); in sge_timer_tx()
3156 q->tx_q = netdevq; in t3_sge_alloc_qset()
/linux/drivers/net/fddi/skfp/
H A Dfplustm.c305 smc->hw.fp.tx[QUEUE_S] = queue = &smc->hw.fp.tx_q[QUEUE_S] ; in init_tx()
316 smc->hw.fp.tx[QUEUE_A0] = queue = &smc->hw.fp.tx_q[QUEUE_A0] ; in init_tx()
H A Dhwmtm.c561 if (smc->hw.fp.tx_q[QUEUE_A0].tx_used) { in mac_drv_repair_descr()
566 if (smc->hw.fp.tx_q[QUEUE_S].tx_used) { in mac_drv_repair_descr()
/linux/include/net/bluetooth/
H A Dl2cap.h604 struct sk_buff_head tx_q; member
H A Dhci_core.h760 struct tx_queue tx_q; member