| /linux/drivers/net/vmxnet3/ |
| H A D | vmxnet3_xdp.c | 22 struct vmxnet3_tx_queue *tq; in vmxnet3_xdp_get_tq() local 29 tq = &adapter->tx_queue[cpu]; in vmxnet3_xdp_get_tq() 31 tq = &adapter->tx_queue[cpu % tq_number]; in vmxnet3_xdp_get_tq() 33 return tq; in vmxnet3_xdp_get_tq() 117 struct vmxnet3_tx_queue *tq, bool dma_map) in vmxnet3_xdp_xmit_frame() argument 127 spin_lock_irq(&tq->tx_lock); in vmxnet3_xdp_xmit_frame() 128 dw2 = (tq->tx_ring.gen ^ 0x1) << VMXNET3_TXD_GEN_SHIFT; in vmxnet3_xdp_xmit_frame() 130 ctx.sop_txd = tq->tx_ring.base + tq->tx_ring.next2fill; in vmxnet3_xdp_xmit_frame() 134 tbi = tq->buf_info + tq->tx_ring.next2fill; in vmxnet3_xdp_xmit_frame() 136 if (vmxnet3_cmd_ring_desc_avail(&tq->tx_ring) == 0) { in vmxnet3_xdp_xmit_frame() [all …]
|
| H A D | vmxnet3_drv.c | 120 vmxnet3_tq_stopped(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) in vmxnet3_tq_stopped() argument 122 return tq->stopped; in vmxnet3_tq_stopped() 127 vmxnet3_tq_start(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) in vmxnet3_tq_start() argument 129 tq->stopped = false; in vmxnet3_tq_start() 130 netif_start_subqueue(adapter->netdev, tq - adapter->tx_queue); in vmxnet3_tq_start() 135 vmxnet3_tq_wake(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) in vmxnet3_tq_wake() argument 137 tq->stopped = false; in vmxnet3_tq_wake() 138 netif_wake_subqueue(adapter->netdev, (tq - adapter->tx_queue)); in vmxnet3_tq_wake() 143 vmxnet3_tq_stop(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) in vmxnet3_tq_stop() argument 145 tq->stopped = true; in vmxnet3_tq_stop() [all …]
|
| /linux/net/rxrpc/ |
| H A D | call_event.c | 63 struct rxrpc_txqueue *tq = req->tq; in rxrpc_retransmit_data() local 65 struct rxrpc_txbuf *txb = tq->bufs[ix]; in rxrpc_retransmit_data() 67 _enter("%x,%x,%x,%x", tq->qbase, req->seq, ix, txb->debug_id); in rxrpc_retransmit_data() 76 req->tq = NULL; in rxrpc_retransmit_data() 92 struct rxrpc_txqueue *tq; in rxrpc_resend() local 99 for (tq = call->tx_queue; tq; tq = tq->next) { in rxrpc_resend() 100 unsigned long lost = tq->segment_lost; in rxrpc_resend() 102 if (after(tq->qbase, call->tx_transmitted)) in rxrpc_resend() 106 tq->segment_acked, tq->nr_reported_acks, call->debug_id, tq->qbase); in rxrpc_resend() 109 trace_rxrpc_resend_lost(call, tq, lost); in rxrpc_resend() [all …]
|
| H A D | input.c | 206 struct rxrpc_txqueue *tq, in rxrpc_add_data_rtt_sample() argument 209 ktime_t xmit_ts = ktime_add_us(tq->xmit_ts_base, tq->segment_xmit_ts[ix]); in rxrpc_add_data_rtt_sample() 214 __clear_bit(ix, &tq->rtt_samples); /* Prevent repeat RTT sample */ in rxrpc_add_data_rtt_sample() 223 struct rxrpc_txqueue *tq = call->tx_queue; in rxrpc_rotate_tx_window() local 230 trace_rxrpc_tq(call, tq, seq, rxrpc_tq_rotate); in rxrpc_rotate_tx_window() 244 call->tx_queue = tq->next; in rxrpc_rotate_tx_window() 245 trace_rxrpc_tq(call, tq, seq, rxrpc_tq_rotate_and_free); in rxrpc_rotate_tx_window() 246 kfree(tq); in rxrpc_rotate_tx_window() 247 tq = call->tx_queue; in rxrpc_rotate_tx_window() 253 _debug("tq=%x seq=%x i=%d f=%x", tq->qbase, seq, ix, tq->bufs[ix]->flags); in rxrpc_rotate_tx_window() [all …]
|
| H A D | output.c | 520 static unsigned int rxrpc_prepare_txqueue(struct rxrpc_txqueue *tq, in rxrpc_prepare_txqueue() argument 523 if (!tq) in rxrpc_prepare_txqueue() 525 if (tq->xmit_ts_base == KTIME_MIN) { in rxrpc_prepare_txqueue() 526 tq->xmit_ts_base = req->now; in rxrpc_prepare_txqueue() 529 return ktime_to_us(ktime_sub(req->now, tq->xmit_ts_base)); in rxrpc_prepare_txqueue() 539 struct rxrpc_txqueue *tq = req->tq; in rxrpc_prepare_data_packet() local 546 trace_rxrpc_tq(call, tq, seq, rxrpc_tq_transmit); in rxrpc_prepare_data_packet() 565 xmit_ts = rxrpc_prepare_txqueue(tq, req); in rxrpc_prepare_data_packet() 566 prefetch(tq->next); in rxrpc_prepare_data_packet() 570 struct rxrpc_txbuf *txb = tq->bufs[seq & RXRPC_TXQ_MASK]; in rxrpc_prepare_data_packet() [all …]
|
| H A D | sendmsg.c | 283 struct rxrpc_txqueue *tq; in rxrpc_alloc_txqueue() local 285 tq = kzalloc_obj(*tq, sk->sk_allocation); in rxrpc_alloc_txqueue() 286 if (!tq) in rxrpc_alloc_txqueue() 289 tq->xmit_ts_base = KTIME_MIN; in rxrpc_alloc_txqueue() 291 tq->segment_xmit_ts[i] = UINT_MAX; in rxrpc_alloc_txqueue() 294 tq->qbase = call->send_top + 1; in rxrpc_alloc_txqueue() 295 call->send_queue->next = tq; in rxrpc_alloc_txqueue() 296 call->send_queue = tq; in rxrpc_alloc_txqueue() 298 kfree(tq); in rxrpc_alloc_txqueue() 302 tq->nr_reported_acks = 1; in rxrpc_alloc_txqueue() [all …]
|
| /linux/drivers/net/wireless/ath/ath5k/ |
| H A D | qcu.c | 286 struct ath5k_txq_info *tq = &ah->ah_txq[queue]; in ath5k_hw_set_tx_retry_limits() local 292 (tq->tqi_cw_min << AR5K_NODCU_RETRY_LMT_CW_MIN_S) in ath5k_hw_set_tx_retry_limits() 326 struct ath5k_txq_info *tq = &ah->ah_txq[queue]; in ath5k_hw_reset_tx_queue() local 333 (tq->tqi_type == AR5K_TX_QUEUE_INACTIVE)) in ath5k_hw_reset_tx_queue() 341 AR5K_REG_SM(tq->tqi_cw_min, AR5K_DCU_LCL_IFS_CW_MIN) | in ath5k_hw_reset_tx_queue() 342 AR5K_REG_SM(tq->tqi_cw_max, AR5K_DCU_LCL_IFS_CW_MAX) | in ath5k_hw_reset_tx_queue() 343 AR5K_REG_SM(tq->tqi_aifs, AR5K_DCU_LCL_IFS_AIFS), in ath5k_hw_reset_tx_queue() 366 if (tq->tqi_cbr_period) { in ath5k_hw_reset_tx_queue() 367 ath5k_hw_reg_write(ah, AR5K_REG_SM(tq->tqi_cbr_period, in ath5k_hw_reset_tx_queue() 369 AR5K_REG_SM(tq->tqi_cbr_overflow_limit, in ath5k_hw_reset_tx_queue() [all …]
|
| /linux/drivers/net/ethernet/chelsio/cxgb4vf/ |
| H A D | sge.c | 233 static inline unsigned int txq_avail(const struct sge_txq *tq) in txq_avail() argument 235 return tq->size - 1 - tq->in_use; in txq_avail() 307 const struct ulptx_sgl *sgl, const struct sge_txq *tq) in unmap_sgl() argument 326 if (likely((u8 *)(p + 1) <= (u8 *)tq->stat)) { in unmap_sgl() 333 } else if ((u8 *)p == (u8 *)tq->stat) { in unmap_sgl() 334 p = (const struct ulptx_sge_pair *)tq->desc; in unmap_sgl() 336 } else if ((u8 *)p + 8 == (u8 *)tq->stat) { in unmap_sgl() 337 const __be64 *addr = (const __be64 *)tq->desc; in unmap_sgl() 345 const __be64 *addr = (const __be64 *)tq->desc; in unmap_sgl() 357 if ((u8 *)p == (u8 *)tq->stat) in unmap_sgl() [all …]
|
| /linux/arch/arm/boot/dts/nxp/imx/ |
| H A D | imx6q-mba6b.dts | 6 * Author: Markus Niebel <Markus.Niebel@tq-group.com> 18 compatible = "tq,imx6q-mba6x-b", "tq,mba6b", 19 "tq,imx6q-tqma6q-b", "fsl,imx6q";
|
| H A D | imx6qp-mba6b.dts | 4 * Author: Markus Niebel <Markus.Niebel@tq-group.com> 16 compatible = "tq,imx6qp-mba6x-b", "tq,mba6b", 17 "tq,imx6qp-tqma6qp-b", "fsl,imx6qp";
|
| H A D | imx6q-mba6a.dts | 6 * Author: Markus Niebel <Markus.Niebel@tq-group.com> 18 compatible = "tq,imx6q-mba6x-a", "tq,mba6a", 19 "tq,imx6q-tqma6q-a", "fsl,imx6q";
|
| H A D | imx6dl-mba6b.dts | 6 * Author: Markus Niebel <Markus.Niebel@tq-group.com> 19 compatible = "tq,imx6dl-mba6x-b", "tq,mba6b", 20 "tq,imx6dl-tqma6dl-b", "fsl,imx6dl";
|
| H A D | imx6dl-mba6a.dts | 6 * Author: Markus Niebel <Markus.Niebel@tq-group.com> 19 compatible = "tq,imx6dl-mba6x-a", "tq,mba6a", 20 "tq,imx6dl-tqma6dl-a", "fsl,imx6dl";
|
| H A D | imx6ull-tqma6ull2-mba6ulx.dts | 4 * Author: Markus Niebel <Markus.Niebel@tq-group.com> 14 compatible = "tq,imx6ull-tqma6ull2-mba6ulx", "tq,imx6ull-tqma6ull2", "fsl,imx6ull";
|
| H A D | imx6ul-tqma6ul2-mba6ulx.dts | 4 * Author: Markus Niebel <Markus.Niebel@tq-group.com> 14 compatible = "tq,imx6ul-tqma6ul2-mba6ulx", "tq,imx6ul-tqma6ul2", "fsl,imx6ul";
|
| H A D | imx6ul-tqma6ul2l-mba6ulx.dts | 4 * Author: Markus Niebel <Markus.Niebel@tq-group.com> 14 compatible = "tq,imx6ul-tqma6ul2l-mba6ulx", "tq,imx6ul-tqma6ul2l", "fsl,imx6ul";
|
| H A D | imx6ull-tqma6ull2l-mba6ulx.dts | 4 * Author: Markus Niebel <Markus.Niebel@tq-group.com> 14 compatible = "tq,imx6ull-tqma6ull2l-mba6ulx", "tq,imx6ull-tqma6ull2l", "fsl,imx6ull";
|
| H A D | imx7s-mba7.dts | 6 * Author: Markus Niebel <Markus.Niebel@tq-group.com> 17 compatible = "tq,imx7s-mba7", "tq,imx7s-tqma7", "fsl,imx7s";
|
| /linux/arch/arm64/boot/dts/freescale/ |
| H A D | imx8qxp-tqma8xqp-mba8xx.dts | 3 * Copyright 2018-2023 TQ-Systems GmbH <linux@ew.tq-group.com>, 15 compatible = "tq,imx8qxp-tqma8xqp-mba8xx", "tq,imx8qxp-tqma8xqp", "fsl,imx8qxp";
|
| H A D | imx8dxp-tqma8xdp-mba8xx.dts | 3 * Copyright 2018-2023 TQ-Systems GmbH <linux@ew.tq-group.com>, 15 compatible = "tq,imx8dxp-tqma8xdp-mba8xx", "tq,imx8dxp-tqma8xdp", "fsl,imx8dxp";
|
| H A D | imx8qxp-tqma8xqps-mb-smarc-2.dts | 3 * Copyright (c) 2018-2025 TQ-Systems GmbH <linux@ew.tq-group.com>, 15 compatible = "tq,imx8qxp-tqma8xqps-mb-smarc-2", "tq,imx8qxp-tqma8xqps", "fsl,imx8qxp";
|
| H A D | imx8dxp-tqma8xdps-mb-smarc-2.dts | 3 * Copyright (c) 2021-2025 TQ-Systems GmbH <linux@ew.tq-group.com>, 15 compatible = "tq,imx8dxp-tqma8xdps-mb-smarc-2", "tq,imx8dxp-tqma8xdps", "fsl,imx8dxp";
|
| H A D | imx8mm-tqma8mqml-mba8mx-lvds-tm070jvhg33.dtso | 3 * Copyright (c) 2022-2023 TQ-Systems GmbH <linux@ew.tq-group.com>, 14 compatible = "tq,imx8mm-tqma8mqml-mba8mx", "tq,imx8mm-tqma8mqml", "fsl,imx8mm";
|
| H A D | imx8mn-tqma8mqnl-mba8mx-lvds-tm070jvhg33.dtso | 3 * Copyright (c) 2022-2023 TQ-Systems GmbH <linux@ew.tq-group.com>, 14 compatible = "tq,imx8mn-tqma8mqnl-mba8mx", "tq,imx8mn-tqma8mqnl", "fsl,imx8mn";
|
| /linux/net/tipc/ |
| H A D | trace.h | 284 TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq), 286 TP_ARGS(r, f, t, tq), 301 __entry->len = skb_queue_len(tq); 303 msg_seqno(buf_msg(skb_peek(tq))) : 0; 305 msg_seqno(buf_msg(skb_peek_tail(tq))) : 0; 314 TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq), 315 TP_ARGS(r, f, t, tq), 320 TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq), 321 TP_ARGS(r, f, t, tq),
|