Lines Matching refs:tx_buffer
272 struct ixgbevf_tx_buffer *tx_buffer; in ixgbevf_clean_tx_irq() local
281 tx_buffer = &tx_ring->tx_buffer_info[i]; in ixgbevf_clean_tx_irq()
286 union ixgbe_adv_tx_desc *eop_desc = tx_buffer->next_to_watch; in ixgbevf_clean_tx_irq()
300 tx_buffer->next_to_watch = NULL; in ixgbevf_clean_tx_irq()
303 total_bytes += tx_buffer->bytecount; in ixgbevf_clean_tx_irq()
304 total_packets += tx_buffer->gso_segs; in ixgbevf_clean_tx_irq()
305 if (tx_buffer->tx_flags & IXGBE_TX_FLAGS_IPSEC) in ixgbevf_clean_tx_irq()
310 page_frag_free(tx_buffer->data); in ixgbevf_clean_tx_irq()
312 napi_consume_skb(tx_buffer->skb, napi_budget); in ixgbevf_clean_tx_irq()
316 dma_unmap_addr(tx_buffer, dma), in ixgbevf_clean_tx_irq()
317 dma_unmap_len(tx_buffer, len), in ixgbevf_clean_tx_irq()
321 dma_unmap_len_set(tx_buffer, len, 0); in ixgbevf_clean_tx_irq()
325 tx_buffer++; in ixgbevf_clean_tx_irq()
330 tx_buffer = tx_ring->tx_buffer_info; in ixgbevf_clean_tx_irq()
335 if (dma_unmap_len(tx_buffer, len)) { in ixgbevf_clean_tx_irq()
337 dma_unmap_addr(tx_buffer, dma), in ixgbevf_clean_tx_irq()
338 dma_unmap_len(tx_buffer, len), in ixgbevf_clean_tx_irq()
340 dma_unmap_len_set(tx_buffer, len, 0); in ixgbevf_clean_tx_irq()
345 tx_buffer++; in ixgbevf_clean_tx_irq()
350 tx_buffer = tx_ring->tx_buffer_info; in ixgbevf_clean_tx_irq()
974 struct ixgbevf_tx_buffer *tx_buffer; in ixgbevf_xmit_xdp_ring() local
991 tx_buffer = &ring->tx_buffer_info[i]; in ixgbevf_xmit_xdp_ring()
993 dma_unmap_len_set(tx_buffer, len, len); in ixgbevf_xmit_xdp_ring()
994 dma_unmap_addr_set(tx_buffer, dma, dma); in ixgbevf_xmit_xdp_ring()
995 tx_buffer->data = xdp->data; in ixgbevf_xmit_xdp_ring()
996 tx_buffer->bytecount = len; in ixgbevf_xmit_xdp_ring()
997 tx_buffer->gso_segs = 1; in ixgbevf_xmit_xdp_ring()
998 tx_buffer->protocol = 0; in ixgbevf_xmit_xdp_ring()
1042 tx_buffer->next_to_watch = tx_desc; in ixgbevf_xmit_xdp_ring()
2410 struct ixgbevf_tx_buffer *tx_buffer = &tx_ring->tx_buffer_info[i]; in ixgbevf_clean_tx_ring() local
2417 page_frag_free(tx_buffer->data); in ixgbevf_clean_tx_ring()
2419 dev_kfree_skb_any(tx_buffer->skb); in ixgbevf_clean_tx_ring()
2423 dma_unmap_addr(tx_buffer, dma), in ixgbevf_clean_tx_ring()
2424 dma_unmap_len(tx_buffer, len), in ixgbevf_clean_tx_ring()
2428 eop_desc = tx_buffer->next_to_watch; in ixgbevf_clean_tx_ring()
2433 tx_buffer++; in ixgbevf_clean_tx_ring()
2438 tx_buffer = tx_ring->tx_buffer_info; in ixgbevf_clean_tx_ring()
2443 if (dma_unmap_len(tx_buffer, len)) in ixgbevf_clean_tx_ring()
2445 dma_unmap_addr(tx_buffer, dma), in ixgbevf_clean_tx_ring()
2446 dma_unmap_len(tx_buffer, len), in ixgbevf_clean_tx_ring()
2451 tx_buffer++; in ixgbevf_clean_tx_ring()
2455 tx_buffer = tx_ring->tx_buffer_info; in ixgbevf_clean_tx_ring()
3970 struct ixgbevf_tx_buffer *tx_buffer; in ixgbevf_tx_map() local
3988 tx_buffer = first; in ixgbevf_tx_map()
3995 dma_unmap_len_set(tx_buffer, len, size); in ixgbevf_tx_map()
3996 dma_unmap_addr_set(tx_buffer, dma, dma); in ixgbevf_tx_map()
4037 tx_buffer = &tx_ring->tx_buffer_info[i]; in ixgbevf_tx_map()
4073 tx_buffer = &tx_ring->tx_buffer_info[i]; in ixgbevf_tx_map()
4076 while (tx_buffer != first) { in ixgbevf_tx_map()
4077 if (dma_unmap_len(tx_buffer, len)) in ixgbevf_tx_map()
4079 dma_unmap_addr(tx_buffer, dma), in ixgbevf_tx_map()
4080 dma_unmap_len(tx_buffer, len), in ixgbevf_tx_map()
4082 dma_unmap_len_set(tx_buffer, len, 0); in ixgbevf_tx_map()
4086 tx_buffer = &tx_ring->tx_buffer_info[i]; in ixgbevf_tx_map()
4089 if (dma_unmap_len(tx_buffer, len)) in ixgbevf_tx_map()
4091 dma_unmap_addr(tx_buffer, dma), in ixgbevf_tx_map()
4092 dma_unmap_len(tx_buffer, len), in ixgbevf_tx_map()
4094 dma_unmap_len_set(tx_buffer, len, 0); in ixgbevf_tx_map()
4096 dev_kfree_skb_any(tx_buffer->skb); in ixgbevf_tx_map()
4097 tx_buffer->skb = NULL; in ixgbevf_tx_map()