| /linux/drivers/ps3/ |
| H A D | ps3-vuart.c | 72 } tx_list; member 486 spin_lock_irqsave(&priv->tx_list.lock, flags); in ps3_vuart_write() 488 if (list_empty(&priv->tx_list.head)) { in ps3_vuart_write() 493 spin_unlock_irqrestore(&priv->tx_list.lock, flags); in ps3_vuart_write() 511 spin_unlock_irqrestore(&priv->tx_list.lock, flags); in ps3_vuart_write() 523 spin_lock_irqsave(&priv->tx_list.lock, flags); in ps3_vuart_write() 524 list_add_tail(&lb->link, &priv->tx_list.head); in ps3_vuart_write() 526 spin_unlock_irqrestore(&priv->tx_list.lock, flags); in ps3_vuart_write() 728 spin_lock_irqsave(&priv->tx_list.lock, flags); in ps3_vuart_handle_interrupt_tx() 730 list_for_each_entry_safe(lb, n, &priv->tx_list.head, link) { in ps3_vuart_handle_interrupt_tx() [all …]
|
| /linux/drivers/dma/ |
| H A D | mmp_pdma.c | 98 struct list_head tx_list; member 508 list_for_each_entry(child, &desc->tx_list, node) { in mmp_pdma_tx_submit() 513 list_splice_tail_init(&desc->tx_list, &chan->chain_pending); in mmp_pdma_tx_submit() 532 INIT_LIST_HEAD(&desc->tx_list); in mmp_pdma_alloc_descriptor() 662 list_add_tail(&new->node, &first->tx_list); in mmp_pdma_prep_memcpy() 678 mmp_pdma_free_desc_list(chan, &first->tx_list); in mmp_pdma_prep_memcpy() 738 list_add_tail(&new->node, &first->tx_list); in mmp_pdma_prep_slave_sg() 760 mmp_pdma_free_desc_list(chan, &first->tx_list); in mmp_pdma_prep_slave_sg() 837 list_add_tail(&new->node, &first->tx_list); in mmp_pdma_prep_dma_cyclic() 851 mmp_pdma_free_desc_list(chan, &first->tx_list); in mmp_pdma_prep_dma_cyclic()
|
| H A D | txx9dmac.c | 182 if (!list_empty(&desc->tx_list)) in txx9dmac_last_child() 183 desc = list_entry(desc->tx_list.prev, typeof(*desc), desc_node); in txx9dmac_last_child() 198 INIT_LIST_HEAD(&desc->tx_list); in txx9dmac_desc_alloc() 247 list_for_each_entry(child, &desc->tx_list, desc_node) in txx9dmac_sync_desc_for_cpu() 269 list_for_each_entry(child, &desc->tx_list, desc_node) in txx9dmac_desc_put() 273 list_splice_init(&desc->tx_list, &dc->free_list); in txx9dmac_desc_put() 412 list_splice_init(&desc->tx_list, &dc->free_list); in txx9dmac_descriptor_complete() 530 list_for_each_entry(child, &bad_desc->tx_list, desc_node) in txx9dmac_handle_error() 572 list_for_each_entry(child, &desc->tx_list, desc_node) in txx9dmac_scan_descriptors() 782 list_add_tail(&desc->desc_node, &first->tx_list); in txx9dmac_prep_dma_memcpy() [all …]
|
| H A D | altera-msgdma.c | 164 struct list_head tx_list; member 220 INIT_LIST_HEAD(&desc->tx_list); in msgdma_get_descriptor() 237 list_for_each_entry_safe(child, next, &desc->tx_list, node) { in msgdma_free_descriptor() 367 list_add_tail(&new->node, &first->tx_list); in msgdma_prep_memcpy() 439 list_add_tail(&new->node, &first->tx_list); in msgdma_prep_slave_sg() 539 list_for_each_entry_safe(sdesc, next, &desc->tx_list, node) in msgdma_copy_desc_to_fifo()
|
| H A D | ep93xx_dma.c | 154 struct list_head tx_list; member 308 while (!list_empty(&desc->tx_list)) { in ep93xx_dma_set_active() 309 struct ep93xx_dma_desc *d = list_first_entry(&desc->tx_list, in ep93xx_dma_set_active() 769 list_splice_init(&desc->tx_list, &edmac->free_list); in ep93xx_dma_desc_put() 975 INIT_LIST_HEAD(&desc->tx_list); in ep93xx_dma_alloc_chan_resources() 1061 list_add_tail(&desc->node, &first->tx_list); in ep93xx_dma_prep_dma_memcpy() 1136 list_add_tail(&desc->node, &first->tx_list); in ep93xx_dma_prep_slave_sg() 1217 list_add_tail(&desc->node, &first->tx_list); in ep93xx_dma_prep_dma_cyclic()
|
| H A D | pch_dma.c | 87 struct list_head tx_list; member 331 if (list_empty(&desc->tx_list)) { in pdc_dostart() 350 list_splice_init(&desc->tx_list, &pd_chan->free_list); in pdc_chain_complete() 429 INIT_LIST_HEAD(&desc->tx_list); in pdc_alloc_desc() 478 list_splice_init(&desc->tx_list, &pd_chan->free_list); in pdc_desc_put() 625 list_add_tail(&desc->desc_node, &first->tx_list); in pd_prep_slave_sg()
|
| H A D | tegra20-apb-dma.c | 169 struct list_head tx_list; member 302 if (!list_empty(&dma_desc->tx_list)) in tegra_dma_desc_put() 303 list_splice_init(&dma_desc->tx_list, &tdc->free_sg_req); in tegra_dma_desc_put() 696 list_splice_tail_init(&dma_desc->tx_list, &tdc->pending_sg_req); in tegra_dma_tx_submit() 1093 INIT_LIST_HEAD(&dma_desc->tx_list); in tegra_dma_prep_slave_sg() 1136 list_add_tail(&sg_req->node, &dma_desc->tx_list); in tegra_dma_prep_slave_sg() 1242 INIT_LIST_HEAD(&dma_desc->tx_list); in tegra_dma_prep_dma_cyclic() 1271 list_add_tail(&sg_req->node, &dma_desc->tx_list); in tegra_dma_prep_dma_cyclic()
|
| H A D | fsldma.c | 404 list_splice_tail_init(&desc->tx_list, &chan->ld_pending); in append_ld_queue() 428 list_for_each_entry(child, &desc->tx_list, node) { in fsl_dma_tx_submit() 470 INIT_LIST_HEAD(&desc->tx_list); in fsl_dma_alloc_descriptor() 805 list_add_tail(&new->node, &first->tx_list); in fsl_dma_prep_memcpy() 820 fsldma_free_desc_list_reverse(chan, &first->tx_list); in fsl_dma_prep_memcpy()
|
| H A D | txx9dmac.h | 230 struct list_head tx_list; member
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | ipoib_tx.c | 412 &txq->tx_list, in hfi1_ipoib_submit_tx_list() 427 if (!list_empty(&txq->tx_list)) { in hfi1_ipoib_flush_tx_list() 546 list_add_tail(&tx->txreq.list, &txq->tx_list); in hfi1_ipoib_send_dma_list() 599 if (netdev_xmit_more() || !list_empty(&txp.txq->tx_list)) in hfi1_ipoib_send() 632 list_add_tail(&txreq->list, &txq->tx_list); in hfi1_ipoib_sdma_sleep() 719 INIT_LIST_HEAD(&txq->tx_list); in hfi1_ipoib_txreq_init() 775 list_for_each_entry_safe(txreq, txreq_tmp, &txq->tx_list, list) { in hfi1_ipoib_drain_tx_list() 866 list_empty(&txq->tx_list)); in hfi1_ipoib_tx_timeout()
|
| H A D | ipoib.h | 112 struct list_head tx_list; member
|
| /linux/drivers/dma/dw/ |
| H A D | core.c | 89 INIT_LIST_HEAD(&desc->tx_list); in dwc_desc_get() 105 list_for_each_entry_safe(child, _next, &desc->tx_list, desc_node) { in dwc_desc_put() 201 dwc->tx_node_active = &first->tx_list; in dwc_dostart() 251 list_for_each_entry(child, &desc->tx_list, desc_node) in dwc_descriptor_complete() 323 head = &desc->tx_list; in dwc_scan_descriptors() 382 list_for_each_entry(child, &desc->tx_list, desc_node) { in dwc_scan_descriptors() 455 list_for_each_entry(child, &bad_desc->tx_list, desc_node) in dwc_handle_error() 593 list_add_tail(&desc->desc_node, &first->tx_list); in dwc_prep_dma_memcpy() 681 list_add_tail(&desc->desc_node, &first->tx_list); in dwc_prep_slave_sg() 730 list_add_tail(&desc->desc_node, &first->tx_list); in dwc_prep_slave_sg()
|
| /linux/drivers/hid/intel-ish-hid/ishtp/ |
| H A D | client.c | 81 INIT_LIST_HEAD(&cl->tx_list.list); in ishtp_cl_init() 737 have_msg_to_send = !list_empty(&cl->tx_list.list); in ishtp_cl_send() 738 list_add_tail(&cl_msg->list, &cl->tx_list.list); in ishtp_cl_send() 806 if (list_empty(&cl->tx_list.list)) { in ipc_tx_send() 823 cl_msg = list_entry(cl->tx_list.list.next, struct ishtp_cl_tx_ring, in ipc_tx_send() 905 if (list_empty(&cl->tx_list.list)) { in ishtp_cl_send_msg_dma() 910 cl_msg = list_entry(cl->tx_list.list.next, struct ishtp_cl_tx_ring, in ishtp_cl_send_msg_dma()
|
| H A D | client-buffers.c | 149 while (!list_empty(&cl->tx_list.list)) { in ishtp_cl_free_tx_ring() 150 tx_buf = list_entry(cl->tx_list.list.next, in ishtp_cl_free_tx_ring()
|
| /linux/drivers/dma/xilinx/ |
| H A D | zynqmp_dma.c | 185 struct list_head tx_list; member 396 if (!list_empty(&desc->tx_list)) in zynqmp_dma_tx_submit() 397 desc = list_last_entry(&desc->tx_list, in zynqmp_dma_tx_submit() 429 INIT_LIST_HEAD(&desc->tx_list); in zynqmp_dma_get_descriptor() 449 list_for_each_entry_safe(child, next, &sdesc->tx_list, node) { in zynqmp_dma_free_descriptor() 870 list_add_tail(&new->node, &first->tx_list); in zynqmp_dma_prep_memcpy()
|
| /linux/drivers/net/wireless/intersil/p54/ |
| H A D | p54spi.c | 435 struct p54s_tx_info, tx_list); in p54spi_wq_tx() 437 list_del_init(&entry->tx_list); in p54spi_wq_tx() 442 tx_list); in p54spi_wq_tx() 473 list_add_tail(&di->tx_list, &priv->tx_pending); in p54spi_op_tx()
|
| H A D | p54spi.h | 89 struct list_head tx_list; member
|
| /linux/include/linux/platform_data/ |
| H A D | dma-iop32x.h | 91 struct list_head tx_list; member
|
| /linux/drivers/dma/sh/ |
| H A D | shdma-base.c | 579 LIST_HEAD(tx_list); in shdma_prep_sg() 627 list_add_tail(&new->node, &tx_list); in shdma_prep_sg() 635 list_splice_tail(&tx_list, &schan->ld_free); in shdma_prep_sg() 642 list_for_each_entry(new, &tx_list, node) in shdma_prep_sg() 644 list_splice(&tx_list, &schan->ld_free); in shdma_prep_sg()
|
| /linux/drivers/net/ethernet/ti/ |
| H A D | tlan.c | 838 priv->tx_list = priv->rx_list + TLAN_NUM_RX_LISTS; in tlan_init() 1054 tail_list = priv->tx_list + priv->tx_tail; in tlan_start_tx() 1093 (priv->tx_list + (TLAN_NUM_TX_LISTS - 1))->forward in tlan_start_tx() 1096 (priv->tx_list + (priv->tx_tail - 1))->forward in tlan_start_tx() 1228 tlan_print_list(priv->tx_list + i, "TX", i); in tlan_get_stats() 1358 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eof() 1381 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eof() 1392 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eof() 1634 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eoc() 1902 list = priv->tx_list + i; in tlan_reset_lists() [all …]
|
| /linux/drivers/infiniband/sw/siw/ |
| H A D | siw_qp_tx.c | 1268 qp = container_of(fifo_list, struct siw_qp, tx_list); in siw_run_sq() 1270 qp->tx_list.next = NULL; in siw_run_sq() 1277 llist_for_each_entry(qp, active, tx_list) { in siw_run_sq() 1278 qp->tx_list.next = NULL; in siw_run_sq() 1301 llist_add(&qp->tx_list, &per_cpu(siw_tx_task_g, qp->tx_cpu).active); in siw_sq_start()
|
| /linux/include/net/ |
| H A D | xdp_sock.h | 66 struct list_head tx_list; member
|
| /linux/net/xdp/ |
| H A D | xsk.c | 64 list_for_each_entry_rcu(xs, &pool->xsk_tx_list, tx_list) { in xsk_set_tx_need_wakeup() 91 list_for_each_entry_rcu(xs, &pool->xsk_tx_list, tx_list) { in xsk_clear_tx_need_wakeup() 424 list_for_each_entry_rcu(xs, &pool->xsk_tx_list, tx_list) in xsk_tx_release() 437 list_for_each_entry_rcu(xs, &pool->xsk_tx_list, tx_list) { in xsk_tx_peek_desc() 465 list_for_each_entry_rcu(xs, &pool->xsk_tx_list, tx_list) in xsk_tx_peek_desc() 501 xs = list_first_or_null_rcu(&pool->xsk_tx_list, struct xdp_sock, tx_list); in xsk_tx_peek_release_desc_batch()
|
| /linux/drivers/net/ethernet/airoha/ |
| H A D | airoha_eth.c | 896 list_add_tail(&e->list, &q->tx_list); in airoha_qdma_tx_napi_poll() 946 INIT_LIST_HEAD(&q->tx_list); in airoha_qdma_init_tx_queue() 961 list_add_tail(&q->entry[i].list, &q->tx_list); in airoha_qdma_init_tx_queue() 1043 list_add_tail(&e->list, &q->tx_list); in airoha_qdma_cleanup_tx_queue() 1912 LIST_HEAD(tx_list); in airoha_dev_xmit() 1968 e = list_first_entry(&q->tx_list, struct airoha_queue_entry, in airoha_dev_xmit() 1983 list_move_tail(&e->list, &tx_list); in airoha_dev_xmit() 1988 e = list_first_entry(&q->tx_list, struct airoha_queue_entry, in airoha_dev_xmit() 2024 while (!list_empty(&tx_list)) { in airoha_dev_xmit() 2025 e = list_first_entry(&tx_list, struct airoha_queue_entry, in airoha_dev_xmit() [all …]
|
| /linux/net/tls/ |
| H A D | tls_sw.c | 416 rec = list_first_entry(&ctx->tx_list, in tls_tx_records() 437 list_for_each_entry_safe(rec, tmp, &ctx->tx_list, list) { in tls_tx_records() 510 first_rec = list_first_entry(&ctx->tx_list, in tls_encrypt_done() 578 list_add_tail((struct list_head *)&rec->list, &ctx->tx_list); in tls_do_encryption() 2555 rec = list_first_entry(&ctx->tx_list, in tls_sw_release_resources_tx() 2562 list_for_each_entry_safe(rec, tmp, &ctx->tx_list, list) { in tls_sw_release_resources_tx() 2661 rec = list_first_entry_or_null(&ctx->tx_list, struct tls_rec, list); in tls_is_tx_ready() 2710 INIT_LIST_HEAD(&sw_ctx_tx->tx_list); in init_ctx_tx()
|