Searched refs:tx_req (Results 1 – 8 of 8) sorted by relevance
/linux/arch/arm/mach-omap1/ |
H A D | devices.c | 123 unsigned rx_req, unsigned tx_req, in omap_mmc_add() argument 143 res[3].start = tx_req; in omap_mmc_add() 178 unsigned rx_req, tx_req; in omap1_init_mmc() local 191 tx_req = 21; in omap1_init_mmc() 199 tx_req = 54; in omap1_init_mmc() 207 rx_req, tx_req, mmc_data[i]); in omap1_init_mmc()
|
/linux/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_ib.c | 275 int ipoib_dma_map_tx(struct ib_device *ca, struct ipoib_tx_buf *tx_req) in ipoib_dma_map_tx() argument 277 struct sk_buff *skb = tx_req->skb; in ipoib_dma_map_tx() 278 u64 *mapping = tx_req->mapping; in ipoib_dma_map_tx() 318 struct ipoib_tx_buf *tx_req) in ipoib_dma_unmap_tx() argument 320 struct sk_buff *skb = tx_req->skb; in ipoib_dma_unmap_tx() 321 u64 *mapping = tx_req->mapping; in ipoib_dma_unmap_tx() 389 struct ipoib_tx_buf *tx_req; in ipoib_ib_handle_tx_wc() local 400 tx_req = &priv->tx_ring[wr_id]; in ipoib_ib_handle_tx_wc() 402 ipoib_dma_unmap_tx(priv, tx_req); in ipoib_ib_handle_tx_wc() 405 dev->stats.tx_bytes += tx_req->skb->len; in ipoib_ib_handle_tx_wc() [all …]
|
H A D | ipoib_cm.c | 699 struct ipoib_tx_buf *tx_req) in post_send() argument 701 ipoib_build_sge(priv, tx_req); in post_send() 711 struct ipoib_tx_buf *tx_req; in ipoib_cm_send() local 750 tx_req = &tx->tx_ring[tx->tx_head & (ipoib_sendq_size - 1)]; in ipoib_cm_send() 751 tx_req->skb = skb; in ipoib_cm_send() 753 if (unlikely(ipoib_dma_map_tx(priv->ca, tx_req))) { in ipoib_cm_send() 778 rc = post_send(priv, tx, tx->tx_head & (ipoib_sendq_size - 1), tx_req); in ipoib_cm_send() 782 ipoib_dma_unmap_tx(priv, tx_req); in ipoib_cm_send() 799 struct ipoib_tx_buf *tx_req; in ipoib_cm_handle_tx_wc() local 811 tx_req = &tx->tx_ring[wr_id]; in ipoib_cm_handle_tx_wc() [all …]
|
H A D | ipoib.h | 536 int ipoib_dma_map_tx(struct ib_device *ca, struct ipoib_tx_buf *tx_req); 538 struct ipoib_tx_buf *tx_req); 543 struct ipoib_tx_buf *tx_req) in ipoib_build_sge() argument 546 struct sk_buff *skb = tx_req->skb; in ipoib_build_sge() 549 u64 *mapping = tx_req->mapping; in ipoib_build_sge()
|
/linux/drivers/net/wwan/t7xx/ |
H A D | t7xx_hif_cldma.c | 843 static int t7xx_cldma_gpd_handle_tx_request(struct cldma_queue *queue, struct cldma_request *tx_req, in t7xx_cldma_gpd_handle_tx_request() argument 847 struct cldma_gpd *gpd = tx_req->gpd; in t7xx_cldma_gpd_handle_tx_request() 851 tx_req->mapped_buff = dma_map_single(md_ctrl->dev, skb->data, skb->len, DMA_TO_DEVICE); in t7xx_cldma_gpd_handle_tx_request() 853 if (dma_mapping_error(md_ctrl->dev, tx_req->mapped_buff)) { in t7xx_cldma_gpd_handle_tx_request() 858 t7xx_cldma_gpd_set_data_ptr(gpd, tx_req->mapped_buff); in t7xx_cldma_gpd_handle_tx_request() 870 tx_req->skb = skb; in t7xx_cldma_gpd_handle_tx_request() 923 struct cldma_request *tx_req; in t7xx_cldma_send_skb() local 948 tx_req = queue->tx_next; in t7xx_cldma_send_skb() 949 if (queue->budget > 0 && !tx_req->skb) { in t7xx_cldma_send_skb() 953 t7xx_cldma_gpd_handle_tx_request(queue, tx_req, skb); in t7xx_cldma_send_skb() [all …]
|
/linux/drivers/usb/gadget/udc/ |
H A D | fsl_qe_udc.h | 285 struct qe_req *tx_req; member
|
/linux/drivers/rapidio/ |
H A D | rio_cm.c | 138 struct tx_req { struct 674 struct tx_req *req, *_req; in rio_txcq_handler() 709 struct tx_req *treq; in riocm_queue_req()
|
/linux/drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
H A D | hclge_main.c | 9960 struct hclge_tx_vlan_type_cfg_cmd *tx_req; in hclge_set_vlan_protocol_type() local 9985 tx_req = (struct hclge_tx_vlan_type_cfg_cmd *)desc.data; in hclge_set_vlan_protocol_type() 9986 tx_req->ot_vlan_type = cpu_to_le16(hdev->vlan_type_cfg.tx_ot_vlan_type); in hclge_set_vlan_protocol_type() 9987 tx_req->in_vlan_type = cpu_to_le16(hdev->vlan_type_cfg.tx_in_vlan_type); in hclge_set_vlan_protocol_type()
|