Home
last modified time | relevance | path

Searched full:tx_desc (Results 1 – 25 of 104) sorted by relevance

12345

/linux/drivers/infiniband/ulp/isert/
H A Dib_isert.c780 isert_login_post_send(struct isert_conn *isert_conn, struct iser_tx_desc *tx_desc) in isert_login_post_send() argument
786 ib_dma_sync_single_for_device(ib_dev, tx_desc->dma_addr, in isert_login_post_send()
789 tx_desc->tx_cqe.done = isert_login_send_done; in isert_login_post_send()
792 send_wr.wr_cqe = &tx_desc->tx_cqe; in isert_login_post_send()
793 send_wr.sg_list = tx_desc->tx_sg; in isert_login_post_send()
794 send_wr.num_sge = tx_desc->num_sge; in isert_login_post_send()
807 struct iser_tx_desc *tx_desc) in __isert_create_send_desc() argument
810 memset(&tx_desc->iser_header, 0, sizeof(struct iser_ctrl)); in __isert_create_send_desc()
811 tx_desc->iser_header.flags = ISCSI_CTRL; in __isert_create_send_desc()
813 tx_desc->num_sge = 1; in __isert_create_send_desc()
[all …]
/linux/drivers/crypto/ccp/
H A Dccp-dmaengine.c111 if (!async_tx_test_ack(&desc->tx_desc)) in ccp_cleanup_desc_resources()
148 desc->tx_desc.cookie, cmd); in ccp_issue_next_cmd()
155 ret, desc->tx_desc.cookie, cmd); in ccp_issue_next_cmd()
170 __func__, desc->tx_desc.cookie, cmd); in ccp_free_active_cmd()
193 struct dma_async_tx_descriptor *tx_desc; in ccp_handle_active_desc() local
212 tx_desc = &desc->tx_desc; in ccp_handle_active_desc()
214 tx_desc = NULL; in ccp_handle_active_desc()
225 desc->tx_desc.cookie, desc->status); in ccp_handle_active_desc()
227 dma_cookie_complete(tx_desc); in ccp_handle_active_desc()
228 dma_descriptor_unmap(tx_desc); in ccp_handle_active_desc()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx4/
H A Den_tx.c242 struct mlx4_en_tx_desc *tx_desc = ring->buf + (index << LOG_TXBB_SIZE); in mlx4_en_stamp_wqe() local
245 __be32 *ptr = (__be32 *)tx_desc; in mlx4_en_stamp_wqe()
249 if (likely((void *)tx_desc + in mlx4_en_stamp_wqe()
282 struct mlx4_en_tx_desc *tx_desc = ring->buf + (index << LOG_TXBB_SIZE); in mlx4_en_free_tx_desc() local
283 struct mlx4_wqe_data_seg *data = (void *) tx_desc + tx_info->data_offset; in mlx4_en_free_tx_desc()
313 if (likely((void *)tx_desc + in mlx4_en_free_tx_desc()
400 struct mlx4_en_tx_desc *tx_desc; in mlx4_en_handle_err_cqe() local
414 tx_desc = ring->buf + (wqe_index << LOG_TXBB_SIZE); in mlx4_en_handle_err_cqe()
415 print_hex_dump(KERN_WARNING, "", DUMP_PREFIX_OFFSET, 16, 1, tx_desc, desc_size, false); in mlx4_en_handle_err_cqe()
687 static void build_inline_wqe(struct mlx4_en_tx_desc *tx_desc, in build_inline_wqe() argument
[all …]
/linux/drivers/net/ethernet/sunplus/
H A Dspl2sw_desc.c37 if (!comm->tx_desc) in spl2sw_tx_descs_clean()
41 comm->tx_desc[i].cmd1 = 0; in spl2sw_tx_descs_clean()
43 comm->tx_desc[i].cmd2 = 0; in spl2sw_tx_descs_clean()
44 comm->tx_desc[i].addr1 = 0; in spl2sw_tx_descs_clean()
45 comm->tx_desc[i].addr2 = 0; in spl2sw_tx_descs_clean()
103 comm->tx_desc = NULL; in spl2sw_descs_free()
119 memset(comm->tx_desc, '\0', sizeof(struct spl2sw_mac_desc) * in spl2sw_tx_descs_init()
187 comm->tx_desc = comm->desc_base; in spl2sw_descs_alloc()
190 comm->rx_desc[0] = &comm->tx_desc[TX_DESC_NUM + MAC_GUARD_DESC_NUM]; in spl2sw_descs_alloc()
214 comm->tx_desc = NULL; in spl2sw_descs_init()
/linux/drivers/infiniband/ulp/iser/
H A Diser_initiator.c144 struct iser_tx_desc *tx_desc, enum iser_desc_type type, in iser_create_send_desc() argument
149 tx_desc->type = type; in iser_create_send_desc()
150 tx_desc->cqe.done = done; in iser_create_send_desc()
153 tx_desc->dma_addr, ISER_HEADERS_LEN, DMA_TO_DEVICE); in iser_create_send_desc()
155 memset(&tx_desc->iser_header, 0, sizeof(struct iser_ctrl)); in iser_create_send_desc()
156 tx_desc->iser_header.flags = ISER_VER; in iser_create_send_desc()
157 tx_desc->num_sge = 1; in iser_create_send_desc()
351 struct iser_tx_desc *tx_desc = &iser_task->desc; in iser_send_command() local
356 iser_create_send_desc(iser_conn, tx_desc, ISCSI_TX_SCSI_COMMAND, in iser_send_command()
397 err = iser_post_send(&iser_conn->ib_conn, tx_desc); in iser_send_command()
[all …]
H A Diser_memory.c253 struct iser_tx_desc *tx_desc = &iser_task->desc; in iser_reg_sig_mr() local
257 struct ib_reg_wr *wr = &tx_desc->reg_wr; in iser_reg_sig_mr()
268 iser_inv_rkey(&tx_desc->inv_wr, mr, cqe, &wr->wr); in iser_reg_sig_mr()
281 wr->wr.next = &tx_desc->send_wr; in iser_reg_sig_mr()
310 struct iser_tx_desc *tx_desc = &iser_task->desc; in iser_fast_reg_mr() local
313 struct ib_reg_wr *wr = &tx_desc->reg_wr; in iser_fast_reg_mr()
317 iser_inv_rkey(&tx_desc->inv_wr, mr, cqe, &wr->wr); in iser_fast_reg_mr()
328 wr->wr.next = &tx_desc->send_wr; in iser_fast_reg_mr()
H A Diser_verbs.c853 * @tx_desc: iSER TX descriptor
857 int iser_post_send(struct ib_conn *ib_conn, struct iser_tx_desc *tx_desc) in iser_post_send() argument
859 struct ib_send_wr *wr = &tx_desc->send_wr; in iser_post_send()
864 tx_desc->dma_addr, ISER_HEADERS_LEN, in iser_post_send()
868 wr->wr_cqe = &tx_desc->cqe; in iser_post_send()
869 wr->sg_list = tx_desc->tx_sg; in iser_post_send()
870 wr->num_sge = tx_desc->num_sge; in iser_post_send()
874 if (tx_desc->inv_wr.next) in iser_post_send()
875 first_wr = &tx_desc->inv_wr; in iser_post_send()
876 else if (tx_desc->reg_wr.wr.next) in iser_post_send()
[all …]
/linux/drivers/dma/amd/ptdma/
H A Dptdma-dmaengine.c199 struct dma_async_tx_descriptor *tx_desc; in pt_handle_active_desc() local
214 tx_desc = &desc->vd.tx; in pt_handle_active_desc()
217 tx_desc = NULL; in pt_handle_active_desc()
227 dma_cookie_complete(tx_desc); in pt_handle_active_desc()
228 dma_descriptor_unmap(tx_desc); in pt_handle_active_desc()
232 tx_desc = NULL; in pt_handle_active_desc()
240 if (pt->ver != AE4_DMA_VERSION && tx_desc) { in pt_handle_active_desc()
241 dmaengine_desc_get_callback_invoke(tx_desc, NULL); in pt_handle_active_desc()
242 dma_run_dependencies(tx_desc); in pt_handle_active_desc()
338 struct dma_async_tx_descriptor *tx_desc; in pt_cmd_callback_work() local
[all …]
/linux/drivers/net/wireless/realtek/rtw88/
H A Dtx.c38 struct rtw_tx_desc *tx_desc = (struct rtw_tx_desc *)skb->data; in rtw_tx_fill_tx_desc() local
44 tx_desc->w0 = le32_encode_bits(pkt_info->tx_pkt_size, RTW_TX_DESC_W0_TXPKTSIZE) | in rtw_tx_fill_tx_desc()
50 tx_desc->w1 = le32_encode_bits(pkt_info->mac_id, RTW_TX_DESC_W1_MACID) | in rtw_tx_fill_tx_desc()
57 tx_desc->w2 = le32_encode_bits(pkt_info->ampdu_en, RTW_TX_DESC_W2_AGG_EN) | in rtw_tx_fill_tx_desc()
62 tx_desc->w3 = le32_encode_bits(pkt_info->hw_ssn_sel, RTW_TX_DESC_W3_HW_SSN_SEL) | in rtw_tx_fill_tx_desc()
69 tx_desc->w4 = le32_encode_bits(pkt_info->rate, RTW_TX_DESC_W4_DATARATE); in rtw_tx_fill_tx_desc()
72 tx_desc->w4 |= le32_encode_bits(0x1f, RTW_TX_DESC_W4_DATARATE_FB_LIMIT); in rtw_tx_fill_tx_desc()
74 tx_desc->w5 = le32_encode_bits(pkt_info->short_gi, RTW_TX_DESC_W5_DATA_SHORT) | in rtw_tx_fill_tx_desc()
79 tx_desc->w6 = le32_encode_bits(pkt_info->sn, RTW_TX_DESC_W6_SW_DEFINE); in rtw_tx_fill_tx_desc()
81 tx_desc->w8 = le32_encode_bits(pkt_info->en_hwseq, RTW_TX_DESC_W8_EN_HWSEQ); in rtw_tx_fill_tx_desc()
[all …]
H A Dtx.h123 struct rtw_tx_desc *tx_desc = (struct rtw_tx_desc *)txdesc; in fill_txdesc_checksum_common() local
125 le32p_replace_bits(&tx_desc->w7, 0, RTW_TX_DESC_W7_TXDESC_CHECKSUM); in fill_txdesc_checksum_common()
130 le32p_replace_bits(&tx_desc->w7, __le16_to_cpu(chksum), in fill_txdesc_checksum_common()
/linux/drivers/net/ethernet/intel/idpf/
H A Didpf_singleq_txrx.c198 struct idpf_base_tx_desc *tx_desc; in idpf_tx_singleq_map() local
211 tx_desc = &tx_q->base_tx[i]; in idpf_tx_singleq_map()
232 tx_desc->buf_addr = cpu_to_le64(dma); in idpf_tx_singleq_map()
238 tx_desc->qw1 = idpf_tx_singleq_build_ctob(td_cmd, in idpf_tx_singleq_map()
244 tx_desc = &tx_q->base_tx[0]; in idpf_tx_singleq_map()
248 tx_desc++; in idpf_tx_singleq_map()
257 tx_desc->buf_addr = cpu_to_le64(dma); in idpf_tx_singleq_map()
263 tx_desc->qw1 = idpf_tx_singleq_build_ctob(td_cmd, offsets, in idpf_tx_singleq_map()
268 tx_desc = &tx_q->base_tx[0]; in idpf_tx_singleq_map()
272 tx_desc++; in idpf_tx_singleq_map()
[all …]
/linux/drivers/net/ethernet/intel/fm10k/
H A Dfm10k_main.c746 struct fm10k_tx_desc *tx_desc; in fm10k_tso() local
775 tx_desc = FM10K_TX_DESC(tx_ring, tx_ring->next_to_use); in fm10k_tso()
776 tx_desc->hdrlen = hdrlen; in fm10k_tso()
777 tx_desc->mss = cpu_to_le16(skb_shinfo(skb)->gso_size); in fm10k_tso()
793 struct fm10k_tx_desc *tx_desc; in fm10k_tx_csum() local
869 tx_desc = FM10K_TX_DESC(tx_ring, tx_ring->next_to_use); in fm10k_tx_csum()
870 tx_desc->hdrlen = 0; in fm10k_tx_csum()
871 tx_desc->mss = 0; in fm10k_tx_csum()
892 struct fm10k_tx_desc *tx_desc, u16 i, in fm10k_tx_desc_push() argument
900 tx_desc->buffer_addr = cpu_to_le64(dma); in fm10k_tx_desc_push()
[all …]
/linux/drivers/net/ethernet/intel/ice/
H A Dice_xsk.c612 struct ice_tx_desc *tx_desc; in ice_clean_xdp_irq_zc() local
621 tx_desc = ICE_TX_DESC(xdp_ring, last_rs); in ice_clean_xdp_irq_zc()
622 if (tx_desc->cmd_type_offset_bsz & in ice_clean_xdp_irq_zc()
655 tx_desc->cmd_type_offset_bsz = 0; in ice_clean_xdp_irq_zc()
685 struct ice_tx_desc *tx_desc; in ice_xmit_xdp_tx_zc() local
706 tx_desc = ICE_TX_DESC(xdp_ring, ntu); in ice_xmit_xdp_tx_zc()
718 tx_desc->buf_addr = cpu_to_le64(dma); in ice_xmit_xdp_tx_zc()
719 tx_desc->cmd_type_offset_bsz = ice_build_ctob(0, 0, size, 0); in ice_xmit_xdp_tx_zc()
729 tx_desc = ICE_TX_DESC(xdp_ring, ntu); in ice_xmit_xdp_tx_zc()
739 tx_desc->cmd_type_offset_bsz |= in ice_xmit_xdp_tx_zc()
[all …]
H A Dice_txrx_lib.c298 struct ice_tx_desc *tx_desc; in ice_clean_xdp_irq() local
307 tx_desc = ICE_TX_DESC(xdp_ring, idx); in ice_clean_xdp_irq()
308 if (tx_desc->cmd_type_offset_bsz & in ice_clean_xdp_irq()
354 tx_desc->cmd_type_offset_bsz = 0; in ice_clean_xdp_irq()
375 struct ice_tx_desc *tx_desc; in __ice_xmit_xdp_ring() local
398 tx_desc = ICE_TX_DESC(xdp_ring, ntu); in __ice_xmit_xdp_ring()
420 tx_desc->buf_addr = cpu_to_le64(dma); in __ice_xmit_xdp_ring()
421 tx_desc->cmd_type_offset_bsz = ice_build_ctob(0, 0, size, 0); in __ice_xmit_xdp_ring()
430 tx_desc = ICE_TX_DESC(xdp_ring, ntu); in __ice_xmit_xdp_ring()
448 tx_desc->cmd_type_offset_bsz |= in __ice_xmit_xdp_ring()
H A Dice_txrx_lib.h104 struct ice_tx_desc *tx_desc; in ice_set_rs_bit() local
106 tx_desc = ICE_TX_DESC(xdp_ring, rs_idx); in ice_set_rs_bit()
107 tx_desc->cmd_type_offset_bsz |= in ice_set_rs_bit()
/linux/drivers/net/ethernet/intel/i40e/
H A Di40e_xsk.c512 struct i40e_tx_desc *tx_desc; in i40e_xmit_pkt() local
518 tx_desc = I40E_TX_DESC(xdp_ring, xdp_ring->next_to_use++); in i40e_xmit_pkt()
519 tx_desc->buffer_addr = cpu_to_le64(dma); in i40e_xmit_pkt()
520 tx_desc->cmd_type_offset_bsz = build_ctob(cmd, 0, desc->len, 0); in i40e_xmit_pkt()
529 struct i40e_tx_desc *tx_desc; in i40e_xmit_pkt_batch() local
540 tx_desc = I40E_TX_DESC(xdp_ring, ntu++); in i40e_xmit_pkt_batch()
541 tx_desc->buffer_addr = cpu_to_le64(dma); in i40e_xmit_pkt_batch()
542 tx_desc->cmd_type_offset_bsz = build_ctob(cmd, 0, desc[i].len, 0); in i40e_xmit_pkt_batch()
566 struct i40e_tx_desc *tx_desc; in i40e_set_rs_bit() local
568 tx_desc = I40E_TX_DESC(xdp_ring, ntu); in i40e_set_rs_bit()
[all …]
/linux/tools/testing/selftests/bpf/prog_tests/
H A Dxdp_metadata.c106 printf("%p: tx_desc[%d] -> %lx\n", xsk, i, addr); in open_xsk()
139 struct xdp_desc *tx_desc; in generate_packet() local
151 tx_desc = xsk_ring_prod__tx_desc(&xsk->tx, idx); in generate_packet()
152 tx_desc->addr = idx % (UMEM_NUM / 2) * UMEM_FRAME_SIZE + sizeof(struct xsk_tx_metadata); in generate_packet()
153 printf("%p: tx_desc[%u]->addr=%llx\n", xsk, idx, tx_desc->addr); in generate_packet()
154 data = xsk_umem__get_data(xsk->umem_area, tx_desc->addr); in generate_packet()
192 tx_desc->len = sizeof(*eth) + sizeof(*iph) + sizeof(*udph) + UDP_PAYLOAD_BYTES; in generate_packet()
193 tx_desc->options |= XDP_TX_METADATA; in generate_packet()
/linux/drivers/net/ethernet/hisilicon/hibmcge/
H A Dhbg_hw.c201 void hbg_hw_set_tx_desc(struct hbg_priv *priv, struct hbg_tx_desc *tx_desc) in hbg_hw_set_tx_desc() argument
203 hbg_reg_write(priv, HBG_REG_TX_CFF_ADDR_0_ADDR, tx_desc->word0); in hbg_hw_set_tx_desc()
204 hbg_reg_write(priv, HBG_REG_TX_CFF_ADDR_1_ADDR, tx_desc->word1); in hbg_hw_set_tx_desc()
205 hbg_reg_write(priv, HBG_REG_TX_CFF_ADDR_2_ADDR, tx_desc->word2); in hbg_hw_set_tx_desc()
206 hbg_reg_write(priv, HBG_REG_TX_CFF_ADDR_3_ADDR, tx_desc->word3); in hbg_hw_set_tx_desc()
H A Dhbg_txrx.c66 struct hbg_tx_desc *tx_desc) in hbg_init_tx_desc() argument
78 tx_desc->word0 = word0; in hbg_init_tx_desc()
79 tx_desc->word1 = FIELD_PREP(HBG_TX_DESC_W1_SEND_LEN_M, in hbg_init_tx_desc()
81 tx_desc->word2 = buffer->skb_dma; in hbg_init_tx_desc()
82 tx_desc->word3 = buffer->state_dma; in hbg_init_tx_desc()
94 struct hbg_tx_desc tx_desc; in hbg_net_start_xmit() local
118 hbg_init_tx_desc(buffer, &tx_desc); in hbg_net_start_xmit()
119 hbg_hw_set_tx_desc(priv, &tx_desc); in hbg_net_start_xmit()
/linux/drivers/spi/
H A Dspi-pxa2xx-dma.c139 struct dma_async_tx_descriptor *tx_desc, *rx_desc; in pxa2xx_spi_dma_prepare() local
142 tx_desc = pxa2xx_spi_dma_prepare_one(drv_data, DMA_MEM_TO_DEV, xfer); in pxa2xx_spi_dma_prepare()
143 if (!tx_desc) { in pxa2xx_spi_dma_prepare()
161 dmaengine_submit(tx_desc); in pxa2xx_spi_dma_prepare()
/linux/drivers/net/ethernet/oki-semi/pch_gbe/
H A Dpch_gbe_main.c1067 struct pch_gbe_tx_desc *tx_desc; in pch_gbe_tx_queue() local
1153 tx_desc = PCH_GBE_TX_DESC(*tx_ring, ring_num); in pch_gbe_tx_queue()
1154 tx_desc->buffer_addr = (buffer_info->dma); in pch_gbe_tx_queue()
1155 tx_desc->length = (tmp_skb->len); in pch_gbe_tx_queue()
1156 tx_desc->tx_words_eob = ((tmp_skb->len + 3)); in pch_gbe_tx_queue()
1157 tx_desc->tx_frame_ctrl = (frame_ctrl); in pch_gbe_tx_queue()
1158 tx_desc->gbec_status = (DSC_INIT16); in pch_gbe_tx_queue()
1426 struct pch_gbe_tx_desc *tx_desc; in pch_gbe_alloc_tx_buffers() local
1436 tx_desc = PCH_GBE_TX_DESC(*tx_ring, i); in pch_gbe_alloc_tx_buffers()
1437 tx_desc->gbec_status = (DSC_INIT16); in pch_gbe_alloc_tx_buffers()
[all …]
/linux/Documentation/devicetree/bindings/net/
H A Daltr,tse.yaml107 - const: tx_desc
123 reg-names = "control_port", "rx_csr", "rx_desc", "rx_resp", "tx_csr", "tx_desc", "pcs";
146 reg-names = "control_port", "rx_csr", "rx_desc", "rx_resp", "tx_csr", "tx_desc";
/linux/drivers/staging/gpib/eastwood/
H A Dfluke_gpib.c387 struct dma_async_tx_descriptor *tx_desc; in fluke_dma_write() local
405 tx_desc = dmaengine_prep_slave_single(e_priv->dma_channel, address, length, DMA_MEM_TO_DEV, in fluke_dma_write()
407 if (!tx_desc) { in fluke_dma_write()
412 tx_desc->callback = fluke_dma_callback; in fluke_dma_write()
413 tx_desc->callback_param = board; in fluke_dma_write()
418 dmaengine_submit(tx_desc); in fluke_dma_write()
565 struct dma_async_tx_descriptor *tx_desc; in fluke_dma_read() local
584 tx_desc = dmaengine_prep_slave_single(e_priv->dma_channel, in fluke_dma_read()
587 if (!tx_desc) { in fluke_dma_read()
592 tx_desc->callback = fluke_dma_callback; in fluke_dma_read()
[all …]
/linux/drivers/net/ethernet/wangxun/libwx/
H A Dwx_lib.c724 union wx_tx_desc *tx_desc; in wx_clean_tx_irq() local
730 tx_desc = WX_TX_DESC(tx_ring, i); in wx_clean_tx_irq()
772 while (tx_desc != eop_desc) { in wx_clean_tx_irq()
774 tx_desc++; in wx_clean_tx_irq()
779 tx_desc = WX_TX_DESC(tx_ring, 0); in wx_clean_tx_irq()
794 tx_desc++; in wx_clean_tx_irq()
799 tx_desc = WX_TX_DESC(tx_ring, 0); in wx_clean_tx_irq()
803 prefetch(tx_desc); in wx_clean_tx_irq()
933 static void wx_tx_olinfo_status(union wx_tx_desc *tx_desc, in wx_tx_olinfo_status() argument
951 tx_desc->read.olinfo_status = cpu_to_le32(olinfo_status); in wx_tx_olinfo_status()
[all …]
/linux/drivers/net/ethernet/marvell/
H A Dmvneta.c984 int tx_desc = txq->next_desc_to_proc; in mvneta_txq_next_desc_get() local
986 txq->next_desc_to_proc = MVNETA_QUEUE_NEXT_DESC(txq, tx_desc); in mvneta_txq_next_desc_get()
987 return txq->descs + tx_desc; in mvneta_txq_next_desc_get()
1882 struct mvneta_tx_desc *tx_desc = txq->descs + in mvneta_txq_bufs_free() local
1890 tx_desc->buf_phys_addr, in mvneta_txq_bufs_free()
1891 tx_desc->data_size, DMA_TO_DEVICE); in mvneta_txq_bufs_free()
2096 struct mvneta_tx_desc *tx_desc; in mvneta_xdp_submit_frame() local
2117 tx_desc = mvneta_txq_next_desc_get(txq); in mvneta_xdp_submit_frame()
2146 tx_desc->command = unlikely(i) ? 0 : MVNETA_TXD_F_DESC; in mvneta_xdp_submit_frame()
2147 tx_desc->buf_phys_addr = dma_addr; in mvneta_xdp_submit_frame()
[all …]

12345