| /linux/drivers/infiniband/ulp/isert/ |
| H A D | ib_isert.c | 780 isert_login_post_send(struct isert_conn *isert_conn, struct iser_tx_desc *tx_desc) in isert_login_post_send() argument 786 ib_dma_sync_single_for_device(ib_dev, tx_desc->dma_addr, in isert_login_post_send() 789 tx_desc->tx_cqe.done = isert_login_send_done; in isert_login_post_send() 792 send_wr.wr_cqe = &tx_desc->tx_cqe; in isert_login_post_send() 793 send_wr.sg_list = tx_desc->tx_sg; in isert_login_post_send() 794 send_wr.num_sge = tx_desc->num_sge; in isert_login_post_send() 807 struct iser_tx_desc *tx_desc) in __isert_create_send_desc() argument 810 memset(&tx_desc->iser_header, 0, sizeof(struct iser_ctrl)); in __isert_create_send_desc() 811 tx_desc->iser_header.flags = ISCSI_CTRL; in __isert_create_send_desc() 813 tx_desc->num_sge = 1; in __isert_create_send_desc() [all …]
|
| /linux/drivers/crypto/ccp/ |
| H A D | ccp-dmaengine.c | 111 if (!async_tx_test_ack(&desc->tx_desc)) in ccp_cleanup_desc_resources() 148 desc->tx_desc.cookie, cmd); in ccp_issue_next_cmd() 155 ret, desc->tx_desc.cookie, cmd); in ccp_issue_next_cmd() 170 __func__, desc->tx_desc.cookie, cmd); in ccp_free_active_cmd() 193 struct dma_async_tx_descriptor *tx_desc; in ccp_handle_active_desc() local 212 tx_desc = &desc->tx_desc; in ccp_handle_active_desc() 214 tx_desc = NULL; in ccp_handle_active_desc() 225 desc->tx_desc.cookie, desc->status); in ccp_handle_active_desc() 227 dma_cookie_complete(tx_desc); in ccp_handle_active_desc() 228 dma_descriptor_unmap(tx_desc); in ccp_handle_active_desc() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | en_tx.c | 242 struct mlx4_en_tx_desc *tx_desc = ring->buf + (index << LOG_TXBB_SIZE); in mlx4_en_stamp_wqe() local 245 __be32 *ptr = (__be32 *)tx_desc; in mlx4_en_stamp_wqe() 249 if (likely((void *)tx_desc + in mlx4_en_stamp_wqe() 282 struct mlx4_en_tx_desc *tx_desc = ring->buf + (index << LOG_TXBB_SIZE); in mlx4_en_free_tx_desc() local 283 struct mlx4_wqe_data_seg *data = (void *) tx_desc + tx_info->data_offset; in mlx4_en_free_tx_desc() 313 if (likely((void *)tx_desc + in mlx4_en_free_tx_desc() 400 struct mlx4_en_tx_desc *tx_desc; in mlx4_en_handle_err_cqe() local 414 tx_desc = ring->buf + (wqe_index << LOG_TXBB_SIZE); in mlx4_en_handle_err_cqe() 415 print_hex_dump(KERN_WARNING, "", DUMP_PREFIX_OFFSET, 16, 1, tx_desc, desc_size, false); in mlx4_en_handle_err_cqe() 687 static void build_inline_wqe(struct mlx4_en_tx_desc *tx_desc, in build_inline_wqe() argument [all …]
|
| /linux/drivers/net/ethernet/sunplus/ |
| H A D | spl2sw_desc.c | 37 if (!comm->tx_desc) in spl2sw_tx_descs_clean() 41 comm->tx_desc[i].cmd1 = 0; in spl2sw_tx_descs_clean() 43 comm->tx_desc[i].cmd2 = 0; in spl2sw_tx_descs_clean() 44 comm->tx_desc[i].addr1 = 0; in spl2sw_tx_descs_clean() 45 comm->tx_desc[i].addr2 = 0; in spl2sw_tx_descs_clean() 103 comm->tx_desc = NULL; in spl2sw_descs_free() 119 memset(comm->tx_desc, '\0', sizeof(struct spl2sw_mac_desc) * in spl2sw_tx_descs_init() 187 comm->tx_desc = comm->desc_base; in spl2sw_descs_alloc() 190 comm->rx_desc[0] = &comm->tx_desc[TX_DESC_NUM + MAC_GUARD_DESC_NUM]; in spl2sw_descs_alloc() 214 comm->tx_desc = NULL; in spl2sw_descs_init()
|
| /linux/drivers/infiniband/ulp/iser/ |
| H A D | iser_initiator.c | 144 struct iser_tx_desc *tx_desc, enum iser_desc_type type, in iser_create_send_desc() argument 149 tx_desc->type = type; in iser_create_send_desc() 150 tx_desc->cqe.done = done; in iser_create_send_desc() 153 tx_desc->dma_addr, ISER_HEADERS_LEN, DMA_TO_DEVICE); in iser_create_send_desc() 155 memset(&tx_desc->iser_header, 0, sizeof(struct iser_ctrl)); in iser_create_send_desc() 156 tx_desc->iser_header.flags = ISER_VER; in iser_create_send_desc() 157 tx_desc->num_sge = 1; in iser_create_send_desc() 351 struct iser_tx_desc *tx_desc = &iser_task->desc; in iser_send_command() local 356 iser_create_send_desc(iser_conn, tx_desc, ISCSI_TX_SCSI_COMMAND, in iser_send_command() 397 err = iser_post_send(&iser_conn->ib_conn, tx_desc); in iser_send_command() [all …]
|
| H A D | iser_memory.c | 253 struct iser_tx_desc *tx_desc = &iser_task->desc; in iser_reg_sig_mr() local 257 struct ib_reg_wr *wr = &tx_desc->reg_wr; in iser_reg_sig_mr() 268 iser_inv_rkey(&tx_desc->inv_wr, mr, cqe, &wr->wr); in iser_reg_sig_mr() 281 wr->wr.next = &tx_desc->send_wr; in iser_reg_sig_mr() 310 struct iser_tx_desc *tx_desc = &iser_task->desc; in iser_fast_reg_mr() local 313 struct ib_reg_wr *wr = &tx_desc->reg_wr; in iser_fast_reg_mr() 317 iser_inv_rkey(&tx_desc->inv_wr, mr, cqe, &wr->wr); in iser_fast_reg_mr() 328 wr->wr.next = &tx_desc->send_wr; in iser_fast_reg_mr()
|
| H A D | iser_verbs.c | 853 * @tx_desc: iSER TX descriptor 857 int iser_post_send(struct ib_conn *ib_conn, struct iser_tx_desc *tx_desc) in iser_post_send() argument 859 struct ib_send_wr *wr = &tx_desc->send_wr; in iser_post_send() 864 tx_desc->dma_addr, ISER_HEADERS_LEN, in iser_post_send() 868 wr->wr_cqe = &tx_desc->cqe; in iser_post_send() 869 wr->sg_list = tx_desc->tx_sg; in iser_post_send() 870 wr->num_sge = tx_desc->num_sge; in iser_post_send() 874 if (tx_desc->inv_wr.next) in iser_post_send() 875 first_wr = &tx_desc->inv_wr; in iser_post_send() 876 else if (tx_desc->reg_wr.wr.next) in iser_post_send() [all …]
|
| /linux/drivers/dma/amd/ptdma/ |
| H A D | ptdma-dmaengine.c | 199 struct dma_async_tx_descriptor *tx_desc; in pt_handle_active_desc() local 214 tx_desc = &desc->vd.tx; in pt_handle_active_desc() 217 tx_desc = NULL; in pt_handle_active_desc() 227 dma_cookie_complete(tx_desc); in pt_handle_active_desc() 228 dma_descriptor_unmap(tx_desc); in pt_handle_active_desc() 232 tx_desc = NULL; in pt_handle_active_desc() 240 if (pt->ver != AE4_DMA_VERSION && tx_desc) { in pt_handle_active_desc() 241 dmaengine_desc_get_callback_invoke(tx_desc, NULL); in pt_handle_active_desc() 242 dma_run_dependencies(tx_desc); in pt_handle_active_desc() 338 struct dma_async_tx_descriptor *tx_desc; in pt_cmd_callback_work() local [all …]
|
| /linux/drivers/net/ethernet/intel/ixgbe/ |
| H A D | ixgbe_xsk.c | 394 union ixgbe_adv_tx_desc *tx_desc = NULL; in ixgbe_xmit_zc() local 421 tx_desc = IXGBE_TX_DESC(xdp_ring, xdp_ring->next_to_use); in ixgbe_xmit_zc() 422 tx_desc->read.buffer_addr = cpu_to_le64(dma); in ixgbe_xmit_zc() 429 tx_desc->read.cmd_type_len = cpu_to_le32(cmd_type); in ixgbe_xmit_zc() 430 tx_desc->read.olinfo_status = in ixgbe_xmit_zc() 440 if (tx_desc) { in ixgbe_xmit_zc() 464 union ixgbe_adv_tx_desc *tx_desc; in ixgbe_clean_xdp_tx_irq() local 469 tx_desc = IXGBE_TX_DESC(tx_ring, ntc); in ixgbe_clean_xdp_tx_irq() 472 if (!(tx_desc->wb.status & cpu_to_le32(IXGBE_TXD_STAT_DD))) in ixgbe_clean_xdp_tx_irq() 486 tx_desc++; in ixgbe_clean_xdp_tx_irq() [all …]
|
| /linux/drivers/net/ethernet/intel/idpf/ |
| H A D | idpf_singleq_txrx.c | 216 union idpf_tx_flex_desc *tx_desc; in idpf_tx_singleq_dma_map_error() local 222 tx_desc = &txq->flex_tx[idx]; in idpf_tx_singleq_dma_map_error() 223 memset(tx_desc, 0, sizeof(*tx_desc)); in idpf_tx_singleq_dma_map_error() 249 struct idpf_base_tx_desc *tx_desc; in idpf_tx_singleq_map() local 262 tx_desc = &tx_q->base_tx[i]; in idpf_tx_singleq_map() 284 tx_desc->buf_addr = cpu_to_le64(dma); in idpf_tx_singleq_map() 290 tx_desc->qw1 = idpf_tx_singleq_build_ctob(td_cmd, in idpf_tx_singleq_map() 296 tx_desc = &tx_q->base_tx[0]; in idpf_tx_singleq_map() 300 tx_desc++; in idpf_tx_singleq_map() 309 tx_desc->buf_addr = cpu_to_le64(dma); in idpf_tx_singleq_map() [all …]
|
| /linux/drivers/net/wireless/realtek/rtw88/ |
| H A D | tx.c | 38 struct rtw_tx_desc *tx_desc = (struct rtw_tx_desc *)skb->data; in rtw_tx_fill_tx_desc() local 44 tx_desc->w0 = le32_encode_bits(pkt_info->tx_pkt_size, RTW_TX_DESC_W0_TXPKTSIZE) | in rtw_tx_fill_tx_desc() 50 tx_desc->w1 = le32_encode_bits(pkt_info->mac_id, RTW_TX_DESC_W1_MACID) | in rtw_tx_fill_tx_desc() 57 tx_desc->w2 = le32_encode_bits(pkt_info->ampdu_en, RTW_TX_DESC_W2_AGG_EN) | in rtw_tx_fill_tx_desc() 62 tx_desc->w3 = le32_encode_bits(pkt_info->hw_ssn_sel, RTW_TX_DESC_W3_HW_SSN_SEL) | in rtw_tx_fill_tx_desc() 69 tx_desc->w4 = le32_encode_bits(pkt_info->rate, RTW_TX_DESC_W4_DATARATE); in rtw_tx_fill_tx_desc() 72 tx_desc->w4 |= le32_encode_bits(0x1f, RTW_TX_DESC_W4_DATARATE_FB_LIMIT); in rtw_tx_fill_tx_desc() 74 tx_desc->w5 = le32_encode_bits(pkt_info->short_gi, RTW_TX_DESC_W5_DATA_SHORT) | in rtw_tx_fill_tx_desc() 79 tx_desc->w6 = le32_encode_bits(pkt_info->sn, RTW_TX_DESC_W6_SW_DEFINE); in rtw_tx_fill_tx_desc() 81 tx_desc->w8 = le32_encode_bits(pkt_info->en_hwseq, RTW_TX_DESC_W8_EN_HWSEQ); in rtw_tx_fill_tx_desc() [all …]
|
| H A D | tx.h | 123 struct rtw_tx_desc *tx_desc = (struct rtw_tx_desc *)txdesc; in fill_txdesc_checksum_common() local 125 le32p_replace_bits(&tx_desc->w7, 0, RTW_TX_DESC_W7_TXDESC_CHECKSUM); in fill_txdesc_checksum_common() 130 le32p_replace_bits(&tx_desc->w7, __le16_to_cpu(chksum), in fill_txdesc_checksum_common()
|
| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_txrx_lib.c | 298 struct ice_tx_desc *tx_desc; in ice_clean_xdp_irq() local 307 tx_desc = ICE_TX_DESC(xdp_ring, idx); in ice_clean_xdp_irq() 308 if (tx_desc->cmd_type_offset_bsz & in ice_clean_xdp_irq() 354 tx_desc->cmd_type_offset_bsz = 0; in ice_clean_xdp_irq() 375 struct ice_tx_desc *tx_desc; in __ice_xmit_xdp_ring() local 398 tx_desc = ICE_TX_DESC(xdp_ring, ntu); in __ice_xmit_xdp_ring() 420 tx_desc->buf_addr = cpu_to_le64(dma); in __ice_xmit_xdp_ring() 421 tx_desc->cmd_type_offset_bsz = ice_build_ctob(0, 0, size, 0); in __ice_xmit_xdp_ring() 430 tx_desc = ICE_TX_DESC(xdp_ring, ntu); in __ice_xmit_xdp_ring() 448 tx_desc->cmd_type_offset_bsz |= in __ice_xmit_xdp_ring()
|
| H A D | ice_txrx.c | 37 struct ice_tx_desc *tx_desc; in ice_prgm_fdir_fltr() local 73 tx_desc = ICE_TX_DESC(tx_ring, i); in ice_prgm_fdir_fltr() 83 tx_desc->buf_addr = cpu_to_le64(dma); in ice_prgm_fdir_fltr() 90 tx_desc->cmd_type_offset_bsz = in ice_prgm_fdir_fltr() 99 first->next_to_watch = tx_desc; in ice_prgm_fdir_fltr() 275 struct ice_tx_desc *tx_desc; in ice_clean_tx_irq() local 282 tx_desc = ICE_TX_DESC(tx_ring, i); in ice_clean_tx_irq() 299 ice_trace(clean_tx_irq, tx_ring, tx_desc, tx_buf); in ice_clean_tx_irq() 326 while (tx_desc != eop_desc) { in ice_clean_tx_irq() 327 ice_trace(clean_tx_irq_unmap, tx_ring, tx_desc, tx_buf); in ice_clean_tx_irq() [all …]
|
| /linux/drivers/net/ethernet/intel/i40e/ |
| H A D | i40e_xsk.c | 512 struct i40e_tx_desc *tx_desc; in i40e_xmit_pkt() local 518 tx_desc = I40E_TX_DESC(xdp_ring, xdp_ring->next_to_use++); in i40e_xmit_pkt() 519 tx_desc->buffer_addr = cpu_to_le64(dma); in i40e_xmit_pkt() 520 tx_desc->cmd_type_offset_bsz = build_ctob(cmd, 0, desc->len, 0); in i40e_xmit_pkt() 529 struct i40e_tx_desc *tx_desc; in i40e_xmit_pkt_batch() local 540 tx_desc = I40E_TX_DESC(xdp_ring, ntu++); in i40e_xmit_pkt_batch() 541 tx_desc->buffer_addr = cpu_to_le64(dma); in i40e_xmit_pkt_batch() 542 tx_desc->cmd_type_offset_bsz = build_ctob(cmd, 0, desc[i].len, 0); in i40e_xmit_pkt_batch() 566 struct i40e_tx_desc *tx_desc; in i40e_set_rs_bit() local 568 tx_desc = I40E_TX_DESC(xdp_ring, ntu); in i40e_set_rs_bit() [all …]
|
| /linux/tools/testing/selftests/bpf/prog_tests/ |
| H A D | xdp_metadata.c | 106 printf("%p: tx_desc[%d] -> %lx\n", xsk, i, addr); in open_xsk() 139 struct xdp_desc *tx_desc; in generate_packet() local 151 tx_desc = xsk_ring_prod__tx_desc(&xsk->tx, idx); in generate_packet() 152 tx_desc->addr = idx % (UMEM_NUM / 2) * UMEM_FRAME_SIZE + sizeof(struct xsk_tx_metadata); in generate_packet() 153 printf("%p: tx_desc[%u]->addr=%llx\n", xsk, idx, tx_desc->addr); in generate_packet() 154 data = xsk_umem__get_data(xsk->umem_area, tx_desc->addr); in generate_packet() 192 tx_desc->len = sizeof(*eth) + sizeof(*iph) + sizeof(*udph) + UDP_PAYLOAD_BYTES; in generate_packet() 193 tx_desc->options |= XDP_TX_METADATA; in generate_packet()
|
| /linux/drivers/net/wireless/ath/ath12k/ |
| H A D | dp_tx.c | 82 struct ath12k_tx_desc_info *tx_desc, in ath12k_dp_tx_release_txbuf() argument 86 tx_desc->skb_ext_desc = NULL; in ath12k_dp_tx_release_txbuf() 87 list_move_tail(&tx_desc->list, &dp->tx_desc_free_list[pool_id]); in ath12k_dp_tx_release_txbuf() 229 struct ath12k_tx_desc_info *tx_desc; in ath12k_dp_tx() local 277 tx_desc = ath12k_dp_tx_assign_buffer(dp, pool_id); in ath12k_dp_tx() 278 if (!tx_desc) in ath12k_dp_tx() 418 tx_desc->skb = skb; in ath12k_dp_tx() 419 tx_desc->mac_id = ar->pdev_idx; in ath12k_dp_tx() 420 ti.desc_id = tx_desc->desc_id; in ath12k_dp_tx() 458 tx_desc->skb_ext_desc = skb_ext_desc; in ath12k_dp_tx() [all …]
|
| /linux/drivers/net/ethernet/hisilicon/hibmcge/ |
| H A D | hbg_txrx.c | 66 struct hbg_tx_desc *tx_desc) in hbg_init_tx_desc() argument 78 tx_desc->word0 = word0; in hbg_init_tx_desc() 79 tx_desc->word1 = FIELD_PREP(HBG_TX_DESC_W1_SEND_LEN_M, in hbg_init_tx_desc() 81 tx_desc->word2 = buffer->skb_dma; in hbg_init_tx_desc() 82 tx_desc->word3 = buffer->state_dma; in hbg_init_tx_desc() 94 struct hbg_tx_desc tx_desc; in hbg_net_start_xmit() local 118 hbg_init_tx_desc(buffer, &tx_desc); in hbg_net_start_xmit() 119 hbg_hw_set_tx_desc(priv, &tx_desc); in hbg_net_start_xmit()
|
| H A D | hbg_hw.c | 218 void hbg_hw_set_tx_desc(struct hbg_priv *priv, struct hbg_tx_desc *tx_desc) in hbg_hw_set_tx_desc() argument 220 hbg_reg_write(priv, HBG_REG_TX_CFF_ADDR_0_ADDR, tx_desc->word0); in hbg_hw_set_tx_desc() 221 hbg_reg_write(priv, HBG_REG_TX_CFF_ADDR_1_ADDR, tx_desc->word1); in hbg_hw_set_tx_desc() 222 hbg_reg_write(priv, HBG_REG_TX_CFF_ADDR_2_ADDR, tx_desc->word2); in hbg_hw_set_tx_desc() 223 hbg_reg_write(priv, HBG_REG_TX_CFF_ADDR_3_ADDR, tx_desc->word3); in hbg_hw_set_tx_desc()
|
| /linux/drivers/spi/ |
| H A D | spi-pxa2xx-dma.c | 139 struct dma_async_tx_descriptor *tx_desc, *rx_desc; in pxa2xx_spi_dma_prepare() local 142 tx_desc = pxa2xx_spi_dma_prepare_one(drv_data, DMA_MEM_TO_DEV, xfer); in pxa2xx_spi_dma_prepare() 143 if (!tx_desc) { in pxa2xx_spi_dma_prepare() 161 dmaengine_submit(tx_desc); in pxa2xx_spi_dma_prepare()
|
| /linux/drivers/net/ethernet/oki-semi/pch_gbe/ |
| H A D | pch_gbe_main.c | 1067 struct pch_gbe_tx_desc *tx_desc; in pch_gbe_tx_queue() local 1153 tx_desc = PCH_GBE_TX_DESC(*tx_ring, ring_num); in pch_gbe_tx_queue() 1154 tx_desc->buffer_addr = (buffer_info->dma); in pch_gbe_tx_queue() 1155 tx_desc->length = (tmp_skb->len); in pch_gbe_tx_queue() 1156 tx_desc->tx_words_eob = ((tmp_skb->len + 3)); in pch_gbe_tx_queue() 1157 tx_desc->tx_frame_ctrl = (frame_ctrl); in pch_gbe_tx_queue() 1158 tx_desc->gbec_status = (DSC_INIT16); in pch_gbe_tx_queue() 1426 struct pch_gbe_tx_desc *tx_desc; in pch_gbe_alloc_tx_buffers() local 1436 tx_desc = PCH_GBE_TX_DESC(*tx_ring, i); in pch_gbe_alloc_tx_buffers() 1437 tx_desc->gbec_status = (DSC_INIT16); in pch_gbe_alloc_tx_buffers() [all …]
|
| /linux/Documentation/devicetree/bindings/net/ |
| H A D | altr,tse.yaml | 107 - const: tx_desc 123 reg-names = "control_port", "rx_csr", "rx_desc", "rx_resp", "tx_csr", "tx_desc", "pcs"; 146 reg-names = "control_port", "rx_csr", "rx_desc", "rx_resp", "tx_csr", "tx_desc";
|
| /linux/drivers/net/ethernet/intel/igb/ |
| H A D | igb_xsk.c | 470 union e1000_adv_tx_desc *tx_desc = NULL; in igb_xmit_zc() local 496 tx_desc = IGB_TX_DESC(tx_ring, tx_ring->next_to_use); in igb_xmit_zc() 497 tx_desc->read.buffer_addr = cpu_to_le64(dma); in igb_xmit_zc() 510 tx_desc->read.cmd_type_len = cpu_to_le32(cmd_type); in igb_xmit_zc() 511 tx_desc->read.olinfo_status = cpu_to_le32(olinfo_status); in igb_xmit_zc() 516 tx_buffer_info->next_to_watch = tx_desc; in igb_xmit_zc()
|
| /linux/drivers/net/ethernet/spacemit/ |
| H A D | k1_emac.c | 487 struct emac_desc *tx_desc; in emac_tx_clean_desc() local 497 tx_desc = &((struct emac_desc *)tx_ring->desc_addr)[i]; in emac_tx_clean_desc() 500 if (READ_ONCE(tx_desc->desc0) & TX_DESC_0_OWN) in emac_tx_clean_desc() 504 memset(tx_desc, 0, sizeof(struct emac_desc)); in emac_tx_clean_desc() 696 static int emac_tx_map_frag(struct device *dev, struct emac_desc *tx_desc, in emac_tx_map_frag() argument 730 tx_desc->buffer_addr_1 = addr; in emac_tx_map_frag() 731 tx_desc->desc1 |= FIELD_PREP(TX_DESC_1_BUFFER_SIZE_1_MASK, len); in emac_tx_map_frag() 733 tx_desc->buffer_addr_2 = addr; in emac_tx_map_frag() 734 tx_desc->desc1 |= FIELD_PREP(TX_DESC_1_BUFFER_SIZE_2_MASK, len); in emac_tx_map_frag() 743 struct emac_desc tx_desc, *tx_desc_addr; in emac_tx_mem_map() local [all …]
|
| /linux/drivers/net/ethernet/intel/igc/ |
| H A D | igc_tsn.h | 61 static inline bool igc_fpe_transmitted_smd_v(union igc_adv_tx_desc *tx_desc) in igc_fpe_transmitted_smd_v() argument 63 u32 olinfo_status = le32_to_cpu(tx_desc->read.olinfo_status); in igc_fpe_transmitted_smd_v()
|