Lines Matching defs:tfd

64 				       struct iwl_tfh_tfd *tfd,
77 ret = iwl_txq_gen2_set_tb(trans, tfd, phys, len);
132 ret = iwl_txq_gen2_set_tb(trans, tfd, phys, len);
162 struct iwl_tfh_tfd *tfd,
180 trace_iwlwifi_dev_tx(trans->dev, skb, tfd, sizeof(*tfd),
252 iwl_txq_gen2_set_tb(trans, tfd, tb_phys, tb_len);
272 ret = iwl_txq_gen2_set_tb_with_wa(trans, skb, tfd,
307 struct iwl_tfh_tfd *tfd = iwl_txq_get_tfd(trans, txq, idx);
319 iwl_txq_gen2_set_tb(trans, tfd, tb_phys, IWL_FIRST_TB_SIZE);
341 iwl_txq_gen2_set_tb(trans, tfd, tb_phys, len);
343 if (iwl_txq_gen2_build_amsdu(trans, skb, tfd, out_meta,
349 return tfd;
352 iwl_txq_gen2_tfd_unmap(trans, out_meta, tfd);
358 struct iwl_tfh_tfd *tfd,
374 ret = iwl_txq_gen2_set_tb_with_wa(trans, skb, tfd, tb_phys,
395 struct iwl_tfh_tfd *tfd = iwl_txq_get_tfd(trans, txq, idx);
411 iwl_txq_gen2_set_tb(trans, tfd, tb_phys, IWL_FIRST_TB_SIZE);
436 iwl_txq_gen2_set_tb(trans, tfd, tb_phys, tb1_len);
437 trace_iwlwifi_dev_tx(trans->dev, skb, tfd, sizeof(*tfd), &dev_cmd->hdr,
448 ret = iwl_txq_gen2_set_tb_with_wa(trans, skb, tfd, tb_phys,
455 if (iwl_txq_gen2_tx_add_frags(trans, skb, tfd, out_meta))
463 ret = iwl_txq_gen2_set_tb_with_wa(trans, skb, tfd, tb_phys,
469 if (iwl_txq_gen2_tx_add_frags(trans, frag, tfd, out_meta))
473 return tfd;
476 iwl_txq_gen2_tfd_unmap(trans, out_meta, tfd);
489 struct iwl_tfh_tfd *tfd = iwl_txq_get_tfd(trans, txq, idx);
503 memset(tfd, 0, sizeof(*tfd));
605 static u8 iwl_txq_gen2_get_num_tbs(struct iwl_tfh_tfd *tfd)
607 return le16_to_cpu(tfd->num_tbs) & 0x1f;
610 int iwl_txq_gen2_set_tb(struct iwl_trans *trans, struct iwl_tfh_tfd *tfd,
614 int idx = iwl_txq_gen2_get_num_tbs(tfd);
629 tb = &tfd->tbs[idx];
632 if (le16_to_cpu(tfd->num_tbs) >= trans_pcie->txqs.tfd.max_tbs) {
634 trans_pcie->txqs.tfd.max_tbs);
641 tfd->num_tbs = cpu_to_le16(idx + 1);
648 struct iwl_tfh_tfd *tfd)
654 num_tbs = iwl_txq_gen2_get_num_tbs(tfd);
656 if (num_tbs > trans_pcie->txqs.tfd.max_tbs) {
669 le64_to_cpu(tfd->tbs[i].addr),
670 le16_to_cpu(tfd->tbs[i].tb_len),
674 le64_to_cpu(tfd->tbs[i].addr),
675 le16_to_cpu(tfd->tbs[i].tb_len),
679 iwl_txq_set_tfd_invalid_gen2(trans, tfd);
734 void *tfd;
782 tfd = iwl_txq_gen2_build_tfd(trans, txq, dev_cmd, skb, out_meta);
783 if (!tfd) {
802 iwl_txq_gen2_get_num_tbs(tfd));
869 trans_pcie->txqs.tfd.size * txq->n_window,
1220 struct iwl_tfh_tfd *tfd;
1296 tfd = iwl_txq_get_tfd(trans, txq, txq->write_ptr);
1297 memset(tfd, 0, sizeof(*tfd));
1377 iwl_txq_gen2_set_tb(trans, tfd, iwl_txq_get_first_tb_dma(txq, idx),
1388 iwl_txq_gen2_tfd_unmap(trans, out_meta, tfd);
1391 iwl_txq_gen2_set_tb(trans, tfd, phys_addr,
1410 iwl_txq_gen2_tfd_unmap(trans, out_meta, tfd);
1413 iwl_txq_gen2_set_tb(trans, tfd, phys_addr, cmdlen[i]);