Lines Matching refs:tfd
148 static inline void iwl_pcie_gen1_tfd_set_tb(struct iwl_tfd *tfd,
151 struct iwl_tfd_tb *tb = &tfd->tbs[idx];
159 tfd->num_tbs = idx + 1;
162 static inline u8 iwl_txq_gen1_tfd_get_num_tbs(struct iwl_tfd *tfd)
164 return tfd->num_tbs & 0x1f;
171 void *tfd;
174 tfd = (u8 *)txq->tfds + trans_pcie->txqs.tfd.size * txq->write_ptr;
177 memset(tfd, 0, trans_pcie->txqs.tfd.size);
179 num_tbs = iwl_txq_gen1_tfd_get_num_tbs(tfd);
182 if (num_tbs >= trans_pcie->txqs.tfd.max_tbs) {
184 trans_pcie->txqs.tfd.max_tbs);
192 iwl_pcie_gen1_tfd_set_tb(tfd, num_tbs, addr, len);
264 iwl_txq_gen1_tfd_tb_get_addr(struct iwl_tfd *tfd, u8 idx)
266 struct iwl_tfd_tb *tb = &tfd->tbs[idx];
286 struct iwl_tfd *tfd)
288 tfd->num_tbs = 0;
290 iwl_pcie_gen1_tfd_set_tb(tfd, 0, trans->invalid_tx_cmd.dma,
300 struct iwl_tfd *tfd = iwl_txq_get_tfd(trans, txq, index);
303 num_tbs = iwl_txq_gen1_tfd_get_num_tbs(tfd);
305 if (num_tbs > trans_pcie->txqs.tfd.max_tbs) {
320 iwl_txq_gen1_tfd_tb_get_addr(tfd, i),
322 tfd, i),
326 iwl_txq_gen1_tfd_tb_get_addr(tfd, i),
328 tfd, i),
334 iwl_txq_set_tfd_invalid_gen1(trans, tfd);
463 trans_pcie->txqs.tfd.size *
743 tfd_sz = trans_pcie->txqs.tfd.size * num_entries;
785 void *tfd = iwl_txq_get_tfd(trans, txq, i);
788 iwl_txq_set_tfd_invalid_gen2(trans, tfd);
790 iwl_txq_set_tfd_invalid_gen1(trans, tfd);
922 "Max tfd queue size must be a power of two, but is %d",
1938 trans_pcie->txqs.tfd.size,
2124 void *tfd;
2243 trans_pcie->txqs.tfd.size,
2275 tfd = iwl_txq_get_tfd(trans, txq, txq->write_ptr);
2278 iwl_txq_gen1_tfd_get_num_tbs(tfd));