Lines Matching defs:trans_pcie

23 	struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
29 page_ptr = (void *)((u8 *)skb->cb + trans_pcie->txqs.page_offs);
564 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
589 WARN_ON(trans_pcie->txqs.bc_table_dword);
597 WARN_ON(!trans_pcie->txqs.bc_table_dword);
613 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
632 if (le16_to_cpu(tfd->num_tbs) >= trans_pcie->txqs.tfd.max_tbs) {
634 trans_pcie->txqs.tfd.max_tbs);
650 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
656 if (num_tbs > trans_pcie->txqs.tfd.max_tbs) {
729 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
731 struct iwl_txq *txq = trans_pcie->txqs.txq[txq_id];
740 if (WARN_ONCE(!test_bit(txq_id, trans_pcie->txqs.queue_used),
745 skb_shinfo(skb)->nr_frags > IWL_TRANS_PCIE_MAX_FRAGS(trans_pcie) &&
759 trans_pcie->txqs.dev_cmd_offs);
826 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
827 struct iwl_txq *txq = trans_pcie->txqs.txq[txq_id];
835 if (txq_id != trans_pcie->txqs.cmd.q_id) {
863 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
869 trans_pcie->txqs.tfd.size * txq->n_window,
878 dma_pool_free(trans_pcie->txqs.bc_pool,
893 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
901 txq = trans_pcie->txqs.txq[txq_id];
909 if (txq_id == trans_pcie->txqs.cmd.q_id)
918 trans_pcie->txqs.txq[txq_id] = NULL;
920 clear_bit(txq_id, trans_pcie->txqs.queue_used);
926 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
931 WARN_ON(!trans_pcie->txqs.bc_tbl_size);
933 bc_tbl_size = trans_pcie->txqs.bc_tbl_size;
943 txq->bc_tbl.addr = dma_pool_alloc(trans_pcie->txqs.bc_pool, GFP_KERNEL,
975 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
990 if (qid >= ARRAY_SIZE(trans_pcie->txqs.txq)) {
996 if (test_and_set_bit(qid, trans_pcie->txqs.queue_used)) {
1002 if (WARN_ONCE(trans_pcie->txqs.txq[qid],
1009 trans_pcie->txqs.txq[qid] = txq;
1030 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
1042 size = min_t(u32, size, trans_pcie->txqs.bc_tbl_size / sizeof(u16));
1067 if (trans_pcie->txqs.queue_alloc_cmd_ver == 0) {
1084 } else if (trans_pcie->txqs.queue_alloc_cmd_ver == 3) {
1115 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
1127 if (!test_and_clear_bit(queue, trans_pcie->txqs.queue_used)) {
1140 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
1143 memset(trans_pcie->txqs.queue_used, 0,
1144 sizeof(trans_pcie->txqs.queue_used));
1147 for (i = 0; i < ARRAY_SIZE(trans_pcie->txqs.txq); i++) {
1148 if (!trans_pcie->txqs.txq[i])
1157 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
1162 if (!trans_pcie->txqs.txq[txq_id]) {
1168 trans_pcie->txqs.txq[txq_id] = queue;
1175 queue = trans_pcie->txqs.txq[txq_id];
1179 (txq_id == trans_pcie->txqs.cmd.q_id));
1184 trans_pcie->txqs.txq[txq_id]->id = txq_id;
1185 set_bit(txq_id, trans_pcie->txqs.queue_used);
1208 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
1209 struct iwl_txq *txq = trans_pcie->txqs.txq[trans_pcie->txqs.cmd.q_id];
1324 cpu_to_le16(QUEUE_TO_SEQ(trans_pcie->txqs.cmd.q_id) |
1372 cmd_size, txq->write_ptr, idx, trans_pcie->txqs.cmd.q_id);
1428 spin_lock(&trans_pcie->reg_lock);
1432 spin_unlock(&trans_pcie->reg_lock);