xref: /linux/drivers/net/wireless/mediatek/mt76/mt7921/pci_mac.c (revision 92481c7d14b8030418f00c4b4ec65556565d892d)
1 // SPDX-License-Identifier: ISC
2 /* Copyright (C) 2021 MediaTek Inc. */
3 
4 #include "mt7921.h"
5 #include "../dma.h"
6 #include "mac.h"
7 
8 int mt7921e_tx_prepare_skb(struct mt76_dev *mdev, void *txwi_ptr,
9 			   enum mt76_txq_id qid, struct mt76_wcid *wcid,
10 			   struct ieee80211_sta *sta,
11 			   struct mt76_tx_info *tx_info)
12 {
13 	struct mt7921_dev *dev = container_of(mdev, struct mt7921_dev, mt76);
14 	struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx_info->skb);
15 	struct ieee80211_key_conf *key = info->control.hw_key;
16 	struct mt76_connac_hw_txp *txp;
17 	struct mt76_txwi_cache *t;
18 	int id, pid;
19 	u8 *txwi = (u8 *)txwi_ptr;
20 
21 	if (unlikely(tx_info->skb->len <= ETH_HLEN))
22 		return -EINVAL;
23 
24 	if (!wcid)
25 		wcid = &dev->mt76.global_wcid;
26 
27 	t = (struct mt76_txwi_cache *)(txwi + mdev->drv->txwi_size);
28 	t->skb = tx_info->skb;
29 
30 	id = mt76_token_consume(mdev, &t);
31 	if (id < 0)
32 		return id;
33 
34 	if (sta) {
35 		struct mt7921_sta *msta = (struct mt7921_sta *)sta->drv_priv;
36 
37 		if (time_after(jiffies, msta->last_txs + HZ / 4)) {
38 			info->flags |= IEEE80211_TX_CTL_REQ_TX_STATUS;
39 			msta->last_txs = jiffies;
40 		}
41 	}
42 
43 	pid = mt76_tx_status_skb_add(mdev, wcid, tx_info->skb);
44 	mt76_connac2_mac_write_txwi(mdev, txwi_ptr, tx_info->skb, wcid, key,
45 				    pid, qid, 0);
46 
47 	txp = (struct mt76_connac_hw_txp *)(txwi + MT_TXD_SIZE);
48 	memset(txp, 0, sizeof(struct mt76_connac_hw_txp));
49 	mt76_connac_write_hw_txp(mdev, tx_info, txp, id);
50 
51 	tx_info->skb = DMA_DUMMY_DATA;
52 
53 	return 0;
54 }
55 
56 static void
57 mt7921_txwi_free(struct mt7921_dev *dev, struct mt76_txwi_cache *t,
58 		 struct ieee80211_sta *sta, bool clear_status,
59 		 struct list_head *free_list)
60 {
61 	struct mt76_dev *mdev = &dev->mt76;
62 	__le32 *txwi;
63 	u16 wcid_idx;
64 
65 	mt76_connac_txp_skb_unmap(mdev, t);
66 	if (!t->skb)
67 		goto out;
68 
69 	txwi = (__le32 *)mt76_get_txwi_ptr(mdev, t);
70 	if (sta) {
71 		struct mt76_wcid *wcid = (struct mt76_wcid *)sta->drv_priv;
72 
73 		if (likely(t->skb->protocol != cpu_to_be16(ETH_P_PAE)))
74 			mt7921_tx_check_aggr(sta, txwi);
75 
76 		wcid_idx = wcid->idx;
77 	} else {
78 		wcid_idx = le32_get_bits(txwi[1], MT_TXD1_WLAN_IDX);
79 	}
80 
81 	__mt76_tx_complete_skb(mdev, wcid_idx, t->skb, free_list);
82 
83 out:
84 	t->skb = NULL;
85 	mt76_put_txwi(mdev, t);
86 }
87 
88 static void
89 mt7921e_mac_tx_free(struct mt7921_dev *dev, void *data, int len)
90 {
91 	struct mt76_connac_tx_free *free = data;
92 	__le32 *tx_info = (__le32 *)(data + sizeof(*free));
93 	struct mt76_dev *mdev = &dev->mt76;
94 	struct mt76_txwi_cache *txwi;
95 	struct ieee80211_sta *sta = NULL;
96 	struct sk_buff *skb, *tmp;
97 	void *end = data + len;
98 	LIST_HEAD(free_list);
99 	bool wake = false;
100 	u8 i, count;
101 
102 	/* clean DMA queues and unmap buffers first */
103 	mt76_queue_tx_cleanup(dev, dev->mphy.q_tx[MT_TXQ_PSD], false);
104 	mt76_queue_tx_cleanup(dev, dev->mphy.q_tx[MT_TXQ_BE], false);
105 
106 	count = le16_get_bits(free->ctrl, MT_TX_FREE_MSDU_CNT);
107 	if (WARN_ON_ONCE((void *)&tx_info[count] > end))
108 		return;
109 
110 	for (i = 0; i < count; i++) {
111 		u32 msdu, info = le32_to_cpu(tx_info[i]);
112 		u8 stat;
113 
114 		/* 1'b1: new wcid pair.
115 		 * 1'b0: msdu_id with the same 'wcid pair' as above.
116 		 */
117 		if (info & MT_TX_FREE_PAIR) {
118 			struct mt7921_sta *msta;
119 			struct mt76_wcid *wcid;
120 			u16 idx;
121 
122 			count++;
123 			idx = FIELD_GET(MT_TX_FREE_WLAN_ID, info);
124 			wcid = rcu_dereference(dev->mt76.wcid[idx]);
125 			sta = wcid_to_sta(wcid);
126 			if (!sta)
127 				continue;
128 
129 			msta = container_of(wcid, struct mt7921_sta, wcid);
130 			spin_lock_bh(&dev->sta_poll_lock);
131 			if (list_empty(&msta->poll_list))
132 				list_add_tail(&msta->poll_list, &dev->sta_poll_list);
133 			spin_unlock_bh(&dev->sta_poll_lock);
134 			continue;
135 		}
136 
137 		msdu = FIELD_GET(MT_TX_FREE_MSDU_ID, info);
138 		stat = FIELD_GET(MT_TX_FREE_STATUS, info);
139 
140 		txwi = mt76_token_release(mdev, msdu, &wake);
141 		if (!txwi)
142 			continue;
143 
144 		mt7921_txwi_free(dev, txwi, sta, stat, &free_list);
145 	}
146 
147 	if (wake)
148 		mt76_set_tx_blocked(&dev->mt76, false);
149 
150 	list_for_each_entry_safe(skb, tmp, &free_list, list) {
151 		skb_list_del_init(skb);
152 		napi_consume_skb(skb, 1);
153 	}
154 
155 	rcu_read_lock();
156 	mt7921_mac_sta_poll(dev);
157 	rcu_read_unlock();
158 
159 	mt76_worker_schedule(&dev->mt76.tx_worker);
160 }
161 
162 bool mt7921e_rx_check(struct mt76_dev *mdev, void *data, int len)
163 {
164 	struct mt7921_dev *dev = container_of(mdev, struct mt7921_dev, mt76);
165 	__le32 *rxd = (__le32 *)data;
166 	__le32 *end = (__le32 *)&rxd[len / 4];
167 	enum rx_pkt_type type;
168 
169 	type = le32_get_bits(rxd[0], MT_RXD0_PKT_TYPE);
170 
171 	switch (type) {
172 	case PKT_TYPE_TXRX_NOTIFY:
173 		mt7921e_mac_tx_free(dev, data, len);
174 		return false;
175 	case PKT_TYPE_TXS:
176 		for (rxd += 2; rxd + 8 <= end; rxd += 8)
177 			mt7921_mac_add_txs(dev, rxd);
178 		return false;
179 	default:
180 		return true;
181 	}
182 }
183 
184 void mt7921e_queue_rx_skb(struct mt76_dev *mdev, enum mt76_rxq_id q,
185 			  struct sk_buff *skb)
186 {
187 	struct mt7921_dev *dev = container_of(mdev, struct mt7921_dev, mt76);
188 	__le32 *rxd = (__le32 *)skb->data;
189 	enum rx_pkt_type type;
190 
191 	type = le32_get_bits(rxd[0], MT_RXD0_PKT_TYPE);
192 
193 	switch (type) {
194 	case PKT_TYPE_TXRX_NOTIFY:
195 		mt7921e_mac_tx_free(dev, skb->data, skb->len);
196 		napi_consume_skb(skb, 1);
197 		break;
198 	default:
199 		mt7921_queue_rx_skb(mdev, q, skb);
200 		break;
201 	}
202 }
203 
204 void mt7921_tx_token_put(struct mt7921_dev *dev)
205 {
206 	struct mt76_txwi_cache *txwi;
207 	int id;
208 
209 	spin_lock_bh(&dev->mt76.token_lock);
210 	idr_for_each_entry(&dev->mt76.token, txwi, id) {
211 		mt7921_txwi_free(dev, txwi, NULL, false, NULL);
212 		dev->mt76.token_count--;
213 	}
214 	spin_unlock_bh(&dev->mt76.token_lock);
215 	idr_destroy(&dev->mt76.token);
216 }
217 
218 int mt7921e_mac_reset(struct mt7921_dev *dev)
219 {
220 	int i, err;
221 
222 	mt7921e_mcu_drv_pmctrl(dev);
223 
224 	mt76_connac_free_pending_tx_skbs(&dev->pm, NULL);
225 
226 	mt76_wr(dev, MT_WFDMA0_HOST_INT_ENA, 0);
227 	mt76_wr(dev, MT_PCIE_MAC_INT_ENABLE, 0x0);
228 
229 	set_bit(MT76_RESET, &dev->mphy.state);
230 	set_bit(MT76_MCU_RESET, &dev->mphy.state);
231 	wake_up(&dev->mt76.mcu.wait);
232 	skb_queue_purge(&dev->mt76.mcu.res_q);
233 
234 	mt76_txq_schedule_all(&dev->mphy);
235 
236 	mt76_worker_disable(&dev->mt76.tx_worker);
237 	napi_disable(&dev->mt76.napi[MT_RXQ_MAIN]);
238 	napi_disable(&dev->mt76.napi[MT_RXQ_MCU]);
239 	napi_disable(&dev->mt76.napi[MT_RXQ_MCU_WA]);
240 	napi_disable(&dev->mt76.tx_napi);
241 
242 	mt7921_tx_token_put(dev);
243 	idr_init(&dev->mt76.token);
244 
245 	mt7921_wpdma_reset(dev, true);
246 
247 	local_bh_disable();
248 	mt76_for_each_q_rx(&dev->mt76, i) {
249 		napi_enable(&dev->mt76.napi[i]);
250 		napi_schedule(&dev->mt76.napi[i]);
251 	}
252 	local_bh_enable();
253 
254 	dev->fw_assert = false;
255 	clear_bit(MT76_MCU_RESET, &dev->mphy.state);
256 
257 	mt76_wr(dev, MT_WFDMA0_HOST_INT_ENA,
258 		MT_INT_RX_DONE_ALL | MT_INT_TX_DONE_ALL |
259 		MT_INT_MCU_CMD);
260 	mt76_wr(dev, MT_PCIE_MAC_INT_ENABLE, 0xff);
261 
262 	err = mt7921e_driver_own(dev);
263 	if (err)
264 		return err;
265 
266 	err = mt7921_run_firmware(dev);
267 	if (err)
268 		goto out;
269 
270 	err = mt7921_mcu_set_eeprom(dev);
271 	if (err)
272 		goto out;
273 
274 	err = mt7921_mac_init(dev);
275 	if (err)
276 		goto out;
277 
278 	err = __mt7921_start(&dev->phy);
279 out:
280 	clear_bit(MT76_RESET, &dev->mphy.state);
281 
282 	local_bh_disable();
283 	napi_enable(&dev->mt76.tx_napi);
284 	napi_schedule(&dev->mt76.tx_napi);
285 	local_bh_enable();
286 
287 	mt76_worker_enable(&dev->mt76.tx_worker);
288 
289 	return err;
290 }
291