xref: /freebsd/sys/contrib/dev/mediatek/mt76/mac80211.c (revision a8089ea5aee578e08acab2438e82fc9a9ae50ed8)
1 // SPDX-License-Identifier: ISC
2 /*
3  * Copyright (C) 2016 Felix Fietkau <nbd@nbd.name>
4  */
5 #include <linux/sched.h>
6 #if defined(CONFIG_OF)
7 #include <linux/of.h>
8 #endif
9 #include <net/page_pool.h>
10 #if defined(__FreeBSD__)
11 #include <linux/math64.h>
12 #include <linux/numa.h>
13 #endif
14 #include "mt76.h"
15 
16 #define CHAN2G(_idx, _freq) {			\
17 	.band = NL80211_BAND_2GHZ,		\
18 	.center_freq = (_freq),			\
19 	.hw_value = (_idx),			\
20 	.max_power = 30,			\
21 }
22 
23 #define CHAN5G(_idx, _freq) {			\
24 	.band = NL80211_BAND_5GHZ,		\
25 	.center_freq = (_freq),			\
26 	.hw_value = (_idx),			\
27 	.max_power = 30,			\
28 }
29 
30 #define CHAN6G(_idx, _freq) {			\
31 	.band = NL80211_BAND_6GHZ,		\
32 	.center_freq = (_freq),			\
33 	.hw_value = (_idx),			\
34 	.max_power = 30,			\
35 }
36 
37 static const struct ieee80211_channel mt76_channels_2ghz[] = {
38 	CHAN2G(1, 2412),
39 	CHAN2G(2, 2417),
40 	CHAN2G(3, 2422),
41 	CHAN2G(4, 2427),
42 	CHAN2G(5, 2432),
43 	CHAN2G(6, 2437),
44 	CHAN2G(7, 2442),
45 	CHAN2G(8, 2447),
46 	CHAN2G(9, 2452),
47 	CHAN2G(10, 2457),
48 	CHAN2G(11, 2462),
49 	CHAN2G(12, 2467),
50 	CHAN2G(13, 2472),
51 	CHAN2G(14, 2484),
52 };
53 
54 static const struct ieee80211_channel mt76_channels_5ghz[] = {
55 	CHAN5G(36, 5180),
56 	CHAN5G(40, 5200),
57 	CHAN5G(44, 5220),
58 	CHAN5G(48, 5240),
59 
60 	CHAN5G(52, 5260),
61 	CHAN5G(56, 5280),
62 	CHAN5G(60, 5300),
63 	CHAN5G(64, 5320),
64 
65 	CHAN5G(100, 5500),
66 	CHAN5G(104, 5520),
67 	CHAN5G(108, 5540),
68 	CHAN5G(112, 5560),
69 	CHAN5G(116, 5580),
70 	CHAN5G(120, 5600),
71 	CHAN5G(124, 5620),
72 	CHAN5G(128, 5640),
73 	CHAN5G(132, 5660),
74 	CHAN5G(136, 5680),
75 	CHAN5G(140, 5700),
76 	CHAN5G(144, 5720),
77 
78 	CHAN5G(149, 5745),
79 	CHAN5G(153, 5765),
80 	CHAN5G(157, 5785),
81 	CHAN5G(161, 5805),
82 	CHAN5G(165, 5825),
83 	CHAN5G(169, 5845),
84 	CHAN5G(173, 5865),
85 	CHAN5G(177, 5885),
86 };
87 
88 static const struct ieee80211_channel mt76_channels_6ghz[] = {
89 	/* UNII-5 */
90 	CHAN6G(1, 5955),
91 	CHAN6G(5, 5975),
92 	CHAN6G(9, 5995),
93 	CHAN6G(13, 6015),
94 	CHAN6G(17, 6035),
95 	CHAN6G(21, 6055),
96 	CHAN6G(25, 6075),
97 	CHAN6G(29, 6095),
98 	CHAN6G(33, 6115),
99 	CHAN6G(37, 6135),
100 	CHAN6G(41, 6155),
101 	CHAN6G(45, 6175),
102 	CHAN6G(49, 6195),
103 	CHAN6G(53, 6215),
104 	CHAN6G(57, 6235),
105 	CHAN6G(61, 6255),
106 	CHAN6G(65, 6275),
107 	CHAN6G(69, 6295),
108 	CHAN6G(73, 6315),
109 	CHAN6G(77, 6335),
110 	CHAN6G(81, 6355),
111 	CHAN6G(85, 6375),
112 	CHAN6G(89, 6395),
113 	CHAN6G(93, 6415),
114 	/* UNII-6 */
115 	CHAN6G(97, 6435),
116 	CHAN6G(101, 6455),
117 	CHAN6G(105, 6475),
118 	CHAN6G(109, 6495),
119 	CHAN6G(113, 6515),
120 	CHAN6G(117, 6535),
121 	/* UNII-7 */
122 	CHAN6G(121, 6555),
123 	CHAN6G(125, 6575),
124 	CHAN6G(129, 6595),
125 	CHAN6G(133, 6615),
126 	CHAN6G(137, 6635),
127 	CHAN6G(141, 6655),
128 	CHAN6G(145, 6675),
129 	CHAN6G(149, 6695),
130 	CHAN6G(153, 6715),
131 	CHAN6G(157, 6735),
132 	CHAN6G(161, 6755),
133 	CHAN6G(165, 6775),
134 	CHAN6G(169, 6795),
135 	CHAN6G(173, 6815),
136 	CHAN6G(177, 6835),
137 	CHAN6G(181, 6855),
138 	CHAN6G(185, 6875),
139 	/* UNII-8 */
140 	CHAN6G(189, 6895),
141 	CHAN6G(193, 6915),
142 	CHAN6G(197, 6935),
143 	CHAN6G(201, 6955),
144 	CHAN6G(205, 6975),
145 	CHAN6G(209, 6995),
146 	CHAN6G(213, 7015),
147 	CHAN6G(217, 7035),
148 	CHAN6G(221, 7055),
149 	CHAN6G(225, 7075),
150 	CHAN6G(229, 7095),
151 	CHAN6G(233, 7115),
152 };
153 
154 #if defined(CONFIG_MT76_LEDS)
155 static const struct ieee80211_tpt_blink mt76_tpt_blink[] = {
156 	{ .throughput =   0 * 1024, .blink_time = 334 },
157 	{ .throughput =   1 * 1024, .blink_time = 260 },
158 	{ .throughput =   5 * 1024, .blink_time = 220 },
159 	{ .throughput =  10 * 1024, .blink_time = 190 },
160 	{ .throughput =  20 * 1024, .blink_time = 170 },
161 	{ .throughput =  50 * 1024, .blink_time = 150 },
162 	{ .throughput =  70 * 1024, .blink_time = 130 },
163 	{ .throughput = 100 * 1024, .blink_time = 110 },
164 	{ .throughput = 200 * 1024, .blink_time =  80 },
165 	{ .throughput = 300 * 1024, .blink_time =  50 },
166 };
167 #endif
168 
169 struct ieee80211_rate mt76_rates[] = {
170 	CCK_RATE(0, 10),
171 	CCK_RATE(1, 20),
172 	CCK_RATE(2, 55),
173 	CCK_RATE(3, 110),
174 	OFDM_RATE(11, 60),
175 	OFDM_RATE(15, 90),
176 	OFDM_RATE(10, 120),
177 	OFDM_RATE(14, 180),
178 	OFDM_RATE(9,  240),
179 	OFDM_RATE(13, 360),
180 	OFDM_RATE(8,  480),
181 	OFDM_RATE(12, 540),
182 };
183 EXPORT_SYMBOL_GPL(mt76_rates);
184 
185 static const struct cfg80211_sar_freq_ranges mt76_sar_freq_ranges[] = {
186 	{ .start_freq = 2402, .end_freq = 2494, },
187 	{ .start_freq = 5150, .end_freq = 5350, },
188 	{ .start_freq = 5350, .end_freq = 5470, },
189 	{ .start_freq = 5470, .end_freq = 5725, },
190 	{ .start_freq = 5725, .end_freq = 5950, },
191 	{ .start_freq = 5945, .end_freq = 6165, },
192 	{ .start_freq = 6165, .end_freq = 6405, },
193 	{ .start_freq = 6405, .end_freq = 6525, },
194 	{ .start_freq = 6525, .end_freq = 6705, },
195 	{ .start_freq = 6705, .end_freq = 6865, },
196 	{ .start_freq = 6865, .end_freq = 7125, },
197 };
198 
199 static const struct cfg80211_sar_capa mt76_sar_capa = {
200 	.type = NL80211_SAR_TYPE_POWER,
201 	.num_freq_ranges = ARRAY_SIZE(mt76_sar_freq_ranges),
202 	.freq_ranges = &mt76_sar_freq_ranges[0],
203 };
204 
205 #if defined(CONFIG_MT76_LEDS)
206 static int mt76_led_init(struct mt76_phy *phy)
207 {
208 	struct mt76_dev *dev = phy->dev;
209 	struct ieee80211_hw *hw = phy->hw;
210 
211 	if (!phy->leds.cdev.brightness_set && !phy->leds.cdev.blink_set)
212 		return 0;
213 
214 	snprintf(phy->leds.name, sizeof(phy->leds.name), "mt76-%s",
215 		 wiphy_name(hw->wiphy));
216 
217 	phy->leds.cdev.name = phy->leds.name;
218 	phy->leds.cdev.default_trigger =
219 		ieee80211_create_tpt_led_trigger(hw,
220 					IEEE80211_TPT_LEDTRIG_FL_RADIO,
221 					mt76_tpt_blink,
222 					ARRAY_SIZE(mt76_tpt_blink));
223 
224 #if defined(CONFIG_OF)
225 	if (phy == &dev->phy) {
226 		struct device_node *np = dev->dev->of_node;
227 
228 		np = of_get_child_by_name(np, "led");
229 		if (np) {
230 			int led_pin;
231 
232 			if (!of_property_read_u32(np, "led-sources", &led_pin))
233 				phy->leds.pin = led_pin;
234 			phy->leds.al = of_property_read_bool(np,
235 							     "led-active-low");
236 			of_node_put(np);
237 		}
238 	}
239 #endif
240 
241 	return led_classdev_register(dev->dev, &phy->leds.cdev);
242 }
243 
244 static void mt76_led_cleanup(struct mt76_phy *phy)
245 {
246 	if (!phy->leds.cdev.brightness_set && !phy->leds.cdev.blink_set)
247 		return;
248 
249 	led_classdev_unregister(&phy->leds.cdev);
250 }
251 #endif
252 
253 static void mt76_init_stream_cap(struct mt76_phy *phy,
254 				 struct ieee80211_supported_band *sband,
255 				 bool vht)
256 {
257 	struct ieee80211_sta_ht_cap *ht_cap = &sband->ht_cap;
258 	int i, nstream = hweight8(phy->antenna_mask);
259 	struct ieee80211_sta_vht_cap *vht_cap;
260 	u16 mcs_map = 0;
261 
262 	if (nstream > 1)
263 		ht_cap->cap |= IEEE80211_HT_CAP_TX_STBC;
264 	else
265 		ht_cap->cap &= ~IEEE80211_HT_CAP_TX_STBC;
266 
267 	for (i = 0; i < IEEE80211_HT_MCS_MASK_LEN; i++)
268 		ht_cap->mcs.rx_mask[i] = i < nstream ? 0xff : 0;
269 
270 	if (!vht)
271 		return;
272 
273 	vht_cap = &sband->vht_cap;
274 	if (nstream > 1)
275 		vht_cap->cap |= IEEE80211_VHT_CAP_TXSTBC;
276 	else
277 		vht_cap->cap &= ~IEEE80211_VHT_CAP_TXSTBC;
278 	vht_cap->cap |= IEEE80211_VHT_CAP_TX_ANTENNA_PATTERN |
279 			IEEE80211_VHT_CAP_RX_ANTENNA_PATTERN;
280 
281 	for (i = 0; i < 8; i++) {
282 		if (i < nstream)
283 			mcs_map |= (IEEE80211_VHT_MCS_SUPPORT_0_9 << (i * 2));
284 		else
285 			mcs_map |=
286 				(IEEE80211_VHT_MCS_NOT_SUPPORTED << (i * 2));
287 	}
288 	vht_cap->vht_mcs.rx_mcs_map = cpu_to_le16(mcs_map);
289 	vht_cap->vht_mcs.tx_mcs_map = cpu_to_le16(mcs_map);
290 	if (ieee80211_hw_check(phy->hw, SUPPORTS_VHT_EXT_NSS_BW))
291 		vht_cap->vht_mcs.tx_highest |=
292 				cpu_to_le16(IEEE80211_VHT_EXT_NSS_BW_CAPABLE);
293 }
294 
295 void mt76_set_stream_caps(struct mt76_phy *phy, bool vht)
296 {
297 	if (phy->cap.has_2ghz)
298 		mt76_init_stream_cap(phy, &phy->sband_2g.sband, false);
299 	if (phy->cap.has_5ghz)
300 		mt76_init_stream_cap(phy, &phy->sband_5g.sband, vht);
301 	if (phy->cap.has_6ghz)
302 		mt76_init_stream_cap(phy, &phy->sband_6g.sband, vht);
303 }
304 EXPORT_SYMBOL_GPL(mt76_set_stream_caps);
305 
306 static int
307 mt76_init_sband(struct mt76_phy *phy, struct mt76_sband *msband,
308 		const struct ieee80211_channel *chan, int n_chan,
309 		struct ieee80211_rate *rates, int n_rates,
310 		bool ht, bool vht)
311 {
312 	struct ieee80211_supported_band *sband = &msband->sband;
313 	struct ieee80211_sta_vht_cap *vht_cap;
314 	struct ieee80211_sta_ht_cap *ht_cap;
315 	struct mt76_dev *dev = phy->dev;
316 	void *chanlist;
317 	int size;
318 
319 	size = n_chan * sizeof(*chan);
320 	chanlist = devm_kmemdup(dev->dev, chan, size, GFP_KERNEL);
321 	if (!chanlist)
322 		return -ENOMEM;
323 
324 	msband->chan = devm_kcalloc(dev->dev, n_chan, sizeof(*msband->chan),
325 				    GFP_KERNEL);
326 	if (!msband->chan)
327 		return -ENOMEM;
328 
329 	sband->channels = chanlist;
330 	sband->n_channels = n_chan;
331 	sband->bitrates = rates;
332 	sband->n_bitrates = n_rates;
333 
334 	if (!ht)
335 		return 0;
336 
337 	ht_cap = &sband->ht_cap;
338 	ht_cap->ht_supported = true;
339 	ht_cap->cap |= IEEE80211_HT_CAP_SUP_WIDTH_20_40 |
340 		       IEEE80211_HT_CAP_GRN_FLD |
341 		       IEEE80211_HT_CAP_SGI_20 |
342 		       IEEE80211_HT_CAP_SGI_40 |
343 		       (1 << IEEE80211_HT_CAP_RX_STBC_SHIFT);
344 
345 	ht_cap->mcs.tx_params = IEEE80211_HT_MCS_TX_DEFINED;
346 	ht_cap->ampdu_factor = IEEE80211_HT_MAX_AMPDU_64K;
347 
348 	mt76_init_stream_cap(phy, sband, vht);
349 
350 	if (!vht)
351 		return 0;
352 
353 	vht_cap = &sband->vht_cap;
354 	vht_cap->vht_supported = true;
355 	vht_cap->cap |= IEEE80211_VHT_CAP_RXLDPC |
356 			IEEE80211_VHT_CAP_RXSTBC_1 |
357 			IEEE80211_VHT_CAP_SHORT_GI_80 |
358 			(3 << IEEE80211_VHT_CAP_MAX_A_MPDU_LENGTH_EXPONENT_SHIFT);
359 
360 	return 0;
361 }
362 
363 static int
364 mt76_init_sband_2g(struct mt76_phy *phy, struct ieee80211_rate *rates,
365 		   int n_rates)
366 {
367 	phy->hw->wiphy->bands[NL80211_BAND_2GHZ] = &phy->sband_2g.sband;
368 
369 	return mt76_init_sband(phy, &phy->sband_2g, mt76_channels_2ghz,
370 			       ARRAY_SIZE(mt76_channels_2ghz), rates,
371 			       n_rates, true, false);
372 }
373 
374 static int
375 mt76_init_sband_5g(struct mt76_phy *phy, struct ieee80211_rate *rates,
376 		   int n_rates, bool vht)
377 {
378 	phy->hw->wiphy->bands[NL80211_BAND_5GHZ] = &phy->sband_5g.sband;
379 
380 	return mt76_init_sband(phy, &phy->sband_5g, mt76_channels_5ghz,
381 			       ARRAY_SIZE(mt76_channels_5ghz), rates,
382 			       n_rates, true, vht);
383 }
384 
385 static int
386 mt76_init_sband_6g(struct mt76_phy *phy, struct ieee80211_rate *rates,
387 		   int n_rates)
388 {
389 	phy->hw->wiphy->bands[NL80211_BAND_6GHZ] = &phy->sband_6g.sband;
390 
391 	return mt76_init_sband(phy, &phy->sband_6g, mt76_channels_6ghz,
392 			       ARRAY_SIZE(mt76_channels_6ghz), rates,
393 			       n_rates, false, false);
394 }
395 
396 static void
397 mt76_check_sband(struct mt76_phy *phy, struct mt76_sband *msband,
398 		 enum nl80211_band band)
399 {
400 	struct ieee80211_supported_band *sband = &msband->sband;
401 	bool found = false;
402 	int i;
403 
404 	if (!sband)
405 		return;
406 
407 	for (i = 0; i < sband->n_channels; i++) {
408 		if (sband->channels[i].flags & IEEE80211_CHAN_DISABLED)
409 			continue;
410 
411 		found = true;
412 		break;
413 	}
414 
415 	if (found) {
416 		phy->chandef.chan = &sband->channels[0];
417 		phy->chan_state = &msband->chan[0];
418 		return;
419 	}
420 
421 	sband->n_channels = 0;
422 	phy->hw->wiphy->bands[band] = NULL;
423 }
424 
425 static int
426 mt76_phy_init(struct mt76_phy *phy, struct ieee80211_hw *hw)
427 {
428 	struct mt76_dev *dev = phy->dev;
429 	struct wiphy *wiphy = hw->wiphy;
430 
431 	SET_IEEE80211_DEV(hw, dev->dev);
432 	SET_IEEE80211_PERM_ADDR(hw, phy->macaddr);
433 
434 	wiphy->features |= NL80211_FEATURE_ACTIVE_MONITOR |
435 			   NL80211_FEATURE_AP_MODE_CHAN_WIDTH_CHANGE;
436 	wiphy->flags |= WIPHY_FLAG_HAS_CHANNEL_SWITCH |
437 			WIPHY_FLAG_SUPPORTS_TDLS |
438 			WIPHY_FLAG_AP_UAPSD;
439 
440 	wiphy_ext_feature_set(wiphy, NL80211_EXT_FEATURE_CQM_RSSI_LIST);
441 	wiphy_ext_feature_set(wiphy, NL80211_EXT_FEATURE_AIRTIME_FAIRNESS);
442 	wiphy_ext_feature_set(wiphy, NL80211_EXT_FEATURE_AQL);
443 
444 	wiphy->available_antennas_tx = phy->antenna_mask;
445 	wiphy->available_antennas_rx = phy->antenna_mask;
446 
447 	wiphy->sar_capa = &mt76_sar_capa;
448 	phy->frp = devm_kcalloc(dev->dev, wiphy->sar_capa->num_freq_ranges,
449 				sizeof(struct mt76_freq_range_power),
450 				GFP_KERNEL);
451 	if (!phy->frp)
452 		return -ENOMEM;
453 
454 	hw->txq_data_size = sizeof(struct mt76_txq);
455 	hw->uapsd_max_sp_len = IEEE80211_WMM_IE_STA_QOSINFO_SP_ALL;
456 
457 	if (!hw->max_tx_fragments)
458 		hw->max_tx_fragments = 16;
459 
460 	ieee80211_hw_set(hw, SIGNAL_DBM);
461 	ieee80211_hw_set(hw, AMPDU_AGGREGATION);
462 	ieee80211_hw_set(hw, SUPPORTS_RC_TABLE);
463 	ieee80211_hw_set(hw, SUPPORT_FAST_XMIT);
464 	ieee80211_hw_set(hw, SUPPORTS_CLONED_SKBS);
465 	ieee80211_hw_set(hw, SUPPORTS_AMSDU_IN_AMPDU);
466 	ieee80211_hw_set(hw, SUPPORTS_REORDERING_BUFFER);
467 
468 	if (!(dev->drv->drv_flags & MT_DRV_AMSDU_OFFLOAD)) {
469 		ieee80211_hw_set(hw, TX_AMSDU);
470 		ieee80211_hw_set(hw, TX_FRAG_LIST);
471 	}
472 
473 	ieee80211_hw_set(hw, MFP_CAPABLE);
474 	ieee80211_hw_set(hw, AP_LINK_PS);
475 	ieee80211_hw_set(hw, REPORTS_TX_ACK_STATUS);
476 
477 	return 0;
478 }
479 
480 struct mt76_phy *
481 mt76_alloc_phy(struct mt76_dev *dev, unsigned int size,
482 	       const struct ieee80211_ops *ops, u8 band_idx)
483 {
484 	struct ieee80211_hw *hw;
485 	unsigned int phy_size;
486 	struct mt76_phy *phy;
487 
488 	phy_size = ALIGN(sizeof(*phy), 8);
489 	hw = ieee80211_alloc_hw(size + phy_size, ops);
490 	if (!hw)
491 		return NULL;
492 
493 	phy = hw->priv;
494 	phy->dev = dev;
495 	phy->hw = hw;
496 #if defined(__linux__)
497 	phy->priv = hw->priv + phy_size;
498 #elif defined(__FreeBSD__)
499 	phy->priv = (u8 *)hw->priv + phy_size;
500 #endif
501 	phy->band_idx = band_idx;
502 
503 	hw->wiphy->flags |= WIPHY_FLAG_IBSS_RSN;
504 	hw->wiphy->interface_modes =
505 		BIT(NL80211_IFTYPE_STATION) |
506 		BIT(NL80211_IFTYPE_AP) |
507 #ifdef CONFIG_MAC80211_MESH
508 		BIT(NL80211_IFTYPE_MESH_POINT) |
509 #endif
510 		BIT(NL80211_IFTYPE_P2P_CLIENT) |
511 		BIT(NL80211_IFTYPE_P2P_GO) |
512 		BIT(NL80211_IFTYPE_ADHOC);
513 
514 	return phy;
515 }
516 EXPORT_SYMBOL_GPL(mt76_alloc_phy);
517 
518 int mt76_register_phy(struct mt76_phy *phy, bool vht,
519 		      struct ieee80211_rate *rates, int n_rates)
520 {
521 	int ret;
522 
523 	ret = mt76_phy_init(phy, phy->hw);
524 	if (ret)
525 		return ret;
526 
527 	if (phy->cap.has_2ghz) {
528 		ret = mt76_init_sband_2g(phy, rates, n_rates);
529 		if (ret)
530 			return ret;
531 	}
532 
533 	if (phy->cap.has_5ghz) {
534 		ret = mt76_init_sband_5g(phy, rates + 4, n_rates - 4, vht);
535 		if (ret)
536 			return ret;
537 	}
538 
539 	if (phy->cap.has_6ghz) {
540 		ret = mt76_init_sband_6g(phy, rates + 4, n_rates - 4);
541 		if (ret)
542 			return ret;
543 	}
544 
545 #if defined(CONFIG_MT76_LEDS)
546 	if (IS_ENABLED(CONFIG_MT76_LEDS)) {
547 		ret = mt76_led_init(phy);
548 		if (ret)
549 			return ret;
550 	}
551 #endif
552 
553 	wiphy_read_of_freq_limits(phy->hw->wiphy);
554 	mt76_check_sband(phy, &phy->sband_2g, NL80211_BAND_2GHZ);
555 	mt76_check_sband(phy, &phy->sband_5g, NL80211_BAND_5GHZ);
556 	mt76_check_sband(phy, &phy->sband_6g, NL80211_BAND_6GHZ);
557 
558 	ret = ieee80211_register_hw(phy->hw);
559 	if (ret)
560 		return ret;
561 
562 	set_bit(MT76_STATE_REGISTERED, &phy->state);
563 	phy->dev->phys[phy->band_idx] = phy;
564 
565 	return 0;
566 }
567 EXPORT_SYMBOL_GPL(mt76_register_phy);
568 
569 void mt76_unregister_phy(struct mt76_phy *phy)
570 {
571 	struct mt76_dev *dev = phy->dev;
572 
573 	if (!test_bit(MT76_STATE_REGISTERED, &phy->state))
574 		return;
575 
576 #if defined(CONFIG_MT76_LEDS)
577 	if (IS_ENABLED(CONFIG_MT76_LEDS))
578 		mt76_led_cleanup(phy);
579 #endif
580 	mt76_tx_status_check(dev, true);
581 	ieee80211_unregister_hw(phy->hw);
582 	dev->phys[phy->band_idx] = NULL;
583 }
584 EXPORT_SYMBOL_GPL(mt76_unregister_phy);
585 
586 int mt76_create_page_pool(struct mt76_dev *dev, struct mt76_queue *q)
587 {
588 	struct page_pool_params pp_params = {
589 		.order = 0,
590 		.flags = PP_FLAG_PAGE_FRAG,
591 		.nid = NUMA_NO_NODE,
592 		.dev = dev->dma_dev,
593 	};
594 	int idx = q - dev->q_rx;
595 
596 	switch (idx) {
597 	case MT_RXQ_MAIN:
598 	case MT_RXQ_BAND1:
599 	case MT_RXQ_BAND2:
600 		pp_params.pool_size = 256;
601 		break;
602 	default:
603 		pp_params.pool_size = 16;
604 		break;
605 	}
606 
607 	if (mt76_is_mmio(dev)) {
608 		/* rely on page_pool for DMA mapping */
609 		pp_params.flags |= PP_FLAG_DMA_MAP | PP_FLAG_DMA_SYNC_DEV;
610 		pp_params.dma_dir = DMA_FROM_DEVICE;
611 		pp_params.max_len = PAGE_SIZE;
612 		pp_params.offset = 0;
613 	}
614 
615 	q->page_pool = page_pool_create(&pp_params);
616 	if (IS_ERR(q->page_pool)) {
617 		int err = PTR_ERR(q->page_pool);
618 
619 		q->page_pool = NULL;
620 		return err;
621 	}
622 
623 	return 0;
624 }
625 EXPORT_SYMBOL_GPL(mt76_create_page_pool);
626 
627 struct mt76_dev *
628 mt76_alloc_device(struct device *pdev, unsigned int size,
629 		  const struct ieee80211_ops *ops,
630 		  const struct mt76_driver_ops *drv_ops)
631 {
632 	struct ieee80211_hw *hw;
633 	struct mt76_phy *phy;
634 	struct mt76_dev *dev;
635 	int i;
636 
637 	hw = ieee80211_alloc_hw(size, ops);
638 	if (!hw)
639 		return NULL;
640 
641 	dev = hw->priv;
642 	dev->hw = hw;
643 	dev->dev = pdev;
644 	dev->drv = drv_ops;
645 	dev->dma_dev = pdev;
646 
647 	phy = &dev->phy;
648 	phy->dev = dev;
649 	phy->hw = hw;
650 	phy->band_idx = MT_BAND0;
651 	dev->phys[phy->band_idx] = phy;
652 
653 	spin_lock_init(&dev->rx_lock);
654 	spin_lock_init(&dev->lock);
655 	spin_lock_init(&dev->cc_lock);
656 	spin_lock_init(&dev->status_lock);
657 	spin_lock_init(&dev->wed_lock);
658 	mutex_init(&dev->mutex);
659 	init_waitqueue_head(&dev->tx_wait);
660 
661 	skb_queue_head_init(&dev->mcu.res_q);
662 	init_waitqueue_head(&dev->mcu.wait);
663 	mutex_init(&dev->mcu.mutex);
664 	dev->tx_worker.fn = mt76_tx_worker;
665 
666 	hw->wiphy->flags |= WIPHY_FLAG_IBSS_RSN;
667 	hw->wiphy->interface_modes =
668 		BIT(NL80211_IFTYPE_STATION) |
669 		BIT(NL80211_IFTYPE_AP) |
670 #ifdef CONFIG_MAC80211_MESH
671 		BIT(NL80211_IFTYPE_MESH_POINT) |
672 #endif
673 		BIT(NL80211_IFTYPE_P2P_CLIENT) |
674 		BIT(NL80211_IFTYPE_P2P_GO) |
675 		BIT(NL80211_IFTYPE_ADHOC);
676 
677 	spin_lock_init(&dev->token_lock);
678 	idr_init(&dev->token);
679 
680 	spin_lock_init(&dev->rx_token_lock);
681 	idr_init(&dev->rx_token);
682 
683 	INIT_LIST_HEAD(&dev->wcid_list);
684 	INIT_LIST_HEAD(&dev->sta_poll_list);
685 	spin_lock_init(&dev->sta_poll_lock);
686 
687 	INIT_LIST_HEAD(&dev->txwi_cache);
688 	INIT_LIST_HEAD(&dev->rxwi_cache);
689 	dev->token_size = dev->drv->token_size;
690 
691 	for (i = 0; i < ARRAY_SIZE(dev->q_rx); i++)
692 		skb_queue_head_init(&dev->rx_skb[i]);
693 
694 	dev->wq = alloc_ordered_workqueue("mt76", 0);
695 	if (!dev->wq) {
696 		ieee80211_free_hw(hw);
697 		return NULL;
698 	}
699 
700 	return dev;
701 }
702 EXPORT_SYMBOL_GPL(mt76_alloc_device);
703 
704 int mt76_register_device(struct mt76_dev *dev, bool vht,
705 			 struct ieee80211_rate *rates, int n_rates)
706 {
707 	struct ieee80211_hw *hw = dev->hw;
708 	struct mt76_phy *phy = &dev->phy;
709 	int ret;
710 
711 	dev_set_drvdata(dev->dev, dev);
712 	ret = mt76_phy_init(phy, hw);
713 	if (ret)
714 		return ret;
715 
716 	if (phy->cap.has_2ghz) {
717 		ret = mt76_init_sband_2g(phy, rates, n_rates);
718 		if (ret)
719 			return ret;
720 	}
721 
722 	if (phy->cap.has_5ghz) {
723 		ret = mt76_init_sband_5g(phy, rates + 4, n_rates - 4, vht);
724 		if (ret)
725 			return ret;
726 	}
727 
728 	if (phy->cap.has_6ghz) {
729 		ret = mt76_init_sband_6g(phy, rates + 4, n_rates - 4);
730 		if (ret)
731 			return ret;
732 	}
733 
734 	wiphy_read_of_freq_limits(hw->wiphy);
735 	mt76_check_sband(&dev->phy, &phy->sband_2g, NL80211_BAND_2GHZ);
736 	mt76_check_sband(&dev->phy, &phy->sband_5g, NL80211_BAND_5GHZ);
737 	mt76_check_sband(&dev->phy, &phy->sband_6g, NL80211_BAND_6GHZ);
738 
739 #if defined(CONFIG_MT76_LEDS)
740 	if (IS_ENABLED(CONFIG_MT76_LEDS)) {
741 		ret = mt76_led_init(phy);
742 		if (ret)
743 			return ret;
744 	}
745 #endif
746 
747 	ret = ieee80211_register_hw(hw);
748 	if (ret)
749 		return ret;
750 
751 	WARN_ON(mt76_worker_setup(hw, &dev->tx_worker, NULL, "tx"));
752 	set_bit(MT76_STATE_REGISTERED, &phy->state);
753 	sched_set_fifo_low(dev->tx_worker.task);
754 
755 	return 0;
756 }
757 EXPORT_SYMBOL_GPL(mt76_register_device);
758 
759 void mt76_unregister_device(struct mt76_dev *dev)
760 {
761 #if defined(__linux__)
762 	struct ieee80211_hw *hw = dev->hw;
763 #endif
764 
765 	if (!test_bit(MT76_STATE_REGISTERED, &dev->phy.state))
766 		return;
767 
768 #if defined(CONFIG_MT76_LEDS)
769 	if (IS_ENABLED(CONFIG_MT76_LEDS))
770 		mt76_led_cleanup(&dev->phy);
771 #endif
772 	mt76_tx_status_check(dev, true);
773 #if defined(__linux__)
774 	ieee80211_unregister_hw(hw);
775 #elif defined(__FreeBSD__)
776 	ieee80211_unregister_hw(dev->hw);
777 #endif
778 }
779 EXPORT_SYMBOL_GPL(mt76_unregister_device);
780 
781 void mt76_free_device(struct mt76_dev *dev)
782 {
783 	mt76_worker_teardown(&dev->tx_worker);
784 	if (dev->wq) {
785 		destroy_workqueue(dev->wq);
786 		dev->wq = NULL;
787 	}
788 	ieee80211_free_hw(dev->hw);
789 }
790 EXPORT_SYMBOL_GPL(mt76_free_device);
791 
792 static void mt76_rx_release_amsdu(struct mt76_phy *phy, enum mt76_rxq_id q)
793 {
794 	struct sk_buff *skb = phy->rx_amsdu[q].head;
795 	struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb;
796 	struct mt76_dev *dev = phy->dev;
797 
798 	phy->rx_amsdu[q].head = NULL;
799 	phy->rx_amsdu[q].tail = NULL;
800 
801 	/*
802 	 * Validate if the amsdu has a proper first subframe.
803 	 * A single MSDU can be parsed as A-MSDU when the unauthenticated A-MSDU
804 	 * flag of the QoS header gets flipped. In such cases, the first
805 	 * subframe has a LLC/SNAP header in the location of the destination
806 	 * address.
807 	 */
808 	if (skb_shinfo(skb)->frag_list) {
809 		int offset = 0;
810 
811 		if (!(status->flag & RX_FLAG_8023)) {
812 			offset = ieee80211_get_hdrlen_from_skb(skb);
813 
814 			if ((status->flag &
815 			     (RX_FLAG_DECRYPTED | RX_FLAG_IV_STRIPPED)) ==
816 			    RX_FLAG_DECRYPTED)
817 				offset += 8;
818 		}
819 
820 		if (ether_addr_equal(skb->data + offset, rfc1042_header)) {
821 			dev_kfree_skb(skb);
822 			return;
823 		}
824 	}
825 	__skb_queue_tail(&dev->rx_skb[q], skb);
826 }
827 
828 static void mt76_rx_release_burst(struct mt76_phy *phy, enum mt76_rxq_id q,
829 				  struct sk_buff *skb)
830 {
831 	struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb;
832 
833 	if (phy->rx_amsdu[q].head &&
834 	    (!status->amsdu || status->first_amsdu ||
835 	     status->seqno != phy->rx_amsdu[q].seqno))
836 		mt76_rx_release_amsdu(phy, q);
837 
838 	if (!phy->rx_amsdu[q].head) {
839 		phy->rx_amsdu[q].tail = &skb_shinfo(skb)->frag_list;
840 		phy->rx_amsdu[q].seqno = status->seqno;
841 		phy->rx_amsdu[q].head = skb;
842 	} else {
843 		*phy->rx_amsdu[q].tail = skb;
844 		phy->rx_amsdu[q].tail = &skb->next;
845 	}
846 
847 	if (!status->amsdu || status->last_amsdu)
848 		mt76_rx_release_amsdu(phy, q);
849 }
850 
851 void mt76_rx(struct mt76_dev *dev, enum mt76_rxq_id q, struct sk_buff *skb)
852 {
853 	struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb;
854 	struct mt76_phy *phy = mt76_dev_phy(dev, status->phy_idx);
855 
856 	if (!test_bit(MT76_STATE_RUNNING, &phy->state)) {
857 		dev_kfree_skb(skb);
858 		return;
859 	}
860 
861 #ifdef CONFIG_NL80211_TESTMODE
862 	if (phy->test.state == MT76_TM_STATE_RX_FRAMES) {
863 		phy->test.rx_stats.packets[q]++;
864 		if (status->flag & RX_FLAG_FAILED_FCS_CRC)
865 			phy->test.rx_stats.fcs_error[q]++;
866 	}
867 #endif
868 
869 	mt76_rx_release_burst(phy, q, skb);
870 }
871 EXPORT_SYMBOL_GPL(mt76_rx);
872 
873 bool mt76_has_tx_pending(struct mt76_phy *phy)
874 {
875 	struct mt76_queue *q;
876 	int i;
877 
878 	for (i = 0; i < __MT_TXQ_MAX; i++) {
879 		q = phy->q_tx[i];
880 		if (q && q->queued)
881 			return true;
882 	}
883 
884 	return false;
885 }
886 EXPORT_SYMBOL_GPL(mt76_has_tx_pending);
887 
888 static struct mt76_channel_state *
889 mt76_channel_state(struct mt76_phy *phy, struct ieee80211_channel *c)
890 {
891 	struct mt76_sband *msband;
892 	int idx;
893 
894 	if (c->band == NL80211_BAND_2GHZ)
895 		msband = &phy->sband_2g;
896 	else if (c->band == NL80211_BAND_6GHZ)
897 		msband = &phy->sband_6g;
898 	else
899 		msband = &phy->sband_5g;
900 
901 	idx = c - &msband->sband.channels[0];
902 	return &msband->chan[idx];
903 }
904 
905 void mt76_update_survey_active_time(struct mt76_phy *phy, ktime_t time)
906 {
907 	struct mt76_channel_state *state = phy->chan_state;
908 
909 	state->cc_active += ktime_to_us(ktime_sub(time,
910 						  phy->survey_time));
911 	phy->survey_time = time;
912 }
913 EXPORT_SYMBOL_GPL(mt76_update_survey_active_time);
914 
915 void mt76_update_survey(struct mt76_phy *phy)
916 {
917 	struct mt76_dev *dev = phy->dev;
918 	ktime_t cur_time;
919 
920 	if (dev->drv->update_survey)
921 		dev->drv->update_survey(phy);
922 
923 	cur_time = ktime_get_boottime();
924 	mt76_update_survey_active_time(phy, cur_time);
925 
926 	if (dev->drv->drv_flags & MT_DRV_SW_RX_AIRTIME) {
927 		struct mt76_channel_state *state = phy->chan_state;
928 
929 		spin_lock_bh(&dev->cc_lock);
930 		state->cc_bss_rx += dev->cur_cc_bss_rx;
931 		dev->cur_cc_bss_rx = 0;
932 		spin_unlock_bh(&dev->cc_lock);
933 	}
934 }
935 EXPORT_SYMBOL_GPL(mt76_update_survey);
936 
937 void mt76_set_channel(struct mt76_phy *phy)
938 {
939 	struct mt76_dev *dev = phy->dev;
940 	struct ieee80211_hw *hw = phy->hw;
941 	struct cfg80211_chan_def *chandef = &hw->conf.chandef;
942 	bool offchannel = hw->conf.flags & IEEE80211_CONF_OFFCHANNEL;
943 	int timeout = HZ / 5;
944 
945 	wait_event_timeout(dev->tx_wait, !mt76_has_tx_pending(phy), timeout);
946 	mt76_update_survey(phy);
947 
948 	if (phy->chandef.chan->center_freq != chandef->chan->center_freq ||
949 	    phy->chandef.width != chandef->width)
950 		phy->dfs_state = MT_DFS_STATE_UNKNOWN;
951 
952 	phy->chandef = *chandef;
953 	phy->chan_state = mt76_channel_state(phy, chandef->chan);
954 
955 	if (!offchannel)
956 		phy->main_chan = chandef->chan;
957 
958 	if (chandef->chan != phy->main_chan)
959 		memset(phy->chan_state, 0, sizeof(*phy->chan_state));
960 }
961 EXPORT_SYMBOL_GPL(mt76_set_channel);
962 
963 int mt76_get_survey(struct ieee80211_hw *hw, int idx,
964 		    struct survey_info *survey)
965 {
966 	struct mt76_phy *phy = hw->priv;
967 	struct mt76_dev *dev = phy->dev;
968 	struct mt76_sband *sband;
969 	struct ieee80211_channel *chan;
970 	struct mt76_channel_state *state;
971 	int ret = 0;
972 
973 	mutex_lock(&dev->mutex);
974 	if (idx == 0 && dev->drv->update_survey)
975 		mt76_update_survey(phy);
976 
977 	if (idx >= phy->sband_2g.sband.n_channels +
978 		   phy->sband_5g.sband.n_channels) {
979 		idx -= (phy->sband_2g.sband.n_channels +
980 			phy->sband_5g.sband.n_channels);
981 		sband = &phy->sband_6g;
982 	} else if (idx >= phy->sband_2g.sband.n_channels) {
983 		idx -= phy->sband_2g.sband.n_channels;
984 		sband = &phy->sband_5g;
985 	} else {
986 		sband = &phy->sband_2g;
987 	}
988 
989 	if (idx >= sband->sband.n_channels) {
990 		ret = -ENOENT;
991 		goto out;
992 	}
993 
994 	chan = &sband->sband.channels[idx];
995 	state = mt76_channel_state(phy, chan);
996 
997 	memset(survey, 0, sizeof(*survey));
998 	survey->channel = chan;
999 	survey->filled = SURVEY_INFO_TIME | SURVEY_INFO_TIME_BUSY;
1000 	survey->filled |= dev->drv->survey_flags;
1001 	if (state->noise)
1002 		survey->filled |= SURVEY_INFO_NOISE_DBM;
1003 
1004 	if (chan == phy->main_chan) {
1005 		survey->filled |= SURVEY_INFO_IN_USE;
1006 
1007 		if (dev->drv->drv_flags & MT_DRV_SW_RX_AIRTIME)
1008 			survey->filled |= SURVEY_INFO_TIME_BSS_RX;
1009 	}
1010 
1011 	survey->time_busy = div_u64(state->cc_busy, 1000);
1012 	survey->time_rx = div_u64(state->cc_rx, 1000);
1013 	survey->time = div_u64(state->cc_active, 1000);
1014 	survey->noise = state->noise;
1015 
1016 	spin_lock_bh(&dev->cc_lock);
1017 	survey->time_bss_rx = div_u64(state->cc_bss_rx, 1000);
1018 	survey->time_tx = div_u64(state->cc_tx, 1000);
1019 	spin_unlock_bh(&dev->cc_lock);
1020 
1021 out:
1022 	mutex_unlock(&dev->mutex);
1023 
1024 	return ret;
1025 }
1026 EXPORT_SYMBOL_GPL(mt76_get_survey);
1027 
1028 void mt76_wcid_key_setup(struct mt76_dev *dev, struct mt76_wcid *wcid,
1029 			 struct ieee80211_key_conf *key)
1030 {
1031 	struct ieee80211_key_seq seq;
1032 	int i;
1033 
1034 	wcid->rx_check_pn = false;
1035 
1036 	if (!key)
1037 		return;
1038 
1039 	if (key->cipher != WLAN_CIPHER_SUITE_CCMP)
1040 		return;
1041 
1042 	wcid->rx_check_pn = true;
1043 
1044 	/* data frame */
1045 	for (i = 0; i < IEEE80211_NUM_TIDS; i++) {
1046 		ieee80211_get_key_rx_seq(key, i, &seq);
1047 		memcpy(wcid->rx_key_pn[i], seq.ccmp.pn, sizeof(seq.ccmp.pn));
1048 	}
1049 
1050 	/* robust management frame */
1051 	ieee80211_get_key_rx_seq(key, -1, &seq);
1052 	memcpy(wcid->rx_key_pn[i], seq.ccmp.pn, sizeof(seq.ccmp.pn));
1053 
1054 }
1055 EXPORT_SYMBOL(mt76_wcid_key_setup);
1056 
1057 int mt76_rx_signal(u8 chain_mask, s8 *chain_signal)
1058 {
1059 	int signal = -128;
1060 	u8 chains;
1061 
1062 	for (chains = chain_mask; chains; chains >>= 1, chain_signal++) {
1063 		int cur, diff;
1064 
1065 		cur = *chain_signal;
1066 		if (!(chains & BIT(0)) ||
1067 		    cur > 0)
1068 			continue;
1069 
1070 		if (cur > signal)
1071 			swap(cur, signal);
1072 
1073 		diff = signal - cur;
1074 		if (diff == 0)
1075 			signal += 3;
1076 		else if (diff <= 2)
1077 			signal += 2;
1078 		else if (diff <= 6)
1079 			signal += 1;
1080 	}
1081 
1082 	return signal;
1083 }
1084 EXPORT_SYMBOL(mt76_rx_signal);
1085 
1086 static void
1087 mt76_rx_convert(struct mt76_dev *dev, struct sk_buff *skb,
1088 		struct ieee80211_hw **hw,
1089 		struct ieee80211_sta **sta)
1090 {
1091 	struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(skb);
1092 	struct ieee80211_hdr *hdr = mt76_skb_get_hdr(skb);
1093 	struct mt76_rx_status mstat;
1094 
1095 	mstat = *((struct mt76_rx_status *)skb->cb);
1096 	memset(status, 0, sizeof(*status));
1097 
1098 	status->flag = mstat.flag;
1099 	status->freq = mstat.freq;
1100 	status->enc_flags = mstat.enc_flags;
1101 	status->encoding = mstat.encoding;
1102 	status->bw = mstat.bw;
1103 	if (status->encoding == RX_ENC_EHT) {
1104 		status->eht.ru = mstat.eht.ru;
1105 		status->eht.gi = mstat.eht.gi;
1106 	} else {
1107 		status->he_ru = mstat.he_ru;
1108 		status->he_gi = mstat.he_gi;
1109 		status->he_dcm = mstat.he_dcm;
1110 	}
1111 	status->rate_idx = mstat.rate_idx;
1112 	status->nss = mstat.nss;
1113 	status->band = mstat.band;
1114 	status->signal = mstat.signal;
1115 	status->chains = mstat.chains;
1116 	status->ampdu_reference = mstat.ampdu_ref;
1117 	status->device_timestamp = mstat.timestamp;
1118 	status->mactime = mstat.timestamp;
1119 	status->signal = mt76_rx_signal(mstat.chains, mstat.chain_signal);
1120 	if (status->signal <= -128)
1121 		status->flag |= RX_FLAG_NO_SIGNAL_VAL;
1122 
1123 	if (ieee80211_is_beacon(hdr->frame_control) ||
1124 	    ieee80211_is_probe_resp(hdr->frame_control))
1125 		status->boottime_ns = ktime_get_boottime_ns();
1126 
1127 	BUILD_BUG_ON(sizeof(mstat) > sizeof(skb->cb));
1128 	BUILD_BUG_ON(sizeof(status->chain_signal) !=
1129 		     sizeof(mstat.chain_signal));
1130 	memcpy(status->chain_signal, mstat.chain_signal,
1131 	       sizeof(mstat.chain_signal));
1132 
1133 	*sta = wcid_to_sta(mstat.wcid);
1134 	*hw = mt76_phy_hw(dev, mstat.phy_idx);
1135 }
1136 
1137 static void
1138 mt76_check_ccmp_pn(struct sk_buff *skb)
1139 {
1140 	struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb;
1141 	struct mt76_wcid *wcid = status->wcid;
1142 	struct ieee80211_hdr *hdr;
1143 	int security_idx;
1144 	int ret;
1145 
1146 	if (!(status->flag & RX_FLAG_DECRYPTED))
1147 		return;
1148 
1149 	if (status->flag & RX_FLAG_ONLY_MONITOR)
1150 		return;
1151 
1152 	if (!wcid || !wcid->rx_check_pn)
1153 		return;
1154 
1155 	security_idx = status->qos_ctl & IEEE80211_QOS_CTL_TID_MASK;
1156 	if (status->flag & RX_FLAG_8023)
1157 		goto skip_hdr_check;
1158 
1159 	hdr = mt76_skb_get_hdr(skb);
1160 	if (!(status->flag & RX_FLAG_IV_STRIPPED)) {
1161 		/*
1162 		 * Validate the first fragment both here and in mac80211
1163 		 * All further fragments will be validated by mac80211 only.
1164 		 */
1165 		if (ieee80211_is_frag(hdr) &&
1166 		    !ieee80211_is_first_frag(hdr->frame_control))
1167 			return;
1168 	}
1169 
1170 	/* IEEE 802.11-2020, 12.5.3.4.4 "PN and replay detection" c):
1171 	 *
1172 	 * the recipient shall maintain a single replay counter for received
1173 	 * individually addressed robust Management frames that are received
1174 	 * with the To DS subfield equal to 0, [...]
1175 	 */
1176 	if (ieee80211_is_mgmt(hdr->frame_control) &&
1177 	    !ieee80211_has_tods(hdr->frame_control))
1178 		security_idx = IEEE80211_NUM_TIDS;
1179 
1180 skip_hdr_check:
1181 	BUILD_BUG_ON(sizeof(status->iv) != sizeof(wcid->rx_key_pn[0]));
1182 	ret = memcmp(status->iv, wcid->rx_key_pn[security_idx],
1183 		     sizeof(status->iv));
1184 	if (ret <= 0) {
1185 		status->flag |= RX_FLAG_ONLY_MONITOR;
1186 		return;
1187 	}
1188 
1189 	memcpy(wcid->rx_key_pn[security_idx], status->iv, sizeof(status->iv));
1190 
1191 	if (status->flag & RX_FLAG_IV_STRIPPED)
1192 		status->flag |= RX_FLAG_PN_VALIDATED;
1193 }
1194 
1195 static void
1196 mt76_airtime_report(struct mt76_dev *dev, struct mt76_rx_status *status,
1197 		    int len)
1198 {
1199 	struct mt76_wcid *wcid = status->wcid;
1200 	struct ieee80211_rx_status info = {
1201 		.enc_flags = status->enc_flags,
1202 		.rate_idx = status->rate_idx,
1203 		.encoding = status->encoding,
1204 		.band = status->band,
1205 		.nss = status->nss,
1206 		.bw = status->bw,
1207 	};
1208 	struct ieee80211_sta *sta;
1209 	u32 airtime;
1210 	u8 tidno = status->qos_ctl & IEEE80211_QOS_CTL_TID_MASK;
1211 
1212 	airtime = ieee80211_calc_rx_airtime(dev->hw, &info, len);
1213 	spin_lock(&dev->cc_lock);
1214 	dev->cur_cc_bss_rx += airtime;
1215 	spin_unlock(&dev->cc_lock);
1216 
1217 	if (!wcid || !wcid->sta)
1218 		return;
1219 
1220 	sta = container_of((void *)wcid, struct ieee80211_sta, drv_priv);
1221 	ieee80211_sta_register_airtime(sta, tidno, 0, airtime);
1222 }
1223 
1224 static void
1225 mt76_airtime_flush_ampdu(struct mt76_dev *dev)
1226 {
1227 	struct mt76_wcid *wcid;
1228 	int wcid_idx;
1229 
1230 	if (!dev->rx_ampdu_len)
1231 		return;
1232 
1233 	wcid_idx = dev->rx_ampdu_status.wcid_idx;
1234 	if (wcid_idx < ARRAY_SIZE(dev->wcid))
1235 		wcid = rcu_dereference(dev->wcid[wcid_idx]);
1236 	else
1237 		wcid = NULL;
1238 	dev->rx_ampdu_status.wcid = wcid;
1239 
1240 	mt76_airtime_report(dev, &dev->rx_ampdu_status, dev->rx_ampdu_len);
1241 
1242 	dev->rx_ampdu_len = 0;
1243 	dev->rx_ampdu_ref = 0;
1244 }
1245 
1246 static void
1247 mt76_airtime_check(struct mt76_dev *dev, struct sk_buff *skb)
1248 {
1249 	struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb;
1250 	struct mt76_wcid *wcid = status->wcid;
1251 
1252 	if (!(dev->drv->drv_flags & MT_DRV_SW_RX_AIRTIME))
1253 		return;
1254 
1255 	if (!wcid || !wcid->sta) {
1256 		struct ieee80211_hdr *hdr = mt76_skb_get_hdr(skb);
1257 
1258 		if (status->flag & RX_FLAG_8023)
1259 			return;
1260 
1261 		if (!ether_addr_equal(hdr->addr1, dev->phy.macaddr))
1262 			return;
1263 
1264 		wcid = NULL;
1265 	}
1266 
1267 	if (!(status->flag & RX_FLAG_AMPDU_DETAILS) ||
1268 	    status->ampdu_ref != dev->rx_ampdu_ref)
1269 		mt76_airtime_flush_ampdu(dev);
1270 
1271 	if (status->flag & RX_FLAG_AMPDU_DETAILS) {
1272 		if (!dev->rx_ampdu_len ||
1273 		    status->ampdu_ref != dev->rx_ampdu_ref) {
1274 			dev->rx_ampdu_status = *status;
1275 			dev->rx_ampdu_status.wcid_idx = wcid ? wcid->idx : 0xff;
1276 			dev->rx_ampdu_ref = status->ampdu_ref;
1277 		}
1278 
1279 		dev->rx_ampdu_len += skb->len;
1280 		return;
1281 	}
1282 
1283 	mt76_airtime_report(dev, status, skb->len);
1284 }
1285 
1286 static void
1287 mt76_check_sta(struct mt76_dev *dev, struct sk_buff *skb)
1288 {
1289 	struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb;
1290 	struct ieee80211_hdr *hdr = mt76_skb_get_hdr(skb);
1291 	struct ieee80211_sta *sta;
1292 	struct ieee80211_hw *hw;
1293 	struct mt76_wcid *wcid = status->wcid;
1294 	u8 tidno = status->qos_ctl & IEEE80211_QOS_CTL_TID_MASK;
1295 	bool ps;
1296 
1297 	hw = mt76_phy_hw(dev, status->phy_idx);
1298 	if (ieee80211_is_pspoll(hdr->frame_control) && !wcid &&
1299 	    !(status->flag & RX_FLAG_8023)) {
1300 		sta = ieee80211_find_sta_by_ifaddr(hw, hdr->addr2, NULL);
1301 		if (sta)
1302 			wcid = status->wcid = (struct mt76_wcid *)sta->drv_priv;
1303 	}
1304 
1305 	mt76_airtime_check(dev, skb);
1306 
1307 	if (!wcid || !wcid->sta)
1308 		return;
1309 
1310 	sta = container_of((void *)wcid, struct ieee80211_sta, drv_priv);
1311 
1312 	if (status->signal <= 0)
1313 		ewma_signal_add(&wcid->rssi, -status->signal);
1314 
1315 	wcid->inactive_count = 0;
1316 
1317 	if (status->flag & RX_FLAG_8023)
1318 		return;
1319 
1320 	if (!test_bit(MT_WCID_FLAG_CHECK_PS, &wcid->flags))
1321 		return;
1322 
1323 	if (ieee80211_is_pspoll(hdr->frame_control)) {
1324 		ieee80211_sta_pspoll(sta);
1325 		return;
1326 	}
1327 
1328 	if (ieee80211_has_morefrags(hdr->frame_control) ||
1329 	    !(ieee80211_is_mgmt(hdr->frame_control) ||
1330 	      ieee80211_is_data(hdr->frame_control)))
1331 		return;
1332 
1333 	ps = ieee80211_has_pm(hdr->frame_control);
1334 
1335 	if (ps && (ieee80211_is_data_qos(hdr->frame_control) ||
1336 		   ieee80211_is_qos_nullfunc(hdr->frame_control)))
1337 		ieee80211_sta_uapsd_trigger(sta, tidno);
1338 
1339 	if (!!test_bit(MT_WCID_FLAG_PS, &wcid->flags) == ps)
1340 		return;
1341 
1342 	if (ps)
1343 		set_bit(MT_WCID_FLAG_PS, &wcid->flags);
1344 
1345 	if (dev->drv->sta_ps)
1346 		dev->drv->sta_ps(dev, sta, ps);
1347 
1348 	if (!ps)
1349 		clear_bit(MT_WCID_FLAG_PS, &wcid->flags);
1350 
1351 	ieee80211_sta_ps_transition(sta, ps);
1352 }
1353 
1354 void mt76_rx_complete(struct mt76_dev *dev, struct sk_buff_head *frames,
1355 		      struct napi_struct *napi)
1356 {
1357 	struct ieee80211_sta *sta;
1358 	struct ieee80211_hw *hw;
1359 	struct sk_buff *skb, *tmp;
1360 #if defined(__linux__)
1361 	LIST_HEAD(list);
1362 #elif defined(__FreeBSD__)
1363 	LINUX_LIST_HEAD(list);
1364 #endif
1365 
1366 	spin_lock(&dev->rx_lock);
1367 	while ((skb = __skb_dequeue(frames)) != NULL) {
1368 		struct sk_buff *nskb = skb_shinfo(skb)->frag_list;
1369 
1370 		mt76_check_ccmp_pn(skb);
1371 		skb_shinfo(skb)->frag_list = NULL;
1372 		mt76_rx_convert(dev, skb, &hw, &sta);
1373 		ieee80211_rx_list(hw, sta, skb, &list);
1374 
1375 		/* subsequent amsdu frames */
1376 		while (nskb) {
1377 			skb = nskb;
1378 			nskb = nskb->next;
1379 			skb->next = NULL;
1380 
1381 			mt76_rx_convert(dev, skb, &hw, &sta);
1382 			ieee80211_rx_list(hw, sta, skb, &list);
1383 		}
1384 	}
1385 	spin_unlock(&dev->rx_lock);
1386 
1387 	if (!napi) {
1388 		netif_receive_skb_list(&list);
1389 		return;
1390 	}
1391 
1392 	list_for_each_entry_safe(skb, tmp, &list, list) {
1393 		skb_list_del_init(skb);
1394 		napi_gro_receive(napi, skb);
1395 	}
1396 }
1397 
1398 void mt76_rx_poll_complete(struct mt76_dev *dev, enum mt76_rxq_id q,
1399 			   struct napi_struct *napi)
1400 {
1401 	struct sk_buff_head frames;
1402 	struct sk_buff *skb;
1403 
1404 	__skb_queue_head_init(&frames);
1405 
1406 	while ((skb = __skb_dequeue(&dev->rx_skb[q])) != NULL) {
1407 		mt76_check_sta(dev, skb);
1408 		if (mtk_wed_device_active(&dev->mmio.wed))
1409 			__skb_queue_tail(&frames, skb);
1410 		else
1411 			mt76_rx_aggr_reorder(skb, &frames);
1412 	}
1413 
1414 	mt76_rx_complete(dev, &frames, napi);
1415 }
1416 EXPORT_SYMBOL_GPL(mt76_rx_poll_complete);
1417 
1418 static int
1419 mt76_sta_add(struct mt76_phy *phy, struct ieee80211_vif *vif,
1420 	     struct ieee80211_sta *sta)
1421 {
1422 	struct mt76_wcid *wcid = (struct mt76_wcid *)sta->drv_priv;
1423 	struct mt76_dev *dev = phy->dev;
1424 	int ret;
1425 	int i;
1426 
1427 	mutex_lock(&dev->mutex);
1428 
1429 	ret = dev->drv->sta_add(dev, vif, sta);
1430 	if (ret)
1431 		goto out;
1432 
1433 	for (i = 0; i < ARRAY_SIZE(sta->txq); i++) {
1434 		struct mt76_txq *mtxq;
1435 
1436 		if (!sta->txq[i])
1437 			continue;
1438 
1439 		mtxq = (struct mt76_txq *)sta->txq[i]->drv_priv;
1440 		mtxq->wcid = wcid->idx;
1441 	}
1442 
1443 	ewma_signal_init(&wcid->rssi);
1444 	if (phy->band_idx == MT_BAND1)
1445 		mt76_wcid_mask_set(dev->wcid_phy_mask, wcid->idx);
1446 	wcid->phy_idx = phy->band_idx;
1447 	rcu_assign_pointer(dev->wcid[wcid->idx], wcid);
1448 
1449 	mt76_packet_id_init(wcid);
1450 out:
1451 	mutex_unlock(&dev->mutex);
1452 
1453 	return ret;
1454 }
1455 
1456 void __mt76_sta_remove(struct mt76_dev *dev, struct ieee80211_vif *vif,
1457 		       struct ieee80211_sta *sta)
1458 {
1459 	struct mt76_wcid *wcid = (struct mt76_wcid *)sta->drv_priv;
1460 	int i, idx = wcid->idx;
1461 
1462 	for (i = 0; i < ARRAY_SIZE(wcid->aggr); i++)
1463 		mt76_rx_aggr_stop(dev, wcid, i);
1464 
1465 	if (dev->drv->sta_remove)
1466 		dev->drv->sta_remove(dev, vif, sta);
1467 
1468 	mt76_packet_id_flush(dev, wcid);
1469 
1470 	mt76_wcid_mask_clear(dev->wcid_mask, idx);
1471 	mt76_wcid_mask_clear(dev->wcid_phy_mask, idx);
1472 }
1473 EXPORT_SYMBOL_GPL(__mt76_sta_remove);
1474 
1475 static void
1476 mt76_sta_remove(struct mt76_dev *dev, struct ieee80211_vif *vif,
1477 		struct ieee80211_sta *sta)
1478 {
1479 	mutex_lock(&dev->mutex);
1480 	__mt76_sta_remove(dev, vif, sta);
1481 	mutex_unlock(&dev->mutex);
1482 }
1483 
1484 int mt76_sta_state(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
1485 		   struct ieee80211_sta *sta,
1486 		   enum ieee80211_sta_state old_state,
1487 		   enum ieee80211_sta_state new_state)
1488 {
1489 	struct mt76_phy *phy = hw->priv;
1490 	struct mt76_dev *dev = phy->dev;
1491 
1492 	if (old_state == IEEE80211_STA_NOTEXIST &&
1493 	    new_state == IEEE80211_STA_NONE)
1494 		return mt76_sta_add(phy, vif, sta);
1495 
1496 	if (old_state == IEEE80211_STA_AUTH &&
1497 	    new_state == IEEE80211_STA_ASSOC &&
1498 	    dev->drv->sta_assoc)
1499 		dev->drv->sta_assoc(dev, vif, sta);
1500 
1501 	if (old_state == IEEE80211_STA_NONE &&
1502 	    new_state == IEEE80211_STA_NOTEXIST)
1503 		mt76_sta_remove(dev, vif, sta);
1504 
1505 	return 0;
1506 }
1507 EXPORT_SYMBOL_GPL(mt76_sta_state);
1508 
1509 void mt76_sta_pre_rcu_remove(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
1510 			     struct ieee80211_sta *sta)
1511 {
1512 	struct mt76_phy *phy = hw->priv;
1513 	struct mt76_dev *dev = phy->dev;
1514 	struct mt76_wcid *wcid = (struct mt76_wcid *)sta->drv_priv;
1515 
1516 	mutex_lock(&dev->mutex);
1517 	spin_lock_bh(&dev->status_lock);
1518 	rcu_assign_pointer(dev->wcid[wcid->idx], NULL);
1519 	spin_unlock_bh(&dev->status_lock);
1520 	mutex_unlock(&dev->mutex);
1521 }
1522 EXPORT_SYMBOL_GPL(mt76_sta_pre_rcu_remove);
1523 
1524 int mt76_get_txpower(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
1525 		     int *dbm)
1526 {
1527 	struct mt76_phy *phy = hw->priv;
1528 	int n_chains = hweight8(phy->antenna_mask);
1529 	int delta = mt76_tx_power_nss_delta(n_chains);
1530 
1531 	*dbm = DIV_ROUND_UP(phy->txpower_cur + delta, 2);
1532 
1533 	return 0;
1534 }
1535 EXPORT_SYMBOL_GPL(mt76_get_txpower);
1536 
1537 int mt76_init_sar_power(struct ieee80211_hw *hw,
1538 			const struct cfg80211_sar_specs *sar)
1539 {
1540 	struct mt76_phy *phy = hw->priv;
1541 	const struct cfg80211_sar_capa *capa = hw->wiphy->sar_capa;
1542 	int i;
1543 
1544 	if (sar->type != NL80211_SAR_TYPE_POWER || !sar->num_sub_specs)
1545 		return -EINVAL;
1546 
1547 	for (i = 0; i < sar->num_sub_specs; i++) {
1548 		u32 index = sar->sub_specs[i].freq_range_index;
1549 		/* SAR specifies power limitaton in 0.25dbm */
1550 		s32 power = sar->sub_specs[i].power >> 1;
1551 
1552 		if (power > 127 || power < -127)
1553 			power = 127;
1554 
1555 		phy->frp[index].range = &capa->freq_ranges[index];
1556 		phy->frp[index].power = power;
1557 	}
1558 
1559 	return 0;
1560 }
1561 EXPORT_SYMBOL_GPL(mt76_init_sar_power);
1562 
1563 int mt76_get_sar_power(struct mt76_phy *phy,
1564 		       struct ieee80211_channel *chan,
1565 		       int power)
1566 {
1567 	const struct cfg80211_sar_capa *capa = phy->hw->wiphy->sar_capa;
1568 	int freq, i;
1569 
1570 	if (!capa || !phy->frp)
1571 		return power;
1572 
1573 	if (power > 127 || power < -127)
1574 		power = 127;
1575 
1576 	freq = ieee80211_channel_to_frequency(chan->hw_value, chan->band);
1577 	for (i = 0 ; i < capa->num_freq_ranges; i++) {
1578 		if (phy->frp[i].range &&
1579 		    freq >= phy->frp[i].range->start_freq &&
1580 		    freq < phy->frp[i].range->end_freq) {
1581 			power = min_t(int, phy->frp[i].power, power);
1582 			break;
1583 		}
1584 	}
1585 
1586 	return power;
1587 }
1588 EXPORT_SYMBOL_GPL(mt76_get_sar_power);
1589 
1590 static void
1591 __mt76_csa_finish(void *priv, u8 *mac, struct ieee80211_vif *vif)
1592 {
1593 	if (vif->bss_conf.csa_active && ieee80211_beacon_cntdwn_is_complete(vif))
1594 		ieee80211_csa_finish(vif);
1595 }
1596 
1597 void mt76_csa_finish(struct mt76_dev *dev)
1598 {
1599 	if (!dev->csa_complete)
1600 		return;
1601 
1602 	ieee80211_iterate_active_interfaces_atomic(dev->hw,
1603 		IEEE80211_IFACE_ITER_RESUME_ALL,
1604 		__mt76_csa_finish, dev);
1605 
1606 	dev->csa_complete = 0;
1607 }
1608 EXPORT_SYMBOL_GPL(mt76_csa_finish);
1609 
1610 static void
1611 __mt76_csa_check(void *priv, u8 *mac, struct ieee80211_vif *vif)
1612 {
1613 	struct mt76_dev *dev = priv;
1614 
1615 	if (!vif->bss_conf.csa_active)
1616 		return;
1617 
1618 	dev->csa_complete |= ieee80211_beacon_cntdwn_is_complete(vif);
1619 }
1620 
1621 void mt76_csa_check(struct mt76_dev *dev)
1622 {
1623 	ieee80211_iterate_active_interfaces_atomic(dev->hw,
1624 		IEEE80211_IFACE_ITER_RESUME_ALL,
1625 		__mt76_csa_check, dev);
1626 }
1627 EXPORT_SYMBOL_GPL(mt76_csa_check);
1628 
1629 int
1630 mt76_set_tim(struct ieee80211_hw *hw, struct ieee80211_sta *sta, bool set)
1631 {
1632 	return 0;
1633 }
1634 EXPORT_SYMBOL_GPL(mt76_set_tim);
1635 
1636 void mt76_insert_ccmp_hdr(struct sk_buff *skb, u8 key_id)
1637 {
1638 	struct mt76_rx_status *status = (struct mt76_rx_status *)skb->cb;
1639 	int hdr_len = ieee80211_get_hdrlen_from_skb(skb);
1640 	u8 *hdr, *pn = status->iv;
1641 
1642 	__skb_push(skb, 8);
1643 	memmove(skb->data, skb->data + 8, hdr_len);
1644 	hdr = skb->data + hdr_len;
1645 
1646 	hdr[0] = pn[5];
1647 	hdr[1] = pn[4];
1648 	hdr[2] = 0;
1649 	hdr[3] = 0x20 | (key_id << 6);
1650 	hdr[4] = pn[3];
1651 	hdr[5] = pn[2];
1652 	hdr[6] = pn[1];
1653 	hdr[7] = pn[0];
1654 
1655 	status->flag &= ~RX_FLAG_IV_STRIPPED;
1656 }
1657 EXPORT_SYMBOL_GPL(mt76_insert_ccmp_hdr);
1658 
1659 int mt76_get_rate(struct mt76_dev *dev,
1660 		  struct ieee80211_supported_band *sband,
1661 		  int idx, bool cck)
1662 {
1663 	int i, offset = 0, len = sband->n_bitrates;
1664 
1665 	if (cck) {
1666 		if (sband != &dev->phy.sband_2g.sband)
1667 			return 0;
1668 
1669 		idx &= ~BIT(2); /* short preamble */
1670 	} else if (sband == &dev->phy.sband_2g.sband) {
1671 		offset = 4;
1672 	}
1673 
1674 	for (i = offset; i < len; i++) {
1675 		if ((sband->bitrates[i].hw_value & GENMASK(7, 0)) == idx)
1676 			return i;
1677 	}
1678 
1679 	return 0;
1680 }
1681 EXPORT_SYMBOL_GPL(mt76_get_rate);
1682 
1683 void mt76_sw_scan(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
1684 		  const u8 *mac)
1685 {
1686 	struct mt76_phy *phy = hw->priv;
1687 
1688 	set_bit(MT76_SCANNING, &phy->state);
1689 }
1690 EXPORT_SYMBOL_GPL(mt76_sw_scan);
1691 
1692 void mt76_sw_scan_complete(struct ieee80211_hw *hw, struct ieee80211_vif *vif)
1693 {
1694 	struct mt76_phy *phy = hw->priv;
1695 
1696 	clear_bit(MT76_SCANNING, &phy->state);
1697 }
1698 EXPORT_SYMBOL_GPL(mt76_sw_scan_complete);
1699 
1700 int mt76_get_antenna(struct ieee80211_hw *hw, u32 *tx_ant, u32 *rx_ant)
1701 {
1702 	struct mt76_phy *phy = hw->priv;
1703 	struct mt76_dev *dev = phy->dev;
1704 
1705 	mutex_lock(&dev->mutex);
1706 	*tx_ant = phy->antenna_mask;
1707 	*rx_ant = phy->antenna_mask;
1708 	mutex_unlock(&dev->mutex);
1709 
1710 	return 0;
1711 }
1712 EXPORT_SYMBOL_GPL(mt76_get_antenna);
1713 
1714 struct mt76_queue *
1715 mt76_init_queue(struct mt76_dev *dev, int qid, int idx, int n_desc,
1716 		int ring_base, u32 flags)
1717 {
1718 	struct mt76_queue *hwq;
1719 	int err;
1720 
1721 	hwq = devm_kzalloc(dev->dev, sizeof(*hwq), GFP_KERNEL);
1722 	if (!hwq)
1723 		return ERR_PTR(-ENOMEM);
1724 
1725 	hwq->flags = flags;
1726 
1727 	err = dev->queue_ops->alloc(dev, hwq, idx, n_desc, 0, ring_base);
1728 	if (err < 0)
1729 		return ERR_PTR(err);
1730 
1731 	return hwq;
1732 }
1733 EXPORT_SYMBOL_GPL(mt76_init_queue);
1734 
1735 u16 mt76_calculate_default_rate(struct mt76_phy *phy, int rateidx)
1736 {
1737 	int offset = 0;
1738 
1739 	if (phy->chandef.chan->band != NL80211_BAND_2GHZ)
1740 		offset = 4;
1741 
1742 	/* pick the lowest rate for hidden nodes */
1743 	if (rateidx < 0)
1744 		rateidx = 0;
1745 
1746 	rateidx += offset;
1747 	if (rateidx >= ARRAY_SIZE(mt76_rates))
1748 		rateidx = offset;
1749 
1750 	return mt76_rates[rateidx].hw_value;
1751 }
1752 EXPORT_SYMBOL_GPL(mt76_calculate_default_rate);
1753 
1754 void mt76_ethtool_worker(struct mt76_ethtool_worker_info *wi,
1755 			 struct mt76_sta_stats *stats, bool eht)
1756 {
1757 	int i, ei = wi->initial_stat_idx;
1758 	u64 *data = wi->data;
1759 
1760 	wi->sta_count++;
1761 
1762 	data[ei++] += stats->tx_mode[MT_PHY_TYPE_CCK];
1763 	data[ei++] += stats->tx_mode[MT_PHY_TYPE_OFDM];
1764 	data[ei++] += stats->tx_mode[MT_PHY_TYPE_HT];
1765 	data[ei++] += stats->tx_mode[MT_PHY_TYPE_HT_GF];
1766 	data[ei++] += stats->tx_mode[MT_PHY_TYPE_VHT];
1767 	data[ei++] += stats->tx_mode[MT_PHY_TYPE_HE_SU];
1768 	data[ei++] += stats->tx_mode[MT_PHY_TYPE_HE_EXT_SU];
1769 	data[ei++] += stats->tx_mode[MT_PHY_TYPE_HE_TB];
1770 	data[ei++] += stats->tx_mode[MT_PHY_TYPE_HE_MU];
1771 	if (eht) {
1772 		data[ei++] += stats->tx_mode[MT_PHY_TYPE_EHT_SU];
1773 		data[ei++] += stats->tx_mode[MT_PHY_TYPE_EHT_TRIG];
1774 		data[ei++] += stats->tx_mode[MT_PHY_TYPE_EHT_MU];
1775 	}
1776 
1777 	for (i = 0; i < (ARRAY_SIZE(stats->tx_bw) - !eht); i++)
1778 		data[ei++] += stats->tx_bw[i];
1779 
1780 	for (i = 0; i < (eht ? 14 : 12); i++)
1781 		data[ei++] += stats->tx_mcs[i];
1782 
1783 	for (i = 0; i < 4; i++)
1784 		data[ei++] += stats->tx_nss[i];
1785 
1786 	wi->worker_stat_count = ei - wi->initial_stat_idx;
1787 }
1788 EXPORT_SYMBOL_GPL(mt76_ethtool_worker);
1789 
1790 void mt76_ethtool_page_pool_stats(struct mt76_dev *dev, u64 *data, int *index)
1791 {
1792 #ifdef CONFIG_PAGE_POOL_STATS
1793 	struct page_pool_stats stats = {};
1794 	int i;
1795 
1796 	mt76_for_each_q_rx(dev, i)
1797 		page_pool_get_stats(dev->q_rx[i].page_pool, &stats);
1798 
1799 	page_pool_ethtool_stats_get(data, &stats);
1800 	*index += page_pool_ethtool_stats_get_count();
1801 #endif
1802 }
1803 EXPORT_SYMBOL_GPL(mt76_ethtool_page_pool_stats);
1804 
1805 enum mt76_dfs_state mt76_phy_dfs_state(struct mt76_phy *phy)
1806 {
1807 	struct ieee80211_hw *hw = phy->hw;
1808 	struct mt76_dev *dev = phy->dev;
1809 
1810 	if (dev->region == NL80211_DFS_UNSET ||
1811 	    test_bit(MT76_SCANNING, &phy->state))
1812 		return MT_DFS_STATE_DISABLED;
1813 
1814 	if (!hw->conf.radar_enabled) {
1815 		if ((hw->conf.flags & IEEE80211_CONF_MONITOR) &&
1816 		    (phy->chandef.chan->flags & IEEE80211_CHAN_RADAR))
1817 			return MT_DFS_STATE_ACTIVE;
1818 
1819 		return MT_DFS_STATE_DISABLED;
1820 	}
1821 
1822 	if (!cfg80211_reg_can_beacon(hw->wiphy, &phy->chandef, NL80211_IFTYPE_AP))
1823 		return MT_DFS_STATE_CAC;
1824 
1825 	return MT_DFS_STATE_ACTIVE;
1826 }
1827 EXPORT_SYMBOL_GPL(mt76_phy_dfs_state);
1828