xref: /freebsd/sys/contrib/dev/rtw88/mac.c (revision e5b786625f7f82a1fa91e41823332459ea5550f9)
1 // SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause
2 /* Copyright(c) 2018-2019  Realtek Corporation
3  */
4 
5 #include "main.h"
6 #include "mac.h"
7 #include "reg.h"
8 #include "fw.h"
9 #include "debug.h"
10 #include "sdio.h"
11 
12 void rtw_set_channel_mac(struct rtw_dev *rtwdev, u8 channel, u8 bw,
13 			 u8 primary_ch_idx)
14 {
15 	u8 txsc40 = 0, txsc20 = 0;
16 	u32 value32;
17 	u8 value8;
18 
19 	txsc20 = primary_ch_idx;
20 	if (bw == RTW_CHANNEL_WIDTH_80) {
21 		if (txsc20 == RTW_SC_20_UPPER || txsc20 == RTW_SC_20_UPMOST)
22 			txsc40 = RTW_SC_40_UPPER;
23 		else
24 			txsc40 = RTW_SC_40_LOWER;
25 	}
26 	rtw_write8(rtwdev, REG_DATA_SC,
27 		   BIT_TXSC_20M(txsc20) | BIT_TXSC_40M(txsc40));
28 
29 	value32 = rtw_read32(rtwdev, REG_WMAC_TRXPTCL_CTL);
30 	value32 &= ~BIT_RFMOD;
31 	switch (bw) {
32 	case RTW_CHANNEL_WIDTH_80:
33 		value32 |= BIT_RFMOD_80M;
34 		break;
35 	case RTW_CHANNEL_WIDTH_40:
36 		value32 |= BIT_RFMOD_40M;
37 		break;
38 	case RTW_CHANNEL_WIDTH_20:
39 	default:
40 		break;
41 	}
42 	rtw_write32(rtwdev, REG_WMAC_TRXPTCL_CTL, value32);
43 
44 	if (rtw_chip_wcpu_11n(rtwdev))
45 		return;
46 
47 	value32 = rtw_read32(rtwdev, REG_AFE_CTRL1) & ~(BIT_MAC_CLK_SEL);
48 	value32 |= (MAC_CLK_HW_DEF_80M << BIT_SHIFT_MAC_CLK_SEL);
49 	rtw_write32(rtwdev, REG_AFE_CTRL1, value32);
50 
51 	rtw_write8(rtwdev, REG_USTIME_TSF, MAC_CLK_SPEED);
52 	rtw_write8(rtwdev, REG_USTIME_EDCA, MAC_CLK_SPEED);
53 
54 	value8 = rtw_read8(rtwdev, REG_CCK_CHECK);
55 	value8 = value8 & ~BIT_CHECK_CCK_EN;
56 	if (IS_CH_5G_BAND(channel))
57 		value8 |= BIT_CHECK_CCK_EN;
58 	rtw_write8(rtwdev, REG_CCK_CHECK, value8);
59 }
60 EXPORT_SYMBOL(rtw_set_channel_mac);
61 
62 static int rtw_mac_pre_system_cfg(struct rtw_dev *rtwdev)
63 {
64 	unsigned int retry;
65 	u32 value32;
66 	u8 value8;
67 
68 	rtw_write8(rtwdev, REG_RSV_CTRL, 0);
69 
70 	if (rtw_chip_wcpu_11n(rtwdev)) {
71 		if (rtw_read32(rtwdev, REG_SYS_CFG1) & BIT_LDO)
72 			rtw_write8(rtwdev, REG_LDO_SWR_CTRL, LDO_SEL);
73 		else
74 			rtw_write8(rtwdev, REG_LDO_SWR_CTRL, SPS_SEL);
75 		return 0;
76 	}
77 
78 	switch (rtw_hci_type(rtwdev)) {
79 	case RTW_HCI_TYPE_PCIE:
80 		rtw_write32_set(rtwdev, REG_HCI_OPT_CTRL, BIT_USB_SUS_DIS);
81 		break;
82 	case RTW_HCI_TYPE_SDIO:
83 		rtw_write8_clr(rtwdev, REG_SDIO_HSUS_CTRL, BIT_HCI_SUS_REQ);
84 
85 		for (retry = 0; retry < RTW_PWR_POLLING_CNT; retry++) {
86 			if (rtw_read8(rtwdev, REG_SDIO_HSUS_CTRL) & BIT_HCI_RESUME_RDY)
87 				break;
88 
89 			usleep_range(10, 50);
90 		}
91 
92 		if (retry == RTW_PWR_POLLING_CNT) {
93 			rtw_err(rtwdev, "failed to poll REG_SDIO_HSUS_CTRL[1]");
94 			return -ETIMEDOUT;
95 		}
96 
97 		if (rtw_sdio_is_sdio30_supported(rtwdev))
98 			rtw_write8_set(rtwdev, REG_HCI_OPT_CTRL + 2,
99 				       BIT_SDIO_PAD_E5 >> 16);
100 		else
101 			rtw_write8_clr(rtwdev, REG_HCI_OPT_CTRL + 2,
102 				       BIT_SDIO_PAD_E5 >> 16);
103 		break;
104 	case RTW_HCI_TYPE_USB:
105 		break;
106 	default:
107 		return -EINVAL;
108 	}
109 
110 	/* config PIN Mux */
111 	value32 = rtw_read32(rtwdev, REG_PAD_CTRL1);
112 	value32 |= BIT_PAPE_WLBT_SEL | BIT_LNAON_WLBT_SEL;
113 	rtw_write32(rtwdev, REG_PAD_CTRL1, value32);
114 
115 	value32 = rtw_read32(rtwdev, REG_LED_CFG);
116 	value32 &= ~(BIT_PAPE_SEL_EN | BIT_LNAON_SEL_EN);
117 	rtw_write32(rtwdev, REG_LED_CFG, value32);
118 
119 	value32 = rtw_read32(rtwdev, REG_GPIO_MUXCFG);
120 	value32 |= BIT_WLRFE_4_5_EN;
121 	rtw_write32(rtwdev, REG_GPIO_MUXCFG, value32);
122 
123 	/* disable BB/RF */
124 	value8 = rtw_read8(rtwdev, REG_SYS_FUNC_EN);
125 	value8 &= ~(BIT_FEN_BB_RSTB | BIT_FEN_BB_GLB_RST);
126 	rtw_write8(rtwdev, REG_SYS_FUNC_EN, value8);
127 
128 	value8 = rtw_read8(rtwdev, REG_RF_CTRL);
129 	value8 &= ~(BIT_RF_SDM_RSTB | BIT_RF_RSTB | BIT_RF_EN);
130 	rtw_write8(rtwdev, REG_RF_CTRL, value8);
131 
132 	value32 = rtw_read32(rtwdev, REG_WLRF1);
133 	value32 &= ~BIT_WLRF1_BBRF_EN;
134 	rtw_write32(rtwdev, REG_WLRF1, value32);
135 
136 	return 0;
137 }
138 
139 static bool do_pwr_poll_cmd(struct rtw_dev *rtwdev, u32 addr, u32 mask, u32 target)
140 {
141 	u32 val;
142 
143 	target &= mask;
144 
145 	return read_poll_timeout_atomic(rtw_read8, val, (val & mask) == target,
146 					50, 50 * RTW_PWR_POLLING_CNT, false,
147 					rtwdev, addr) == 0;
148 }
149 
150 static int rtw_pwr_cmd_polling(struct rtw_dev *rtwdev,
151 			       const struct rtw_pwr_seq_cmd *cmd)
152 {
153 	u8 value;
154 	u32 offset;
155 
156 	if (cmd->base == RTW_PWR_ADDR_SDIO)
157 		offset = cmd->offset | SDIO_LOCAL_OFFSET;
158 	else
159 		offset = cmd->offset;
160 
161 	if (do_pwr_poll_cmd(rtwdev, offset, cmd->mask, cmd->value))
162 		return 0;
163 
164 	if (rtw_hci_type(rtwdev) != RTW_HCI_TYPE_PCIE)
165 		goto err;
166 
167 	/* if PCIE, toggle BIT_PFM_WOWL and try again */
168 	value = rtw_read8(rtwdev, REG_SYS_PW_CTRL);
169 	if (rtwdev->chip->id == RTW_CHIP_TYPE_8723D)
170 		rtw_write8(rtwdev, REG_SYS_PW_CTRL, value & ~BIT_PFM_WOWL);
171 	rtw_write8(rtwdev, REG_SYS_PW_CTRL, value | BIT_PFM_WOWL);
172 	rtw_write8(rtwdev, REG_SYS_PW_CTRL, value & ~BIT_PFM_WOWL);
173 	if (rtwdev->chip->id == RTW_CHIP_TYPE_8723D)
174 		rtw_write8(rtwdev, REG_SYS_PW_CTRL, value | BIT_PFM_WOWL);
175 
176 	if (do_pwr_poll_cmd(rtwdev, offset, cmd->mask, cmd->value))
177 		return 0;
178 
179 err:
180 	rtw_err(rtwdev, "failed to poll offset=0x%x mask=0x%x value=0x%x\n",
181 		offset, cmd->mask, cmd->value);
182 	return -EBUSY;
183 }
184 
185 static int rtw_sub_pwr_seq_parser(struct rtw_dev *rtwdev, u8 intf_mask,
186 				  u8 cut_mask,
187 				  const struct rtw_pwr_seq_cmd *cmd)
188 {
189 	const struct rtw_pwr_seq_cmd *cur_cmd;
190 	u32 offset;
191 	u8 value;
192 
193 	for (cur_cmd = cmd; cur_cmd->cmd != RTW_PWR_CMD_END; cur_cmd++) {
194 		if (!(cur_cmd->intf_mask & intf_mask) ||
195 		    !(cur_cmd->cut_mask & cut_mask))
196 			continue;
197 
198 		switch (cur_cmd->cmd) {
199 		case RTW_PWR_CMD_WRITE:
200 			offset = cur_cmd->offset;
201 
202 			if (cur_cmd->base == RTW_PWR_ADDR_SDIO)
203 				offset |= SDIO_LOCAL_OFFSET;
204 
205 			value = rtw_read8(rtwdev, offset);
206 			value &= ~cur_cmd->mask;
207 			value |= (cur_cmd->value & cur_cmd->mask);
208 			rtw_write8(rtwdev, offset, value);
209 			break;
210 		case RTW_PWR_CMD_POLLING:
211 			if (rtw_pwr_cmd_polling(rtwdev, cur_cmd))
212 				return -EBUSY;
213 			break;
214 		case RTW_PWR_CMD_DELAY:
215 			if (cur_cmd->value == RTW_PWR_DELAY_US)
216 				udelay(cur_cmd->offset);
217 			else
218 				mdelay(cur_cmd->offset);
219 			break;
220 		case RTW_PWR_CMD_READ:
221 			break;
222 		default:
223 			return -EINVAL;
224 		}
225 	}
226 
227 	return 0;
228 }
229 
230 static int rtw_pwr_seq_parser(struct rtw_dev *rtwdev,
231 			      const struct rtw_pwr_seq_cmd **cmd_seq)
232 {
233 	u8 cut_mask;
234 	u8 intf_mask;
235 	u8 cut;
236 	u32 idx = 0;
237 	const struct rtw_pwr_seq_cmd *cmd;
238 	int ret;
239 
240 	cut = rtwdev->hal.cut_version;
241 	cut_mask = cut_version_to_mask(cut);
242 	switch (rtw_hci_type(rtwdev)) {
243 	case RTW_HCI_TYPE_PCIE:
244 		intf_mask = RTW_PWR_INTF_PCI_MSK;
245 		break;
246 	case RTW_HCI_TYPE_USB:
247 		intf_mask = RTW_PWR_INTF_USB_MSK;
248 		break;
249 	case RTW_HCI_TYPE_SDIO:
250 		intf_mask = RTW_PWR_INTF_SDIO_MSK;
251 		break;
252 	default:
253 		return -EINVAL;
254 	}
255 
256 	do {
257 		cmd = cmd_seq[idx];
258 		if (!cmd)
259 			break;
260 
261 		ret = rtw_sub_pwr_seq_parser(rtwdev, intf_mask, cut_mask, cmd);
262 		if (ret)
263 			return ret;
264 
265 		idx++;
266 	} while (1);
267 
268 	return 0;
269 }
270 
271 static int rtw_mac_power_switch(struct rtw_dev *rtwdev, bool pwr_on)
272 {
273 	const struct rtw_chip_info *chip = rtwdev->chip;
274 	const struct rtw_pwr_seq_cmd **pwr_seq;
275 	u32 imr = 0;
276 	u8 rpwm;
277 	bool cur_pwr;
278 	int ret;
279 
280 	if (rtw_chip_wcpu_11ac(rtwdev)) {
281 		rpwm = rtw_read8(rtwdev, rtwdev->hci.rpwm_addr);
282 
283 		/* Check FW still exist or not */
284 		if (rtw_read16(rtwdev, REG_MCUFW_CTRL) == 0xC078) {
285 			rpwm = (rpwm ^ BIT_RPWM_TOGGLE) & BIT_RPWM_TOGGLE;
286 			rtw_write8(rtwdev, rtwdev->hci.rpwm_addr, rpwm);
287 		}
288 	}
289 
290 	if (rtw_read8(rtwdev, REG_CR) == 0xea)
291 		cur_pwr = false;
292 	else if (rtw_hci_type(rtwdev) == RTW_HCI_TYPE_USB &&
293 		 (rtw_read8(rtwdev, REG_SYS_STATUS1 + 1) & BIT(0)))
294 		cur_pwr = false;
295 	else
296 		cur_pwr = true;
297 
298 	if (pwr_on == cur_pwr)
299 		return -EALREADY;
300 
301 	if (rtw_hci_type(rtwdev) == RTW_HCI_TYPE_SDIO) {
302 		imr = rtw_read32(rtwdev, REG_SDIO_HIMR);
303 		rtw_write32(rtwdev, REG_SDIO_HIMR, 0);
304 	}
305 
306 	if (!pwr_on)
307 		clear_bit(RTW_FLAG_POWERON, rtwdev->flags);
308 
309 	pwr_seq = pwr_on ? chip->pwr_on_seq : chip->pwr_off_seq;
310 	ret = rtw_pwr_seq_parser(rtwdev, pwr_seq);
311 
312 	if (rtw_hci_type(rtwdev) == RTW_HCI_TYPE_SDIO)
313 		rtw_write32(rtwdev, REG_SDIO_HIMR, imr);
314 
315 	if (!ret && pwr_on)
316 		set_bit(RTW_FLAG_POWERON, rtwdev->flags);
317 
318 	return ret;
319 }
320 
321 static int __rtw_mac_init_system_cfg(struct rtw_dev *rtwdev)
322 {
323 	u8 sys_func_en = rtwdev->chip->sys_func_en;
324 	u8 value8;
325 	u32 value, tmp;
326 
327 	value = rtw_read32(rtwdev, REG_CPU_DMEM_CON);
328 	value |= BIT_WL_PLATFORM_RST | BIT_DDMA_EN;
329 	rtw_write32(rtwdev, REG_CPU_DMEM_CON, value);
330 
331 	rtw_write8_set(rtwdev, REG_SYS_FUNC_EN + 1, sys_func_en);
332 	value8 = (rtw_read8(rtwdev, REG_CR_EXT + 3) & 0xF0) | 0x0C;
333 	rtw_write8(rtwdev, REG_CR_EXT + 3, value8);
334 
335 	/* disable boot-from-flash for driver's DL FW */
336 	tmp = rtw_read32(rtwdev, REG_MCUFW_CTRL);
337 	if (tmp & BIT_BOOT_FSPI_EN) {
338 		rtw_write32(rtwdev, REG_MCUFW_CTRL, tmp & (~BIT_BOOT_FSPI_EN));
339 		value = rtw_read32(rtwdev, REG_GPIO_MUXCFG) & (~BIT_FSPI_EN);
340 		rtw_write32(rtwdev, REG_GPIO_MUXCFG, value);
341 	}
342 
343 	return 0;
344 }
345 
346 static int __rtw_mac_init_system_cfg_legacy(struct rtw_dev *rtwdev)
347 {
348 	rtw_write8(rtwdev, REG_CR, 0xff);
349 	mdelay(2);
350 	rtw_write8(rtwdev, REG_HWSEQ_CTRL, 0x7f);
351 	mdelay(2);
352 
353 	rtw_write8_set(rtwdev, REG_SYS_CLKR, BIT_WAKEPAD_EN);
354 	rtw_write16_clr(rtwdev, REG_GPIO_MUXCFG, BIT_EN_SIC);
355 
356 	rtw_write16(rtwdev, REG_CR, 0x2ff);
357 
358 	return 0;
359 }
360 
361 static int rtw_mac_init_system_cfg(struct rtw_dev *rtwdev)
362 {
363 	if (rtw_chip_wcpu_11n(rtwdev))
364 		return __rtw_mac_init_system_cfg_legacy(rtwdev);
365 
366 	return __rtw_mac_init_system_cfg(rtwdev);
367 }
368 
369 int rtw_mac_power_on(struct rtw_dev *rtwdev)
370 {
371 	int ret = 0;
372 
373 	ret = rtw_mac_pre_system_cfg(rtwdev);
374 	if (ret)
375 		goto err;
376 
377 	ret = rtw_mac_power_switch(rtwdev, true);
378 	if (ret == -EALREADY) {
379 		rtw_mac_power_switch(rtwdev, false);
380 
381 		ret = rtw_mac_pre_system_cfg(rtwdev);
382 		if (ret)
383 			goto err;
384 
385 		ret = rtw_mac_power_switch(rtwdev, true);
386 		if (ret)
387 			goto err;
388 	} else if (ret) {
389 		goto err;
390 	}
391 
392 	ret = rtw_mac_init_system_cfg(rtwdev);
393 	if (ret)
394 		goto err;
395 
396 	return 0;
397 
398 err:
399 	rtw_err(rtwdev, "mac power on failed");
400 	return ret;
401 }
402 
403 void rtw_mac_power_off(struct rtw_dev *rtwdev)
404 {
405 	rtw_mac_power_switch(rtwdev, false);
406 }
407 
408 static bool check_firmware_size(const u8 *data, u32 size)
409 {
410 	const struct rtw_fw_hdr *fw_hdr = (const struct rtw_fw_hdr *)data;
411 	u32 dmem_size;
412 	u32 imem_size;
413 	u32 emem_size;
414 	u32 real_size;
415 
416 	dmem_size = le32_to_cpu(fw_hdr->dmem_size);
417 	imem_size = le32_to_cpu(fw_hdr->imem_size);
418 	emem_size = (fw_hdr->mem_usage & BIT(4)) ?
419 		    le32_to_cpu(fw_hdr->emem_size) : 0;
420 
421 	dmem_size += FW_HDR_CHKSUM_SIZE;
422 	imem_size += FW_HDR_CHKSUM_SIZE;
423 	emem_size += emem_size ? FW_HDR_CHKSUM_SIZE : 0;
424 	real_size = FW_HDR_SIZE + dmem_size + imem_size + emem_size;
425 	if (real_size != size)
426 		return false;
427 
428 	return true;
429 }
430 
431 static void wlan_cpu_enable(struct rtw_dev *rtwdev, bool enable)
432 {
433 	if (enable) {
434 		/* cpu io interface enable */
435 		rtw_write8_set(rtwdev, REG_RSV_CTRL + 1, BIT_WLMCU_IOIF);
436 
437 		/* cpu enable */
438 		rtw_write8_set(rtwdev, REG_SYS_FUNC_EN + 1, BIT_FEN_CPUEN);
439 	} else {
440 		/* cpu io interface disable */
441 		rtw_write8_clr(rtwdev, REG_SYS_FUNC_EN + 1, BIT_FEN_CPUEN);
442 
443 		/* cpu disable */
444 		rtw_write8_clr(rtwdev, REG_RSV_CTRL + 1, BIT_WLMCU_IOIF);
445 	}
446 }
447 
448 #define DLFW_RESTORE_REG_NUM 6
449 
450 static void download_firmware_reg_backup(struct rtw_dev *rtwdev,
451 					 struct rtw_backup_info *bckp)
452 {
453 	u8 tmp;
454 	u8 bckp_idx = 0;
455 
456 	/* set HIQ to hi priority */
457 	bckp[bckp_idx].len = 1;
458 	bckp[bckp_idx].reg = REG_TXDMA_PQ_MAP + 1;
459 	bckp[bckp_idx].val = rtw_read8(rtwdev, REG_TXDMA_PQ_MAP + 1);
460 	bckp_idx++;
461 	tmp = RTW_DMA_MAPPING_HIGH << 6;
462 	rtw_write8(rtwdev, REG_TXDMA_PQ_MAP + 1, tmp);
463 
464 	/* DLFW only use HIQ, map HIQ to hi priority */
465 	bckp[bckp_idx].len = 1;
466 	bckp[bckp_idx].reg = REG_CR;
467 	bckp[bckp_idx].val = rtw_read8(rtwdev, REG_CR);
468 	bckp_idx++;
469 	bckp[bckp_idx].len = 4;
470 	bckp[bckp_idx].reg = REG_H2CQ_CSR;
471 	bckp[bckp_idx].val = BIT_H2CQ_FULL;
472 	bckp_idx++;
473 	tmp = BIT_HCI_TXDMA_EN | BIT_TXDMA_EN;
474 	rtw_write8(rtwdev, REG_CR, tmp);
475 	rtw_write32(rtwdev, REG_H2CQ_CSR, BIT_H2CQ_FULL);
476 
477 	/* Config hi priority queue and public priority queue page number */
478 	bckp[bckp_idx].len = 2;
479 	bckp[bckp_idx].reg = REG_FIFOPAGE_INFO_1;
480 	bckp[bckp_idx].val = rtw_read16(rtwdev, REG_FIFOPAGE_INFO_1);
481 	bckp_idx++;
482 	bckp[bckp_idx].len = 4;
483 	bckp[bckp_idx].reg = REG_RQPN_CTRL_2;
484 	bckp[bckp_idx].val = rtw_read32(rtwdev, REG_RQPN_CTRL_2) | BIT_LD_RQPN;
485 	bckp_idx++;
486 	rtw_write16(rtwdev, REG_FIFOPAGE_INFO_1, 0x200);
487 	rtw_write32(rtwdev, REG_RQPN_CTRL_2, bckp[bckp_idx - 1].val);
488 
489 	if (rtw_hci_type(rtwdev) == RTW_HCI_TYPE_SDIO)
490 		rtw_read32(rtwdev, REG_SDIO_FREE_TXPG);
491 
492 	/* Disable beacon related functions */
493 	tmp = rtw_read8(rtwdev, REG_BCN_CTRL);
494 	bckp[bckp_idx].len = 1;
495 	bckp[bckp_idx].reg = REG_BCN_CTRL;
496 	bckp[bckp_idx].val = tmp;
497 	bckp_idx++;
498 	tmp = (u8)((tmp & (~BIT_EN_BCN_FUNCTION)) | BIT_DIS_TSF_UDT);
499 	rtw_write8(rtwdev, REG_BCN_CTRL, tmp);
500 
501 	WARN(bckp_idx != DLFW_RESTORE_REG_NUM, "wrong backup number\n");
502 }
503 
504 static void download_firmware_reset_platform(struct rtw_dev *rtwdev)
505 {
506 	rtw_write8_clr(rtwdev, REG_CPU_DMEM_CON + 2, BIT_WL_PLATFORM_RST >> 16);
507 	rtw_write8_clr(rtwdev, REG_SYS_CLK_CTRL + 1, BIT_CPU_CLK_EN >> 8);
508 	rtw_write8_set(rtwdev, REG_CPU_DMEM_CON + 2, BIT_WL_PLATFORM_RST >> 16);
509 	rtw_write8_set(rtwdev, REG_SYS_CLK_CTRL + 1, BIT_CPU_CLK_EN >> 8);
510 }
511 
512 static void download_firmware_reg_restore(struct rtw_dev *rtwdev,
513 					  struct rtw_backup_info *bckp,
514 					  u8 bckp_num)
515 {
516 	rtw_restore_reg(rtwdev, bckp, bckp_num);
517 }
518 
519 #define TX_DESC_SIZE 48
520 
521 static int send_firmware_pkt_rsvd_page(struct rtw_dev *rtwdev, u16 pg_addr,
522 				       const u8 *data, u32 size)
523 {
524 	u8 *buf;
525 	int ret;
526 
527 	buf = kmemdup(data, size, GFP_KERNEL);
528 	if (!buf)
529 		return -ENOMEM;
530 
531 	ret = rtw_fw_write_data_rsvd_page(rtwdev, pg_addr, buf, size);
532 	kfree(buf);
533 	return ret;
534 }
535 
536 static int
537 send_firmware_pkt(struct rtw_dev *rtwdev, u16 pg_addr, const u8 *data, u32 size)
538 {
539 	int ret;
540 
541 	if (rtw_hci_type(rtwdev) == RTW_HCI_TYPE_USB &&
542 	    !((size + TX_DESC_SIZE) & (512 - 1)))
543 		size += 1;
544 
545 	ret = send_firmware_pkt_rsvd_page(rtwdev, pg_addr, data, size);
546 	if (ret)
547 		rtw_err(rtwdev, "failed to download rsvd page\n");
548 
549 	return ret;
550 }
551 
552 static int
553 iddma_enable(struct rtw_dev *rtwdev, u32 src, u32 dst, u32 ctrl)
554 {
555 	rtw_write32(rtwdev, REG_DDMA_CH0SA, src);
556 	rtw_write32(rtwdev, REG_DDMA_CH0DA, dst);
557 	rtw_write32(rtwdev, REG_DDMA_CH0CTRL, ctrl);
558 
559 	if (!check_hw_ready(rtwdev, REG_DDMA_CH0CTRL, BIT_DDMACH0_OWN, 0))
560 		return -EBUSY;
561 
562 	return 0;
563 }
564 
565 static int iddma_download_firmware(struct rtw_dev *rtwdev, u32 src, u32 dst,
566 				   u32 len, u8 first)
567 {
568 	u32 ch0_ctrl = BIT_DDMACH0_CHKSUM_EN | BIT_DDMACH0_OWN;
569 
570 	if (!check_hw_ready(rtwdev, REG_DDMA_CH0CTRL, BIT_DDMACH0_OWN, 0))
571 		return -EBUSY;
572 
573 	ch0_ctrl |= len & BIT_MASK_DDMACH0_DLEN;
574 	if (!first)
575 		ch0_ctrl |= BIT_DDMACH0_CHKSUM_CONT;
576 
577 	if (iddma_enable(rtwdev, src, dst, ch0_ctrl))
578 		return -EBUSY;
579 
580 	return 0;
581 }
582 
583 int rtw_ddma_to_fw_fifo(struct rtw_dev *rtwdev, u32 ocp_src, u32 size)
584 {
585 	u32 ch0_ctrl = BIT_DDMACH0_OWN | BIT_DDMACH0_DDMA_MODE;
586 
587 	if (!check_hw_ready(rtwdev, REG_DDMA_CH0CTRL, BIT_DDMACH0_OWN, 0)) {
588 		rtw_dbg(rtwdev, RTW_DBG_FW, "busy to start ddma\n");
589 		return -EBUSY;
590 	}
591 
592 	ch0_ctrl |= size & BIT_MASK_DDMACH0_DLEN;
593 
594 	if (iddma_enable(rtwdev, ocp_src, OCPBASE_RXBUF_FW_88XX, ch0_ctrl)) {
595 		rtw_dbg(rtwdev, RTW_DBG_FW, "busy to complete ddma\n");
596 		return -EBUSY;
597 	}
598 
599 	return 0;
600 }
601 
602 static bool
603 check_fw_checksum(struct rtw_dev *rtwdev, u32 addr)
604 {
605 	u8 fw_ctrl;
606 
607 	fw_ctrl = rtw_read8(rtwdev, REG_MCUFW_CTRL);
608 
609 	if (rtw_read32(rtwdev, REG_DDMA_CH0CTRL) & BIT_DDMACH0_CHKSUM_STS) {
610 		if (addr < OCPBASE_DMEM_88XX) {
611 			fw_ctrl |= BIT_IMEM_DW_OK;
612 			fw_ctrl &= ~BIT_IMEM_CHKSUM_OK;
613 			rtw_write8(rtwdev, REG_MCUFW_CTRL, fw_ctrl);
614 		} else {
615 			fw_ctrl |= BIT_DMEM_DW_OK;
616 			fw_ctrl &= ~BIT_DMEM_CHKSUM_OK;
617 			rtw_write8(rtwdev, REG_MCUFW_CTRL, fw_ctrl);
618 		}
619 
620 		rtw_err(rtwdev, "invalid fw checksum\n");
621 
622 		return false;
623 	}
624 
625 	if (addr < OCPBASE_DMEM_88XX) {
626 		fw_ctrl |= (BIT_IMEM_DW_OK | BIT_IMEM_CHKSUM_OK);
627 		rtw_write8(rtwdev, REG_MCUFW_CTRL, fw_ctrl);
628 	} else {
629 		fw_ctrl |= (BIT_DMEM_DW_OK | BIT_DMEM_CHKSUM_OK);
630 		rtw_write8(rtwdev, REG_MCUFW_CTRL, fw_ctrl);
631 	}
632 
633 	return true;
634 }
635 
636 static int
637 download_firmware_to_mem(struct rtw_dev *rtwdev, const u8 *data,
638 			 u32 src, u32 dst, u32 size)
639 {
640 	const struct rtw_chip_info *chip = rtwdev->chip;
641 	u32 desc_size = chip->tx_pkt_desc_sz;
642 	u8 first_part;
643 	u32 mem_offset;
644 	u32 residue_size;
645 	u32 pkt_size;
646 	u32 max_size = 0x1000;
647 	u32 val;
648 	int ret;
649 
650 	mem_offset = 0;
651 	first_part = 1;
652 	residue_size = size;
653 
654 	val = rtw_read32(rtwdev, REG_DDMA_CH0CTRL);
655 	val |= BIT_DDMACH0_RESET_CHKSUM_STS;
656 	rtw_write32(rtwdev, REG_DDMA_CH0CTRL, val);
657 
658 	while (residue_size) {
659 		if (residue_size >= max_size)
660 			pkt_size = max_size;
661 		else
662 			pkt_size = residue_size;
663 
664 		ret = send_firmware_pkt(rtwdev, (u16)(src >> 7),
665 					data + mem_offset, pkt_size);
666 		if (ret)
667 			return ret;
668 
669 		ret = iddma_download_firmware(rtwdev, OCPBASE_TXBUF_88XX +
670 					      src + desc_size,
671 					      dst + mem_offset, pkt_size,
672 					      first_part);
673 		if (ret)
674 			return ret;
675 
676 		first_part = 0;
677 		mem_offset += pkt_size;
678 		residue_size -= pkt_size;
679 	}
680 
681 	if (!check_fw_checksum(rtwdev, dst))
682 		return -EINVAL;
683 
684 	return 0;
685 }
686 
687 static int
688 start_download_firmware(struct rtw_dev *rtwdev, const u8 *data, u32 size)
689 {
690 	const struct rtw_fw_hdr *fw_hdr = (const struct rtw_fw_hdr *)data;
691 	const u8 *cur_fw;
692 	u16 val;
693 	u32 imem_size;
694 	u32 dmem_size;
695 	u32 emem_size;
696 	u32 addr;
697 	int ret;
698 
699 	dmem_size = le32_to_cpu(fw_hdr->dmem_size);
700 	imem_size = le32_to_cpu(fw_hdr->imem_size);
701 	emem_size = (fw_hdr->mem_usage & BIT(4)) ?
702 		    le32_to_cpu(fw_hdr->emem_size) : 0;
703 	dmem_size += FW_HDR_CHKSUM_SIZE;
704 	imem_size += FW_HDR_CHKSUM_SIZE;
705 	emem_size += emem_size ? FW_HDR_CHKSUM_SIZE : 0;
706 
707 	val = (u16)(rtw_read16(rtwdev, REG_MCUFW_CTRL) & 0x3800);
708 	val |= BIT_MCUFWDL_EN;
709 	rtw_write16(rtwdev, REG_MCUFW_CTRL, val);
710 
711 	cur_fw = data + FW_HDR_SIZE;
712 	addr = le32_to_cpu(fw_hdr->dmem_addr);
713 	addr &= ~BIT(31);
714 	ret = download_firmware_to_mem(rtwdev, cur_fw, 0, addr, dmem_size);
715 	if (ret)
716 		return ret;
717 
718 	cur_fw = data + FW_HDR_SIZE + dmem_size;
719 	addr = le32_to_cpu(fw_hdr->imem_addr);
720 	addr &= ~BIT(31);
721 	ret = download_firmware_to_mem(rtwdev, cur_fw, 0, addr, imem_size);
722 	if (ret)
723 		return ret;
724 
725 	if (emem_size) {
726 		cur_fw = data + FW_HDR_SIZE + dmem_size + imem_size;
727 		addr = le32_to_cpu(fw_hdr->emem_addr);
728 		addr &= ~BIT(31);
729 		ret = download_firmware_to_mem(rtwdev, cur_fw, 0, addr,
730 					       emem_size);
731 		if (ret)
732 			return ret;
733 	}
734 
735 	return 0;
736 }
737 
738 static int download_firmware_validate(struct rtw_dev *rtwdev)
739 {
740 	u32 fw_key;
741 
742 	if (!check_hw_ready(rtwdev, REG_MCUFW_CTRL, FW_READY_MASK, FW_READY)) {
743 		fw_key = rtw_read32(rtwdev, REG_FW_DBG7) & FW_KEY_MASK;
744 		if (fw_key == ILLEGAL_KEY_GROUP)
745 			rtw_err(rtwdev, "invalid fw key\n");
746 		return -EINVAL;
747 	}
748 
749 	return 0;
750 }
751 
752 static void download_firmware_end_flow(struct rtw_dev *rtwdev)
753 {
754 	u16 fw_ctrl;
755 
756 	rtw_write32(rtwdev, REG_TXDMA_STATUS, BTI_PAGE_OVF);
757 
758 	/* Check IMEM & DMEM checksum is OK or not */
759 	fw_ctrl = rtw_read16(rtwdev, REG_MCUFW_CTRL);
760 	if ((fw_ctrl & BIT_CHECK_SUM_OK) != BIT_CHECK_SUM_OK)
761 		return;
762 
763 	fw_ctrl = (fw_ctrl | BIT_FW_DW_RDY) & ~BIT_MCUFWDL_EN;
764 	rtw_write16(rtwdev, REG_MCUFW_CTRL, fw_ctrl);
765 }
766 
767 static int __rtw_download_firmware(struct rtw_dev *rtwdev,
768 				   struct rtw_fw_state *fw)
769 {
770 	struct rtw_backup_info bckp[DLFW_RESTORE_REG_NUM];
771 	const u8 *data = fw->firmware->data;
772 	u32 size = fw->firmware->size;
773 	u32 ltecoex_bckp;
774 	int ret;
775 
776 	if (!check_firmware_size(data, size))
777 		return -EINVAL;
778 
779 	if (!ltecoex_read_reg(rtwdev, 0x38, &ltecoex_bckp))
780 		return -EBUSY;
781 
782 	wlan_cpu_enable(rtwdev, false);
783 
784 	download_firmware_reg_backup(rtwdev, bckp);
785 	download_firmware_reset_platform(rtwdev);
786 
787 	ret = start_download_firmware(rtwdev, data, size);
788 	if (ret)
789 		goto dlfw_fail;
790 
791 	download_firmware_reg_restore(rtwdev, bckp, DLFW_RESTORE_REG_NUM);
792 
793 	download_firmware_end_flow(rtwdev);
794 
795 	wlan_cpu_enable(rtwdev, true);
796 
797 	if (!ltecoex_reg_write(rtwdev, 0x38, ltecoex_bckp)) {
798 		ret = -EBUSY;
799 		goto dlfw_fail;
800 	}
801 
802 	ret = download_firmware_validate(rtwdev);
803 	if (ret)
804 		goto dlfw_fail;
805 
806 	/* reset desc and index */
807 	rtw_hci_setup(rtwdev);
808 
809 	rtwdev->h2c.last_box_num = 0;
810 	rtwdev->h2c.seq = 0;
811 
812 	set_bit(RTW_FLAG_FW_RUNNING, rtwdev->flags);
813 
814 	return 0;
815 
816 dlfw_fail:
817 	/* Disable FWDL_EN */
818 	rtw_write8_clr(rtwdev, REG_MCUFW_CTRL, BIT_MCUFWDL_EN);
819 	rtw_write8_set(rtwdev, REG_SYS_FUNC_EN + 1, BIT_FEN_CPUEN);
820 
821 	return ret;
822 }
823 
824 static void en_download_firmware_legacy(struct rtw_dev *rtwdev, bool en)
825 {
826 	int try;
827 
828 	if (en) {
829 		wlan_cpu_enable(rtwdev, false);
830 		wlan_cpu_enable(rtwdev, true);
831 
832 		rtw_write8_set(rtwdev, REG_MCUFW_CTRL, BIT_MCUFWDL_EN);
833 
834 		for (try = 0; try < 10; try++) {
835 			if (rtw_read8(rtwdev, REG_MCUFW_CTRL) & BIT_MCUFWDL_EN)
836 				goto fwdl_ready;
837 			rtw_write8_set(rtwdev, REG_MCUFW_CTRL, BIT_MCUFWDL_EN);
838 			msleep(20);
839 		}
840 		rtw_err(rtwdev, "failed to check fw download ready\n");
841 fwdl_ready:
842 		rtw_write32_clr(rtwdev, REG_MCUFW_CTRL, BIT_ROM_DLEN);
843 	} else {
844 		rtw_write8_clr(rtwdev, REG_MCUFW_CTRL, BIT_MCUFWDL_EN);
845 	}
846 }
847 
848 static void
849 write_firmware_page(struct rtw_dev *rtwdev, u32 page, const u8 *data, u32 size)
850 {
851 	u32 val32;
852 	u32 block_nr;
853 	u32 remain_size;
854 	u32 write_addr = FW_START_ADDR_LEGACY;
855 	const __le32 *ptr = (const __le32 *)data;
856 	u32 block;
857 	__le32 remain_data = 0;
858 
859 	block_nr = size >> DLFW_BLK_SIZE_SHIFT_LEGACY;
860 	remain_size = size & (DLFW_BLK_SIZE_LEGACY - 1);
861 
862 	val32 = rtw_read32(rtwdev, REG_MCUFW_CTRL);
863 	val32 &= ~BIT_ROM_PGE;
864 	val32 |= (page << BIT_SHIFT_ROM_PGE) & BIT_ROM_PGE;
865 	rtw_write32(rtwdev, REG_MCUFW_CTRL, val32);
866 
867 	for (block = 0; block < block_nr; block++) {
868 		rtw_write32(rtwdev, write_addr, le32_to_cpu(*ptr));
869 
870 		write_addr += DLFW_BLK_SIZE_LEGACY;
871 		ptr++;
872 	}
873 
874 	if (remain_size) {
875 		memcpy(&remain_data, ptr, remain_size);
876 		rtw_write32(rtwdev, write_addr, le32_to_cpu(remain_data));
877 	}
878 }
879 
880 static int
881 download_firmware_legacy(struct rtw_dev *rtwdev, const u8 *data, u32 size)
882 {
883 	u32 page;
884 	u32 total_page;
885 	u32 last_page_size;
886 
887 	data += sizeof(struct rtw_fw_hdr_legacy);
888 	size -= sizeof(struct rtw_fw_hdr_legacy);
889 
890 	total_page = size >> DLFW_PAGE_SIZE_SHIFT_LEGACY;
891 	last_page_size = size & (DLFW_PAGE_SIZE_LEGACY - 1);
892 
893 	rtw_write8_set(rtwdev, REG_MCUFW_CTRL, BIT_FWDL_CHK_RPT);
894 
895 	for (page = 0; page < total_page; page++) {
896 		write_firmware_page(rtwdev, page, data, DLFW_PAGE_SIZE_LEGACY);
897 		data += DLFW_PAGE_SIZE_LEGACY;
898 	}
899 	if (last_page_size)
900 		write_firmware_page(rtwdev, page, data, last_page_size);
901 
902 	if (!check_hw_ready(rtwdev, REG_MCUFW_CTRL, BIT_FWDL_CHK_RPT, 1)) {
903 		rtw_err(rtwdev, "failed to check download firmware report\n");
904 		return -EINVAL;
905 	}
906 
907 	return 0;
908 }
909 
910 static int download_firmware_validate_legacy(struct rtw_dev *rtwdev)
911 {
912 	u32 val32;
913 	int try;
914 
915 	val32 = rtw_read32(rtwdev, REG_MCUFW_CTRL);
916 	val32 |= BIT_MCUFWDL_RDY;
917 	val32 &= ~BIT_WINTINI_RDY;
918 	rtw_write32(rtwdev, REG_MCUFW_CTRL, val32);
919 
920 	wlan_cpu_enable(rtwdev, false);
921 	wlan_cpu_enable(rtwdev, true);
922 
923 	for (try = 0; try < 10; try++) {
924 		val32 = rtw_read32(rtwdev, REG_MCUFW_CTRL);
925 		if ((val32 & FW_READY_LEGACY) == FW_READY_LEGACY)
926 			return 0;
927 		msleep(20);
928 	}
929 
930 	rtw_err(rtwdev, "failed to validate firmware\n");
931 	return -EINVAL;
932 }
933 
934 static int __rtw_download_firmware_legacy(struct rtw_dev *rtwdev,
935 					  struct rtw_fw_state *fw)
936 {
937 	int ret = 0;
938 
939 	en_download_firmware_legacy(rtwdev, true);
940 	ret = download_firmware_legacy(rtwdev, fw->firmware->data, fw->firmware->size);
941 	en_download_firmware_legacy(rtwdev, false);
942 	if (ret)
943 		goto out;
944 
945 	ret = download_firmware_validate_legacy(rtwdev);
946 	if (ret)
947 		goto out;
948 
949 	/* reset desc and index */
950 	rtw_hci_setup(rtwdev);
951 
952 	rtwdev->h2c.last_box_num = 0;
953 	rtwdev->h2c.seq = 0;
954 
955 	set_bit(RTW_FLAG_FW_RUNNING, rtwdev->flags);
956 
957 out:
958 	return ret;
959 }
960 
961 static
962 int _rtw_download_firmware(struct rtw_dev *rtwdev, struct rtw_fw_state *fw)
963 {
964 	if (rtw_chip_wcpu_11n(rtwdev))
965 		return __rtw_download_firmware_legacy(rtwdev, fw);
966 
967 	return __rtw_download_firmware(rtwdev, fw);
968 }
969 
970 int rtw_download_firmware(struct rtw_dev *rtwdev, struct rtw_fw_state *fw)
971 {
972 	int ret;
973 
974 	ret = _rtw_download_firmware(rtwdev, fw);
975 	if (ret)
976 		return ret;
977 
978 	if (rtw_hci_type(rtwdev) == RTW_HCI_TYPE_PCIE &&
979 	    rtwdev->chip->id == RTW_CHIP_TYPE_8821C)
980 		rtw_fw_set_recover_bt_device(rtwdev);
981 
982 	return 0;
983 }
984 
985 static u32 get_priority_queues(struct rtw_dev *rtwdev, u32 queues)
986 {
987 	const struct rtw_rqpn *rqpn = rtwdev->fifo.rqpn;
988 	u32 prio_queues = 0;
989 
990 	if (queues & BIT(IEEE80211_AC_VO))
991 		prio_queues |= BIT(rqpn->dma_map_vo);
992 	if (queues & BIT(IEEE80211_AC_VI))
993 		prio_queues |= BIT(rqpn->dma_map_vi);
994 	if (queues & BIT(IEEE80211_AC_BE))
995 		prio_queues |= BIT(rqpn->dma_map_be);
996 	if (queues & BIT(IEEE80211_AC_BK))
997 		prio_queues |= BIT(rqpn->dma_map_bk);
998 
999 	return prio_queues;
1000 }
1001 
1002 static void __rtw_mac_flush_prio_queue(struct rtw_dev *rtwdev,
1003 				       u32 prio_queue, bool drop)
1004 {
1005 	const struct rtw_chip_info *chip = rtwdev->chip;
1006 	const struct rtw_prioq_addr *addr;
1007 	bool wsize;
1008 	u16 avail_page, rsvd_page;
1009 	int i;
1010 
1011 	if (prio_queue >= RTW_DMA_MAPPING_MAX)
1012 		return;
1013 
1014 	addr = &chip->prioq_addrs->prio[prio_queue];
1015 	wsize = chip->prioq_addrs->wsize;
1016 
1017 	/* check if all of the reserved pages are available for 100 msecs */
1018 	for (i = 0; i < 5; i++) {
1019 		rsvd_page = wsize ? rtw_read16(rtwdev, addr->rsvd) :
1020 				     rtw_read8(rtwdev, addr->rsvd);
1021 		avail_page = wsize ? rtw_read16(rtwdev, addr->avail) :
1022 				      rtw_read8(rtwdev, addr->avail);
1023 		if (rsvd_page == avail_page)
1024 			return;
1025 
1026 		msleep(20);
1027 	}
1028 
1029 	/* priority queue is still not empty, throw a warning,
1030 	 *
1031 	 * Note that if we want to flush the tx queue when having a lot of
1032 	 * traffic (ex, 100Mbps up), some of the packets could be dropped.
1033 	 * And it requires like ~2secs to flush the full priority queue.
1034 	 */
1035 	if (!drop)
1036 		rtw_warn(rtwdev, "timed out to flush queue %d\n", prio_queue);
1037 }
1038 
1039 static void rtw_mac_flush_prio_queues(struct rtw_dev *rtwdev,
1040 				      u32 prio_queues, bool drop)
1041 {
1042 	u32 q;
1043 
1044 	for (q = 0; q < RTW_DMA_MAPPING_MAX; q++)
1045 		if (prio_queues & BIT(q))
1046 			__rtw_mac_flush_prio_queue(rtwdev, q, drop);
1047 }
1048 
1049 void rtw_mac_flush_queues(struct rtw_dev *rtwdev, u32 queues, bool drop)
1050 {
1051 	u32 prio_queues = 0;
1052 
1053 	/* If all of the hardware queues are requested to flush,
1054 	 * or the priority queues are not mapped yet,
1055 	 * flush all of the priority queues
1056 	 */
1057 	if (queues == BIT(rtwdev->hw->queues) - 1 || !rtwdev->fifo.rqpn)
1058 		prio_queues = BIT(RTW_DMA_MAPPING_MAX) - 1;
1059 	else
1060 		prio_queues = get_priority_queues(rtwdev, queues);
1061 
1062 	rtw_mac_flush_prio_queues(rtwdev, prio_queues, drop);
1063 }
1064 
1065 static int txdma_queue_mapping(struct rtw_dev *rtwdev)
1066 {
1067 	const struct rtw_chip_info *chip = rtwdev->chip;
1068 	const struct rtw_rqpn *rqpn = NULL;
1069 	u16 txdma_pq_map = 0;
1070 
1071 	switch (rtw_hci_type(rtwdev)) {
1072 	case RTW_HCI_TYPE_PCIE:
1073 		rqpn = &chip->rqpn_table[1];
1074 		break;
1075 	case RTW_HCI_TYPE_USB:
1076 		if (rtwdev->hci.bulkout_num == 2)
1077 			rqpn = &chip->rqpn_table[2];
1078 		else if (rtwdev->hci.bulkout_num == 3)
1079 			rqpn = &chip->rqpn_table[3];
1080 		else if (rtwdev->hci.bulkout_num == 4)
1081 			rqpn = &chip->rqpn_table[4];
1082 		else
1083 			return -EINVAL;
1084 		break;
1085 	case RTW_HCI_TYPE_SDIO:
1086 		rqpn = &chip->rqpn_table[0];
1087 		break;
1088 	default:
1089 		return -EINVAL;
1090 	}
1091 
1092 	rtwdev->fifo.rqpn = rqpn;
1093 	txdma_pq_map |= BIT_TXDMA_HIQ_MAP(rqpn->dma_map_hi);
1094 	txdma_pq_map |= BIT_TXDMA_MGQ_MAP(rqpn->dma_map_mg);
1095 	txdma_pq_map |= BIT_TXDMA_BKQ_MAP(rqpn->dma_map_bk);
1096 	txdma_pq_map |= BIT_TXDMA_BEQ_MAP(rqpn->dma_map_be);
1097 	txdma_pq_map |= BIT_TXDMA_VIQ_MAP(rqpn->dma_map_vi);
1098 	txdma_pq_map |= BIT_TXDMA_VOQ_MAP(rqpn->dma_map_vo);
1099 	rtw_write16(rtwdev, REG_TXDMA_PQ_MAP, txdma_pq_map);
1100 
1101 	rtw_write8(rtwdev, REG_CR, 0);
1102 	rtw_write8(rtwdev, REG_CR, MAC_TRX_ENABLE);
1103 	if (rtw_chip_wcpu_11ac(rtwdev))
1104 		rtw_write32(rtwdev, REG_H2CQ_CSR, BIT_H2CQ_FULL);
1105 
1106 	if (rtw_hci_type(rtwdev) == RTW_HCI_TYPE_SDIO) {
1107 		rtw_read32(rtwdev, REG_SDIO_FREE_TXPG);
1108 		rtw_write32(rtwdev, REG_SDIO_TX_CTRL, 0);
1109 	} else if (rtw_hci_type(rtwdev) == RTW_HCI_TYPE_USB) {
1110 		rtw_write8_set(rtwdev, REG_TXDMA_PQ_MAP, BIT_RXDMA_ARBBW_EN);
1111 	}
1112 
1113 	return 0;
1114 }
1115 
1116 static int set_trx_fifo_info(struct rtw_dev *rtwdev)
1117 {
1118 	const struct rtw_chip_info *chip = rtwdev->chip;
1119 	struct rtw_fifo_conf *fifo = &rtwdev->fifo;
1120 	u16 cur_pg_addr;
1121 	u8 csi_buf_pg_num = chip->csi_buf_pg_num;
1122 
1123 	/* config rsvd page num */
1124 	fifo->rsvd_drv_pg_num = chip->rsvd_drv_pg_num;
1125 	fifo->txff_pg_num = chip->txff_size >> 7;
1126 	if (rtw_chip_wcpu_11n(rtwdev))
1127 		fifo->rsvd_pg_num = fifo->rsvd_drv_pg_num;
1128 	else
1129 		fifo->rsvd_pg_num = fifo->rsvd_drv_pg_num +
1130 				   RSVD_PG_H2C_EXTRAINFO_NUM +
1131 				   RSVD_PG_H2C_STATICINFO_NUM +
1132 				   RSVD_PG_H2CQ_NUM +
1133 				   RSVD_PG_CPU_INSTRUCTION_NUM +
1134 				   RSVD_PG_FW_TXBUF_NUM +
1135 				   csi_buf_pg_num;
1136 
1137 	if (fifo->rsvd_pg_num > fifo->txff_pg_num)
1138 		return -ENOMEM;
1139 
1140 	fifo->acq_pg_num = fifo->txff_pg_num - fifo->rsvd_pg_num;
1141 	fifo->rsvd_boundary = fifo->txff_pg_num - fifo->rsvd_pg_num;
1142 
1143 	cur_pg_addr = fifo->txff_pg_num;
1144 	if (rtw_chip_wcpu_11ac(rtwdev)) {
1145 		cur_pg_addr -= csi_buf_pg_num;
1146 		fifo->rsvd_csibuf_addr = cur_pg_addr;
1147 		cur_pg_addr -= RSVD_PG_FW_TXBUF_NUM;
1148 		fifo->rsvd_fw_txbuf_addr = cur_pg_addr;
1149 		cur_pg_addr -= RSVD_PG_CPU_INSTRUCTION_NUM;
1150 		fifo->rsvd_cpu_instr_addr = cur_pg_addr;
1151 		cur_pg_addr -= RSVD_PG_H2CQ_NUM;
1152 		fifo->rsvd_h2cq_addr = cur_pg_addr;
1153 		cur_pg_addr -= RSVD_PG_H2C_STATICINFO_NUM;
1154 		fifo->rsvd_h2c_sta_info_addr = cur_pg_addr;
1155 		cur_pg_addr -= RSVD_PG_H2C_EXTRAINFO_NUM;
1156 		fifo->rsvd_h2c_info_addr = cur_pg_addr;
1157 	}
1158 	cur_pg_addr -= fifo->rsvd_drv_pg_num;
1159 	fifo->rsvd_drv_addr = cur_pg_addr;
1160 
1161 	if (fifo->rsvd_boundary != fifo->rsvd_drv_addr) {
1162 		rtw_err(rtwdev, "wrong rsvd driver address\n");
1163 		return -EINVAL;
1164 	}
1165 
1166 	return 0;
1167 }
1168 
1169 static int __priority_queue_cfg(struct rtw_dev *rtwdev,
1170 				const struct rtw_page_table *pg_tbl,
1171 				u16 pubq_num)
1172 {
1173 	const struct rtw_chip_info *chip = rtwdev->chip;
1174 	struct rtw_fifo_conf *fifo = &rtwdev->fifo;
1175 
1176 	rtw_write16(rtwdev, REG_FIFOPAGE_INFO_1, pg_tbl->hq_num);
1177 	rtw_write16(rtwdev, REG_FIFOPAGE_INFO_2, pg_tbl->lq_num);
1178 	rtw_write16(rtwdev, REG_FIFOPAGE_INFO_3, pg_tbl->nq_num);
1179 	rtw_write16(rtwdev, REG_FIFOPAGE_INFO_4, pg_tbl->exq_num);
1180 	rtw_write16(rtwdev, REG_FIFOPAGE_INFO_5, pubq_num);
1181 	rtw_write32_set(rtwdev, REG_RQPN_CTRL_2, BIT_LD_RQPN);
1182 
1183 	rtw_write16(rtwdev, REG_FIFOPAGE_CTRL_2, fifo->rsvd_boundary);
1184 	rtw_write8_set(rtwdev, REG_FWHW_TXQ_CTRL + 2, BIT_EN_WR_FREE_TAIL >> 16);
1185 
1186 	rtw_write16(rtwdev, REG_BCNQ_BDNY_V1, fifo->rsvd_boundary);
1187 	rtw_write16(rtwdev, REG_FIFOPAGE_CTRL_2 + 2, fifo->rsvd_boundary);
1188 	rtw_write16(rtwdev, REG_BCNQ1_BDNY_V1, fifo->rsvd_boundary);
1189 	rtw_write32(rtwdev, REG_RXFF_BNDY, chip->rxff_size - C2H_PKT_BUF - 1);
1190 	rtw_write8_set(rtwdev, REG_AUTO_LLT_V1, BIT_AUTO_INIT_LLT_V1);
1191 
1192 	if (!check_hw_ready(rtwdev, REG_AUTO_LLT_V1, BIT_AUTO_INIT_LLT_V1, 0))
1193 		return -EBUSY;
1194 
1195 	rtw_write8(rtwdev, REG_CR + 3, 0);
1196 
1197 	return 0;
1198 }
1199 
1200 static int __priority_queue_cfg_legacy(struct rtw_dev *rtwdev,
1201 				       const struct rtw_page_table *pg_tbl,
1202 				       u16 pubq_num)
1203 {
1204 	const struct rtw_chip_info *chip = rtwdev->chip;
1205 	struct rtw_fifo_conf *fifo = &rtwdev->fifo;
1206 	u32 val32;
1207 
1208 	val32 = BIT_RQPN_NE(pg_tbl->nq_num, pg_tbl->exq_num);
1209 	rtw_write32(rtwdev, REG_RQPN_NPQ, val32);
1210 	val32 = BIT_RQPN_HLP(pg_tbl->hq_num, pg_tbl->lq_num, pubq_num);
1211 	rtw_write32(rtwdev, REG_RQPN, val32);
1212 
1213 	rtw_write8(rtwdev, REG_TRXFF_BNDY, fifo->rsvd_boundary);
1214 	rtw_write16(rtwdev, REG_TRXFF_BNDY + 2, chip->rxff_size - REPORT_BUF - 1);
1215 	rtw_write8(rtwdev, REG_DWBCN0_CTRL + 1, fifo->rsvd_boundary);
1216 	rtw_write8(rtwdev, REG_BCNQ_BDNY, fifo->rsvd_boundary);
1217 	rtw_write8(rtwdev, REG_MGQ_BDNY, fifo->rsvd_boundary);
1218 	rtw_write8(rtwdev, REG_WMAC_LBK_BF_HD, fifo->rsvd_boundary);
1219 
1220 	rtw_write32_set(rtwdev, REG_AUTO_LLT, BIT_AUTO_INIT_LLT);
1221 
1222 	if (!check_hw_ready(rtwdev, REG_AUTO_LLT, BIT_AUTO_INIT_LLT, 0))
1223 		return -EBUSY;
1224 
1225 	return 0;
1226 }
1227 
1228 static int priority_queue_cfg(struct rtw_dev *rtwdev)
1229 {
1230 	const struct rtw_chip_info *chip = rtwdev->chip;
1231 	struct rtw_fifo_conf *fifo = &rtwdev->fifo;
1232 	const struct rtw_page_table *pg_tbl = NULL;
1233 	u16 pubq_num;
1234 	int ret;
1235 
1236 	ret = set_trx_fifo_info(rtwdev);
1237 	if (ret)
1238 		return ret;
1239 
1240 	switch (rtw_hci_type(rtwdev)) {
1241 	case RTW_HCI_TYPE_PCIE:
1242 		pg_tbl = &chip->page_table[1];
1243 		break;
1244 	case RTW_HCI_TYPE_USB:
1245 		if (rtwdev->hci.bulkout_num == 2)
1246 			pg_tbl = &chip->page_table[2];
1247 		else if (rtwdev->hci.bulkout_num == 3)
1248 			pg_tbl = &chip->page_table[3];
1249 		else if (rtwdev->hci.bulkout_num == 4)
1250 			pg_tbl = &chip->page_table[4];
1251 		else
1252 			return -EINVAL;
1253 		break;
1254 	case RTW_HCI_TYPE_SDIO:
1255 		pg_tbl = &chip->page_table[0];
1256 		break;
1257 	default:
1258 		return -EINVAL;
1259 	}
1260 
1261 	pubq_num = fifo->acq_pg_num - pg_tbl->hq_num - pg_tbl->lq_num -
1262 		   pg_tbl->nq_num - pg_tbl->exq_num - pg_tbl->gapq_num;
1263 	if (rtw_chip_wcpu_11n(rtwdev))
1264 		return __priority_queue_cfg_legacy(rtwdev, pg_tbl, pubq_num);
1265 	else
1266 		return __priority_queue_cfg(rtwdev, pg_tbl, pubq_num);
1267 }
1268 
1269 static int init_h2c(struct rtw_dev *rtwdev)
1270 {
1271 	struct rtw_fifo_conf *fifo = &rtwdev->fifo;
1272 	u8 value8;
1273 	u32 value32;
1274 	u32 h2cq_addr;
1275 	u32 h2cq_size;
1276 	u32 h2cq_free;
1277 	u32 wp, rp;
1278 
1279 	if (rtw_chip_wcpu_11n(rtwdev))
1280 		return 0;
1281 
1282 	h2cq_addr = fifo->rsvd_h2cq_addr << TX_PAGE_SIZE_SHIFT;
1283 	h2cq_size = RSVD_PG_H2CQ_NUM << TX_PAGE_SIZE_SHIFT;
1284 
1285 	value32 = rtw_read32(rtwdev, REG_H2C_HEAD);
1286 	value32 = (value32 & 0xFFFC0000) | h2cq_addr;
1287 	rtw_write32(rtwdev, REG_H2C_HEAD, value32);
1288 
1289 	value32 = rtw_read32(rtwdev, REG_H2C_READ_ADDR);
1290 	value32 = (value32 & 0xFFFC0000) | h2cq_addr;
1291 	rtw_write32(rtwdev, REG_H2C_READ_ADDR, value32);
1292 
1293 	value32 = rtw_read32(rtwdev, REG_H2C_TAIL);
1294 	value32 &= 0xFFFC0000;
1295 	value32 |= (h2cq_addr + h2cq_size);
1296 	rtw_write32(rtwdev, REG_H2C_TAIL, value32);
1297 
1298 	value8 = rtw_read8(rtwdev, REG_H2C_INFO);
1299 	value8 = (u8)((value8 & 0xFC) | 0x01);
1300 	rtw_write8(rtwdev, REG_H2C_INFO, value8);
1301 
1302 	value8 = rtw_read8(rtwdev, REG_H2C_INFO);
1303 	value8 = (u8)((value8 & 0xFB) | 0x04);
1304 	rtw_write8(rtwdev, REG_H2C_INFO, value8);
1305 
1306 	value8 = rtw_read8(rtwdev, REG_TXDMA_OFFSET_CHK + 1);
1307 	value8 = (u8)((value8 & 0x7f) | 0x80);
1308 	rtw_write8(rtwdev, REG_TXDMA_OFFSET_CHK + 1, value8);
1309 
1310 	wp = rtw_read32(rtwdev, REG_H2C_PKT_WRITEADDR) & 0x3FFFF;
1311 	rp = rtw_read32(rtwdev, REG_H2C_PKT_READADDR) & 0x3FFFF;
1312 	h2cq_free = wp >= rp ? h2cq_size - (wp - rp) : rp - wp;
1313 
1314 	if (h2cq_size != h2cq_free) {
1315 		rtw_err(rtwdev, "H2C queue mismatch\n");
1316 		return -EINVAL;
1317 	}
1318 
1319 	return 0;
1320 }
1321 
1322 static int rtw_init_trx_cfg(struct rtw_dev *rtwdev)
1323 {
1324 	int ret;
1325 
1326 	ret = txdma_queue_mapping(rtwdev);
1327 	if (ret)
1328 		return ret;
1329 
1330 	ret = priority_queue_cfg(rtwdev);
1331 	if (ret)
1332 		return ret;
1333 
1334 	ret = init_h2c(rtwdev);
1335 	if (ret)
1336 		return ret;
1337 
1338 	return 0;
1339 }
1340 
1341 static int rtw_drv_info_cfg(struct rtw_dev *rtwdev)
1342 {
1343 	u8 value8;
1344 
1345 	rtw_write8(rtwdev, REG_RX_DRVINFO_SZ, PHY_STATUS_SIZE);
1346 	if (rtw_chip_wcpu_11ac(rtwdev)) {
1347 		value8 = rtw_read8(rtwdev, REG_TRXFF_BNDY + 1);
1348 		value8 &= 0xF0;
1349 		/* For rxdesc len = 0 issue */
1350 		value8 |= 0xF;
1351 		rtw_write8(rtwdev, REG_TRXFF_BNDY + 1, value8);
1352 	}
1353 	rtw_write32_set(rtwdev, REG_RCR, BIT_APP_PHYSTS);
1354 	rtw_write32_clr(rtwdev, REG_WMAC_OPTION_FUNCTION + 4, BIT(8) | BIT(9));
1355 
1356 	return 0;
1357 }
1358 
1359 int rtw_mac_init(struct rtw_dev *rtwdev)
1360 {
1361 	const struct rtw_chip_info *chip = rtwdev->chip;
1362 	int ret;
1363 
1364 	ret = rtw_init_trx_cfg(rtwdev);
1365 	if (ret)
1366 		return ret;
1367 
1368 	ret = chip->ops->mac_init(rtwdev);
1369 	if (ret)
1370 		return ret;
1371 
1372 	ret = rtw_drv_info_cfg(rtwdev);
1373 	if (ret)
1374 		return ret;
1375 
1376 	rtw_hci_interface_cfg(rtwdev);
1377 
1378 	return 0;
1379 }
1380