xref: /linux/drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c (revision 9c736ace0666efe68efd53fcdfa2c6653c3e0e72)
1 // SPDX-License-Identifier: (GPL-2.0 OR MIT)
2 /*
3  * Copyright (c) 2018 Synopsys, Inc. and/or its affiliates.
4  * stmmac XGMAC support.
5  */
6 
7 #include <linux/bitrev.h>
8 #include <linux/crc32.h>
9 #include <linux/iopoll.h>
10 #include "stmmac.h"
11 #include "stmmac_fpe.h"
12 #include "stmmac_ptp.h"
13 #include "stmmac_vlan.h"
14 #include "dwxlgmac2.h"
15 #include "dwxgmac2.h"
16 
dwxgmac2_core_init(struct mac_device_info * hw,struct net_device * dev)17 static void dwxgmac2_core_init(struct mac_device_info *hw,
18 			       struct net_device *dev)
19 {
20 	void __iomem *ioaddr = hw->pcsr;
21 	u32 tx, rx;
22 
23 	tx = readl(ioaddr + XGMAC_TX_CONFIG);
24 	rx = readl(ioaddr + XGMAC_RX_CONFIG);
25 
26 	tx |= XGMAC_CORE_INIT_TX;
27 	rx |= XGMAC_CORE_INIT_RX;
28 
29 	if (hw->ps) {
30 		tx |= XGMAC_CONFIG_TE;
31 		tx &= ~hw->link.speed_mask;
32 
33 		switch (hw->ps) {
34 		case SPEED_10000:
35 			tx |= hw->link.xgmii.speed10000;
36 			break;
37 		case SPEED_2500:
38 			tx |= hw->link.speed2500;
39 			break;
40 		case SPEED_1000:
41 		default:
42 			tx |= hw->link.speed1000;
43 			break;
44 		}
45 	}
46 
47 	writel(tx, ioaddr + XGMAC_TX_CONFIG);
48 	writel(rx, ioaddr + XGMAC_RX_CONFIG);
49 	writel(XGMAC_INT_DEFAULT_EN, ioaddr + XGMAC_INT_EN);
50 }
51 
dwxgmac2_update_caps(struct stmmac_priv * priv)52 static void dwxgmac2_update_caps(struct stmmac_priv *priv)
53 {
54 	if (!priv->dma_cap.mbps_10_100)
55 		priv->hw->link.caps &= ~(MAC_10 | MAC_100);
56 	else if (!priv->dma_cap.half_duplex)
57 		priv->hw->link.caps &= ~(MAC_10HD | MAC_100HD);
58 }
59 
dwxgmac2_set_mac(void __iomem * ioaddr,bool enable)60 static void dwxgmac2_set_mac(void __iomem *ioaddr, bool enable)
61 {
62 	u32 tx = readl(ioaddr + XGMAC_TX_CONFIG);
63 	u32 rx = readl(ioaddr + XGMAC_RX_CONFIG);
64 
65 	if (enable) {
66 		tx |= XGMAC_CONFIG_TE;
67 		rx |= XGMAC_CONFIG_RE;
68 	} else {
69 		tx &= ~XGMAC_CONFIG_TE;
70 		rx &= ~XGMAC_CONFIG_RE;
71 	}
72 
73 	writel(tx, ioaddr + XGMAC_TX_CONFIG);
74 	writel(rx, ioaddr + XGMAC_RX_CONFIG);
75 }
76 
dwxgmac2_rx_ipc(struct mac_device_info * hw)77 static int dwxgmac2_rx_ipc(struct mac_device_info *hw)
78 {
79 	void __iomem *ioaddr = hw->pcsr;
80 	u32 value;
81 
82 	value = readl(ioaddr + XGMAC_RX_CONFIG);
83 	if (hw->rx_csum)
84 		value |= XGMAC_CONFIG_IPC;
85 	else
86 		value &= ~XGMAC_CONFIG_IPC;
87 	writel(value, ioaddr + XGMAC_RX_CONFIG);
88 
89 	return !!(readl(ioaddr + XGMAC_RX_CONFIG) & XGMAC_CONFIG_IPC);
90 }
91 
dwxgmac2_rx_queue_enable(struct mac_device_info * hw,u8 mode,u32 queue)92 static void dwxgmac2_rx_queue_enable(struct mac_device_info *hw, u8 mode,
93 				     u32 queue)
94 {
95 	void __iomem *ioaddr = hw->pcsr;
96 	u32 value;
97 
98 	value = readl(ioaddr + XGMAC_RXQ_CTRL0) & ~XGMAC_RXQEN(queue);
99 	if (mode == MTL_QUEUE_AVB)
100 		value |= 0x1 << XGMAC_RXQEN_SHIFT(queue);
101 	else if (mode == MTL_QUEUE_DCB)
102 		value |= 0x2 << XGMAC_RXQEN_SHIFT(queue);
103 	writel(value, ioaddr + XGMAC_RXQ_CTRL0);
104 }
105 
dwxgmac2_rx_queue_prio(struct mac_device_info * hw,u32 prio,u32 queue)106 static void dwxgmac2_rx_queue_prio(struct mac_device_info *hw, u32 prio,
107 				   u32 queue)
108 {
109 	void __iomem *ioaddr = hw->pcsr;
110 	u32 clear_mask = 0;
111 	u32 ctrl2, ctrl3;
112 	int i;
113 
114 	ctrl2 = readl(ioaddr + XGMAC_RXQ_CTRL2);
115 	ctrl3 = readl(ioaddr + XGMAC_RXQ_CTRL3);
116 
117 	/* The software must ensure that the same priority
118 	 * is not mapped to multiple Rx queues
119 	 */
120 	for (i = 0; i < 4; i++)
121 		clear_mask |= ((prio << XGMAC_PSRQ_SHIFT(i)) &
122 						XGMAC_PSRQ(i));
123 
124 	ctrl2 &= ~clear_mask;
125 	ctrl3 &= ~clear_mask;
126 
127 	/* First assign new priorities to a queue, then
128 	 * clear them from others queues
129 	 */
130 	if (queue < 4) {
131 		ctrl2 |= (prio << XGMAC_PSRQ_SHIFT(queue)) &
132 						XGMAC_PSRQ(queue);
133 
134 		writel(ctrl2, ioaddr + XGMAC_RXQ_CTRL2);
135 		writel(ctrl3, ioaddr + XGMAC_RXQ_CTRL3);
136 	} else {
137 		queue -= 4;
138 
139 		ctrl3 |= (prio << XGMAC_PSRQ_SHIFT(queue)) &
140 						XGMAC_PSRQ(queue);
141 
142 		writel(ctrl3, ioaddr + XGMAC_RXQ_CTRL3);
143 		writel(ctrl2, ioaddr + XGMAC_RXQ_CTRL2);
144 	}
145 }
146 
dwxgmac2_tx_queue_prio(struct mac_device_info * hw,u32 prio,u32 queue)147 static void dwxgmac2_tx_queue_prio(struct mac_device_info *hw, u32 prio,
148 				   u32 queue)
149 {
150 	void __iomem *ioaddr = hw->pcsr;
151 	u32 value, reg;
152 
153 	reg = (queue < 4) ? XGMAC_TC_PRTY_MAP0 : XGMAC_TC_PRTY_MAP1;
154 	if (queue >= 4)
155 		queue -= 4;
156 
157 	value = readl(ioaddr + reg);
158 	value &= ~XGMAC_PSTC(queue);
159 	value |= (prio << XGMAC_PSTC_SHIFT(queue)) & XGMAC_PSTC(queue);
160 
161 	writel(value, ioaddr + reg);
162 }
163 
dwxgmac2_rx_queue_routing(struct mac_device_info * hw,u8 packet,u32 queue)164 static void dwxgmac2_rx_queue_routing(struct mac_device_info *hw,
165 				      u8 packet, u32 queue)
166 {
167 	void __iomem *ioaddr = hw->pcsr;
168 	u32 value;
169 
170 	static const struct stmmac_rx_routing dwxgmac2_route_possibilities[] = {
171 		{ XGMAC_AVCPQ, XGMAC_AVCPQ_SHIFT },
172 		{ XGMAC_PTPQ, XGMAC_PTPQ_SHIFT },
173 		{ XGMAC_DCBCPQ, XGMAC_DCBCPQ_SHIFT },
174 		{ XGMAC_UPQ, XGMAC_UPQ_SHIFT },
175 		{ XGMAC_MCBCQ, XGMAC_MCBCQ_SHIFT },
176 	};
177 
178 	value = readl(ioaddr + XGMAC_RXQ_CTRL1);
179 
180 	/* routing configuration */
181 	value &= ~dwxgmac2_route_possibilities[packet - 1].reg_mask;
182 	value |= (queue << dwxgmac2_route_possibilities[packet - 1].reg_shift) &
183 		 dwxgmac2_route_possibilities[packet - 1].reg_mask;
184 
185 	/* some packets require extra ops */
186 	if (packet == PACKET_AVCPQ)
187 		value |= FIELD_PREP(XGMAC_TACPQE, 1);
188 	else if (packet == PACKET_MCBCQ)
189 		value |= FIELD_PREP(XGMAC_MCBCQEN, 1);
190 
191 	writel(value, ioaddr + XGMAC_RXQ_CTRL1);
192 }
193 
dwxgmac2_prog_mtl_rx_algorithms(struct mac_device_info * hw,u32 rx_alg)194 static void dwxgmac2_prog_mtl_rx_algorithms(struct mac_device_info *hw,
195 					    u32 rx_alg)
196 {
197 	void __iomem *ioaddr = hw->pcsr;
198 	u32 value;
199 
200 	value = readl(ioaddr + XGMAC_MTL_OPMODE);
201 	value &= ~XGMAC_RAA;
202 
203 	switch (rx_alg) {
204 	case MTL_RX_ALGORITHM_SP:
205 		break;
206 	case MTL_RX_ALGORITHM_WSP:
207 		value |= XGMAC_RAA;
208 		break;
209 	default:
210 		break;
211 	}
212 
213 	writel(value, ioaddr + XGMAC_MTL_OPMODE);
214 }
215 
dwxgmac2_prog_mtl_tx_algorithms(struct mac_device_info * hw,u32 tx_alg)216 static void dwxgmac2_prog_mtl_tx_algorithms(struct mac_device_info *hw,
217 					    u32 tx_alg)
218 {
219 	void __iomem *ioaddr = hw->pcsr;
220 	bool ets = true;
221 	u32 value;
222 	int i;
223 
224 	value = readl(ioaddr + XGMAC_MTL_OPMODE);
225 	value &= ~XGMAC_ETSALG;
226 
227 	switch (tx_alg) {
228 	case MTL_TX_ALGORITHM_WRR:
229 		value |= XGMAC_WRR;
230 		break;
231 	case MTL_TX_ALGORITHM_WFQ:
232 		value |= XGMAC_WFQ;
233 		break;
234 	case MTL_TX_ALGORITHM_DWRR:
235 		value |= XGMAC_DWRR;
236 		break;
237 	default:
238 		ets = false;
239 		break;
240 	}
241 
242 	writel(value, ioaddr + XGMAC_MTL_OPMODE);
243 
244 	/* Set ETS if desired */
245 	for (i = 0; i < MTL_MAX_TX_QUEUES; i++) {
246 		value = readl(ioaddr + XGMAC_MTL_TCx_ETS_CONTROL(i));
247 		value &= ~XGMAC_TSA;
248 		if (ets)
249 			value |= XGMAC_ETS;
250 		writel(value, ioaddr + XGMAC_MTL_TCx_ETS_CONTROL(i));
251 	}
252 }
253 
dwxgmac2_set_mtl_tx_queue_weight(struct stmmac_priv * priv,struct mac_device_info * hw,u32 weight,u32 queue)254 static void dwxgmac2_set_mtl_tx_queue_weight(struct stmmac_priv *priv,
255 					     struct mac_device_info *hw,
256 					     u32 weight, u32 queue)
257 {
258 	void __iomem *ioaddr = hw->pcsr;
259 
260 	writel(weight, ioaddr + XGMAC_MTL_TCx_QUANTUM_WEIGHT(queue));
261 }
262 
dwxgmac2_map_mtl_to_dma(struct mac_device_info * hw,u32 queue,u32 chan)263 static void dwxgmac2_map_mtl_to_dma(struct mac_device_info *hw, u32 queue,
264 				    u32 chan)
265 {
266 	void __iomem *ioaddr = hw->pcsr;
267 	u32 value, reg;
268 
269 	reg = (queue < 4) ? XGMAC_MTL_RXQ_DMA_MAP0 : XGMAC_MTL_RXQ_DMA_MAP1;
270 	if (queue >= 4)
271 		queue -= 4;
272 
273 	value = readl(ioaddr + reg);
274 	value &= ~XGMAC_QxMDMACH(queue);
275 	value |= (chan << XGMAC_QxMDMACH_SHIFT(queue)) & XGMAC_QxMDMACH(queue);
276 
277 	writel(value, ioaddr + reg);
278 }
279 
dwxgmac2_config_cbs(struct stmmac_priv * priv,struct mac_device_info * hw,u32 send_slope,u32 idle_slope,u32 high_credit,u32 low_credit,u32 queue)280 static void dwxgmac2_config_cbs(struct stmmac_priv *priv,
281 				struct mac_device_info *hw,
282 				u32 send_slope, u32 idle_slope,
283 				u32 high_credit, u32 low_credit, u32 queue)
284 {
285 	void __iomem *ioaddr = hw->pcsr;
286 	u32 value;
287 
288 	writel(send_slope, ioaddr + XGMAC_MTL_TCx_SENDSLOPE(queue));
289 	writel(idle_slope, ioaddr + XGMAC_MTL_TCx_QUANTUM_WEIGHT(queue));
290 	writel(high_credit, ioaddr + XGMAC_MTL_TCx_HICREDIT(queue));
291 	writel(low_credit, ioaddr + XGMAC_MTL_TCx_LOCREDIT(queue));
292 
293 	value = readl(ioaddr + XGMAC_MTL_TCx_ETS_CONTROL(queue));
294 	value &= ~XGMAC_TSA;
295 	value |= XGMAC_CC | XGMAC_CBS;
296 	writel(value, ioaddr + XGMAC_MTL_TCx_ETS_CONTROL(queue));
297 }
298 
dwxgmac2_dump_regs(struct mac_device_info * hw,u32 * reg_space)299 static void dwxgmac2_dump_regs(struct mac_device_info *hw, u32 *reg_space)
300 {
301 	void __iomem *ioaddr = hw->pcsr;
302 	int i;
303 
304 	for (i = 0; i < XGMAC_MAC_REGSIZE; i++)
305 		reg_space[i] = readl(ioaddr + i * 4);
306 }
307 
dwxgmac2_host_irq_status(struct mac_device_info * hw,struct stmmac_extra_stats * x)308 static int dwxgmac2_host_irq_status(struct mac_device_info *hw,
309 				    struct stmmac_extra_stats *x)
310 {
311 	void __iomem *ioaddr = hw->pcsr;
312 	u32 stat, en;
313 	int ret = 0;
314 
315 	en = readl(ioaddr + XGMAC_INT_EN);
316 	stat = readl(ioaddr + XGMAC_INT_STATUS);
317 
318 	stat &= en;
319 
320 	if (stat & XGMAC_PMTIS) {
321 		x->irq_receive_pmt_irq_n++;
322 		readl(ioaddr + XGMAC_PMT);
323 	}
324 
325 	if (stat & XGMAC_LPIIS) {
326 		u32 lpi = readl(ioaddr + XGMAC_LPI_CTRL);
327 
328 		if (lpi & LPI_CTRL_STATUS_TLPIEN) {
329 			ret |= CORE_IRQ_TX_PATH_IN_LPI_MODE;
330 			x->irq_tx_path_in_lpi_mode_n++;
331 		}
332 		if (lpi & LPI_CTRL_STATUS_TLPIEX) {
333 			ret |= CORE_IRQ_TX_PATH_EXIT_LPI_MODE;
334 			x->irq_tx_path_exit_lpi_mode_n++;
335 		}
336 		if (lpi & LPI_CTRL_STATUS_RLPIEN)
337 			x->irq_rx_path_in_lpi_mode_n++;
338 		if (lpi & LPI_CTRL_STATUS_RLPIEX)
339 			x->irq_rx_path_exit_lpi_mode_n++;
340 	}
341 
342 	return ret;
343 }
344 
dwxgmac2_host_mtl_irq_status(struct stmmac_priv * priv,struct mac_device_info * hw,u32 chan)345 static int dwxgmac2_host_mtl_irq_status(struct stmmac_priv *priv,
346 					struct mac_device_info *hw, u32 chan)
347 {
348 	void __iomem *ioaddr = hw->pcsr;
349 	int ret = 0;
350 	u32 status;
351 
352 	status = readl(ioaddr + XGMAC_MTL_INT_STATUS);
353 	if (status & BIT(chan)) {
354 		u32 chan_status = readl(ioaddr + XGMAC_MTL_QINT_STATUS(chan));
355 
356 		if (chan_status & XGMAC_RXOVFIS)
357 			ret |= CORE_IRQ_MTL_RX_OVERFLOW;
358 
359 		writel(~0x0, ioaddr + XGMAC_MTL_QINT_STATUS(chan));
360 	}
361 
362 	return ret;
363 }
364 
dwxgmac2_flow_ctrl(struct mac_device_info * hw,unsigned int duplex,unsigned int fc,unsigned int pause_time,u32 tx_cnt)365 static void dwxgmac2_flow_ctrl(struct mac_device_info *hw, unsigned int duplex,
366 			       unsigned int fc, unsigned int pause_time,
367 			       u32 tx_cnt)
368 {
369 	void __iomem *ioaddr = hw->pcsr;
370 	u32 i;
371 
372 	if (fc & FLOW_RX)
373 		writel(XGMAC_RFE, ioaddr + XGMAC_RX_FLOW_CTRL);
374 	if (fc & FLOW_TX) {
375 		for (i = 0; i < tx_cnt; i++) {
376 			u32 value = XGMAC_TFE;
377 
378 			if (duplex)
379 				value |= pause_time << XGMAC_PT_SHIFT;
380 
381 			writel(value, ioaddr + XGMAC_Qx_TX_FLOW_CTRL(i));
382 		}
383 	}
384 }
385 
dwxgmac2_pmt(struct mac_device_info * hw,unsigned long mode)386 static void dwxgmac2_pmt(struct mac_device_info *hw, unsigned long mode)
387 {
388 	void __iomem *ioaddr = hw->pcsr;
389 	u32 val = 0x0;
390 
391 	if (mode & WAKE_MAGIC)
392 		val |= XGMAC_PWRDWN | XGMAC_MGKPKTEN;
393 	if (mode & WAKE_UCAST)
394 		val |= XGMAC_PWRDWN | XGMAC_GLBLUCAST | XGMAC_RWKPKTEN;
395 	if (val) {
396 		u32 cfg = readl(ioaddr + XGMAC_RX_CONFIG);
397 		cfg |= XGMAC_CONFIG_RE;
398 		writel(cfg, ioaddr + XGMAC_RX_CONFIG);
399 	}
400 
401 	writel(val, ioaddr + XGMAC_PMT);
402 }
403 
dwxgmac2_set_umac_addr(struct mac_device_info * hw,const unsigned char * addr,unsigned int reg_n)404 static void dwxgmac2_set_umac_addr(struct mac_device_info *hw,
405 				   const unsigned char *addr,
406 				   unsigned int reg_n)
407 {
408 	void __iomem *ioaddr = hw->pcsr;
409 	u32 value;
410 
411 	value = (addr[5] << 8) | addr[4];
412 	writel(value | XGMAC_AE, ioaddr + XGMAC_ADDRx_HIGH(reg_n));
413 
414 	value = (addr[3] << 24) | (addr[2] << 16) | (addr[1] << 8) | addr[0];
415 	writel(value, ioaddr + XGMAC_ADDRx_LOW(reg_n));
416 }
417 
dwxgmac2_get_umac_addr(struct mac_device_info * hw,unsigned char * addr,unsigned int reg_n)418 static void dwxgmac2_get_umac_addr(struct mac_device_info *hw,
419 				   unsigned char *addr, unsigned int reg_n)
420 {
421 	void __iomem *ioaddr = hw->pcsr;
422 	u32 hi_addr, lo_addr;
423 
424 	/* Read the MAC address from the hardware */
425 	hi_addr = readl(ioaddr + XGMAC_ADDRx_HIGH(reg_n));
426 	lo_addr = readl(ioaddr + XGMAC_ADDRx_LOW(reg_n));
427 
428 	/* Extract the MAC address from the high and low words */
429 	addr[0] = lo_addr & 0xff;
430 	addr[1] = (lo_addr >> 8) & 0xff;
431 	addr[2] = (lo_addr >> 16) & 0xff;
432 	addr[3] = (lo_addr >> 24) & 0xff;
433 	addr[4] = hi_addr & 0xff;
434 	addr[5] = (hi_addr >> 8) & 0xff;
435 }
436 
dwxgmac2_set_lpi_mode(struct mac_device_info * hw,enum stmmac_lpi_mode mode,bool en_tx_lpi_clockgating,u32 et)437 static int dwxgmac2_set_lpi_mode(struct mac_device_info *hw,
438 				 enum stmmac_lpi_mode mode,
439 				 bool en_tx_lpi_clockgating, u32 et)
440 {
441 	void __iomem *ioaddr = hw->pcsr;
442 	u32 value;
443 
444 	if (mode == STMMAC_LPI_TIMER)
445 		return -EOPNOTSUPP;
446 
447 	value = readl(ioaddr + XGMAC_LPI_CTRL);
448 	if (mode == STMMAC_LPI_FORCED) {
449 		value |= LPI_CTRL_STATUS_LPIEN | LPI_CTRL_STATUS_LPITXA;
450 		if (en_tx_lpi_clockgating)
451 			value |= LPI_CTRL_STATUS_LPITCSE;
452 	} else {
453 		value &= ~(LPI_CTRL_STATUS_LPIEN | LPI_CTRL_STATUS_LPITXA |
454 			   LPI_CTRL_STATUS_LPITCSE);
455 	}
456 	writel(value, ioaddr + XGMAC_LPI_CTRL);
457 
458 	return 0;
459 }
460 
dwxgmac2_set_eee_pls(struct mac_device_info * hw,int link)461 static void dwxgmac2_set_eee_pls(struct mac_device_info *hw, int link)
462 {
463 	void __iomem *ioaddr = hw->pcsr;
464 	u32 value;
465 
466 	value = readl(ioaddr + XGMAC_LPI_CTRL);
467 	if (link)
468 		value |= LPI_CTRL_STATUS_PLS;
469 	else
470 		value &= ~LPI_CTRL_STATUS_PLS;
471 	writel(value, ioaddr + XGMAC_LPI_CTRL);
472 }
473 
dwxgmac2_set_eee_timer(struct mac_device_info * hw,int ls,int tw)474 static void dwxgmac2_set_eee_timer(struct mac_device_info *hw, int ls, int tw)
475 {
476 	void __iomem *ioaddr = hw->pcsr;
477 	u32 value;
478 
479 	value = (tw & 0xffff) | ((ls & 0x3ff) << 16);
480 	writel(value, ioaddr + XGMAC_LPI_TIMER_CTRL);
481 }
482 
dwxgmac2_set_mchash(void __iomem * ioaddr,u32 * mcfilterbits,int mcbitslog2)483 static void dwxgmac2_set_mchash(void __iomem *ioaddr, u32 *mcfilterbits,
484 				int mcbitslog2)
485 {
486 	int numhashregs, regs;
487 
488 	switch (mcbitslog2) {
489 	case 6:
490 		numhashregs = 2;
491 		break;
492 	case 7:
493 		numhashregs = 4;
494 		break;
495 	case 8:
496 		numhashregs = 8;
497 		break;
498 	default:
499 		return;
500 	}
501 
502 	for (regs = 0; regs < numhashregs; regs++)
503 		writel(mcfilterbits[regs], ioaddr + XGMAC_HASH_TABLE(regs));
504 }
505 
dwxgmac2_set_filter(struct mac_device_info * hw,struct net_device * dev)506 static void dwxgmac2_set_filter(struct mac_device_info *hw,
507 				struct net_device *dev)
508 {
509 	void __iomem *ioaddr = (void __iomem *)dev->base_addr;
510 	u32 value = readl(ioaddr + XGMAC_PACKET_FILTER);
511 	int mcbitslog2 = hw->mcast_bits_log2;
512 	u32 mc_filter[8];
513 	int i;
514 
515 	value &= ~(XGMAC_FILTER_PR | XGMAC_FILTER_HMC | XGMAC_FILTER_PM);
516 	value |= XGMAC_FILTER_HPF;
517 
518 	memset(mc_filter, 0, sizeof(mc_filter));
519 
520 	if (dev->flags & IFF_PROMISC) {
521 		value |= XGMAC_FILTER_PR;
522 		value |= XGMAC_FILTER_PCF;
523 	} else if ((dev->flags & IFF_ALLMULTI) ||
524 		   (netdev_mc_count(dev) > hw->multicast_filter_bins)) {
525 		value |= XGMAC_FILTER_PM;
526 
527 		for (i = 0; i < XGMAC_MAX_HASH_TABLE; i++)
528 			writel(~0x0, ioaddr + XGMAC_HASH_TABLE(i));
529 	} else if (!netdev_mc_empty(dev) && (dev->flags & IFF_MULTICAST)) {
530 		struct netdev_hw_addr *ha;
531 
532 		value |= XGMAC_FILTER_HMC;
533 
534 		netdev_for_each_mc_addr(ha, dev) {
535 			u32 nr = (bitrev32(~crc32_le(~0, ha->addr, 6)) >>
536 					(32 - mcbitslog2));
537 			mc_filter[nr >> 5] |= (1 << (nr & 0x1F));
538 		}
539 	}
540 
541 	dwxgmac2_set_mchash(ioaddr, mc_filter, mcbitslog2);
542 
543 	/* Handle multiple unicast addresses */
544 	if (netdev_uc_count(dev) > hw->unicast_filter_entries) {
545 		value |= XGMAC_FILTER_PR;
546 	} else {
547 		struct netdev_hw_addr *ha;
548 		int reg = 1;
549 
550 		netdev_for_each_uc_addr(ha, dev) {
551 			dwxgmac2_set_umac_addr(hw, ha->addr, reg);
552 			reg++;
553 		}
554 
555 		for ( ; reg < XGMAC_ADDR_MAX; reg++) {
556 			writel(0, ioaddr + XGMAC_ADDRx_HIGH(reg));
557 			writel(0, ioaddr + XGMAC_ADDRx_LOW(reg));
558 		}
559 	}
560 
561 	writel(value, ioaddr + XGMAC_PACKET_FILTER);
562 }
563 
dwxgmac2_set_mac_loopback(void __iomem * ioaddr,bool enable)564 static void dwxgmac2_set_mac_loopback(void __iomem *ioaddr, bool enable)
565 {
566 	u32 value = readl(ioaddr + XGMAC_RX_CONFIG);
567 
568 	if (enable)
569 		value |= XGMAC_CONFIG_LM;
570 	else
571 		value &= ~XGMAC_CONFIG_LM;
572 
573 	writel(value, ioaddr + XGMAC_RX_CONFIG);
574 }
575 
dwxgmac2_rss_write_reg(void __iomem * ioaddr,bool is_key,int idx,u32 val)576 static int dwxgmac2_rss_write_reg(void __iomem *ioaddr, bool is_key, int idx,
577 				  u32 val)
578 {
579 	u32 ctrl = 0;
580 
581 	writel(val, ioaddr + XGMAC_RSS_DATA);
582 	ctrl |= idx << XGMAC_RSSIA_SHIFT;
583 	ctrl |= is_key ? XGMAC_ADDRT : 0x0;
584 	ctrl |= XGMAC_OB;
585 	writel(ctrl, ioaddr + XGMAC_RSS_ADDR);
586 
587 	return readl_poll_timeout(ioaddr + XGMAC_RSS_ADDR, ctrl,
588 				  !(ctrl & XGMAC_OB), 100, 10000);
589 }
590 
dwxgmac2_rss_configure(struct mac_device_info * hw,struct stmmac_rss * cfg,u32 num_rxq)591 static int dwxgmac2_rss_configure(struct mac_device_info *hw,
592 				  struct stmmac_rss *cfg, u32 num_rxq)
593 {
594 	void __iomem *ioaddr = hw->pcsr;
595 	u32 value, *key;
596 	int i, ret;
597 
598 	value = readl(ioaddr + XGMAC_RSS_CTRL);
599 	if (!cfg || !cfg->enable) {
600 		value &= ~XGMAC_RSSE;
601 		writel(value, ioaddr + XGMAC_RSS_CTRL);
602 		return 0;
603 	}
604 
605 	key = (u32 *)cfg->key;
606 	for (i = 0; i < (ARRAY_SIZE(cfg->key) / sizeof(u32)); i++) {
607 		ret = dwxgmac2_rss_write_reg(ioaddr, true, i, key[i]);
608 		if (ret)
609 			return ret;
610 	}
611 
612 	for (i = 0; i < ARRAY_SIZE(cfg->table); i++) {
613 		ret = dwxgmac2_rss_write_reg(ioaddr, false, i, cfg->table[i]);
614 		if (ret)
615 			return ret;
616 	}
617 
618 	for (i = 0; i < num_rxq; i++)
619 		dwxgmac2_map_mtl_to_dma(hw, i, XGMAC_QDDMACH);
620 
621 	value |= XGMAC_UDP4TE | XGMAC_TCP4TE | XGMAC_IP2TE | XGMAC_RSSE;
622 	writel(value, ioaddr + XGMAC_RSS_CTRL);
623 	return 0;
624 }
625 
626 struct dwxgmac3_error_desc {
627 	bool valid;
628 	const char *desc;
629 	const char *detailed_desc;
630 };
631 
632 #define STAT_OFF(field)		offsetof(struct stmmac_safety_stats, field)
633 
dwxgmac3_log_error(struct net_device * ndev,u32 value,bool corr,const char * module_name,const struct dwxgmac3_error_desc * desc,unsigned long field_offset,struct stmmac_safety_stats * stats)634 static void dwxgmac3_log_error(struct net_device *ndev, u32 value, bool corr,
635 			       const char *module_name,
636 			       const struct dwxgmac3_error_desc *desc,
637 			       unsigned long field_offset,
638 			       struct stmmac_safety_stats *stats)
639 {
640 	unsigned long loc, mask;
641 	u8 *bptr = (u8 *)stats;
642 	unsigned long *ptr;
643 
644 	ptr = (unsigned long *)(bptr + field_offset);
645 
646 	mask = value;
647 	for_each_set_bit(loc, &mask, 32) {
648 		netdev_err(ndev, "Found %s error in %s: '%s: %s'\n", corr ?
649 				"correctable" : "uncorrectable", module_name,
650 				desc[loc].desc, desc[loc].detailed_desc);
651 
652 		/* Update counters */
653 		ptr[loc]++;
654 	}
655 }
656 
657 static const struct dwxgmac3_error_desc dwxgmac3_mac_errors[32]= {
658 	{ true, "ATPES", "Application Transmit Interface Parity Check Error" },
659 	{ true, "DPES", "Descriptor Cache Data Path Parity Check Error" },
660 	{ true, "TPES", "TSO Data Path Parity Check Error" },
661 	{ true, "TSOPES", "TSO Header Data Path Parity Check Error" },
662 	{ true, "MTPES", "MTL Data Path Parity Check Error" },
663 	{ true, "MTSPES", "MTL TX Status Data Path Parity Check Error" },
664 	{ true, "MTBUPES", "MAC TBU Data Path Parity Check Error" },
665 	{ true, "MTFCPES", "MAC TFC Data Path Parity Check Error" },
666 	{ true, "ARPES", "Application Receive Interface Data Path Parity Check Error" },
667 	{ true, "MRWCPES", "MTL RWC Data Path Parity Check Error" },
668 	{ true, "MRRCPES", "MTL RCC Data Path Parity Check Error" },
669 	{ true, "CWPES", "CSR Write Data Path Parity Check Error" },
670 	{ true, "ASRPES", "AXI Slave Read Data Path Parity Check Error" },
671 	{ true, "TTES", "TX FSM Timeout Error" },
672 	{ true, "RTES", "RX FSM Timeout Error" },
673 	{ true, "CTES", "CSR FSM Timeout Error" },
674 	{ true, "ATES", "APP FSM Timeout Error" },
675 	{ true, "PTES", "PTP FSM Timeout Error" },
676 	{ false, "UNKNOWN", "Unknown Error" }, /* 18 */
677 	{ false, "UNKNOWN", "Unknown Error" }, /* 19 */
678 	{ false, "UNKNOWN", "Unknown Error" }, /* 20 */
679 	{ true, "MSTTES", "Master Read/Write Timeout Error" },
680 	{ true, "SLVTES", "Slave Read/Write Timeout Error" },
681 	{ true, "ATITES", "Application Timeout on ATI Interface Error" },
682 	{ true, "ARITES", "Application Timeout on ARI Interface Error" },
683 	{ true, "FSMPES", "FSM State Parity Error" },
684 	{ false, "UNKNOWN", "Unknown Error" }, /* 26 */
685 	{ false, "UNKNOWN", "Unknown Error" }, /* 27 */
686 	{ false, "UNKNOWN", "Unknown Error" }, /* 28 */
687 	{ false, "UNKNOWN", "Unknown Error" }, /* 29 */
688 	{ false, "UNKNOWN", "Unknown Error" }, /* 30 */
689 	{ true, "CPI", "Control Register Parity Check Error" },
690 };
691 
dwxgmac3_handle_mac_err(struct net_device * ndev,void __iomem * ioaddr,bool correctable,struct stmmac_safety_stats * stats)692 static void dwxgmac3_handle_mac_err(struct net_device *ndev,
693 				    void __iomem *ioaddr, bool correctable,
694 				    struct stmmac_safety_stats *stats)
695 {
696 	u32 value;
697 
698 	value = readl(ioaddr + XGMAC_MAC_DPP_FSM_INT_STATUS);
699 	writel(value, ioaddr + XGMAC_MAC_DPP_FSM_INT_STATUS);
700 
701 	dwxgmac3_log_error(ndev, value, correctable, "MAC",
702 			   dwxgmac3_mac_errors, STAT_OFF(mac_errors), stats);
703 }
704 
705 static const struct dwxgmac3_error_desc dwxgmac3_mtl_errors[32]= {
706 	{ true, "TXCES", "MTL TX Memory Error" },
707 	{ true, "TXAMS", "MTL TX Memory Address Mismatch Error" },
708 	{ true, "TXUES", "MTL TX Memory Error" },
709 	{ false, "UNKNOWN", "Unknown Error" }, /* 3 */
710 	{ true, "RXCES", "MTL RX Memory Error" },
711 	{ true, "RXAMS", "MTL RX Memory Address Mismatch Error" },
712 	{ true, "RXUES", "MTL RX Memory Error" },
713 	{ false, "UNKNOWN", "Unknown Error" }, /* 7 */
714 	{ true, "ECES", "MTL EST Memory Error" },
715 	{ true, "EAMS", "MTL EST Memory Address Mismatch Error" },
716 	{ true, "EUES", "MTL EST Memory Error" },
717 	{ false, "UNKNOWN", "Unknown Error" }, /* 11 */
718 	{ true, "RPCES", "MTL RX Parser Memory Error" },
719 	{ true, "RPAMS", "MTL RX Parser Memory Address Mismatch Error" },
720 	{ true, "RPUES", "MTL RX Parser Memory Error" },
721 	{ false, "UNKNOWN", "Unknown Error" }, /* 15 */
722 	{ false, "UNKNOWN", "Unknown Error" }, /* 16 */
723 	{ false, "UNKNOWN", "Unknown Error" }, /* 17 */
724 	{ false, "UNKNOWN", "Unknown Error" }, /* 18 */
725 	{ false, "UNKNOWN", "Unknown Error" }, /* 19 */
726 	{ false, "UNKNOWN", "Unknown Error" }, /* 20 */
727 	{ false, "UNKNOWN", "Unknown Error" }, /* 21 */
728 	{ false, "UNKNOWN", "Unknown Error" }, /* 22 */
729 	{ false, "UNKNOWN", "Unknown Error" }, /* 23 */
730 	{ false, "UNKNOWN", "Unknown Error" }, /* 24 */
731 	{ false, "UNKNOWN", "Unknown Error" }, /* 25 */
732 	{ false, "UNKNOWN", "Unknown Error" }, /* 26 */
733 	{ false, "UNKNOWN", "Unknown Error" }, /* 27 */
734 	{ false, "UNKNOWN", "Unknown Error" }, /* 28 */
735 	{ false, "UNKNOWN", "Unknown Error" }, /* 29 */
736 	{ false, "UNKNOWN", "Unknown Error" }, /* 30 */
737 	{ false, "UNKNOWN", "Unknown Error" }, /* 31 */
738 };
739 
dwxgmac3_handle_mtl_err(struct net_device * ndev,void __iomem * ioaddr,bool correctable,struct stmmac_safety_stats * stats)740 static void dwxgmac3_handle_mtl_err(struct net_device *ndev,
741 				    void __iomem *ioaddr, bool correctable,
742 				    struct stmmac_safety_stats *stats)
743 {
744 	u32 value;
745 
746 	value = readl(ioaddr + XGMAC_MTL_ECC_INT_STATUS);
747 	writel(value, ioaddr + XGMAC_MTL_ECC_INT_STATUS);
748 
749 	dwxgmac3_log_error(ndev, value, correctable, "MTL",
750 			   dwxgmac3_mtl_errors, STAT_OFF(mtl_errors), stats);
751 }
752 
753 static const struct dwxgmac3_error_desc dwxgmac3_dma_errors[32]= {
754 	{ true, "TCES", "DMA TSO Memory Error" },
755 	{ true, "TAMS", "DMA TSO Memory Address Mismatch Error" },
756 	{ true, "TUES", "DMA TSO Memory Error" },
757 	{ false, "UNKNOWN", "Unknown Error" }, /* 3 */
758 	{ true, "DCES", "DMA DCACHE Memory Error" },
759 	{ true, "DAMS", "DMA DCACHE Address Mismatch Error" },
760 	{ true, "DUES", "DMA DCACHE Memory Error" },
761 	{ false, "UNKNOWN", "Unknown Error" }, /* 7 */
762 	{ false, "UNKNOWN", "Unknown Error" }, /* 8 */
763 	{ false, "UNKNOWN", "Unknown Error" }, /* 9 */
764 	{ false, "UNKNOWN", "Unknown Error" }, /* 10 */
765 	{ false, "UNKNOWN", "Unknown Error" }, /* 11 */
766 	{ false, "UNKNOWN", "Unknown Error" }, /* 12 */
767 	{ false, "UNKNOWN", "Unknown Error" }, /* 13 */
768 	{ false, "UNKNOWN", "Unknown Error" }, /* 14 */
769 	{ false, "UNKNOWN", "Unknown Error" }, /* 15 */
770 	{ false, "UNKNOWN", "Unknown Error" }, /* 16 */
771 	{ false, "UNKNOWN", "Unknown Error" }, /* 17 */
772 	{ false, "UNKNOWN", "Unknown Error" }, /* 18 */
773 	{ false, "UNKNOWN", "Unknown Error" }, /* 19 */
774 	{ false, "UNKNOWN", "Unknown Error" }, /* 20 */
775 	{ false, "UNKNOWN", "Unknown Error" }, /* 21 */
776 	{ false, "UNKNOWN", "Unknown Error" }, /* 22 */
777 	{ false, "UNKNOWN", "Unknown Error" }, /* 23 */
778 	{ false, "UNKNOWN", "Unknown Error" }, /* 24 */
779 	{ false, "UNKNOWN", "Unknown Error" }, /* 25 */
780 	{ false, "UNKNOWN", "Unknown Error" }, /* 26 */
781 	{ false, "UNKNOWN", "Unknown Error" }, /* 27 */
782 	{ false, "UNKNOWN", "Unknown Error" }, /* 28 */
783 	{ false, "UNKNOWN", "Unknown Error" }, /* 29 */
784 	{ false, "UNKNOWN", "Unknown Error" }, /* 30 */
785 	{ false, "UNKNOWN", "Unknown Error" }, /* 31 */
786 };
787 
788 static const char dpp_rx_err[] = "Read Rx Descriptor Parity checker Error";
789 static const char dpp_tx_err[] = "Read Tx Descriptor Parity checker Error";
790 static const struct dwxgmac3_error_desc dwxgmac3_dma_dpp_errors[32] = {
791 	{ true, "TDPES0", dpp_tx_err },
792 	{ true, "TDPES1", dpp_tx_err },
793 	{ true, "TDPES2", dpp_tx_err },
794 	{ true, "TDPES3", dpp_tx_err },
795 	{ true, "TDPES4", dpp_tx_err },
796 	{ true, "TDPES5", dpp_tx_err },
797 	{ true, "TDPES6", dpp_tx_err },
798 	{ true, "TDPES7", dpp_tx_err },
799 	{ true, "TDPES8", dpp_tx_err },
800 	{ true, "TDPES9", dpp_tx_err },
801 	{ true, "TDPES10", dpp_tx_err },
802 	{ true, "TDPES11", dpp_tx_err },
803 	{ true, "TDPES12", dpp_tx_err },
804 	{ true, "TDPES13", dpp_tx_err },
805 	{ true, "TDPES14", dpp_tx_err },
806 	{ true, "TDPES15", dpp_tx_err },
807 	{ true, "RDPES0", dpp_rx_err },
808 	{ true, "RDPES1", dpp_rx_err },
809 	{ true, "RDPES2", dpp_rx_err },
810 	{ true, "RDPES3", dpp_rx_err },
811 	{ true, "RDPES4", dpp_rx_err },
812 	{ true, "RDPES5", dpp_rx_err },
813 	{ true, "RDPES6", dpp_rx_err },
814 	{ true, "RDPES7", dpp_rx_err },
815 	{ true, "RDPES8", dpp_rx_err },
816 	{ true, "RDPES9", dpp_rx_err },
817 	{ true, "RDPES10", dpp_rx_err },
818 	{ true, "RDPES11", dpp_rx_err },
819 	{ true, "RDPES12", dpp_rx_err },
820 	{ true, "RDPES13", dpp_rx_err },
821 	{ true, "RDPES14", dpp_rx_err },
822 	{ true, "RDPES15", dpp_rx_err },
823 };
824 
dwxgmac3_handle_dma_err(struct net_device * ndev,void __iomem * ioaddr,bool correctable,struct stmmac_safety_stats * stats)825 static void dwxgmac3_handle_dma_err(struct net_device *ndev,
826 				    void __iomem *ioaddr, bool correctable,
827 				    struct stmmac_safety_stats *stats)
828 {
829 	u32 value;
830 
831 	value = readl(ioaddr + XGMAC_DMA_ECC_INT_STATUS);
832 	writel(value, ioaddr + XGMAC_DMA_ECC_INT_STATUS);
833 
834 	dwxgmac3_log_error(ndev, value, correctable, "DMA",
835 			   dwxgmac3_dma_errors, STAT_OFF(dma_errors), stats);
836 
837 	value = readl(ioaddr + XGMAC_DMA_DPP_INT_STATUS);
838 	writel(value, ioaddr + XGMAC_DMA_DPP_INT_STATUS);
839 
840 	dwxgmac3_log_error(ndev, value, false, "DMA_DPP",
841 			   dwxgmac3_dma_dpp_errors,
842 			   STAT_OFF(dma_dpp_errors), stats);
843 }
844 
845 static int
dwxgmac3_safety_feat_config(void __iomem * ioaddr,unsigned int asp,struct stmmac_safety_feature_cfg * safety_cfg)846 dwxgmac3_safety_feat_config(void __iomem *ioaddr, unsigned int asp,
847 			    struct stmmac_safety_feature_cfg *safety_cfg)
848 {
849 	u32 value;
850 
851 	if (!asp)
852 		return -EINVAL;
853 
854 	/* 1. Enable Safety Features */
855 	writel(0x0, ioaddr + XGMAC_MTL_ECC_CONTROL);
856 
857 	/* 2. Enable MTL Safety Interrupts */
858 	value = readl(ioaddr + XGMAC_MTL_ECC_INT_ENABLE);
859 	value |= XGMAC_RPCEIE; /* RX Parser Memory Correctable Error */
860 	value |= XGMAC_ECEIE; /* EST Memory Correctable Error */
861 	value |= XGMAC_RXCEIE; /* RX Memory Correctable Error */
862 	value |= XGMAC_TXCEIE; /* TX Memory Correctable Error */
863 	writel(value, ioaddr + XGMAC_MTL_ECC_INT_ENABLE);
864 
865 	/* 3. Enable DMA Safety Interrupts */
866 	value = readl(ioaddr + XGMAC_DMA_ECC_INT_ENABLE);
867 	value |= XGMAC_DCEIE; /* Descriptor Cache Memory Correctable Error */
868 	value |= XGMAC_TCEIE; /* TSO Memory Correctable Error */
869 	writel(value, ioaddr + XGMAC_DMA_ECC_INT_ENABLE);
870 
871 	/* 0x2: Without ECC or Parity Ports on External Application Interface
872 	 * 0x4: Only ECC Protection for External Memory feature is selected
873 	 */
874 	if (asp == 0x2 || asp == 0x4)
875 		return 0;
876 
877 	/* 4. Enable Parity and Timeout for FSM */
878 	value = readl(ioaddr + XGMAC_MAC_FSM_CONTROL);
879 	value |= XGMAC_PRTYEN; /* FSM Parity Feature */
880 	value |= XGMAC_TMOUTEN; /* FSM Timeout Feature */
881 	writel(value, ioaddr + XGMAC_MAC_FSM_CONTROL);
882 
883 	/* 5. Enable Data Path Parity Protection */
884 	value = readl(ioaddr + XGMAC_MTL_DPP_CONTROL);
885 	/* already enabled by default, explicit enable it again */
886 	value &= ~XGMAC_DPP_DISABLE;
887 	writel(value, ioaddr + XGMAC_MTL_DPP_CONTROL);
888 
889 	return 0;
890 }
891 
dwxgmac3_safety_feat_irq_status(struct net_device * ndev,void __iomem * ioaddr,unsigned int asp,struct stmmac_safety_stats * stats)892 static int dwxgmac3_safety_feat_irq_status(struct net_device *ndev,
893 					   void __iomem *ioaddr,
894 					   unsigned int asp,
895 					   struct stmmac_safety_stats *stats)
896 {
897 	bool err, corr;
898 	u32 mtl, dma;
899 	int ret = 0;
900 
901 	if (!asp)
902 		return -EINVAL;
903 
904 	mtl = readl(ioaddr + XGMAC_MTL_SAFETY_INT_STATUS);
905 	dma = readl(ioaddr + XGMAC_DMA_SAFETY_INT_STATUS);
906 
907 	err = (mtl & XGMAC_MCSIS) || (dma & XGMAC_MCSIS);
908 	corr = false;
909 	if (err) {
910 		dwxgmac3_handle_mac_err(ndev, ioaddr, corr, stats);
911 		ret |= !corr;
912 	}
913 
914 	err = (mtl & (XGMAC_MEUIS | XGMAC_MECIS)) ||
915 	      (dma & (XGMAC_MSUIS | XGMAC_MSCIS));
916 	corr = (mtl & XGMAC_MECIS) || (dma & XGMAC_MSCIS);
917 	if (err) {
918 		dwxgmac3_handle_mtl_err(ndev, ioaddr, corr, stats);
919 		ret |= !corr;
920 	}
921 
922 	/* DMA_DPP_Interrupt_Status is indicated by MCSIS bit in
923 	 * DMA_Safety_Interrupt_Status, so we handle DMA Data Path
924 	 * Parity Errors here
925 	 */
926 	err = dma & (XGMAC_DEUIS | XGMAC_DECIS | XGMAC_MCSIS);
927 	corr = dma & XGMAC_DECIS;
928 	if (err) {
929 		dwxgmac3_handle_dma_err(ndev, ioaddr, corr, stats);
930 		ret |= !corr;
931 	}
932 
933 	return ret;
934 }
935 
936 static const struct dwxgmac3_error {
937 	const struct dwxgmac3_error_desc *desc;
938 } dwxgmac3_all_errors[] = {
939 	{ dwxgmac3_mac_errors },
940 	{ dwxgmac3_mtl_errors },
941 	{ dwxgmac3_dma_errors },
942 	{ dwxgmac3_dma_dpp_errors },
943 };
944 
dwxgmac3_safety_feat_dump(struct stmmac_safety_stats * stats,int index,unsigned long * count,const char ** desc)945 static int dwxgmac3_safety_feat_dump(struct stmmac_safety_stats *stats,
946 				     int index, unsigned long *count,
947 				     const char **desc)
948 {
949 	int module = index / 32, offset = index % 32;
950 	unsigned long *ptr = (unsigned long *)stats;
951 
952 	if (module >= ARRAY_SIZE(dwxgmac3_all_errors))
953 		return -EINVAL;
954 	if (!dwxgmac3_all_errors[module].desc[offset].valid)
955 		return -EINVAL;
956 	if (count)
957 		*count = *(ptr + index);
958 	if (desc)
959 		*desc = dwxgmac3_all_errors[module].desc[offset].desc;
960 	return 0;
961 }
962 
dwxgmac3_rxp_disable(void __iomem * ioaddr)963 static int dwxgmac3_rxp_disable(void __iomem *ioaddr)
964 {
965 	u32 val = readl(ioaddr + XGMAC_MTL_OPMODE);
966 
967 	val &= ~XGMAC_FRPE;
968 	writel(val, ioaddr + XGMAC_MTL_OPMODE);
969 
970 	return 0;
971 }
972 
dwxgmac3_rxp_enable(void __iomem * ioaddr)973 static void dwxgmac3_rxp_enable(void __iomem *ioaddr)
974 {
975 	u32 val;
976 
977 	val = readl(ioaddr + XGMAC_MTL_OPMODE);
978 	val |= XGMAC_FRPE;
979 	writel(val, ioaddr + XGMAC_MTL_OPMODE);
980 }
981 
dwxgmac3_rxp_update_single_entry(void __iomem * ioaddr,struct stmmac_tc_entry * entry,int pos)982 static int dwxgmac3_rxp_update_single_entry(void __iomem *ioaddr,
983 					    struct stmmac_tc_entry *entry,
984 					    int pos)
985 {
986 	int ret, i;
987 
988 	for (i = 0; i < (sizeof(entry->val) / sizeof(u32)); i++) {
989 		int real_pos = pos * (sizeof(entry->val) / sizeof(u32)) + i;
990 		u32 val;
991 
992 		/* Wait for ready */
993 		ret = readl_poll_timeout(ioaddr + XGMAC_MTL_RXP_IACC_CTRL_ST,
994 					 val, !(val & XGMAC_STARTBUSY), 1, 10000);
995 		if (ret)
996 			return ret;
997 
998 		/* Write data */
999 		val = *((u32 *)&entry->val + i);
1000 		writel(val, ioaddr + XGMAC_MTL_RXP_IACC_DATA);
1001 
1002 		/* Write pos */
1003 		val = real_pos & XGMAC_ADDR;
1004 		writel(val, ioaddr + XGMAC_MTL_RXP_IACC_CTRL_ST);
1005 
1006 		/* Write OP */
1007 		val |= XGMAC_WRRDN;
1008 		writel(val, ioaddr + XGMAC_MTL_RXP_IACC_CTRL_ST);
1009 
1010 		/* Start Write */
1011 		val |= XGMAC_STARTBUSY;
1012 		writel(val, ioaddr + XGMAC_MTL_RXP_IACC_CTRL_ST);
1013 
1014 		/* Wait for done */
1015 		ret = readl_poll_timeout(ioaddr + XGMAC_MTL_RXP_IACC_CTRL_ST,
1016 					 val, !(val & XGMAC_STARTBUSY), 1, 10000);
1017 		if (ret)
1018 			return ret;
1019 	}
1020 
1021 	return 0;
1022 }
1023 
1024 static struct stmmac_tc_entry *
dwxgmac3_rxp_get_next_entry(struct stmmac_tc_entry * entries,unsigned int count,u32 curr_prio)1025 dwxgmac3_rxp_get_next_entry(struct stmmac_tc_entry *entries,
1026 			    unsigned int count, u32 curr_prio)
1027 {
1028 	struct stmmac_tc_entry *entry;
1029 	u32 min_prio = ~0x0;
1030 	int i, min_prio_idx;
1031 	bool found = false;
1032 
1033 	for (i = count - 1; i >= 0; i--) {
1034 		entry = &entries[i];
1035 
1036 		/* Do not update unused entries */
1037 		if (!entry->in_use)
1038 			continue;
1039 		/* Do not update already updated entries (i.e. fragments) */
1040 		if (entry->in_hw)
1041 			continue;
1042 		/* Let last entry be updated last */
1043 		if (entry->is_last)
1044 			continue;
1045 		/* Do not return fragments */
1046 		if (entry->is_frag)
1047 			continue;
1048 		/* Check if we already checked this prio */
1049 		if (entry->prio < curr_prio)
1050 			continue;
1051 		/* Check if this is the minimum prio */
1052 		if (entry->prio < min_prio) {
1053 			min_prio = entry->prio;
1054 			min_prio_idx = i;
1055 			found = true;
1056 		}
1057 	}
1058 
1059 	if (found)
1060 		return &entries[min_prio_idx];
1061 	return NULL;
1062 }
1063 
dwxgmac3_rxp_config(void __iomem * ioaddr,struct stmmac_tc_entry * entries,unsigned int count)1064 static int dwxgmac3_rxp_config(void __iomem *ioaddr,
1065 			       struct stmmac_tc_entry *entries,
1066 			       unsigned int count)
1067 {
1068 	struct stmmac_tc_entry *entry, *frag;
1069 	int i, ret, nve = 0;
1070 	u32 curr_prio = 0;
1071 	u32 old_val, val;
1072 
1073 	/* Force disable RX */
1074 	old_val = readl(ioaddr + XGMAC_RX_CONFIG);
1075 	val = old_val & ~XGMAC_CONFIG_RE;
1076 	writel(val, ioaddr + XGMAC_RX_CONFIG);
1077 
1078 	/* Disable RX Parser */
1079 	ret = dwxgmac3_rxp_disable(ioaddr);
1080 	if (ret)
1081 		goto re_enable;
1082 
1083 	/* Set all entries as NOT in HW */
1084 	for (i = 0; i < count; i++) {
1085 		entry = &entries[i];
1086 		entry->in_hw = false;
1087 	}
1088 
1089 	/* Update entries by reverse order */
1090 	while (1) {
1091 		entry = dwxgmac3_rxp_get_next_entry(entries, count, curr_prio);
1092 		if (!entry)
1093 			break;
1094 
1095 		curr_prio = entry->prio;
1096 		frag = entry->frag_ptr;
1097 
1098 		/* Set special fragment requirements */
1099 		if (frag) {
1100 			entry->val.af = 0;
1101 			entry->val.rf = 0;
1102 			entry->val.nc = 1;
1103 			entry->val.ok_index = nve + 2;
1104 		}
1105 
1106 		ret = dwxgmac3_rxp_update_single_entry(ioaddr, entry, nve);
1107 		if (ret)
1108 			goto re_enable;
1109 
1110 		entry->table_pos = nve++;
1111 		entry->in_hw = true;
1112 
1113 		if (frag && !frag->in_hw) {
1114 			ret = dwxgmac3_rxp_update_single_entry(ioaddr, frag, nve);
1115 			if (ret)
1116 				goto re_enable;
1117 			frag->table_pos = nve++;
1118 			frag->in_hw = true;
1119 		}
1120 	}
1121 
1122 	if (!nve)
1123 		goto re_enable;
1124 
1125 	/* Update all pass entry */
1126 	for (i = 0; i < count; i++) {
1127 		entry = &entries[i];
1128 		if (!entry->is_last)
1129 			continue;
1130 
1131 		ret = dwxgmac3_rxp_update_single_entry(ioaddr, entry, nve);
1132 		if (ret)
1133 			goto re_enable;
1134 
1135 		entry->table_pos = nve++;
1136 	}
1137 
1138 	/* Assume n. of parsable entries == n. of valid entries */
1139 	val = (nve << 16) & XGMAC_NPE;
1140 	val |= nve & XGMAC_NVE;
1141 	writel(val, ioaddr + XGMAC_MTL_RXP_CONTROL_STATUS);
1142 
1143 	/* Enable RX Parser */
1144 	dwxgmac3_rxp_enable(ioaddr);
1145 
1146 re_enable:
1147 	/* Re-enable RX */
1148 	writel(old_val, ioaddr + XGMAC_RX_CONFIG);
1149 	return ret;
1150 }
1151 
dwxgmac2_get_mac_tx_timestamp(struct mac_device_info * hw,u64 * ts)1152 static int dwxgmac2_get_mac_tx_timestamp(struct mac_device_info *hw, u64 *ts)
1153 {
1154 	void __iomem *ioaddr = hw->pcsr;
1155 	u32 value;
1156 
1157 	if (readl_poll_timeout_atomic(ioaddr + XGMAC_TIMESTAMP_STATUS,
1158 				      value, value & XGMAC_TXTSC, 100, 10000))
1159 		return -EBUSY;
1160 
1161 	*ts = readl(ioaddr + XGMAC_TXTIMESTAMP_NSEC) & XGMAC_TXTSSTSLO;
1162 	*ts += readl(ioaddr + XGMAC_TXTIMESTAMP_SEC) * 1000000000ULL;
1163 	return 0;
1164 }
1165 
dwxgmac2_flex_pps_config(void __iomem * ioaddr,int index,struct stmmac_pps_cfg * cfg,bool enable,u32 sub_second_inc,u32 systime_flags)1166 static int dwxgmac2_flex_pps_config(void __iomem *ioaddr, int index,
1167 				    struct stmmac_pps_cfg *cfg, bool enable,
1168 				    u32 sub_second_inc, u32 systime_flags)
1169 {
1170 	u32 tnsec = readl(ioaddr + XGMAC_PPSx_TARGET_TIME_NSEC(index));
1171 	u32 val = readl(ioaddr + XGMAC_PPS_CONTROL);
1172 	u64 period;
1173 
1174 	if (!cfg->available)
1175 		return -EINVAL;
1176 	if (tnsec & XGMAC_TRGTBUSY0)
1177 		return -EBUSY;
1178 	if (!sub_second_inc || !systime_flags)
1179 		return -EINVAL;
1180 
1181 	val &= ~XGMAC_PPSx_MASK(index);
1182 
1183 	if (!enable) {
1184 		val |= XGMAC_PPSCMDx(index, XGMAC_PPSCMD_STOP);
1185 		writel(val, ioaddr + XGMAC_PPS_CONTROL);
1186 		return 0;
1187 	}
1188 
1189 	val |= XGMAC_PPSCMDx(index, XGMAC_PPSCMD_START);
1190 	val |= XGMAC_TRGTMODSELx(index, XGMAC_PPSCMD_START);
1191 
1192 	/* XGMAC Core has 4 PPS outputs at most.
1193 	 *
1194 	 * Prior XGMAC Core 3.20, Fixed mode or Flexible mode are selectable for
1195 	 * PPS0 only via PPSEN0. PPS{1,2,3} are in Flexible mode by default,
1196 	 * and can not be switched to Fixed mode, since PPSEN{1,2,3} are
1197 	 * read-only reserved to 0.
1198 	 * But we always set PPSEN{1,2,3} do not make things worse ;-)
1199 	 *
1200 	 * From XGMAC Core 3.20 and later, PPSEN{0,1,2,3} are writable and must
1201 	 * be set, or the PPS outputs stay in Fixed PPS mode by default.
1202 	 */
1203 	val |= XGMAC_PPSENx(index);
1204 
1205 	writel(cfg->start.tv_sec, ioaddr + XGMAC_PPSx_TARGET_TIME_SEC(index));
1206 
1207 	if (!(systime_flags & PTP_TCR_TSCTRLSSR))
1208 		cfg->start.tv_nsec = (cfg->start.tv_nsec * 1000) / 465;
1209 	writel(cfg->start.tv_nsec, ioaddr + XGMAC_PPSx_TARGET_TIME_NSEC(index));
1210 
1211 	period = cfg->period.tv_sec * 1000000000;
1212 	period += cfg->period.tv_nsec;
1213 
1214 	do_div(period, sub_second_inc);
1215 
1216 	if (period <= 1)
1217 		return -EINVAL;
1218 
1219 	writel(period - 1, ioaddr + XGMAC_PPSx_INTERVAL(index));
1220 
1221 	period >>= 1;
1222 	if (period <= 1)
1223 		return -EINVAL;
1224 
1225 	writel(period - 1, ioaddr + XGMAC_PPSx_WIDTH(index));
1226 
1227 	/* Finally, activate it */
1228 	writel(val, ioaddr + XGMAC_PPS_CONTROL);
1229 	return 0;
1230 }
1231 
dwxgmac2_sarc_configure(void __iomem * ioaddr,int val)1232 static void dwxgmac2_sarc_configure(void __iomem *ioaddr, int val)
1233 {
1234 	u32 value = readl(ioaddr + XGMAC_TX_CONFIG);
1235 
1236 	value &= ~XGMAC_CONFIG_SARC;
1237 	value |= val << XGMAC_CONFIG_SARC_SHIFT;
1238 
1239 	writel(value, ioaddr + XGMAC_TX_CONFIG);
1240 }
1241 
dwxgmac2_filter_wait(struct mac_device_info * hw)1242 static int dwxgmac2_filter_wait(struct mac_device_info *hw)
1243 {
1244 	void __iomem *ioaddr = hw->pcsr;
1245 	u32 value;
1246 
1247 	if (readl_poll_timeout(ioaddr + XGMAC_L3L4_ADDR_CTRL, value,
1248 			       !(value & XGMAC_XB), 100, 10000))
1249 		return -EBUSY;
1250 	return 0;
1251 }
1252 
dwxgmac2_filter_read(struct mac_device_info * hw,u32 filter_no,u8 reg,u32 * data)1253 static int dwxgmac2_filter_read(struct mac_device_info *hw, u32 filter_no,
1254 				u8 reg, u32 *data)
1255 {
1256 	void __iomem *ioaddr = hw->pcsr;
1257 	u32 value;
1258 	int ret;
1259 
1260 	ret = dwxgmac2_filter_wait(hw);
1261 	if (ret)
1262 		return ret;
1263 
1264 	value = ((filter_no << XGMAC_IDDR_FNUM) | reg) << XGMAC_IDDR_SHIFT;
1265 	value |= XGMAC_TT | XGMAC_XB;
1266 	writel(value, ioaddr + XGMAC_L3L4_ADDR_CTRL);
1267 
1268 	ret = dwxgmac2_filter_wait(hw);
1269 	if (ret)
1270 		return ret;
1271 
1272 	*data = readl(ioaddr + XGMAC_L3L4_DATA);
1273 	return 0;
1274 }
1275 
dwxgmac2_filter_write(struct mac_device_info * hw,u32 filter_no,u8 reg,u32 data)1276 static int dwxgmac2_filter_write(struct mac_device_info *hw, u32 filter_no,
1277 				 u8 reg, u32 data)
1278 {
1279 	void __iomem *ioaddr = hw->pcsr;
1280 	u32 value;
1281 	int ret;
1282 
1283 	ret = dwxgmac2_filter_wait(hw);
1284 	if (ret)
1285 		return ret;
1286 
1287 	writel(data, ioaddr + XGMAC_L3L4_DATA);
1288 
1289 	value = ((filter_no << XGMAC_IDDR_FNUM) | reg) << XGMAC_IDDR_SHIFT;
1290 	value |= XGMAC_XB;
1291 	writel(value, ioaddr + XGMAC_L3L4_ADDR_CTRL);
1292 
1293 	return dwxgmac2_filter_wait(hw);
1294 }
1295 
dwxgmac2_config_l3_filter(struct mac_device_info * hw,u32 filter_no,bool en,bool ipv6,bool sa,bool inv,u32 match)1296 static int dwxgmac2_config_l3_filter(struct mac_device_info *hw, u32 filter_no,
1297 				     bool en, bool ipv6, bool sa, bool inv,
1298 				     u32 match)
1299 {
1300 	void __iomem *ioaddr = hw->pcsr;
1301 	u32 value;
1302 	int ret;
1303 
1304 	value = readl(ioaddr + XGMAC_PACKET_FILTER);
1305 	value |= XGMAC_FILTER_IPFE;
1306 	writel(value, ioaddr + XGMAC_PACKET_FILTER);
1307 
1308 	ret = dwxgmac2_filter_read(hw, filter_no, XGMAC_L3L4_CTRL, &value);
1309 	if (ret)
1310 		return ret;
1311 
1312 	/* For IPv6 not both SA/DA filters can be active */
1313 	if (ipv6) {
1314 		value |= XGMAC_L3PEN0;
1315 		value &= ~(XGMAC_L3SAM0 | XGMAC_L3SAIM0);
1316 		value &= ~(XGMAC_L3DAM0 | XGMAC_L3DAIM0);
1317 		if (sa) {
1318 			value |= XGMAC_L3SAM0;
1319 			if (inv)
1320 				value |= XGMAC_L3SAIM0;
1321 		} else {
1322 			value |= XGMAC_L3DAM0;
1323 			if (inv)
1324 				value |= XGMAC_L3DAIM0;
1325 		}
1326 	} else {
1327 		value &= ~XGMAC_L3PEN0;
1328 		if (sa) {
1329 			value |= XGMAC_L3SAM0;
1330 			if (inv)
1331 				value |= XGMAC_L3SAIM0;
1332 		} else {
1333 			value |= XGMAC_L3DAM0;
1334 			if (inv)
1335 				value |= XGMAC_L3DAIM0;
1336 		}
1337 	}
1338 
1339 	ret = dwxgmac2_filter_write(hw, filter_no, XGMAC_L3L4_CTRL, value);
1340 	if (ret)
1341 		return ret;
1342 
1343 	if (sa) {
1344 		ret = dwxgmac2_filter_write(hw, filter_no, XGMAC_L3_ADDR0, match);
1345 		if (ret)
1346 			return ret;
1347 	} else {
1348 		ret = dwxgmac2_filter_write(hw, filter_no, XGMAC_L3_ADDR1, match);
1349 		if (ret)
1350 			return ret;
1351 	}
1352 
1353 	if (!en)
1354 		return dwxgmac2_filter_write(hw, filter_no, XGMAC_L3L4_CTRL, 0);
1355 
1356 	return 0;
1357 }
1358 
dwxgmac2_config_l4_filter(struct mac_device_info * hw,u32 filter_no,bool en,bool udp,bool sa,bool inv,u32 match)1359 static int dwxgmac2_config_l4_filter(struct mac_device_info *hw, u32 filter_no,
1360 				     bool en, bool udp, bool sa, bool inv,
1361 				     u32 match)
1362 {
1363 	void __iomem *ioaddr = hw->pcsr;
1364 	u32 value;
1365 	int ret;
1366 
1367 	value = readl(ioaddr + XGMAC_PACKET_FILTER);
1368 	value |= XGMAC_FILTER_IPFE;
1369 	writel(value, ioaddr + XGMAC_PACKET_FILTER);
1370 
1371 	ret = dwxgmac2_filter_read(hw, filter_no, XGMAC_L3L4_CTRL, &value);
1372 	if (ret)
1373 		return ret;
1374 
1375 	if (udp) {
1376 		value |= XGMAC_L4PEN0;
1377 	} else {
1378 		value &= ~XGMAC_L4PEN0;
1379 	}
1380 
1381 	value &= ~(XGMAC_L4SPM0 | XGMAC_L4SPIM0);
1382 	value &= ~(XGMAC_L4DPM0 | XGMAC_L4DPIM0);
1383 	if (sa) {
1384 		value |= XGMAC_L4SPM0;
1385 		if (inv)
1386 			value |= XGMAC_L4SPIM0;
1387 	} else {
1388 		value |= XGMAC_L4DPM0;
1389 		if (inv)
1390 			value |= XGMAC_L4DPIM0;
1391 	}
1392 
1393 	ret = dwxgmac2_filter_write(hw, filter_no, XGMAC_L3L4_CTRL, value);
1394 	if (ret)
1395 		return ret;
1396 
1397 	if (sa) {
1398 		value = match & XGMAC_L4SP0;
1399 
1400 		ret = dwxgmac2_filter_write(hw, filter_no, XGMAC_L4_ADDR, value);
1401 		if (ret)
1402 			return ret;
1403 	} else {
1404 		value = (match << XGMAC_L4DP0_SHIFT) & XGMAC_L4DP0;
1405 
1406 		ret = dwxgmac2_filter_write(hw, filter_no, XGMAC_L4_ADDR, value);
1407 		if (ret)
1408 			return ret;
1409 	}
1410 
1411 	if (!en)
1412 		return dwxgmac2_filter_write(hw, filter_no, XGMAC_L3L4_CTRL, 0);
1413 
1414 	return 0;
1415 }
1416 
dwxgmac2_set_arp_offload(struct mac_device_info * hw,bool en,u32 addr)1417 static void dwxgmac2_set_arp_offload(struct mac_device_info *hw, bool en,
1418 				     u32 addr)
1419 {
1420 	void __iomem *ioaddr = hw->pcsr;
1421 	u32 value;
1422 
1423 	writel(addr, ioaddr + XGMAC_ARP_ADDR);
1424 
1425 	value = readl(ioaddr + XGMAC_RX_CONFIG);
1426 	if (en)
1427 		value |= XGMAC_CONFIG_ARPEN;
1428 	else
1429 		value &= ~XGMAC_CONFIG_ARPEN;
1430 	writel(value, ioaddr + XGMAC_RX_CONFIG);
1431 }
1432 
1433 const struct stmmac_ops dwxgmac210_ops = {
1434 	.core_init = dwxgmac2_core_init,
1435 	.update_caps = dwxgmac2_update_caps,
1436 	.set_mac = dwxgmac2_set_mac,
1437 	.rx_ipc = dwxgmac2_rx_ipc,
1438 	.rx_queue_enable = dwxgmac2_rx_queue_enable,
1439 	.rx_queue_prio = dwxgmac2_rx_queue_prio,
1440 	.tx_queue_prio = dwxgmac2_tx_queue_prio,
1441 	.rx_queue_routing = dwxgmac2_rx_queue_routing,
1442 	.prog_mtl_rx_algorithms = dwxgmac2_prog_mtl_rx_algorithms,
1443 	.prog_mtl_tx_algorithms = dwxgmac2_prog_mtl_tx_algorithms,
1444 	.set_mtl_tx_queue_weight = dwxgmac2_set_mtl_tx_queue_weight,
1445 	.map_mtl_to_dma = dwxgmac2_map_mtl_to_dma,
1446 	.config_cbs = dwxgmac2_config_cbs,
1447 	.dump_regs = dwxgmac2_dump_regs,
1448 	.host_irq_status = dwxgmac2_host_irq_status,
1449 	.host_mtl_irq_status = dwxgmac2_host_mtl_irq_status,
1450 	.flow_ctrl = dwxgmac2_flow_ctrl,
1451 	.pmt = dwxgmac2_pmt,
1452 	.set_umac_addr = dwxgmac2_set_umac_addr,
1453 	.get_umac_addr = dwxgmac2_get_umac_addr,
1454 	.set_lpi_mode = dwxgmac2_set_lpi_mode,
1455 	.set_eee_timer = dwxgmac2_set_eee_timer,
1456 	.set_eee_pls = dwxgmac2_set_eee_pls,
1457 	.debug = NULL,
1458 	.set_filter = dwxgmac2_set_filter,
1459 	.safety_feat_config = dwxgmac3_safety_feat_config,
1460 	.safety_feat_irq_status = dwxgmac3_safety_feat_irq_status,
1461 	.safety_feat_dump = dwxgmac3_safety_feat_dump,
1462 	.set_mac_loopback = dwxgmac2_set_mac_loopback,
1463 	.rss_configure = dwxgmac2_rss_configure,
1464 	.rxp_config = dwxgmac3_rxp_config,
1465 	.get_mac_tx_timestamp = dwxgmac2_get_mac_tx_timestamp,
1466 	.flex_pps_config = dwxgmac2_flex_pps_config,
1467 	.sarc_configure = dwxgmac2_sarc_configure,
1468 	.config_l3_filter = dwxgmac2_config_l3_filter,
1469 	.config_l4_filter = dwxgmac2_config_l4_filter,
1470 	.set_arp_offload = dwxgmac2_set_arp_offload,
1471 	.fpe_map_preemption_class = dwxgmac3_fpe_map_preemption_class,
1472 };
1473 
dwxlgmac2_rx_queue_enable(struct mac_device_info * hw,u8 mode,u32 queue)1474 static void dwxlgmac2_rx_queue_enable(struct mac_device_info *hw, u8 mode,
1475 				      u32 queue)
1476 {
1477 	void __iomem *ioaddr = hw->pcsr;
1478 	u32 value;
1479 
1480 	value = readl(ioaddr + XLGMAC_RXQ_ENABLE_CTRL0) & ~XGMAC_RXQEN(queue);
1481 	if (mode == MTL_QUEUE_AVB)
1482 		value |= 0x1 << XGMAC_RXQEN_SHIFT(queue);
1483 	else if (mode == MTL_QUEUE_DCB)
1484 		value |= 0x2 << XGMAC_RXQEN_SHIFT(queue);
1485 	writel(value, ioaddr + XLGMAC_RXQ_ENABLE_CTRL0);
1486 }
1487 
1488 const struct stmmac_ops dwxlgmac2_ops = {
1489 	.core_init = dwxgmac2_core_init,
1490 	.set_mac = dwxgmac2_set_mac,
1491 	.rx_ipc = dwxgmac2_rx_ipc,
1492 	.rx_queue_enable = dwxlgmac2_rx_queue_enable,
1493 	.rx_queue_prio = dwxgmac2_rx_queue_prio,
1494 	.tx_queue_prio = dwxgmac2_tx_queue_prio,
1495 	.rx_queue_routing = dwxgmac2_rx_queue_routing,
1496 	.prog_mtl_rx_algorithms = dwxgmac2_prog_mtl_rx_algorithms,
1497 	.prog_mtl_tx_algorithms = dwxgmac2_prog_mtl_tx_algorithms,
1498 	.set_mtl_tx_queue_weight = dwxgmac2_set_mtl_tx_queue_weight,
1499 	.map_mtl_to_dma = dwxgmac2_map_mtl_to_dma,
1500 	.config_cbs = dwxgmac2_config_cbs,
1501 	.dump_regs = dwxgmac2_dump_regs,
1502 	.host_irq_status = dwxgmac2_host_irq_status,
1503 	.host_mtl_irq_status = dwxgmac2_host_mtl_irq_status,
1504 	.flow_ctrl = dwxgmac2_flow_ctrl,
1505 	.pmt = dwxgmac2_pmt,
1506 	.set_umac_addr = dwxgmac2_set_umac_addr,
1507 	.get_umac_addr = dwxgmac2_get_umac_addr,
1508 	.set_lpi_mode = dwxgmac2_set_lpi_mode,
1509 	.set_eee_timer = dwxgmac2_set_eee_timer,
1510 	.set_eee_pls = dwxgmac2_set_eee_pls,
1511 	.debug = NULL,
1512 	.set_filter = dwxgmac2_set_filter,
1513 	.safety_feat_config = dwxgmac3_safety_feat_config,
1514 	.safety_feat_irq_status = dwxgmac3_safety_feat_irq_status,
1515 	.safety_feat_dump = dwxgmac3_safety_feat_dump,
1516 	.set_mac_loopback = dwxgmac2_set_mac_loopback,
1517 	.rss_configure = dwxgmac2_rss_configure,
1518 	.rxp_config = dwxgmac3_rxp_config,
1519 	.get_mac_tx_timestamp = dwxgmac2_get_mac_tx_timestamp,
1520 	.flex_pps_config = dwxgmac2_flex_pps_config,
1521 	.sarc_configure = dwxgmac2_sarc_configure,
1522 	.config_l3_filter = dwxgmac2_config_l3_filter,
1523 	.config_l4_filter = dwxgmac2_config_l4_filter,
1524 	.set_arp_offload = dwxgmac2_set_arp_offload,
1525 	.fpe_map_preemption_class = dwxgmac3_fpe_map_preemption_class,
1526 };
1527 
dwxgmac2_setup(struct stmmac_priv * priv)1528 int dwxgmac2_setup(struct stmmac_priv *priv)
1529 {
1530 	struct mac_device_info *mac = priv->hw;
1531 
1532 	dev_info(priv->device, "\tXGMAC2\n");
1533 
1534 	priv->dev->priv_flags |= IFF_UNICAST_FLT;
1535 	mac->pcsr = priv->ioaddr;
1536 	mac->multicast_filter_bins = priv->plat->multicast_filter_bins;
1537 	mac->unicast_filter_entries = priv->plat->unicast_filter_entries;
1538 	mac->mcast_bits_log2 = 0;
1539 
1540 	if (mac->multicast_filter_bins)
1541 		mac->mcast_bits_log2 = ilog2(mac->multicast_filter_bins);
1542 
1543 	mac->link.caps = MAC_ASYM_PAUSE | MAC_SYM_PAUSE |
1544 			 MAC_10 | MAC_100 | MAC_1000FD |
1545 			 MAC_2500FD | MAC_5000FD | MAC_10000FD;
1546 	mac->link.duplex = 0;
1547 	mac->link.speed10 = XGMAC_CONFIG_SS_10_MII;
1548 	mac->link.speed100 = XGMAC_CONFIG_SS_100_MII;
1549 	mac->link.speed1000 = XGMAC_CONFIG_SS_1000_GMII;
1550 	mac->link.speed2500 = XGMAC_CONFIG_SS_2500_GMII;
1551 	mac->link.xgmii.speed2500 = XGMAC_CONFIG_SS_2500;
1552 	mac->link.xgmii.speed5000 = XGMAC_CONFIG_SS_5000;
1553 	mac->link.xgmii.speed10000 = XGMAC_CONFIG_SS_10000;
1554 	mac->link.speed_mask = XGMAC_CONFIG_SS_MASK;
1555 
1556 	mac->mii.addr = XGMAC_MDIO_ADDR;
1557 	mac->mii.data = XGMAC_MDIO_DATA;
1558 	mac->mii.addr_shift = 16;
1559 	mac->mii.addr_mask = GENMASK(20, 16);
1560 	mac->mii.reg_shift = 0;
1561 	mac->mii.reg_mask = GENMASK(15, 0);
1562 	mac->mii.clk_csr_shift = 19;
1563 	mac->mii.clk_csr_mask = GENMASK(21, 19);
1564 	mac->num_vlan = stmmac_get_num_vlan(priv->ioaddr);
1565 
1566 	return 0;
1567 }
1568 
dwxlgmac2_setup(struct stmmac_priv * priv)1569 int dwxlgmac2_setup(struct stmmac_priv *priv)
1570 {
1571 	struct mac_device_info *mac = priv->hw;
1572 
1573 	dev_info(priv->device, "\tXLGMAC\n");
1574 
1575 	priv->dev->priv_flags |= IFF_UNICAST_FLT;
1576 	mac->pcsr = priv->ioaddr;
1577 	mac->multicast_filter_bins = priv->plat->multicast_filter_bins;
1578 	mac->unicast_filter_entries = priv->plat->unicast_filter_entries;
1579 	mac->mcast_bits_log2 = 0;
1580 
1581 	if (mac->multicast_filter_bins)
1582 		mac->mcast_bits_log2 = ilog2(mac->multicast_filter_bins);
1583 
1584 	mac->link.caps = MAC_ASYM_PAUSE | MAC_SYM_PAUSE |
1585 			 MAC_1000FD | MAC_2500FD | MAC_5000FD |
1586 			 MAC_10000FD | MAC_25000FD |
1587 			 MAC_40000FD | MAC_50000FD |
1588 			 MAC_100000FD;
1589 	mac->link.duplex = 0;
1590 	mac->link.speed1000 = XLGMAC_CONFIG_SS_1000;
1591 	mac->link.speed2500 = XLGMAC_CONFIG_SS_2500;
1592 	mac->link.xgmii.speed10000 = XLGMAC_CONFIG_SS_10G;
1593 	mac->link.xlgmii.speed25000 = XLGMAC_CONFIG_SS_25G;
1594 	mac->link.xlgmii.speed40000 = XLGMAC_CONFIG_SS_40G;
1595 	mac->link.xlgmii.speed50000 = XLGMAC_CONFIG_SS_50G;
1596 	mac->link.xlgmii.speed100000 = XLGMAC_CONFIG_SS_100G;
1597 	mac->link.speed_mask = XLGMAC_CONFIG_SS;
1598 
1599 	mac->mii.addr = XGMAC_MDIO_ADDR;
1600 	mac->mii.data = XGMAC_MDIO_DATA;
1601 	mac->mii.addr_shift = 16;
1602 	mac->mii.addr_mask = GENMASK(20, 16);
1603 	mac->mii.reg_shift = 0;
1604 	mac->mii.reg_mask = GENMASK(15, 0);
1605 	mac->mii.clk_csr_shift = 19;
1606 	mac->mii.clk_csr_mask = GENMASK(21, 19);
1607 
1608 	return 0;
1609 }
1610