1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * sni_ave.c - Socionext UniPhier AVE ethernet driver
4 * Copyright 2014 Panasonic Corporation
5 * Copyright 2015-2017 Socionext Inc.
6 */
7
8 #include <linux/bitops.h>
9 #include <linux/clk.h>
10 #include <linux/etherdevice.h>
11 #include <linux/interrupt.h>
12 #include <linux/io.h>
13 #include <linux/iopoll.h>
14 #include <linux/mfd/syscon.h>
15 #include <linux/mii.h>
16 #include <linux/module.h>
17 #include <linux/netdevice.h>
18 #include <linux/of.h>
19 #include <linux/of_net.h>
20 #include <linux/of_mdio.h>
21 #include <linux/phy.h>
22 #include <linux/platform_device.h>
23 #include <linux/regmap.h>
24 #include <linux/reset.h>
25 #include <linux/types.h>
26 #include <linux/u64_stats_sync.h>
27
28 /* General Register Group */
29 #define AVE_IDR 0x000 /* ID */
30 #define AVE_VR 0x004 /* Version */
31 #define AVE_GRR 0x008 /* Global Reset */
32 #define AVE_CFGR 0x00c /* Configuration */
33
34 /* Interrupt Register Group */
35 #define AVE_GIMR 0x100 /* Global Interrupt Mask */
36 #define AVE_GISR 0x104 /* Global Interrupt Status */
37
38 /* MAC Register Group */
39 #define AVE_TXCR 0x200 /* TX Setup */
40 #define AVE_RXCR 0x204 /* RX Setup */
41 #define AVE_RXMAC1R 0x208 /* MAC address (lower) */
42 #define AVE_RXMAC2R 0x20c /* MAC address (upper) */
43 #define AVE_MDIOCTR 0x214 /* MDIO Control */
44 #define AVE_MDIOAR 0x218 /* MDIO Address */
45 #define AVE_MDIOWDR 0x21c /* MDIO Data */
46 #define AVE_MDIOSR 0x220 /* MDIO Status */
47 #define AVE_MDIORDR 0x224 /* MDIO Rd Data */
48
49 /* Descriptor Control Register Group */
50 #define AVE_DESCC 0x300 /* Descriptor Control */
51 #define AVE_TXDC 0x304 /* TX Descriptor Configuration */
52 #define AVE_RXDC0 0x308 /* RX Descriptor Ring0 Configuration */
53 #define AVE_IIRQC 0x34c /* Interval IRQ Control */
54
55 /* Packet Filter Register Group */
56 #define AVE_PKTF_BASE 0x800 /* PF Base Address */
57 #define AVE_PFMBYTE_BASE 0xd00 /* PF Mask Byte Base Address */
58 #define AVE_PFMBIT_BASE 0xe00 /* PF Mask Bit Base Address */
59 #define AVE_PFSEL_BASE 0xf00 /* PF Selector Base Address */
60 #define AVE_PFEN 0xffc /* Packet Filter Enable */
61 #define AVE_PKTF(ent) (AVE_PKTF_BASE + (ent) * 0x40)
62 #define AVE_PFMBYTE(ent) (AVE_PFMBYTE_BASE + (ent) * 8)
63 #define AVE_PFMBIT(ent) (AVE_PFMBIT_BASE + (ent) * 4)
64 #define AVE_PFSEL(ent) (AVE_PFSEL_BASE + (ent) * 4)
65
66 /* 64bit descriptor memory */
67 #define AVE_DESC_SIZE_64 12 /* Descriptor Size */
68
69 #define AVE_TXDM_64 0x1000 /* Tx Descriptor Memory */
70 #define AVE_RXDM_64 0x1c00 /* Rx Descriptor Memory */
71
72 #define AVE_TXDM_SIZE_64 0x0ba0 /* Tx Descriptor Memory Size 3KB */
73 #define AVE_RXDM_SIZE_64 0x6000 /* Rx Descriptor Memory Size 24KB */
74
75 /* 32bit descriptor memory */
76 #define AVE_DESC_SIZE_32 8 /* Descriptor Size */
77
78 #define AVE_TXDM_32 0x1000 /* Tx Descriptor Memory */
79 #define AVE_RXDM_32 0x1800 /* Rx Descriptor Memory */
80
81 #define AVE_TXDM_SIZE_32 0x07c0 /* Tx Descriptor Memory Size 2KB */
82 #define AVE_RXDM_SIZE_32 0x4000 /* Rx Descriptor Memory Size 16KB */
83
84 /* RMII Bridge Register Group */
85 #define AVE_RSTCTRL 0x8028 /* Reset control */
86 #define AVE_RSTCTRL_RMIIRST BIT(16)
87 #define AVE_LINKSEL 0x8034 /* Link speed setting */
88 #define AVE_LINKSEL_100M BIT(0)
89
90 /* AVE_GRR */
91 #define AVE_GRR_RXFFR BIT(5) /* Reset RxFIFO */
92 #define AVE_GRR_PHYRST BIT(4) /* Reset external PHY */
93 #define AVE_GRR_GRST BIT(0) /* Reset all MAC */
94
95 /* AVE_CFGR */
96 #define AVE_CFGR_FLE BIT(31) /* Filter Function */
97 #define AVE_CFGR_CHE BIT(30) /* Checksum Function */
98 #define AVE_CFGR_MII BIT(27) /* Func mode (1:MII/RMII, 0:RGMII) */
99 #define AVE_CFGR_IPFCEN BIT(24) /* IP fragment sum Enable */
100
101 /* AVE_GISR (common with GIMR) */
102 #define AVE_GI_PHY BIT(24) /* PHY interrupt */
103 #define AVE_GI_TX BIT(16) /* Tx complete */
104 #define AVE_GI_RXERR BIT(8) /* Receive frame more than max size */
105 #define AVE_GI_RXOVF BIT(7) /* Overflow at the RxFIFO */
106 #define AVE_GI_RXDROP BIT(6) /* Drop packet */
107 #define AVE_GI_RXIINT BIT(5) /* Interval interrupt */
108
109 /* AVE_TXCR */
110 #define AVE_TXCR_FLOCTR BIT(18) /* Flow control */
111 #define AVE_TXCR_TXSPD_1G BIT(17)
112 #define AVE_TXCR_TXSPD_100 BIT(16)
113
114 /* AVE_RXCR */
115 #define AVE_RXCR_RXEN BIT(30) /* Rx enable */
116 #define AVE_RXCR_FDUPEN BIT(22) /* Interface mode */
117 #define AVE_RXCR_FLOCTR BIT(21) /* Flow control */
118 #define AVE_RXCR_AFEN BIT(19) /* MAC address filter */
119 #define AVE_RXCR_DRPEN BIT(18) /* Drop pause frame */
120 #define AVE_RXCR_MPSIZ_MASK GENMASK(10, 0)
121
122 /* AVE_MDIOCTR */
123 #define AVE_MDIOCTR_RREQ BIT(3) /* Read request */
124 #define AVE_MDIOCTR_WREQ BIT(2) /* Write request */
125
126 /* AVE_MDIOSR */
127 #define AVE_MDIOSR_STS BIT(0) /* access status */
128
129 /* AVE_DESCC */
130 #define AVE_DESCC_STATUS_MASK GENMASK(31, 16)
131 #define AVE_DESCC_RD0 BIT(8) /* Enable Rx descriptor Ring0 */
132 #define AVE_DESCC_RDSTP BIT(4) /* Pause Rx descriptor */
133 #define AVE_DESCC_TD BIT(0) /* Enable Tx descriptor */
134
135 /* AVE_TXDC */
136 #define AVE_TXDC_SIZE GENMASK(27, 16) /* Size of Tx descriptor */
137 #define AVE_TXDC_ADDR GENMASK(11, 0) /* Start address */
138 #define AVE_TXDC_ADDR_START 0
139
140 /* AVE_RXDC0 */
141 #define AVE_RXDC0_SIZE GENMASK(30, 16) /* Size of Rx descriptor */
142 #define AVE_RXDC0_ADDR GENMASK(14, 0) /* Start address */
143 #define AVE_RXDC0_ADDR_START 0
144
145 /* AVE_IIRQC */
146 #define AVE_IIRQC_EN0 BIT(27) /* Enable interval interrupt Ring0 */
147 #define AVE_IIRQC_BSCK GENMASK(15, 0) /* Interval count unit */
148
149 /* Command status for descriptor */
150 #define AVE_STS_OWN BIT(31) /* Descriptor ownership */
151 #define AVE_STS_INTR BIT(29) /* Request for interrupt */
152 #define AVE_STS_OK BIT(27) /* Normal transmit */
153 /* TX */
154 #define AVE_STS_NOCSUM BIT(28) /* No use HW checksum */
155 #define AVE_STS_1ST BIT(26) /* Head of buffer chain */
156 #define AVE_STS_LAST BIT(25) /* Tail of buffer chain */
157 #define AVE_STS_OWC BIT(21) /* Out of window,Late Collision */
158 #define AVE_STS_EC BIT(20) /* Excess collision occurred */
159 #define AVE_STS_PKTLEN_TX_MASK GENMASK(15, 0)
160 /* RX */
161 #define AVE_STS_CSSV BIT(21) /* Checksum check performed */
162 #define AVE_STS_CSER BIT(20) /* Checksum error detected */
163 #define AVE_STS_PKTLEN_RX_MASK GENMASK(10, 0)
164
165 /* Packet filter */
166 #define AVE_PFMBYTE_MASK0 (GENMASK(31, 8) | GENMASK(5, 0))
167 #define AVE_PFMBYTE_MASK1 GENMASK(25, 0)
168 #define AVE_PFMBIT_MASK GENMASK(15, 0)
169
170 #define AVE_PF_SIZE 17 /* Number of all packet filter */
171 #define AVE_PF_MULTICAST_SIZE 7 /* Number of multicast filter */
172
173 #define AVE_PFNUM_FILTER 0 /* No.0 */
174 #define AVE_PFNUM_UNICAST 1 /* No.1 */
175 #define AVE_PFNUM_BROADCAST 2 /* No.2 */
176 #define AVE_PFNUM_MULTICAST 11 /* No.11-17 */
177
178 /* NETIF Message control */
179 #define AVE_DEFAULT_MSG_ENABLE (NETIF_MSG_DRV | \
180 NETIF_MSG_PROBE | \
181 NETIF_MSG_LINK | \
182 NETIF_MSG_TIMER | \
183 NETIF_MSG_IFDOWN | \
184 NETIF_MSG_IFUP | \
185 NETIF_MSG_RX_ERR | \
186 NETIF_MSG_TX_ERR)
187
188 /* Parameter for descriptor */
189 #define AVE_NR_TXDESC 64 /* Tx descriptor */
190 #define AVE_NR_RXDESC 256 /* Rx descriptor */
191
192 #define AVE_DESC_OFS_CMDSTS 0
193 #define AVE_DESC_OFS_ADDRL 4
194 #define AVE_DESC_OFS_ADDRU 8
195
196 /* Parameter for ethernet frame */
197 #define AVE_MAX_ETHFRAME 1518
198 #define AVE_FRAME_HEADROOM 2
199
200 /* Parameter for interrupt */
201 #define AVE_INTM_COUNT 20
202 #define AVE_FORCE_TXINTCNT 1
203
204 /* SG */
205 #define SG_ETPINMODE 0x540
206 #define SG_ETPINMODE_EXTPHY BIT(1) /* for LD11 */
207 #define SG_ETPINMODE_RMII(ins) BIT(ins)
208
209 #define IS_DESC_64BIT(p) ((p)->data->is_desc_64bit)
210
211 #define AVE_MAX_CLKS 4
212 #define AVE_MAX_RSTS 2
213
214 enum desc_id {
215 AVE_DESCID_RX,
216 AVE_DESCID_TX,
217 };
218
219 enum desc_state {
220 AVE_DESC_RX_PERMIT,
221 AVE_DESC_RX_SUSPEND,
222 AVE_DESC_START,
223 AVE_DESC_STOP,
224 };
225
226 struct ave_desc {
227 struct sk_buff *skbs;
228 dma_addr_t skbs_dma;
229 size_t skbs_dmalen;
230 };
231
232 struct ave_desc_info {
233 u32 ndesc; /* number of descriptor */
234 u32 daddr; /* start address of descriptor */
235 u32 proc_idx; /* index of processing packet */
236 u32 done_idx; /* index of processed packet */
237 struct ave_desc *desc; /* skb info related descriptor */
238 };
239
240 struct ave_stats {
241 struct u64_stats_sync syncp;
242 u64 packets;
243 u64 bytes;
244 u64 errors;
245 u64 dropped;
246 u64 collisions;
247 u64 fifo_errors;
248 };
249
250 struct ave_private {
251 void __iomem *base;
252 int irq;
253 int phy_id;
254 unsigned int desc_size;
255 u32 msg_enable;
256 int nclks;
257 struct clk *clk[AVE_MAX_CLKS];
258 int nrsts;
259 struct reset_control *rst[AVE_MAX_RSTS];
260 phy_interface_t phy_mode;
261 struct phy_device *phydev;
262 struct mii_bus *mdio;
263 struct regmap *regmap;
264 unsigned int pinmode_mask;
265 unsigned int pinmode_val;
266 u32 wolopts;
267
268 /* stats */
269 struct ave_stats stats_rx;
270 struct ave_stats stats_tx;
271
272 /* NAPI support */
273 struct net_device *ndev;
274 struct napi_struct napi_rx;
275 struct napi_struct napi_tx;
276
277 /* descriptor */
278 struct ave_desc_info rx;
279 struct ave_desc_info tx;
280
281 /* flow control */
282 int pause_auto;
283 int pause_rx;
284 int pause_tx;
285
286 const struct ave_soc_data *data;
287 };
288
289 struct ave_soc_data {
290 bool is_desc_64bit;
291 const char *clock_names[AVE_MAX_CLKS];
292 const char *reset_names[AVE_MAX_RSTS];
293 int (*get_pinmode)(struct ave_private *priv,
294 phy_interface_t phy_mode, u32 arg);
295 };
296
ave_desc_read(struct net_device * ndev,enum desc_id id,int entry,int offset)297 static u32 ave_desc_read(struct net_device *ndev, enum desc_id id, int entry,
298 int offset)
299 {
300 struct ave_private *priv = netdev_priv(ndev);
301 u32 addr;
302
303 addr = ((id == AVE_DESCID_TX) ? priv->tx.daddr : priv->rx.daddr)
304 + entry * priv->desc_size + offset;
305
306 return readl(priv->base + addr);
307 }
308
ave_desc_read_cmdsts(struct net_device * ndev,enum desc_id id,int entry)309 static u32 ave_desc_read_cmdsts(struct net_device *ndev, enum desc_id id,
310 int entry)
311 {
312 return ave_desc_read(ndev, id, entry, AVE_DESC_OFS_CMDSTS);
313 }
314
ave_desc_write(struct net_device * ndev,enum desc_id id,int entry,int offset,u32 val)315 static void ave_desc_write(struct net_device *ndev, enum desc_id id,
316 int entry, int offset, u32 val)
317 {
318 struct ave_private *priv = netdev_priv(ndev);
319 u32 addr;
320
321 addr = ((id == AVE_DESCID_TX) ? priv->tx.daddr : priv->rx.daddr)
322 + entry * priv->desc_size + offset;
323
324 writel(val, priv->base + addr);
325 }
326
ave_desc_write_cmdsts(struct net_device * ndev,enum desc_id id,int entry,u32 val)327 static void ave_desc_write_cmdsts(struct net_device *ndev, enum desc_id id,
328 int entry, u32 val)
329 {
330 ave_desc_write(ndev, id, entry, AVE_DESC_OFS_CMDSTS, val);
331 }
332
ave_desc_write_addr(struct net_device * ndev,enum desc_id id,int entry,dma_addr_t paddr)333 static void ave_desc_write_addr(struct net_device *ndev, enum desc_id id,
334 int entry, dma_addr_t paddr)
335 {
336 struct ave_private *priv = netdev_priv(ndev);
337
338 ave_desc_write(ndev, id, entry, AVE_DESC_OFS_ADDRL,
339 lower_32_bits(paddr));
340 if (IS_DESC_64BIT(priv))
341 ave_desc_write(ndev, id,
342 entry, AVE_DESC_OFS_ADDRU,
343 upper_32_bits(paddr));
344 }
345
ave_irq_disable_all(struct net_device * ndev)346 static u32 ave_irq_disable_all(struct net_device *ndev)
347 {
348 struct ave_private *priv = netdev_priv(ndev);
349 u32 ret;
350
351 ret = readl(priv->base + AVE_GIMR);
352 writel(0, priv->base + AVE_GIMR);
353
354 return ret;
355 }
356
ave_irq_restore(struct net_device * ndev,u32 val)357 static void ave_irq_restore(struct net_device *ndev, u32 val)
358 {
359 struct ave_private *priv = netdev_priv(ndev);
360
361 writel(val, priv->base + AVE_GIMR);
362 }
363
ave_irq_enable(struct net_device * ndev,u32 bitflag)364 static void ave_irq_enable(struct net_device *ndev, u32 bitflag)
365 {
366 struct ave_private *priv = netdev_priv(ndev);
367
368 writel(readl(priv->base + AVE_GIMR) | bitflag, priv->base + AVE_GIMR);
369 writel(bitflag, priv->base + AVE_GISR);
370 }
371
ave_hw_write_macaddr(struct net_device * ndev,const unsigned char * mac_addr,int reg1,int reg2)372 static void ave_hw_write_macaddr(struct net_device *ndev,
373 const unsigned char *mac_addr,
374 int reg1, int reg2)
375 {
376 struct ave_private *priv = netdev_priv(ndev);
377
378 writel(mac_addr[0] | mac_addr[1] << 8 |
379 mac_addr[2] << 16 | mac_addr[3] << 24, priv->base + reg1);
380 writel(mac_addr[4] | mac_addr[5] << 8, priv->base + reg2);
381 }
382
ave_hw_read_version(struct net_device * ndev,char * buf,int len)383 static void ave_hw_read_version(struct net_device *ndev, char *buf, int len)
384 {
385 struct ave_private *priv = netdev_priv(ndev);
386 u32 major, minor, vr;
387
388 vr = readl(priv->base + AVE_VR);
389 major = (vr & GENMASK(15, 8)) >> 8;
390 minor = (vr & GENMASK(7, 0));
391 snprintf(buf, len, "v%u.%u", major, minor);
392 }
393
ave_ethtool_get_drvinfo(struct net_device * ndev,struct ethtool_drvinfo * info)394 static void ave_ethtool_get_drvinfo(struct net_device *ndev,
395 struct ethtool_drvinfo *info)
396 {
397 struct device *dev = ndev->dev.parent;
398
399 strscpy(info->driver, dev->driver->name, sizeof(info->driver));
400 strscpy(info->bus_info, dev_name(dev), sizeof(info->bus_info));
401 ave_hw_read_version(ndev, info->fw_version, sizeof(info->fw_version));
402 }
403
ave_ethtool_get_msglevel(struct net_device * ndev)404 static u32 ave_ethtool_get_msglevel(struct net_device *ndev)
405 {
406 struct ave_private *priv = netdev_priv(ndev);
407
408 return priv->msg_enable;
409 }
410
ave_ethtool_set_msglevel(struct net_device * ndev,u32 val)411 static void ave_ethtool_set_msglevel(struct net_device *ndev, u32 val)
412 {
413 struct ave_private *priv = netdev_priv(ndev);
414
415 priv->msg_enable = val;
416 }
417
ave_ethtool_get_wol(struct net_device * ndev,struct ethtool_wolinfo * wol)418 static void ave_ethtool_get_wol(struct net_device *ndev,
419 struct ethtool_wolinfo *wol)
420 {
421 wol->supported = 0;
422 wol->wolopts = 0;
423
424 if (ndev->phydev)
425 phy_ethtool_get_wol(ndev->phydev, wol);
426 }
427
__ave_ethtool_set_wol(struct net_device * ndev,struct ethtool_wolinfo * wol)428 static int __ave_ethtool_set_wol(struct net_device *ndev,
429 struct ethtool_wolinfo *wol)
430 {
431 if (!ndev->phydev ||
432 (wol->wolopts & (WAKE_ARP | WAKE_MAGICSECURE)))
433 return -EOPNOTSUPP;
434
435 return phy_ethtool_set_wol(ndev->phydev, wol);
436 }
437
ave_ethtool_set_wol(struct net_device * ndev,struct ethtool_wolinfo * wol)438 static int ave_ethtool_set_wol(struct net_device *ndev,
439 struct ethtool_wolinfo *wol)
440 {
441 int ret;
442
443 ret = __ave_ethtool_set_wol(ndev, wol);
444 if (!ret)
445 device_set_wakeup_enable(&ndev->dev, !!wol->wolopts);
446
447 return ret;
448 }
449
ave_ethtool_get_pauseparam(struct net_device * ndev,struct ethtool_pauseparam * pause)450 static void ave_ethtool_get_pauseparam(struct net_device *ndev,
451 struct ethtool_pauseparam *pause)
452 {
453 struct ave_private *priv = netdev_priv(ndev);
454
455 pause->autoneg = priv->pause_auto;
456 pause->rx_pause = priv->pause_rx;
457 pause->tx_pause = priv->pause_tx;
458 }
459
ave_ethtool_set_pauseparam(struct net_device * ndev,struct ethtool_pauseparam * pause)460 static int ave_ethtool_set_pauseparam(struct net_device *ndev,
461 struct ethtool_pauseparam *pause)
462 {
463 struct ave_private *priv = netdev_priv(ndev);
464 struct phy_device *phydev = ndev->phydev;
465
466 if (!phydev)
467 return -EINVAL;
468
469 priv->pause_auto = pause->autoneg;
470 priv->pause_rx = pause->rx_pause;
471 priv->pause_tx = pause->tx_pause;
472
473 phy_set_asym_pause(phydev, pause->rx_pause, pause->tx_pause);
474
475 return 0;
476 }
477
478 static const struct ethtool_ops ave_ethtool_ops = {
479 .get_link_ksettings = phy_ethtool_get_link_ksettings,
480 .set_link_ksettings = phy_ethtool_set_link_ksettings,
481 .get_drvinfo = ave_ethtool_get_drvinfo,
482 .nway_reset = phy_ethtool_nway_reset,
483 .get_link = ethtool_op_get_link,
484 .get_msglevel = ave_ethtool_get_msglevel,
485 .set_msglevel = ave_ethtool_set_msglevel,
486 .get_wol = ave_ethtool_get_wol,
487 .set_wol = ave_ethtool_set_wol,
488 .get_pauseparam = ave_ethtool_get_pauseparam,
489 .set_pauseparam = ave_ethtool_set_pauseparam,
490 };
491
ave_mdiobus_read(struct mii_bus * bus,int phyid,int regnum)492 static int ave_mdiobus_read(struct mii_bus *bus, int phyid, int regnum)
493 {
494 struct net_device *ndev = bus->priv;
495 struct ave_private *priv;
496 u32 mdioctl, mdiosr;
497 int ret;
498
499 priv = netdev_priv(ndev);
500
501 /* write address */
502 writel((phyid << 8) | regnum, priv->base + AVE_MDIOAR);
503
504 /* read request */
505 mdioctl = readl(priv->base + AVE_MDIOCTR);
506 writel((mdioctl | AVE_MDIOCTR_RREQ) & ~AVE_MDIOCTR_WREQ,
507 priv->base + AVE_MDIOCTR);
508
509 ret = readl_poll_timeout(priv->base + AVE_MDIOSR, mdiosr,
510 !(mdiosr & AVE_MDIOSR_STS), 20, 2000);
511 if (ret) {
512 netdev_err(ndev, "failed to read (phy:%d reg:%x)\n",
513 phyid, regnum);
514 return ret;
515 }
516
517 return readl(priv->base + AVE_MDIORDR) & GENMASK(15, 0);
518 }
519
ave_mdiobus_write(struct mii_bus * bus,int phyid,int regnum,u16 val)520 static int ave_mdiobus_write(struct mii_bus *bus, int phyid, int regnum,
521 u16 val)
522 {
523 struct net_device *ndev = bus->priv;
524 struct ave_private *priv;
525 u32 mdioctl, mdiosr;
526 int ret;
527
528 priv = netdev_priv(ndev);
529
530 /* write address */
531 writel((phyid << 8) | regnum, priv->base + AVE_MDIOAR);
532
533 /* write data */
534 writel(val, priv->base + AVE_MDIOWDR);
535
536 /* write request */
537 mdioctl = readl(priv->base + AVE_MDIOCTR);
538 writel((mdioctl | AVE_MDIOCTR_WREQ) & ~AVE_MDIOCTR_RREQ,
539 priv->base + AVE_MDIOCTR);
540
541 ret = readl_poll_timeout(priv->base + AVE_MDIOSR, mdiosr,
542 !(mdiosr & AVE_MDIOSR_STS), 20, 2000);
543 if (ret)
544 netdev_err(ndev, "failed to write (phy:%d reg:%x)\n",
545 phyid, regnum);
546
547 return ret;
548 }
549
ave_dma_map(struct net_device * ndev,struct ave_desc * desc,void * ptr,size_t len,enum dma_data_direction dir,dma_addr_t * paddr)550 static int ave_dma_map(struct net_device *ndev, struct ave_desc *desc,
551 void *ptr, size_t len, enum dma_data_direction dir,
552 dma_addr_t *paddr)
553 {
554 dma_addr_t map_addr;
555
556 map_addr = dma_map_single(ndev->dev.parent, ptr, len, dir);
557 if (unlikely(dma_mapping_error(ndev->dev.parent, map_addr)))
558 return -ENOMEM;
559
560 desc->skbs_dma = map_addr;
561 desc->skbs_dmalen = len;
562 *paddr = map_addr;
563
564 return 0;
565 }
566
ave_dma_unmap(struct net_device * ndev,struct ave_desc * desc,enum dma_data_direction dir)567 static void ave_dma_unmap(struct net_device *ndev, struct ave_desc *desc,
568 enum dma_data_direction dir)
569 {
570 if (!desc->skbs_dma)
571 return;
572
573 dma_unmap_single(ndev->dev.parent,
574 desc->skbs_dma, desc->skbs_dmalen, dir);
575 desc->skbs_dma = 0;
576 }
577
578 /* Prepare Rx descriptor and memory */
ave_rxdesc_prepare(struct net_device * ndev,int entry)579 static int ave_rxdesc_prepare(struct net_device *ndev, int entry)
580 {
581 struct ave_private *priv = netdev_priv(ndev);
582 struct sk_buff *skb;
583 dma_addr_t paddr;
584 int ret;
585
586 skb = priv->rx.desc[entry].skbs;
587 if (!skb) {
588 skb = netdev_alloc_skb(ndev, AVE_MAX_ETHFRAME);
589 if (!skb)
590 return -ENOMEM;
591 skb->data += AVE_FRAME_HEADROOM;
592 skb->tail += AVE_FRAME_HEADROOM;
593 }
594
595 /* set disable to cmdsts */
596 ave_desc_write_cmdsts(ndev, AVE_DESCID_RX, entry,
597 AVE_STS_INTR | AVE_STS_OWN);
598
599 /* map Rx buffer
600 * Rx buffer set to the Rx descriptor has two restrictions:
601 * - Rx buffer address is 4 byte aligned.
602 * - Rx buffer begins with 2 byte headroom, and data will be put from
603 * (buffer + 2).
604 * To satisfy this, specify the address to put back the buffer
605 * pointer advanced by AVE_FRAME_HEADROOM, and expand the map size
606 * by AVE_FRAME_HEADROOM.
607 */
608 ret = ave_dma_map(ndev, &priv->rx.desc[entry],
609 skb->data - AVE_FRAME_HEADROOM,
610 AVE_MAX_ETHFRAME + AVE_FRAME_HEADROOM,
611 DMA_FROM_DEVICE, &paddr);
612 if (ret) {
613 netdev_err(ndev, "can't map skb for Rx\n");
614 dev_kfree_skb_any(skb);
615 return ret;
616 }
617 priv->rx.desc[entry].skbs = skb;
618
619 /* set buffer pointer */
620 ave_desc_write_addr(ndev, AVE_DESCID_RX, entry, paddr);
621
622 /* set enable to cmdsts */
623 ave_desc_write_cmdsts(ndev, AVE_DESCID_RX, entry,
624 AVE_STS_INTR | AVE_MAX_ETHFRAME);
625
626 return ret;
627 }
628
629 /* Switch state of descriptor */
ave_desc_switch(struct net_device * ndev,enum desc_state state)630 static int ave_desc_switch(struct net_device *ndev, enum desc_state state)
631 {
632 struct ave_private *priv = netdev_priv(ndev);
633 int ret = 0;
634 u32 val;
635
636 switch (state) {
637 case AVE_DESC_START:
638 writel(AVE_DESCC_TD | AVE_DESCC_RD0, priv->base + AVE_DESCC);
639 break;
640
641 case AVE_DESC_STOP:
642 writel(0, priv->base + AVE_DESCC);
643 if (readl_poll_timeout(priv->base + AVE_DESCC, val, !val,
644 150, 15000)) {
645 netdev_err(ndev, "can't stop descriptor\n");
646 ret = -EBUSY;
647 }
648 break;
649
650 case AVE_DESC_RX_SUSPEND:
651 val = readl(priv->base + AVE_DESCC);
652 val |= AVE_DESCC_RDSTP;
653 val &= ~AVE_DESCC_STATUS_MASK;
654 writel(val, priv->base + AVE_DESCC);
655 if (readl_poll_timeout(priv->base + AVE_DESCC, val,
656 val & (AVE_DESCC_RDSTP << 16),
657 150, 150000)) {
658 netdev_err(ndev, "can't suspend descriptor\n");
659 ret = -EBUSY;
660 }
661 break;
662
663 case AVE_DESC_RX_PERMIT:
664 val = readl(priv->base + AVE_DESCC);
665 val &= ~AVE_DESCC_RDSTP;
666 val &= ~AVE_DESCC_STATUS_MASK;
667 writel(val, priv->base + AVE_DESCC);
668 break;
669
670 default:
671 ret = -EINVAL;
672 break;
673 }
674
675 return ret;
676 }
677
ave_tx_complete(struct net_device * ndev)678 static int ave_tx_complete(struct net_device *ndev)
679 {
680 struct ave_private *priv = netdev_priv(ndev);
681 u32 proc_idx, done_idx, ndesc, cmdsts;
682 unsigned int nr_freebuf = 0;
683 unsigned int tx_packets = 0;
684 unsigned int tx_bytes = 0;
685
686 proc_idx = priv->tx.proc_idx;
687 done_idx = priv->tx.done_idx;
688 ndesc = priv->tx.ndesc;
689
690 /* free pre-stored skb from done_idx to proc_idx */
691 while (proc_idx != done_idx) {
692 cmdsts = ave_desc_read_cmdsts(ndev, AVE_DESCID_TX, done_idx);
693
694 /* do nothing if owner is HW (==1 for Tx) */
695 if (cmdsts & AVE_STS_OWN)
696 break;
697
698 /* check Tx status and updates statistics */
699 if (cmdsts & AVE_STS_OK) {
700 tx_bytes += cmdsts & AVE_STS_PKTLEN_TX_MASK;
701 /* success */
702 if (cmdsts & AVE_STS_LAST)
703 tx_packets++;
704 } else {
705 /* error */
706 if (cmdsts & AVE_STS_LAST) {
707 priv->stats_tx.errors++;
708 if (cmdsts & (AVE_STS_OWC | AVE_STS_EC))
709 priv->stats_tx.collisions++;
710 }
711 }
712
713 /* release skb */
714 if (priv->tx.desc[done_idx].skbs) {
715 ave_dma_unmap(ndev, &priv->tx.desc[done_idx],
716 DMA_TO_DEVICE);
717 dev_consume_skb_any(priv->tx.desc[done_idx].skbs);
718 priv->tx.desc[done_idx].skbs = NULL;
719 nr_freebuf++;
720 }
721 done_idx = (done_idx + 1) % ndesc;
722 }
723
724 priv->tx.done_idx = done_idx;
725
726 /* update stats */
727 u64_stats_update_begin(&priv->stats_tx.syncp);
728 priv->stats_tx.packets += tx_packets;
729 priv->stats_tx.bytes += tx_bytes;
730 u64_stats_update_end(&priv->stats_tx.syncp);
731
732 /* wake queue for freeing buffer */
733 if (unlikely(netif_queue_stopped(ndev)) && nr_freebuf)
734 netif_wake_queue(ndev);
735
736 return nr_freebuf;
737 }
738
ave_rx_receive(struct net_device * ndev,int num)739 static int ave_rx_receive(struct net_device *ndev, int num)
740 {
741 struct ave_private *priv = netdev_priv(ndev);
742 unsigned int rx_packets = 0;
743 unsigned int rx_bytes = 0;
744 u32 proc_idx, done_idx;
745 struct sk_buff *skb;
746 unsigned int pktlen;
747 int restpkt, npkts;
748 u32 ndesc, cmdsts;
749
750 proc_idx = priv->rx.proc_idx;
751 done_idx = priv->rx.done_idx;
752 ndesc = priv->rx.ndesc;
753 restpkt = ((proc_idx + ndesc - 1) - done_idx) % ndesc;
754
755 for (npkts = 0; npkts < num; npkts++) {
756 /* we can't receive more packet, so fill desc quickly */
757 if (--restpkt < 0)
758 break;
759
760 cmdsts = ave_desc_read_cmdsts(ndev, AVE_DESCID_RX, proc_idx);
761
762 /* do nothing if owner is HW (==0 for Rx) */
763 if (!(cmdsts & AVE_STS_OWN))
764 break;
765
766 if (!(cmdsts & AVE_STS_OK)) {
767 priv->stats_rx.errors++;
768 proc_idx = (proc_idx + 1) % ndesc;
769 continue;
770 }
771
772 pktlen = cmdsts & AVE_STS_PKTLEN_RX_MASK;
773
774 /* get skbuff for rx */
775 skb = priv->rx.desc[proc_idx].skbs;
776 priv->rx.desc[proc_idx].skbs = NULL;
777
778 ave_dma_unmap(ndev, &priv->rx.desc[proc_idx], DMA_FROM_DEVICE);
779
780 skb->dev = ndev;
781 skb_put(skb, pktlen);
782 skb->protocol = eth_type_trans(skb, ndev);
783
784 if ((cmdsts & AVE_STS_CSSV) && (!(cmdsts & AVE_STS_CSER)))
785 skb->ip_summed = CHECKSUM_UNNECESSARY;
786
787 rx_packets++;
788 rx_bytes += pktlen;
789
790 netif_receive_skb(skb);
791
792 proc_idx = (proc_idx + 1) % ndesc;
793 }
794
795 priv->rx.proc_idx = proc_idx;
796
797 /* update stats */
798 u64_stats_update_begin(&priv->stats_rx.syncp);
799 priv->stats_rx.packets += rx_packets;
800 priv->stats_rx.bytes += rx_bytes;
801 u64_stats_update_end(&priv->stats_rx.syncp);
802
803 /* refill the Rx buffers */
804 while (proc_idx != done_idx) {
805 if (ave_rxdesc_prepare(ndev, done_idx))
806 break;
807 done_idx = (done_idx + 1) % ndesc;
808 }
809
810 priv->rx.done_idx = done_idx;
811
812 return npkts;
813 }
814
ave_napi_poll_rx(struct napi_struct * napi,int budget)815 static int ave_napi_poll_rx(struct napi_struct *napi, int budget)
816 {
817 struct ave_private *priv;
818 struct net_device *ndev;
819 int num;
820
821 priv = container_of(napi, struct ave_private, napi_rx);
822 ndev = priv->ndev;
823
824 num = ave_rx_receive(ndev, budget);
825 if (num < budget) {
826 napi_complete_done(napi, num);
827
828 /* enable Rx interrupt when NAPI finishes */
829 ave_irq_enable(ndev, AVE_GI_RXIINT);
830 }
831
832 return num;
833 }
834
ave_napi_poll_tx(struct napi_struct * napi,int budget)835 static int ave_napi_poll_tx(struct napi_struct *napi, int budget)
836 {
837 struct ave_private *priv;
838 struct net_device *ndev;
839 int num;
840
841 priv = container_of(napi, struct ave_private, napi_tx);
842 ndev = priv->ndev;
843
844 num = ave_tx_complete(ndev);
845 napi_complete(napi);
846
847 /* enable Tx interrupt when NAPI finishes */
848 ave_irq_enable(ndev, AVE_GI_TX);
849
850 return num;
851 }
852
ave_global_reset(struct net_device * ndev)853 static void ave_global_reset(struct net_device *ndev)
854 {
855 struct ave_private *priv = netdev_priv(ndev);
856 u32 val;
857
858 /* set config register */
859 val = AVE_CFGR_FLE | AVE_CFGR_IPFCEN | AVE_CFGR_CHE;
860 if (!phy_interface_mode_is_rgmii(priv->phy_mode))
861 val |= AVE_CFGR_MII;
862 writel(val, priv->base + AVE_CFGR);
863
864 /* reset RMII register */
865 val = readl(priv->base + AVE_RSTCTRL);
866 val &= ~AVE_RSTCTRL_RMIIRST;
867 writel(val, priv->base + AVE_RSTCTRL);
868
869 /* assert reset */
870 writel(AVE_GRR_GRST | AVE_GRR_PHYRST, priv->base + AVE_GRR);
871 msleep(20);
872
873 /* 1st, negate PHY reset only */
874 writel(AVE_GRR_GRST, priv->base + AVE_GRR);
875 msleep(40);
876
877 /* negate reset */
878 writel(0, priv->base + AVE_GRR);
879 msleep(40);
880
881 /* negate RMII register */
882 val = readl(priv->base + AVE_RSTCTRL);
883 val |= AVE_RSTCTRL_RMIIRST;
884 writel(val, priv->base + AVE_RSTCTRL);
885
886 ave_irq_disable_all(ndev);
887 }
888
ave_rxfifo_reset(struct net_device * ndev)889 static void ave_rxfifo_reset(struct net_device *ndev)
890 {
891 struct ave_private *priv = netdev_priv(ndev);
892 u32 rxcr_org;
893
894 /* save and disable MAC receive op */
895 rxcr_org = readl(priv->base + AVE_RXCR);
896 writel(rxcr_org & (~AVE_RXCR_RXEN), priv->base + AVE_RXCR);
897
898 /* suspend Rx descriptor */
899 ave_desc_switch(ndev, AVE_DESC_RX_SUSPEND);
900
901 /* receive all packets before descriptor starts */
902 ave_rx_receive(ndev, priv->rx.ndesc);
903
904 /* assert reset */
905 writel(AVE_GRR_RXFFR, priv->base + AVE_GRR);
906 udelay(50);
907
908 /* negate reset */
909 writel(0, priv->base + AVE_GRR);
910 udelay(20);
911
912 /* negate interrupt status */
913 writel(AVE_GI_RXOVF, priv->base + AVE_GISR);
914
915 /* permit descriptor */
916 ave_desc_switch(ndev, AVE_DESC_RX_PERMIT);
917
918 /* restore MAC reccieve op */
919 writel(rxcr_org, priv->base + AVE_RXCR);
920 }
921
ave_irq_handler(int irq,void * netdev)922 static irqreturn_t ave_irq_handler(int irq, void *netdev)
923 {
924 struct net_device *ndev = (struct net_device *)netdev;
925 struct ave_private *priv = netdev_priv(ndev);
926 u32 gimr_val, gisr_val;
927
928 gimr_val = ave_irq_disable_all(ndev);
929
930 /* get interrupt status */
931 gisr_val = readl(priv->base + AVE_GISR);
932
933 /* PHY */
934 if (gisr_val & AVE_GI_PHY)
935 writel(AVE_GI_PHY, priv->base + AVE_GISR);
936
937 /* check exceeding packet */
938 if (gisr_val & AVE_GI_RXERR) {
939 writel(AVE_GI_RXERR, priv->base + AVE_GISR);
940 netdev_err(ndev, "receive a packet exceeding frame buffer\n");
941 }
942
943 gisr_val &= gimr_val;
944 if (!gisr_val)
945 goto exit_isr;
946
947 /* RxFIFO overflow */
948 if (gisr_val & AVE_GI_RXOVF) {
949 priv->stats_rx.fifo_errors++;
950 ave_rxfifo_reset(ndev);
951 goto exit_isr;
952 }
953
954 /* Rx drop */
955 if (gisr_val & AVE_GI_RXDROP) {
956 priv->stats_rx.dropped++;
957 writel(AVE_GI_RXDROP, priv->base + AVE_GISR);
958 }
959
960 /* Rx interval */
961 if (gisr_val & AVE_GI_RXIINT) {
962 napi_schedule(&priv->napi_rx);
963 /* still force to disable Rx interrupt until NAPI finishes */
964 gimr_val &= ~AVE_GI_RXIINT;
965 }
966
967 /* Tx completed */
968 if (gisr_val & AVE_GI_TX) {
969 napi_schedule(&priv->napi_tx);
970 /* still force to disable Tx interrupt until NAPI finishes */
971 gimr_val &= ~AVE_GI_TX;
972 }
973
974 exit_isr:
975 ave_irq_restore(ndev, gimr_val);
976
977 return IRQ_HANDLED;
978 }
979
ave_pfsel_start(struct net_device * ndev,unsigned int entry)980 static int ave_pfsel_start(struct net_device *ndev, unsigned int entry)
981 {
982 struct ave_private *priv = netdev_priv(ndev);
983 u32 val;
984
985 if (WARN_ON(entry > AVE_PF_SIZE))
986 return -EINVAL;
987
988 val = readl(priv->base + AVE_PFEN);
989 writel(val | BIT(entry), priv->base + AVE_PFEN);
990
991 return 0;
992 }
993
ave_pfsel_stop(struct net_device * ndev,unsigned int entry)994 static int ave_pfsel_stop(struct net_device *ndev, unsigned int entry)
995 {
996 struct ave_private *priv = netdev_priv(ndev);
997 u32 val;
998
999 if (WARN_ON(entry > AVE_PF_SIZE))
1000 return -EINVAL;
1001
1002 val = readl(priv->base + AVE_PFEN);
1003 writel(val & ~BIT(entry), priv->base + AVE_PFEN);
1004
1005 return 0;
1006 }
1007
ave_pfsel_set_macaddr(struct net_device * ndev,unsigned int entry,const unsigned char * mac_addr,unsigned int set_size)1008 static int ave_pfsel_set_macaddr(struct net_device *ndev,
1009 unsigned int entry,
1010 const unsigned char *mac_addr,
1011 unsigned int set_size)
1012 {
1013 struct ave_private *priv = netdev_priv(ndev);
1014
1015 if (WARN_ON(entry > AVE_PF_SIZE))
1016 return -EINVAL;
1017 if (WARN_ON(set_size > 6))
1018 return -EINVAL;
1019
1020 ave_pfsel_stop(ndev, entry);
1021
1022 /* set MAC address for the filter */
1023 ave_hw_write_macaddr(ndev, mac_addr,
1024 AVE_PKTF(entry), AVE_PKTF(entry) + 4);
1025
1026 /* set byte mask */
1027 writel(GENMASK(31, set_size) & AVE_PFMBYTE_MASK0,
1028 priv->base + AVE_PFMBYTE(entry));
1029 writel(AVE_PFMBYTE_MASK1, priv->base + AVE_PFMBYTE(entry) + 4);
1030
1031 /* set bit mask filter */
1032 writel(AVE_PFMBIT_MASK, priv->base + AVE_PFMBIT(entry));
1033
1034 /* set selector to ring 0 */
1035 writel(0, priv->base + AVE_PFSEL(entry));
1036
1037 /* restart filter */
1038 ave_pfsel_start(ndev, entry);
1039
1040 return 0;
1041 }
1042
ave_pfsel_set_promisc(struct net_device * ndev,unsigned int entry,u32 rxring)1043 static void ave_pfsel_set_promisc(struct net_device *ndev,
1044 unsigned int entry, u32 rxring)
1045 {
1046 struct ave_private *priv = netdev_priv(ndev);
1047
1048 if (WARN_ON(entry > AVE_PF_SIZE))
1049 return;
1050
1051 ave_pfsel_stop(ndev, entry);
1052
1053 /* set byte mask */
1054 writel(AVE_PFMBYTE_MASK0, priv->base + AVE_PFMBYTE(entry));
1055 writel(AVE_PFMBYTE_MASK1, priv->base + AVE_PFMBYTE(entry) + 4);
1056
1057 /* set bit mask filter */
1058 writel(AVE_PFMBIT_MASK, priv->base + AVE_PFMBIT(entry));
1059
1060 /* set selector to rxring */
1061 writel(rxring, priv->base + AVE_PFSEL(entry));
1062
1063 ave_pfsel_start(ndev, entry);
1064 }
1065
ave_pfsel_init(struct net_device * ndev)1066 static void ave_pfsel_init(struct net_device *ndev)
1067 {
1068 unsigned char bcast_mac[ETH_ALEN];
1069 int i;
1070
1071 eth_broadcast_addr(bcast_mac);
1072
1073 for (i = 0; i < AVE_PF_SIZE; i++)
1074 ave_pfsel_stop(ndev, i);
1075
1076 /* promiscious entry, select ring 0 */
1077 ave_pfsel_set_promisc(ndev, AVE_PFNUM_FILTER, 0);
1078
1079 /* unicast entry */
1080 ave_pfsel_set_macaddr(ndev, AVE_PFNUM_UNICAST, ndev->dev_addr, 6);
1081
1082 /* broadcast entry */
1083 ave_pfsel_set_macaddr(ndev, AVE_PFNUM_BROADCAST, bcast_mac, 6);
1084 }
1085
ave_phy_adjust_link(struct net_device * ndev)1086 static void ave_phy_adjust_link(struct net_device *ndev)
1087 {
1088 struct ave_private *priv = netdev_priv(ndev);
1089 struct phy_device *phydev = ndev->phydev;
1090 u32 val, txcr, rxcr, rxcr_org;
1091 u16 rmt_adv = 0, lcl_adv = 0;
1092 u8 cap;
1093
1094 /* set RGMII speed */
1095 val = readl(priv->base + AVE_TXCR);
1096 val &= ~(AVE_TXCR_TXSPD_100 | AVE_TXCR_TXSPD_1G);
1097
1098 if (phy_interface_is_rgmii(phydev) && phydev->speed == SPEED_1000)
1099 val |= AVE_TXCR_TXSPD_1G;
1100 else if (phydev->speed == SPEED_100)
1101 val |= AVE_TXCR_TXSPD_100;
1102
1103 writel(val, priv->base + AVE_TXCR);
1104
1105 /* set RMII speed (100M/10M only) */
1106 if (!phy_interface_is_rgmii(phydev)) {
1107 val = readl(priv->base + AVE_LINKSEL);
1108 if (phydev->speed == SPEED_10)
1109 val &= ~AVE_LINKSEL_100M;
1110 else
1111 val |= AVE_LINKSEL_100M;
1112 writel(val, priv->base + AVE_LINKSEL);
1113 }
1114
1115 /* check current RXCR/TXCR */
1116 rxcr = readl(priv->base + AVE_RXCR);
1117 txcr = readl(priv->base + AVE_TXCR);
1118 rxcr_org = rxcr;
1119
1120 if (phydev->duplex) {
1121 rxcr |= AVE_RXCR_FDUPEN;
1122
1123 if (phydev->pause)
1124 rmt_adv |= LPA_PAUSE_CAP;
1125 if (phydev->asym_pause)
1126 rmt_adv |= LPA_PAUSE_ASYM;
1127
1128 lcl_adv = linkmode_adv_to_lcl_adv_t(phydev->advertising);
1129 cap = mii_resolve_flowctrl_fdx(lcl_adv, rmt_adv);
1130 if (cap & FLOW_CTRL_TX)
1131 txcr |= AVE_TXCR_FLOCTR;
1132 else
1133 txcr &= ~AVE_TXCR_FLOCTR;
1134 if (cap & FLOW_CTRL_RX)
1135 rxcr |= AVE_RXCR_FLOCTR;
1136 else
1137 rxcr &= ~AVE_RXCR_FLOCTR;
1138 } else {
1139 rxcr &= ~AVE_RXCR_FDUPEN;
1140 rxcr &= ~AVE_RXCR_FLOCTR;
1141 txcr &= ~AVE_TXCR_FLOCTR;
1142 }
1143
1144 if (rxcr_org != rxcr) {
1145 /* disable Rx mac */
1146 writel(rxcr & ~AVE_RXCR_RXEN, priv->base + AVE_RXCR);
1147 /* change and enable TX/Rx mac */
1148 writel(txcr, priv->base + AVE_TXCR);
1149 writel(rxcr, priv->base + AVE_RXCR);
1150 }
1151
1152 phy_print_status(phydev);
1153 }
1154
ave_macaddr_init(struct net_device * ndev)1155 static void ave_macaddr_init(struct net_device *ndev)
1156 {
1157 ave_hw_write_macaddr(ndev, ndev->dev_addr, AVE_RXMAC1R, AVE_RXMAC2R);
1158
1159 /* pfsel unicast entry */
1160 ave_pfsel_set_macaddr(ndev, AVE_PFNUM_UNICAST, ndev->dev_addr, 6);
1161 }
1162
ave_init(struct net_device * ndev)1163 static int ave_init(struct net_device *ndev)
1164 {
1165 struct ethtool_wolinfo wol = { .cmd = ETHTOOL_GWOL };
1166 struct ave_private *priv = netdev_priv(ndev);
1167 struct device *dev = ndev->dev.parent;
1168 struct device_node *np = dev->of_node;
1169 struct device_node *mdio_np;
1170 struct phy_device *phydev;
1171 int nc, nr, ret;
1172
1173 /* enable clk because of hw access until ndo_open */
1174 for (nc = 0; nc < priv->nclks; nc++) {
1175 ret = clk_prepare_enable(priv->clk[nc]);
1176 if (ret) {
1177 dev_err(dev, "can't enable clock\n");
1178 goto out_clk_disable;
1179 }
1180 }
1181
1182 for (nr = 0; nr < priv->nrsts; nr++) {
1183 ret = reset_control_deassert(priv->rst[nr]);
1184 if (ret) {
1185 dev_err(dev, "can't deassert reset\n");
1186 goto out_reset_assert;
1187 }
1188 }
1189
1190 ret = regmap_update_bits(priv->regmap, SG_ETPINMODE,
1191 priv->pinmode_mask, priv->pinmode_val);
1192 if (ret)
1193 goto out_reset_assert;
1194
1195 ave_global_reset(ndev);
1196
1197 mdio_np = of_get_child_by_name(np, "mdio");
1198 if (!mdio_np) {
1199 dev_err(dev, "mdio node not found\n");
1200 ret = -EINVAL;
1201 goto out_reset_assert;
1202 }
1203 ret = of_mdiobus_register(priv->mdio, mdio_np);
1204 of_node_put(mdio_np);
1205 if (ret) {
1206 dev_err(dev, "failed to register mdiobus\n");
1207 goto out_reset_assert;
1208 }
1209
1210 phydev = of_phy_get_and_connect(ndev, np, ave_phy_adjust_link);
1211 if (!phydev) {
1212 dev_err(dev, "could not attach to PHY\n");
1213 ret = -ENODEV;
1214 goto out_mdio_unregister;
1215 }
1216
1217 priv->phydev = phydev;
1218
1219 ave_ethtool_get_wol(ndev, &wol);
1220 device_set_wakeup_capable(&ndev->dev, !!wol.supported);
1221
1222 /* set wol initial state disabled */
1223 wol.wolopts = 0;
1224 __ave_ethtool_set_wol(ndev, &wol);
1225
1226 if (!phy_interface_is_rgmii(phydev))
1227 phy_set_max_speed(phydev, SPEED_100);
1228
1229 phy_support_asym_pause(phydev);
1230
1231 phydev->mac_managed_pm = true;
1232
1233 phy_attached_info(phydev);
1234
1235 return 0;
1236
1237 out_mdio_unregister:
1238 mdiobus_unregister(priv->mdio);
1239 out_reset_assert:
1240 while (--nr >= 0)
1241 reset_control_assert(priv->rst[nr]);
1242 out_clk_disable:
1243 while (--nc >= 0)
1244 clk_disable_unprepare(priv->clk[nc]);
1245
1246 return ret;
1247 }
1248
ave_uninit(struct net_device * ndev)1249 static void ave_uninit(struct net_device *ndev)
1250 {
1251 struct ave_private *priv = netdev_priv(ndev);
1252 int i;
1253
1254 phy_disconnect(priv->phydev);
1255 mdiobus_unregister(priv->mdio);
1256
1257 /* disable clk because of hw access after ndo_stop */
1258 for (i = 0; i < priv->nrsts; i++)
1259 reset_control_assert(priv->rst[i]);
1260 for (i = 0; i < priv->nclks; i++)
1261 clk_disable_unprepare(priv->clk[i]);
1262 }
1263
ave_open(struct net_device * ndev)1264 static int ave_open(struct net_device *ndev)
1265 {
1266 struct ave_private *priv = netdev_priv(ndev);
1267 int entry;
1268 int ret;
1269 u32 val;
1270
1271 ret = request_irq(priv->irq, ave_irq_handler, IRQF_SHARED, ndev->name,
1272 ndev);
1273 if (ret)
1274 return ret;
1275
1276 priv->tx.desc = kzalloc_objs(*priv->tx.desc, priv->tx.ndesc);
1277 if (!priv->tx.desc) {
1278 ret = -ENOMEM;
1279 goto out_free_irq;
1280 }
1281
1282 priv->rx.desc = kzalloc_objs(*priv->rx.desc, priv->rx.ndesc);
1283 if (!priv->rx.desc) {
1284 kfree(priv->tx.desc);
1285 ret = -ENOMEM;
1286 goto out_free_irq;
1287 }
1288
1289 /* initialize Tx work and descriptor */
1290 priv->tx.proc_idx = 0;
1291 priv->tx.done_idx = 0;
1292 for (entry = 0; entry < priv->tx.ndesc; entry++) {
1293 ave_desc_write_cmdsts(ndev, AVE_DESCID_TX, entry, 0);
1294 ave_desc_write_addr(ndev, AVE_DESCID_TX, entry, 0);
1295 }
1296 writel(AVE_TXDC_ADDR_START |
1297 (((priv->tx.ndesc * priv->desc_size) << 16) & AVE_TXDC_SIZE),
1298 priv->base + AVE_TXDC);
1299
1300 /* initialize Rx work and descriptor */
1301 priv->rx.proc_idx = 0;
1302 priv->rx.done_idx = 0;
1303 for (entry = 0; entry < priv->rx.ndesc; entry++) {
1304 if (ave_rxdesc_prepare(ndev, entry))
1305 break;
1306 }
1307 writel(AVE_RXDC0_ADDR_START |
1308 (((priv->rx.ndesc * priv->desc_size) << 16) & AVE_RXDC0_SIZE),
1309 priv->base + AVE_RXDC0);
1310
1311 ave_desc_switch(ndev, AVE_DESC_START);
1312
1313 ave_pfsel_init(ndev);
1314 ave_macaddr_init(ndev);
1315
1316 /* set Rx configuration */
1317 /* full duplex, enable pause drop, enalbe flow control */
1318 val = AVE_RXCR_RXEN | AVE_RXCR_FDUPEN | AVE_RXCR_DRPEN |
1319 AVE_RXCR_FLOCTR | (AVE_MAX_ETHFRAME & AVE_RXCR_MPSIZ_MASK);
1320 writel(val, priv->base + AVE_RXCR);
1321
1322 /* set Tx configuration */
1323 /* enable flow control, disable loopback */
1324 writel(AVE_TXCR_FLOCTR, priv->base + AVE_TXCR);
1325
1326 /* enable timer, clear EN,INTM, and mask interval unit(BSCK) */
1327 val = readl(priv->base + AVE_IIRQC) & AVE_IIRQC_BSCK;
1328 val |= AVE_IIRQC_EN0 | (AVE_INTM_COUNT << 16);
1329 writel(val, priv->base + AVE_IIRQC);
1330
1331 val = AVE_GI_RXIINT | AVE_GI_RXOVF | AVE_GI_TX | AVE_GI_RXDROP;
1332 ave_irq_restore(ndev, val);
1333
1334 napi_enable(&priv->napi_rx);
1335 napi_enable(&priv->napi_tx);
1336
1337 phy_start(ndev->phydev);
1338 phy_start_aneg(ndev->phydev);
1339 netif_start_queue(ndev);
1340
1341 return 0;
1342
1343 out_free_irq:
1344 disable_irq(priv->irq);
1345 free_irq(priv->irq, ndev);
1346
1347 return ret;
1348 }
1349
ave_stop(struct net_device * ndev)1350 static int ave_stop(struct net_device *ndev)
1351 {
1352 struct ave_private *priv = netdev_priv(ndev);
1353 int entry;
1354
1355 ave_irq_disable_all(ndev);
1356 disable_irq(priv->irq);
1357 free_irq(priv->irq, ndev);
1358
1359 netif_tx_disable(ndev);
1360 phy_stop(ndev->phydev);
1361 napi_disable(&priv->napi_tx);
1362 napi_disable(&priv->napi_rx);
1363
1364 ave_desc_switch(ndev, AVE_DESC_STOP);
1365
1366 /* free Tx buffer */
1367 for (entry = 0; entry < priv->tx.ndesc; entry++) {
1368 if (!priv->tx.desc[entry].skbs)
1369 continue;
1370
1371 ave_dma_unmap(ndev, &priv->tx.desc[entry], DMA_TO_DEVICE);
1372 dev_kfree_skb_any(priv->tx.desc[entry].skbs);
1373 priv->tx.desc[entry].skbs = NULL;
1374 }
1375 priv->tx.proc_idx = 0;
1376 priv->tx.done_idx = 0;
1377
1378 /* free Rx buffer */
1379 for (entry = 0; entry < priv->rx.ndesc; entry++) {
1380 if (!priv->rx.desc[entry].skbs)
1381 continue;
1382
1383 ave_dma_unmap(ndev, &priv->rx.desc[entry], DMA_FROM_DEVICE);
1384 dev_kfree_skb_any(priv->rx.desc[entry].skbs);
1385 priv->rx.desc[entry].skbs = NULL;
1386 }
1387 priv->rx.proc_idx = 0;
1388 priv->rx.done_idx = 0;
1389
1390 kfree(priv->tx.desc);
1391 kfree(priv->rx.desc);
1392
1393 return 0;
1394 }
1395
ave_start_xmit(struct sk_buff * skb,struct net_device * ndev)1396 static netdev_tx_t ave_start_xmit(struct sk_buff *skb, struct net_device *ndev)
1397 {
1398 struct ave_private *priv = netdev_priv(ndev);
1399 u32 proc_idx, done_idx, ndesc, cmdsts;
1400 int ret, freepkt;
1401 dma_addr_t paddr;
1402
1403 proc_idx = priv->tx.proc_idx;
1404 done_idx = priv->tx.done_idx;
1405 ndesc = priv->tx.ndesc;
1406 freepkt = ((done_idx + ndesc - 1) - proc_idx) % ndesc;
1407
1408 /* stop queue when not enough entry */
1409 if (unlikely(freepkt < 1)) {
1410 netif_stop_queue(ndev);
1411 return NETDEV_TX_BUSY;
1412 }
1413
1414 /* add padding for short packet */
1415 if (skb_put_padto(skb, ETH_ZLEN)) {
1416 priv->stats_tx.dropped++;
1417 return NETDEV_TX_OK;
1418 }
1419
1420 /* map Tx buffer
1421 * Tx buffer set to the Tx descriptor doesn't have any restriction.
1422 */
1423 ret = ave_dma_map(ndev, &priv->tx.desc[proc_idx],
1424 skb->data, skb->len, DMA_TO_DEVICE, &paddr);
1425 if (ret) {
1426 dev_kfree_skb_any(skb);
1427 priv->stats_tx.dropped++;
1428 return NETDEV_TX_OK;
1429 }
1430
1431 priv->tx.desc[proc_idx].skbs = skb;
1432
1433 ave_desc_write_addr(ndev, AVE_DESCID_TX, proc_idx, paddr);
1434
1435 cmdsts = AVE_STS_OWN | AVE_STS_1ST | AVE_STS_LAST |
1436 (skb->len & AVE_STS_PKTLEN_TX_MASK);
1437
1438 /* set interrupt per AVE_FORCE_TXINTCNT or when queue is stopped */
1439 if (!(proc_idx % AVE_FORCE_TXINTCNT) || netif_queue_stopped(ndev))
1440 cmdsts |= AVE_STS_INTR;
1441
1442 /* disable checksum calculation when skb doesn't calurate checksum */
1443 if (skb->ip_summed == CHECKSUM_NONE ||
1444 skb->ip_summed == CHECKSUM_UNNECESSARY)
1445 cmdsts |= AVE_STS_NOCSUM;
1446
1447 ave_desc_write_cmdsts(ndev, AVE_DESCID_TX, proc_idx, cmdsts);
1448
1449 priv->tx.proc_idx = (proc_idx + 1) % ndesc;
1450
1451 return NETDEV_TX_OK;
1452 }
1453
ave_ioctl(struct net_device * ndev,struct ifreq * ifr,int cmd)1454 static int ave_ioctl(struct net_device *ndev, struct ifreq *ifr, int cmd)
1455 {
1456 return phy_mii_ioctl(ndev->phydev, ifr, cmd);
1457 }
1458
1459 static const u8 v4multi_macadr[] = { 0x01, 0x00, 0x00, 0x00, 0x00, 0x00 };
1460 static const u8 v6multi_macadr[] = { 0x33, 0x00, 0x00, 0x00, 0x00, 0x00 };
1461
ave_set_rx_mode(struct net_device * ndev)1462 static void ave_set_rx_mode(struct net_device *ndev)
1463 {
1464 struct ave_private *priv = netdev_priv(ndev);
1465 struct netdev_hw_addr *hw_adr;
1466 int count, mc_cnt;
1467 u32 val;
1468
1469 /* MAC addr filter enable for promiscious mode */
1470 mc_cnt = netdev_mc_count(ndev);
1471 val = readl(priv->base + AVE_RXCR);
1472 if (ndev->flags & IFF_PROMISC || !mc_cnt)
1473 val &= ~AVE_RXCR_AFEN;
1474 else
1475 val |= AVE_RXCR_AFEN;
1476 writel(val, priv->base + AVE_RXCR);
1477
1478 /* set all multicast address */
1479 if ((ndev->flags & IFF_ALLMULTI) || mc_cnt > AVE_PF_MULTICAST_SIZE) {
1480 ave_pfsel_set_macaddr(ndev, AVE_PFNUM_MULTICAST,
1481 v4multi_macadr, 1);
1482 ave_pfsel_set_macaddr(ndev, AVE_PFNUM_MULTICAST + 1,
1483 v6multi_macadr, 1);
1484 } else {
1485 /* stop all multicast filter */
1486 for (count = 0; count < AVE_PF_MULTICAST_SIZE; count++)
1487 ave_pfsel_stop(ndev, AVE_PFNUM_MULTICAST + count);
1488
1489 /* set multicast addresses */
1490 count = 0;
1491 netdev_for_each_mc_addr(hw_adr, ndev) {
1492 if (count == mc_cnt)
1493 break;
1494 ave_pfsel_set_macaddr(ndev, AVE_PFNUM_MULTICAST + count,
1495 hw_adr->addr, 6);
1496 count++;
1497 }
1498 }
1499 }
1500
ave_get_stats64(struct net_device * ndev,struct rtnl_link_stats64 * stats)1501 static void ave_get_stats64(struct net_device *ndev,
1502 struct rtnl_link_stats64 *stats)
1503 {
1504 struct ave_private *priv = netdev_priv(ndev);
1505 unsigned int start;
1506
1507 do {
1508 start = u64_stats_fetch_begin(&priv->stats_rx.syncp);
1509 stats->rx_packets = priv->stats_rx.packets;
1510 stats->rx_bytes = priv->stats_rx.bytes;
1511 } while (u64_stats_fetch_retry(&priv->stats_rx.syncp, start));
1512
1513 do {
1514 start = u64_stats_fetch_begin(&priv->stats_tx.syncp);
1515 stats->tx_packets = priv->stats_tx.packets;
1516 stats->tx_bytes = priv->stats_tx.bytes;
1517 } while (u64_stats_fetch_retry(&priv->stats_tx.syncp, start));
1518
1519 stats->rx_errors = priv->stats_rx.errors;
1520 stats->tx_errors = priv->stats_tx.errors;
1521 stats->rx_dropped = priv->stats_rx.dropped;
1522 stats->tx_dropped = priv->stats_tx.dropped;
1523 stats->rx_fifo_errors = priv->stats_rx.fifo_errors;
1524 stats->collisions = priv->stats_tx.collisions;
1525 }
1526
ave_set_mac_address(struct net_device * ndev,void * p)1527 static int ave_set_mac_address(struct net_device *ndev, void *p)
1528 {
1529 int ret = eth_mac_addr(ndev, p);
1530
1531 if (ret)
1532 return ret;
1533
1534 ave_macaddr_init(ndev);
1535
1536 return 0;
1537 }
1538
1539 static const struct net_device_ops ave_netdev_ops = {
1540 .ndo_init = ave_init,
1541 .ndo_uninit = ave_uninit,
1542 .ndo_open = ave_open,
1543 .ndo_stop = ave_stop,
1544 .ndo_start_xmit = ave_start_xmit,
1545 .ndo_eth_ioctl = ave_ioctl,
1546 .ndo_set_rx_mode = ave_set_rx_mode,
1547 .ndo_get_stats64 = ave_get_stats64,
1548 .ndo_set_mac_address = ave_set_mac_address,
1549 };
1550
ave_probe(struct platform_device * pdev)1551 static int ave_probe(struct platform_device *pdev)
1552 {
1553 const struct ave_soc_data *data;
1554 struct device *dev = &pdev->dev;
1555 char buf[ETHTOOL_FWVERS_LEN];
1556 struct of_phandle_args args;
1557 phy_interface_t phy_mode;
1558 struct ave_private *priv;
1559 struct net_device *ndev;
1560 struct device_node *np;
1561 void __iomem *base;
1562 const char *name;
1563 int i, irq, ret;
1564 u64 dma_mask;
1565 u32 ave_id;
1566
1567 data = of_device_get_match_data(dev);
1568 if (WARN_ON(!data))
1569 return -EINVAL;
1570
1571 np = dev->of_node;
1572 ret = of_get_phy_mode(np, &phy_mode);
1573 if (ret) {
1574 dev_err(dev, "phy-mode not found\n");
1575 return ret;
1576 }
1577
1578 irq = platform_get_irq(pdev, 0);
1579 if (irq < 0)
1580 return irq;
1581
1582 base = devm_platform_ioremap_resource(pdev, 0);
1583 if (IS_ERR(base))
1584 return PTR_ERR(base);
1585
1586 ndev = devm_alloc_etherdev(dev, sizeof(struct ave_private));
1587 if (!ndev) {
1588 dev_err(dev, "can't allocate ethernet device\n");
1589 return -ENOMEM;
1590 }
1591
1592 ndev->netdev_ops = &ave_netdev_ops;
1593 ndev->ethtool_ops = &ave_ethtool_ops;
1594 SET_NETDEV_DEV(ndev, dev);
1595
1596 ndev->features |= (NETIF_F_IP_CSUM | NETIF_F_RXCSUM);
1597 ndev->hw_features |= (NETIF_F_IP_CSUM | NETIF_F_RXCSUM);
1598
1599 ndev->max_mtu = AVE_MAX_ETHFRAME - (ETH_HLEN + ETH_FCS_LEN);
1600
1601 ret = of_get_ethdev_address(np, ndev);
1602 if (ret) {
1603 /* if the mac address is invalid, use random mac address */
1604 eth_hw_addr_random(ndev);
1605 dev_warn(dev, "Using random MAC address: %pM\n",
1606 ndev->dev_addr);
1607 }
1608
1609 priv = netdev_priv(ndev);
1610 priv->base = base;
1611 priv->irq = irq;
1612 priv->ndev = ndev;
1613 priv->msg_enable = netif_msg_init(-1, AVE_DEFAULT_MSG_ENABLE);
1614 priv->phy_mode = phy_mode;
1615 priv->data = data;
1616
1617 if (IS_DESC_64BIT(priv)) {
1618 priv->desc_size = AVE_DESC_SIZE_64;
1619 priv->tx.daddr = AVE_TXDM_64;
1620 priv->rx.daddr = AVE_RXDM_64;
1621 dma_mask = DMA_BIT_MASK(64);
1622 } else {
1623 priv->desc_size = AVE_DESC_SIZE_32;
1624 priv->tx.daddr = AVE_TXDM_32;
1625 priv->rx.daddr = AVE_RXDM_32;
1626 dma_mask = DMA_BIT_MASK(32);
1627 }
1628 ret = dma_set_mask(dev, dma_mask);
1629 if (ret)
1630 return ret;
1631
1632 priv->tx.ndesc = AVE_NR_TXDESC;
1633 priv->rx.ndesc = AVE_NR_RXDESC;
1634
1635 u64_stats_init(&priv->stats_tx.syncp);
1636 u64_stats_init(&priv->stats_rx.syncp);
1637
1638 for (i = 0; i < AVE_MAX_CLKS; i++) {
1639 name = priv->data->clock_names[i];
1640 if (!name)
1641 break;
1642 priv->clk[i] = devm_clk_get(dev, name);
1643 if (IS_ERR(priv->clk[i]))
1644 return PTR_ERR(priv->clk[i]);
1645 priv->nclks++;
1646 }
1647
1648 for (i = 0; i < AVE_MAX_RSTS; i++) {
1649 name = priv->data->reset_names[i];
1650 if (!name)
1651 break;
1652 priv->rst[i] = devm_reset_control_get_shared(dev, name);
1653 if (IS_ERR(priv->rst[i]))
1654 return PTR_ERR(priv->rst[i]);
1655 priv->nrsts++;
1656 }
1657
1658 ret = of_parse_phandle_with_fixed_args(np,
1659 "socionext,syscon-phy-mode",
1660 1, 0, &args);
1661 if (ret) {
1662 dev_err(dev, "can't get syscon-phy-mode property\n");
1663 return ret;
1664 }
1665 priv->regmap = syscon_node_to_regmap(args.np);
1666 of_node_put(args.np);
1667 if (IS_ERR(priv->regmap)) {
1668 dev_err(dev, "can't map syscon-phy-mode\n");
1669 return PTR_ERR(priv->regmap);
1670 }
1671 ret = priv->data->get_pinmode(priv, phy_mode, args.args[0]);
1672 if (ret) {
1673 dev_err(dev, "invalid phy-mode setting\n");
1674 return ret;
1675 }
1676
1677 priv->mdio = devm_mdiobus_alloc(dev);
1678 if (!priv->mdio)
1679 return -ENOMEM;
1680 priv->mdio->priv = ndev;
1681 priv->mdio->parent = dev;
1682 priv->mdio->read = ave_mdiobus_read;
1683 priv->mdio->write = ave_mdiobus_write;
1684 priv->mdio->name = "uniphier-mdio";
1685 snprintf(priv->mdio->id, MII_BUS_ID_SIZE, "%s-%x",
1686 pdev->name, pdev->id);
1687
1688 /* Register as a NAPI supported driver */
1689 netif_napi_add(ndev, &priv->napi_rx, ave_napi_poll_rx);
1690 netif_napi_add_tx(ndev, &priv->napi_tx, ave_napi_poll_tx);
1691
1692 platform_set_drvdata(pdev, ndev);
1693
1694 ret = register_netdev(ndev);
1695 if (ret) {
1696 dev_err(dev, "failed to register netdevice\n");
1697 goto out_del_napi;
1698 }
1699
1700 /* get ID and version */
1701 ave_id = readl(priv->base + AVE_IDR);
1702 ave_hw_read_version(ndev, buf, sizeof(buf));
1703
1704 dev_info(dev, "Socionext %c%c%c%c Ethernet IP %s (irq=%d, phy=%s)\n",
1705 (ave_id >> 24) & 0xff, (ave_id >> 16) & 0xff,
1706 (ave_id >> 8) & 0xff, (ave_id >> 0) & 0xff,
1707 buf, priv->irq, phy_modes(phy_mode));
1708
1709 return 0;
1710
1711 out_del_napi:
1712 netif_napi_del(&priv->napi_rx);
1713 netif_napi_del(&priv->napi_tx);
1714
1715 return ret;
1716 }
1717
ave_remove(struct platform_device * pdev)1718 static void ave_remove(struct platform_device *pdev)
1719 {
1720 struct net_device *ndev = platform_get_drvdata(pdev);
1721 struct ave_private *priv = netdev_priv(ndev);
1722
1723 unregister_netdev(ndev);
1724 netif_napi_del(&priv->napi_rx);
1725 netif_napi_del(&priv->napi_tx);
1726 }
1727
1728 #ifdef CONFIG_PM_SLEEP
ave_suspend(struct device * dev)1729 static int ave_suspend(struct device *dev)
1730 {
1731 struct ethtool_wolinfo wol = { .cmd = ETHTOOL_GWOL };
1732 struct net_device *ndev = dev_get_drvdata(dev);
1733 struct ave_private *priv = netdev_priv(ndev);
1734 int ret = 0;
1735
1736 if (netif_running(ndev)) {
1737 ret = ave_stop(ndev);
1738 netif_device_detach(ndev);
1739 }
1740
1741 ave_ethtool_get_wol(ndev, &wol);
1742 priv->wolopts = wol.wolopts;
1743
1744 return ret;
1745 }
1746
ave_resume(struct device * dev)1747 static int ave_resume(struct device *dev)
1748 {
1749 struct ethtool_wolinfo wol = { .cmd = ETHTOOL_GWOL };
1750 struct net_device *ndev = dev_get_drvdata(dev);
1751 struct ave_private *priv = netdev_priv(ndev);
1752 int ret = 0;
1753
1754 ave_global_reset(ndev);
1755
1756 ret = phy_init_hw(ndev->phydev);
1757 if (ret)
1758 return ret;
1759
1760 ave_ethtool_get_wol(ndev, &wol);
1761 wol.wolopts = priv->wolopts;
1762 __ave_ethtool_set_wol(ndev, &wol);
1763
1764 if (netif_running(ndev)) {
1765 ret = ave_open(ndev);
1766 netif_device_attach(ndev);
1767 }
1768
1769 return ret;
1770 }
1771
1772 static SIMPLE_DEV_PM_OPS(ave_pm_ops, ave_suspend, ave_resume);
1773 #define AVE_PM_OPS (&ave_pm_ops)
1774 #else
1775 #define AVE_PM_OPS NULL
1776 #endif
1777
ave_pro4_get_pinmode(struct ave_private * priv,phy_interface_t phy_mode,u32 arg)1778 static int ave_pro4_get_pinmode(struct ave_private *priv,
1779 phy_interface_t phy_mode, u32 arg)
1780 {
1781 if (arg > 0)
1782 return -EINVAL;
1783
1784 priv->pinmode_mask = SG_ETPINMODE_RMII(0);
1785
1786 switch (phy_mode) {
1787 case PHY_INTERFACE_MODE_RMII:
1788 priv->pinmode_val = SG_ETPINMODE_RMII(0);
1789 break;
1790 case PHY_INTERFACE_MODE_MII:
1791 case PHY_INTERFACE_MODE_RGMII:
1792 case PHY_INTERFACE_MODE_RGMII_ID:
1793 case PHY_INTERFACE_MODE_RGMII_RXID:
1794 case PHY_INTERFACE_MODE_RGMII_TXID:
1795 priv->pinmode_val = 0;
1796 break;
1797 default:
1798 return -EINVAL;
1799 }
1800
1801 return 0;
1802 }
1803
ave_ld11_get_pinmode(struct ave_private * priv,phy_interface_t phy_mode,u32 arg)1804 static int ave_ld11_get_pinmode(struct ave_private *priv,
1805 phy_interface_t phy_mode, u32 arg)
1806 {
1807 if (arg > 0)
1808 return -EINVAL;
1809
1810 priv->pinmode_mask = SG_ETPINMODE_EXTPHY | SG_ETPINMODE_RMII(0);
1811
1812 switch (phy_mode) {
1813 case PHY_INTERFACE_MODE_INTERNAL:
1814 priv->pinmode_val = 0;
1815 break;
1816 case PHY_INTERFACE_MODE_RMII:
1817 priv->pinmode_val = SG_ETPINMODE_EXTPHY | SG_ETPINMODE_RMII(0);
1818 break;
1819 default:
1820 return -EINVAL;
1821 }
1822
1823 return 0;
1824 }
1825
ave_ld20_get_pinmode(struct ave_private * priv,phy_interface_t phy_mode,u32 arg)1826 static int ave_ld20_get_pinmode(struct ave_private *priv,
1827 phy_interface_t phy_mode, u32 arg)
1828 {
1829 if (arg > 0)
1830 return -EINVAL;
1831
1832 priv->pinmode_mask = SG_ETPINMODE_RMII(0);
1833
1834 switch (phy_mode) {
1835 case PHY_INTERFACE_MODE_RMII:
1836 priv->pinmode_val = SG_ETPINMODE_RMII(0);
1837 break;
1838 case PHY_INTERFACE_MODE_RGMII:
1839 case PHY_INTERFACE_MODE_RGMII_ID:
1840 case PHY_INTERFACE_MODE_RGMII_RXID:
1841 case PHY_INTERFACE_MODE_RGMII_TXID:
1842 priv->pinmode_val = 0;
1843 break;
1844 default:
1845 return -EINVAL;
1846 }
1847
1848 return 0;
1849 }
1850
ave_pxs3_get_pinmode(struct ave_private * priv,phy_interface_t phy_mode,u32 arg)1851 static int ave_pxs3_get_pinmode(struct ave_private *priv,
1852 phy_interface_t phy_mode, u32 arg)
1853 {
1854 if (arg > 1)
1855 return -EINVAL;
1856
1857 priv->pinmode_mask = SG_ETPINMODE_RMII(arg);
1858
1859 switch (phy_mode) {
1860 case PHY_INTERFACE_MODE_RMII:
1861 priv->pinmode_val = SG_ETPINMODE_RMII(arg);
1862 break;
1863 case PHY_INTERFACE_MODE_RGMII:
1864 case PHY_INTERFACE_MODE_RGMII_ID:
1865 case PHY_INTERFACE_MODE_RGMII_RXID:
1866 case PHY_INTERFACE_MODE_RGMII_TXID:
1867 priv->pinmode_val = 0;
1868 break;
1869 default:
1870 return -EINVAL;
1871 }
1872
1873 return 0;
1874 }
1875
1876 static const struct ave_soc_data ave_pro4_data = {
1877 .is_desc_64bit = false,
1878 .clock_names = {
1879 "gio", "ether", "ether-gb", "ether-phy",
1880 },
1881 .reset_names = {
1882 "gio", "ether",
1883 },
1884 .get_pinmode = ave_pro4_get_pinmode,
1885 };
1886
1887 static const struct ave_soc_data ave_pxs2_data = {
1888 .is_desc_64bit = false,
1889 .clock_names = {
1890 "ether",
1891 },
1892 .reset_names = {
1893 "ether",
1894 },
1895 .get_pinmode = ave_pro4_get_pinmode,
1896 };
1897
1898 static const struct ave_soc_data ave_ld11_data = {
1899 .is_desc_64bit = false,
1900 .clock_names = {
1901 "ether",
1902 },
1903 .reset_names = {
1904 "ether",
1905 },
1906 .get_pinmode = ave_ld11_get_pinmode,
1907 };
1908
1909 static const struct ave_soc_data ave_ld20_data = {
1910 .is_desc_64bit = true,
1911 .clock_names = {
1912 "ether",
1913 },
1914 .reset_names = {
1915 "ether",
1916 },
1917 .get_pinmode = ave_ld20_get_pinmode,
1918 };
1919
1920 static const struct ave_soc_data ave_pxs3_data = {
1921 .is_desc_64bit = false,
1922 .clock_names = {
1923 "ether",
1924 },
1925 .reset_names = {
1926 "ether",
1927 },
1928 .get_pinmode = ave_pxs3_get_pinmode,
1929 };
1930
1931 static const struct ave_soc_data ave_nx1_data = {
1932 .is_desc_64bit = true,
1933 .clock_names = {
1934 "ether",
1935 },
1936 .reset_names = {
1937 "ether",
1938 },
1939 .get_pinmode = ave_pxs3_get_pinmode,
1940 };
1941
1942 static const struct of_device_id of_ave_match[] = {
1943 {
1944 .compatible = "socionext,uniphier-pro4-ave4",
1945 .data = &ave_pro4_data,
1946 },
1947 {
1948 .compatible = "socionext,uniphier-pxs2-ave4",
1949 .data = &ave_pxs2_data,
1950 },
1951 {
1952 .compatible = "socionext,uniphier-ld11-ave4",
1953 .data = &ave_ld11_data,
1954 },
1955 {
1956 .compatible = "socionext,uniphier-ld20-ave4",
1957 .data = &ave_ld20_data,
1958 },
1959 {
1960 .compatible = "socionext,uniphier-pxs3-ave4",
1961 .data = &ave_pxs3_data,
1962 },
1963 {
1964 .compatible = "socionext,uniphier-nx1-ave4",
1965 .data = &ave_nx1_data,
1966 },
1967 { /* Sentinel */ }
1968 };
1969 MODULE_DEVICE_TABLE(of, of_ave_match);
1970
1971 static struct platform_driver ave_driver = {
1972 .probe = ave_probe,
1973 .remove = ave_remove,
1974 .driver = {
1975 .name = "ave",
1976 .pm = AVE_PM_OPS,
1977 .of_match_table = of_ave_match,
1978 },
1979 };
1980 module_platform_driver(ave_driver);
1981
1982 MODULE_AUTHOR("Kunihiko Hayashi <hayashi.kunihiko@socionext.com>");
1983 MODULE_DESCRIPTION("Socionext UniPhier AVE ethernet driver");
1984 MODULE_LICENSE("GPL v2");
1985