xref: /freebsd/sys/powerpc/ps3/if_glc.c (revision 595e514d0df2bac5b813d35f83e32875dbf16a83)
1 /*-
2  * Copyright (C) 2010 Nathan Whitehorn
3  * All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions
7  * are met:
8  * 1. Redistributions of source code must retain the above copyright
9  *    notice, this list of conditions and the following disclaimer.
10  * 2. Redistributions in binary form must reproduce the above copyright
11  *    notice, this list of conditions and the following disclaimer in the
12  *    documentation and/or other materials provided with the distribution.
13  *
14  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
15  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
16  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
17  * IN NO EVENT SHALL TOOLS GMBH BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
18  * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
19  * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS;
20  * OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
21  * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR
22  * OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF
23  * ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
24  *
25  * $FreeBSD$
26  */
27 
28 #include <sys/param.h>
29 #include <sys/systm.h>
30 #include <sys/sockio.h>
31 #include <sys/endian.h>
32 #include <sys/mbuf.h>
33 #include <sys/module.h>
34 #include <sys/malloc.h>
35 #include <sys/kernel.h>
36 #include <sys/socket.h>
37 
38 #include <vm/vm.h>
39 #include <vm/pmap.h>
40 
41 #include <net/bpf.h>
42 #include <net/if.h>
43 #include <net/if_arp.h>
44 #include <net/ethernet.h>
45 #include <net/if_dl.h>
46 #include <net/if_media.h>
47 #include <net/if_types.h>
48 #include <net/if_vlan_var.h>
49 
50 #include <machine/pio.h>
51 #include <machine/bus.h>
52 #include <machine/platform.h>
53 #include <machine/pmap.h>
54 #include <machine/resource.h>
55 #include <sys/bus.h>
56 #include <sys/rman.h>
57 
58 #include "ps3bus.h"
59 #include "ps3-hvcall.h"
60 #include "if_glcreg.h"
61 
62 static int	glc_probe(device_t);
63 static int	glc_attach(device_t);
64 static void	glc_init(void *xsc);
65 static void	glc_start(struct ifnet *ifp);
66 static int	glc_ioctl(struct ifnet *ifp, u_long cmd, caddr_t data);
67 static void	glc_set_multicast(struct glc_softc *sc);
68 static int	glc_add_rxbuf(struct glc_softc *sc, int idx);
69 static int	glc_add_rxbuf_dma(struct glc_softc *sc, int idx);
70 static int	glc_encap(struct glc_softc *sc, struct mbuf **m_head,
71 		    bus_addr_t *pktdesc);
72 static int	glc_intr_filter(void *xsc);
73 static void	glc_intr(void *xsc);
74 static void	glc_tick(void *xsc);
75 static void	glc_media_status(struct ifnet *ifp, struct ifmediareq *ifmr);
76 static int	glc_media_change(struct ifnet *ifp);
77 
78 static MALLOC_DEFINE(M_GLC, "gelic", "PS3 GELIC ethernet");
79 
80 static device_method_t glc_methods[] = {
81 	/* Device interface */
82 	DEVMETHOD(device_probe,		glc_probe),
83 	DEVMETHOD(device_attach,	glc_attach),
84 
85 	{ 0, 0 }
86 };
87 
88 static driver_t glc_driver = {
89 	"glc",
90 	glc_methods,
91 	sizeof(struct glc_softc)
92 };
93 
94 static devclass_t glc_devclass;
95 
96 DRIVER_MODULE(glc, ps3bus, glc_driver, glc_devclass, 0, 0);
97 
98 static int
99 glc_probe(device_t dev)
100 {
101 
102 	if (ps3bus_get_bustype(dev) != PS3_BUSTYPE_SYSBUS ||
103 	    ps3bus_get_devtype(dev) != PS3_DEVTYPE_GELIC)
104 		return (ENXIO);
105 
106 	device_set_desc(dev, "Playstation 3 GELIC Network Controller");
107 	return (BUS_PROBE_SPECIFIC);
108 }
109 
110 static void
111 glc_getphys(void *xaddr, bus_dma_segment_t *segs, int nsegs, int error)
112 {
113 	if (error != 0)
114 		return;
115 
116 	*(bus_addr_t *)xaddr = segs[0].ds_addr;
117 }
118 
119 static int
120 glc_attach(device_t dev)
121 {
122 	struct glc_softc *sc;
123 	struct glc_txsoft *txs;
124 	uint64_t mac64, val, junk;
125 	int i, err;
126 
127 	sc = device_get_softc(dev);
128 
129 	sc->sc_bus = ps3bus_get_bus(dev);
130 	sc->sc_dev = ps3bus_get_device(dev);
131 	sc->sc_self = dev;
132 
133 	mtx_init(&sc->sc_mtx, device_get_nameunit(dev), MTX_NETWORK_LOCK,
134 	    MTX_DEF);
135 	callout_init_mtx(&sc->sc_tick_ch, &sc->sc_mtx, 0);
136 	sc->next_txdma_slot = 0;
137 	sc->bsy_txdma_slots = 0;
138 	sc->sc_next_rxdma_slot = 0;
139 	sc->first_used_txdma_slot = -1;
140 
141 	/*
142 	 * Shut down existing tasks.
143 	 */
144 
145 	lv1_net_stop_tx_dma(sc->sc_bus, sc->sc_dev, 0);
146 	lv1_net_stop_rx_dma(sc->sc_bus, sc->sc_dev, 0);
147 
148 	sc->sc_ifp = if_alloc(IFT_ETHER);
149 	sc->sc_ifp->if_softc = sc;
150 
151 	/*
152 	 * Get MAC address and VLAN id
153 	 */
154 
155 	lv1_net_control(sc->sc_bus, sc->sc_dev, GELIC_GET_MAC_ADDRESS,
156 	    0, 0, 0, &mac64, &junk);
157 	memcpy(sc->sc_enaddr, &((uint8_t *)&mac64)[2], sizeof(sc->sc_enaddr));
158 	sc->sc_tx_vlan = sc->sc_rx_vlan = -1;
159 	err = lv1_net_control(sc->sc_bus, sc->sc_dev, GELIC_GET_VLAN_ID,
160 	    GELIC_VLAN_TX_ETHERNET, 0, 0, &val, &junk);
161 	if (err == 0)
162 		sc->sc_tx_vlan = val;
163 	err = lv1_net_control(sc->sc_bus, sc->sc_dev, GELIC_GET_VLAN_ID,
164 	    GELIC_VLAN_RX_ETHERNET, 0, 0, &val, &junk);
165 	if (err == 0)
166 		sc->sc_rx_vlan = val;
167 
168 	/*
169 	 * Set up interrupt handler
170 	 */
171 	sc->sc_irqid = 0;
172 	sc->sc_irq = bus_alloc_resource_any(dev, SYS_RES_IRQ, &sc->sc_irqid,
173 	    RF_ACTIVE);
174 	if (sc->sc_irq == NULL) {
175 		device_printf(dev, "Could not allocate IRQ!\n");
176 		mtx_destroy(&sc->sc_mtx);
177 		return (ENXIO);
178 	}
179 
180 	bus_setup_intr(dev, sc->sc_irq,
181 	    INTR_TYPE_NET | INTR_MPSAFE | INTR_ENTROPY,
182 	    glc_intr_filter, glc_intr, sc, &sc->sc_irqctx);
183 	sc->sc_hwirq_status = (uint64_t *)contigmalloc(8, M_GLC, M_ZERO, 0,
184 	    BUS_SPACE_MAXADDR_32BIT, 8, PAGE_SIZE);
185 	lv1_net_set_interrupt_status_indicator(sc->sc_bus, sc->sc_dev,
186 	    vtophys(sc->sc_hwirq_status), 0);
187 	lv1_net_set_interrupt_mask(sc->sc_bus, sc->sc_dev,
188 	    GELIC_INT_RXDONE | GELIC_INT_RXFRAME | GELIC_INT_PHY |
189 	    GELIC_INT_TX_CHAIN_END, 0);
190 
191 	/*
192 	 * Set up DMA.
193 	 */
194 
195 	err = bus_dma_tag_create(bus_get_dma_tag(dev), 32, 0,
196 	    BUS_SPACE_MAXADDR, BUS_SPACE_MAXADDR, NULL, NULL,
197 	    129*sizeof(struct glc_dmadesc), 1, 128*sizeof(struct glc_dmadesc),
198 	    0, NULL,NULL, &sc->sc_dmadesc_tag);
199 
200 	err = bus_dmamem_alloc(sc->sc_dmadesc_tag, (void **)&sc->sc_txdmadesc,
201 	    BUS_DMA_WAITOK | BUS_DMA_COHERENT | BUS_DMA_ZERO,
202 	    &sc->sc_txdmadesc_map);
203 	err = bus_dmamap_load(sc->sc_dmadesc_tag, sc->sc_txdmadesc_map,
204 	    sc->sc_txdmadesc, 128*sizeof(struct glc_dmadesc), glc_getphys,
205 	    &sc->sc_txdmadesc_phys, 0);
206 	err = bus_dmamem_alloc(sc->sc_dmadesc_tag, (void **)&sc->sc_rxdmadesc,
207 	    BUS_DMA_WAITOK | BUS_DMA_COHERENT | BUS_DMA_ZERO,
208 	    &sc->sc_rxdmadesc_map);
209 	err = bus_dmamap_load(sc->sc_dmadesc_tag, sc->sc_rxdmadesc_map,
210 	    sc->sc_rxdmadesc, 128*sizeof(struct glc_dmadesc), glc_getphys,
211 	    &sc->sc_rxdmadesc_phys, 0);
212 
213 	err = bus_dma_tag_create(bus_get_dma_tag(dev), 128, 0,
214 	    BUS_SPACE_MAXADDR, BUS_SPACE_MAXADDR, NULL, NULL,
215 	    BUS_SPACE_MAXSIZE_32BIT, 0, BUS_SPACE_MAXSIZE_32BIT, 0, NULL,NULL,
216 	    &sc->sc_rxdma_tag);
217 	err = bus_dma_tag_create(bus_get_dma_tag(dev), 1, 0,
218 	    BUS_SPACE_MAXADDR, BUS_SPACE_MAXADDR, NULL, NULL,
219 	    BUS_SPACE_MAXSIZE_32BIT, 16, BUS_SPACE_MAXSIZE_32BIT, 0, NULL,NULL,
220 	    &sc->sc_txdma_tag);
221 
222 	/* init transmit descriptors */
223 	STAILQ_INIT(&sc->sc_txfreeq);
224 	STAILQ_INIT(&sc->sc_txdirtyq);
225 
226 	/* create TX DMA maps */
227 	err = ENOMEM;
228 	for (i = 0; i < GLC_MAX_TX_PACKETS; i++) {
229 		txs = &sc->sc_txsoft[i];
230 		txs->txs_mbuf = NULL;
231 		err = bus_dmamap_create(sc->sc_txdma_tag, 0, &txs->txs_dmamap);
232 		if (err) {
233 			device_printf(dev,
234 			    "unable to create TX DMA map %d, error = %d\n",
235 			    i, err);
236 		}
237 		STAILQ_INSERT_TAIL(&sc->sc_txfreeq, txs, txs_q);
238 	}
239 
240 	/* Create the receive buffer DMA maps. */
241 	for (i = 0; i < GLC_MAX_RX_PACKETS; i++) {
242 		err = bus_dmamap_create(sc->sc_rxdma_tag, 0,
243 		    &sc->sc_rxsoft[i].rxs_dmamap);
244 		if (err) {
245 			device_printf(dev,
246 			    "unable to create RX DMA map %d, error = %d\n",
247 			    i, err);
248 		}
249 		sc->sc_rxsoft[i].rxs_mbuf = NULL;
250 	}
251 
252 	/*
253 	 * Attach to network stack
254 	 */
255 
256 	if_initname(sc->sc_ifp, device_get_name(dev), device_get_unit(dev));
257 	sc->sc_ifp->if_mtu = ETHERMTU;
258 	sc->sc_ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST;
259 	sc->sc_ifp->if_hwassist = CSUM_TCP | CSUM_UDP;
260 	sc->sc_ifp->if_capabilities = IFCAP_HWCSUM | IFCAP_RXCSUM;
261 	sc->sc_ifp->if_capenable = IFCAP_HWCSUM | IFCAP_RXCSUM;
262 	sc->sc_ifp->if_start = glc_start;
263 	sc->sc_ifp->if_ioctl = glc_ioctl;
264 	sc->sc_ifp->if_init = glc_init;
265 
266 	ifmedia_init(&sc->sc_media, IFM_IMASK, glc_media_change,
267 	    glc_media_status);
268 	ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_10_T, 0, NULL);
269 	ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_10_T | IFM_FDX, 0, NULL);
270 	ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_100_TX, 0, NULL);
271 	ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_100_TX | IFM_FDX, 0, NULL);
272 	ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_1000_T | IFM_FDX, 0, NULL);
273 	ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_AUTO, 0, NULL);
274 	ifmedia_set(&sc->sc_media, IFM_ETHER | IFM_AUTO);
275 
276 	IFQ_SET_MAXLEN(&sc->sc_ifp->if_snd, GLC_MAX_TX_PACKETS);
277 	sc->sc_ifp->if_snd.ifq_drv_maxlen = GLC_MAX_TX_PACKETS;
278 	IFQ_SET_READY(&sc->sc_ifp->if_snd);
279 
280 	ether_ifattach(sc->sc_ifp, sc->sc_enaddr);
281 	sc->sc_ifp->if_hwassist = 0;
282 
283 	return (0);
284 
285 	mtx_destroy(&sc->sc_mtx);
286 	if_free(sc->sc_ifp);
287 	return (ENXIO);
288 }
289 
290 static void
291 glc_init_locked(struct glc_softc *sc)
292 {
293 	int i, error;
294 	struct glc_rxsoft *rxs;
295 	struct glc_txsoft *txs;
296 
297 	mtx_assert(&sc->sc_mtx, MA_OWNED);
298 
299 	lv1_net_stop_tx_dma(sc->sc_bus, sc->sc_dev, 0);
300 	lv1_net_stop_rx_dma(sc->sc_bus, sc->sc_dev, 0);
301 
302 	glc_set_multicast(sc);
303 
304 	for (i = 0; i < GLC_MAX_RX_PACKETS; i++) {
305 		rxs = &sc->sc_rxsoft[i];
306 		rxs->rxs_desc_slot = i;
307 
308 		if (rxs->rxs_mbuf == NULL) {
309 			glc_add_rxbuf(sc, i);
310 
311 			if (rxs->rxs_mbuf == NULL) {
312 				rxs->rxs_desc_slot = -1;
313 				break;
314 			}
315 		}
316 
317 		glc_add_rxbuf_dma(sc, i);
318 		bus_dmamap_sync(sc->sc_dmadesc_tag, sc->sc_rxdmadesc_map,
319 		    BUS_DMASYNC_PREREAD);
320 	}
321 
322 	/* Clear TX dirty queue */
323 	while ((txs = STAILQ_FIRST(&sc->sc_txdirtyq)) != NULL) {
324 		STAILQ_REMOVE_HEAD(&sc->sc_txdirtyq, txs_q);
325 		bus_dmamap_unload(sc->sc_txdma_tag, txs->txs_dmamap);
326 
327 		if (txs->txs_mbuf != NULL) {
328 			m_freem(txs->txs_mbuf);
329 			txs->txs_mbuf = NULL;
330 		}
331 
332 		STAILQ_INSERT_TAIL(&sc->sc_txfreeq, txs, txs_q);
333 	}
334 	sc->first_used_txdma_slot = -1;
335 	sc->bsy_txdma_slots = 0;
336 
337 	error = lv1_net_start_rx_dma(sc->sc_bus, sc->sc_dev,
338 	    sc->sc_rxsoft[0].rxs_desc, 0);
339 	if (error != 0)
340 		device_printf(sc->sc_self,
341 		    "lv1_net_start_rx_dma error: %d\n", error);
342 
343 	sc->sc_ifp->if_drv_flags |= IFF_DRV_RUNNING;
344 	sc->sc_ifp->if_drv_flags &= ~IFF_DRV_OACTIVE;
345 	sc->sc_ifpflags = sc->sc_ifp->if_flags;
346 
347 	sc->sc_wdog_timer = 0;
348 	callout_reset(&sc->sc_tick_ch, hz, glc_tick, sc);
349 }
350 
351 static void
352 glc_stop(void *xsc)
353 {
354 	struct glc_softc *sc = xsc;
355 
356 	mtx_assert(&sc->sc_mtx, MA_OWNED);
357 
358 	lv1_net_stop_tx_dma(sc->sc_bus, sc->sc_dev, 0);
359 	lv1_net_stop_rx_dma(sc->sc_bus, sc->sc_dev, 0);
360 }
361 
362 static void
363 glc_init(void *xsc)
364 {
365 	struct glc_softc *sc = xsc;
366 
367 	mtx_lock(&sc->sc_mtx);
368 	glc_init_locked(sc);
369 	mtx_unlock(&sc->sc_mtx);
370 }
371 
372 static void
373 glc_tick(void *xsc)
374 {
375 	struct glc_softc *sc = xsc;
376 
377 	mtx_assert(&sc->sc_mtx, MA_OWNED);
378 
379 	/*
380 	 * XXX: Sometimes the RX queue gets stuck. Poke it periodically until
381 	 * we figure out why. This will fail harmlessly if the RX queue is
382 	 * already running.
383 	 */
384 	lv1_net_start_rx_dma(sc->sc_bus, sc->sc_dev,
385 	    sc->sc_rxsoft[sc->sc_next_rxdma_slot].rxs_desc, 0);
386 
387 	if (sc->sc_wdog_timer == 0 || --sc->sc_wdog_timer != 0) {
388 		callout_reset(&sc->sc_tick_ch, hz, glc_tick, sc);
389 		return;
390 	}
391 
392 	/* Problems */
393 	device_printf(sc->sc_self, "device timeout\n");
394 
395 	glc_init_locked(sc);
396 }
397 
398 static void
399 glc_start_locked(struct ifnet *ifp)
400 {
401 	struct glc_softc *sc = ifp->if_softc;
402 	bus_addr_t first, pktdesc;
403 	int kickstart = 0;
404 	int error;
405 	struct mbuf *mb_head;
406 
407 	mtx_assert(&sc->sc_mtx, MA_OWNED);
408 	first = 0;
409 
410 	if ((ifp->if_drv_flags & (IFF_DRV_RUNNING | IFF_DRV_OACTIVE)) !=
411 	    IFF_DRV_RUNNING)
412 		return;
413 
414 	if (STAILQ_EMPTY(&sc->sc_txdirtyq))
415 		kickstart = 1;
416 
417 	while (!IFQ_DRV_IS_EMPTY(&ifp->if_snd)) {
418 		IFQ_DRV_DEQUEUE(&ifp->if_snd, mb_head);
419 
420 		if (mb_head == NULL)
421 			break;
422 
423 		/* Check if the ring buffer is full */
424 		if (sc->bsy_txdma_slots > 125) {
425 			/* Put the packet back and stop */
426 			ifp->if_drv_flags |= IFF_DRV_OACTIVE;
427 			IFQ_DRV_PREPEND(&ifp->if_snd, mb_head);
428 			break;
429 		}
430 
431 		BPF_MTAP(ifp, mb_head);
432 
433 		if (sc->sc_tx_vlan >= 0)
434 			mb_head = ether_vlanencap(mb_head, sc->sc_tx_vlan);
435 
436 		if (glc_encap(sc, &mb_head, &pktdesc)) {
437 			ifp->if_drv_flags |= IFF_DRV_OACTIVE;
438 			break;
439 		}
440 
441 		if (first == 0)
442 			first = pktdesc;
443 	}
444 
445 	if (kickstart && first != 0) {
446 		error = lv1_net_start_tx_dma(sc->sc_bus, sc->sc_dev, first, 0);
447 		if (error != 0)
448 			device_printf(sc->sc_self,
449 			    "lv1_net_start_tx_dma error: %d\n", error);
450 		sc->sc_wdog_timer = 5;
451 	}
452 }
453 
454 static void
455 glc_start(struct ifnet *ifp)
456 {
457 	struct glc_softc *sc = ifp->if_softc;
458 
459 	mtx_lock(&sc->sc_mtx);
460 	glc_start_locked(ifp);
461 	mtx_unlock(&sc->sc_mtx);
462 }
463 
464 static int
465 glc_ioctl(struct ifnet *ifp, u_long cmd, caddr_t data)
466 {
467 	struct glc_softc *sc = ifp->if_softc;
468 	struct ifreq *ifr = (struct ifreq *)data;
469 	int err = 0;
470 
471 	switch (cmd) {
472 	case SIOCSIFFLAGS:
473                 mtx_lock(&sc->sc_mtx);
474 		if ((ifp->if_flags & IFF_UP) != 0) {
475 			if ((ifp->if_drv_flags & IFF_DRV_RUNNING) != 0 &&
476 			   ((ifp->if_flags ^ sc->sc_ifpflags) &
477 			    (IFF_ALLMULTI | IFF_PROMISC)) != 0)
478 				glc_set_multicast(sc);
479 			else
480 				glc_init_locked(sc);
481 		}
482 		else if ((ifp->if_drv_flags & IFF_DRV_RUNNING) != 0)
483 			glc_stop(sc);
484 		sc->sc_ifpflags = ifp->if_flags;
485 		mtx_unlock(&sc->sc_mtx);
486 		break;
487 	case SIOCADDMULTI:
488 	case SIOCDELMULTI:
489                 mtx_lock(&sc->sc_mtx);
490 		glc_set_multicast(sc);
491                 mtx_unlock(&sc->sc_mtx);
492 		break;
493 	case SIOCGIFMEDIA:
494 	case SIOCSIFMEDIA:
495 		err = ifmedia_ioctl(ifp, ifr, &sc->sc_media, cmd);
496 		break;
497 	default:
498 		err = ether_ioctl(ifp, cmd, data);
499 		break;
500 	}
501 
502 	return (err);
503 }
504 
505 static void
506 glc_set_multicast(struct glc_softc *sc)
507 {
508 	struct ifnet *ifp = sc->sc_ifp;
509 	struct ifmultiaddr *inm;
510 	uint64_t addr;
511 	int naddrs;
512 
513 	/* Clear multicast filter */
514 	lv1_net_remove_multicast_address(sc->sc_bus, sc->sc_dev, 0, 1);
515 
516 	/* Add broadcast */
517 	lv1_net_add_multicast_address(sc->sc_bus, sc->sc_dev,
518 	    0xffffffffffffL, 0);
519 
520 	if ((ifp->if_flags & IFF_ALLMULTI) != 0) {
521 		lv1_net_add_multicast_address(sc->sc_bus, sc->sc_dev, 0, 1);
522 	} else {
523 		if_maddr_rlock(ifp);
524 		naddrs = 1; /* Include broadcast */
525 		TAILQ_FOREACH(inm, &ifp->if_multiaddrs, ifma_link) {
526 			if (inm->ifma_addr->sa_family != AF_LINK)
527 				continue;
528 			addr = 0;
529 			memcpy(&((uint8_t *)(&addr))[2],
530 			    LLADDR((struct sockaddr_dl *)inm->ifma_addr),
531 			    ETHER_ADDR_LEN);
532 
533 			lv1_net_add_multicast_address(sc->sc_bus, sc->sc_dev,
534 			    addr, 0);
535 
536 			/*
537 			 * Filter can only hold 32 addresses, so fall back to
538 			 * the IFF_ALLMULTI case if we have too many.
539 			 */
540 			if (++naddrs >= 32) {
541 				lv1_net_add_multicast_address(sc->sc_bus,
542 				    sc->sc_dev, 0, 1);
543 				break;
544 			}
545 		}
546 		if_maddr_runlock(ifp);
547 	}
548 }
549 
550 static int
551 glc_add_rxbuf(struct glc_softc *sc, int idx)
552 {
553 	struct glc_rxsoft *rxs = &sc->sc_rxsoft[idx];
554 	struct mbuf *m;
555 	bus_dma_segment_t segs[1];
556 	int error, nsegs;
557 
558 	m = m_getcl(M_NOWAIT, MT_DATA, M_PKTHDR);
559 	if (m == NULL)
560 		return (ENOBUFS);
561 	m->m_len = m->m_pkthdr.len = m->m_ext.ext_size;
562 
563 	if (rxs->rxs_mbuf != NULL) {
564 		bus_dmamap_sync(sc->sc_rxdma_tag, rxs->rxs_dmamap,
565 		    BUS_DMASYNC_POSTREAD);
566 		bus_dmamap_unload(sc->sc_rxdma_tag, rxs->rxs_dmamap);
567 	}
568 
569 	error = bus_dmamap_load_mbuf_sg(sc->sc_rxdma_tag, rxs->rxs_dmamap, m,
570 	    segs, &nsegs, BUS_DMA_NOWAIT);
571 	if (error != 0) {
572 		device_printf(sc->sc_self,
573 		    "cannot load RS DMA map %d, error = %d\n", idx, error);
574 		m_freem(m);
575 		return (error);
576 	}
577 	/* If nsegs is wrong then the stack is corrupt. */
578 	KASSERT(nsegs == 1,
579 	    ("%s: too many DMA segments (%d)", __func__, nsegs));
580 	rxs->rxs_mbuf = m;
581 	rxs->segment = segs[0];
582 
583 	bus_dmamap_sync(sc->sc_rxdma_tag, rxs->rxs_dmamap, BUS_DMASYNC_PREREAD);
584 
585 	return (0);
586 }
587 
588 static int
589 glc_add_rxbuf_dma(struct glc_softc *sc, int idx)
590 {
591 	struct glc_rxsoft *rxs = &sc->sc_rxsoft[idx];
592 
593 	bzero(&sc->sc_rxdmadesc[idx], sizeof(sc->sc_rxdmadesc[idx]));
594 	sc->sc_rxdmadesc[idx].paddr = rxs->segment.ds_addr;
595 	sc->sc_rxdmadesc[idx].len = rxs->segment.ds_len;
596 	sc->sc_rxdmadesc[idx].next = sc->sc_rxdmadesc_phys +
597 	    ((idx + 1) % GLC_MAX_RX_PACKETS)*sizeof(sc->sc_rxdmadesc[idx]);
598 	sc->sc_rxdmadesc[idx].cmd_stat = GELIC_DESCR_OWNED;
599 
600 	rxs->rxs_desc_slot = idx;
601 	rxs->rxs_desc = sc->sc_rxdmadesc_phys + idx*sizeof(struct glc_dmadesc);
602 
603         return (0);
604 }
605 
606 static int
607 glc_encap(struct glc_softc *sc, struct mbuf **m_head, bus_addr_t *pktdesc)
608 {
609 	bus_dma_segment_t segs[16];
610 	struct glc_txsoft *txs;
611 	struct mbuf *m;
612 	bus_addr_t firstslotphys;
613 	int i, idx, nsegs, nsegs_max;
614 	int err = 0;
615 
616 	/* Max number of segments is the number of free DMA slots */
617 	nsegs_max = 128 - sc->bsy_txdma_slots;
618 
619 	if (nsegs_max > 16 || sc->first_used_txdma_slot < 0)
620 		nsegs_max = 16;
621 
622 	/* Get a work queue entry. */
623 	if ((txs = STAILQ_FIRST(&sc->sc_txfreeq)) == NULL) {
624 		/* Ran out of descriptors. */
625 		return (ENOBUFS);
626 	}
627 
628 	nsegs = 0;
629 	for (m = *m_head; m != NULL; m = m->m_next)
630 		nsegs++;
631 
632 	if (nsegs > nsegs_max) {
633 		m = m_collapse(*m_head, M_NOWAIT, nsegs_max);
634 		if (m == NULL) {
635 			m_freem(*m_head);
636 			*m_head = NULL;
637 			return (ENOBUFS);
638 		}
639 		*m_head = m;
640 	}
641 
642 	err = bus_dmamap_load_mbuf_sg(sc->sc_txdma_tag, txs->txs_dmamap,
643 	    *m_head, segs, &nsegs, BUS_DMA_NOWAIT);
644 	if (err != 0) {
645 		m_freem(*m_head);
646 		*m_head = NULL;
647 		return (err);
648 	}
649 
650 	KASSERT(nsegs <= 128 - sc->bsy_txdma_slots,
651 	    ("GLC: Mapped too many (%d) DMA segments with %d available",
652 	    nsegs, 128 - sc->bsy_txdma_slots));
653 
654 	if (nsegs == 0) {
655 		m_freem(*m_head);
656 		*m_head = NULL;
657 		return (EIO);
658 	}
659 
660 	txs->txs_ndescs = nsegs;
661 	txs->txs_firstdesc = sc->next_txdma_slot;
662 
663 	idx = txs->txs_firstdesc;
664 	firstslotphys = sc->sc_txdmadesc_phys +
665 	    txs->txs_firstdesc*sizeof(struct glc_dmadesc);
666 
667 	for (i = 0; i < nsegs; i++) {
668 		bzero(&sc->sc_txdmadesc[idx], sizeof(sc->sc_txdmadesc[idx]));
669 		sc->sc_txdmadesc[idx].paddr = segs[i].ds_addr;
670 		sc->sc_txdmadesc[idx].len = segs[i].ds_len;
671 		sc->sc_txdmadesc[idx].next = sc->sc_txdmadesc_phys +
672 		    ((idx + 1) % GLC_MAX_TX_PACKETS)*sizeof(struct glc_dmadesc);
673 		sc->sc_txdmadesc[idx].cmd_stat |= GELIC_CMDSTAT_NOIPSEC;
674 
675 		if (i+1 == nsegs) {
676 			txs->txs_lastdesc = idx;
677 			sc->sc_txdmadesc[idx].next = 0;
678 			sc->sc_txdmadesc[idx].cmd_stat |= GELIC_CMDSTAT_LAST;
679 		}
680 
681 		if ((*m_head)->m_pkthdr.csum_flags & CSUM_TCP)
682 			sc->sc_txdmadesc[idx].cmd_stat |= GELIC_CMDSTAT_CSUM_TCP;
683 		if ((*m_head)->m_pkthdr.csum_flags & CSUM_UDP)
684 			sc->sc_txdmadesc[idx].cmd_stat |= GELIC_CMDSTAT_CSUM_UDP;
685 		sc->sc_txdmadesc[idx].cmd_stat |= GELIC_DESCR_OWNED;
686 
687 		idx = (idx + 1) % GLC_MAX_TX_PACKETS;
688 	}
689 	sc->next_txdma_slot = idx;
690 	sc->bsy_txdma_slots += nsegs;
691 	if (txs->txs_firstdesc != 0)
692 		idx = txs->txs_firstdesc - 1;
693 	else
694 		idx = GLC_MAX_TX_PACKETS - 1;
695 
696 	if (sc->first_used_txdma_slot < 0)
697 		sc->first_used_txdma_slot = txs->txs_firstdesc;
698 
699 	bus_dmamap_sync(sc->sc_txdma_tag, txs->txs_dmamap,
700 	    BUS_DMASYNC_PREWRITE);
701 	sc->sc_txdmadesc[idx].next = firstslotphys;
702 
703 	STAILQ_REMOVE_HEAD(&sc->sc_txfreeq, txs_q);
704 	STAILQ_INSERT_TAIL(&sc->sc_txdirtyq, txs, txs_q);
705 	txs->txs_mbuf = *m_head;
706 	*pktdesc = firstslotphys;
707 
708 	return (0);
709 }
710 
711 static void
712 glc_rxintr(struct glc_softc *sc)
713 {
714 	int i, restart_rxdma, error;
715 	struct mbuf *m;
716 	struct ifnet *ifp = sc->sc_ifp;
717 
718 	bus_dmamap_sync(sc->sc_dmadesc_tag, sc->sc_rxdmadesc_map,
719 	    BUS_DMASYNC_POSTREAD);
720 
721 	restart_rxdma = 0;
722 	while ((sc->sc_rxdmadesc[sc->sc_next_rxdma_slot].cmd_stat &
723 	   GELIC_DESCR_OWNED) == 0) {
724 		i = sc->sc_next_rxdma_slot;
725 		sc->sc_next_rxdma_slot++;
726 		if (sc->sc_next_rxdma_slot >= GLC_MAX_RX_PACKETS)
727 			sc->sc_next_rxdma_slot = 0;
728 
729 		if (sc->sc_rxdmadesc[i].cmd_stat & GELIC_CMDSTAT_CHAIN_END)
730 			restart_rxdma = 1;
731 
732 		if (sc->sc_rxdmadesc[i].rxerror & GELIC_RXERRORS) {
733 			ifp->if_ierrors++;
734 			goto requeue;
735 		}
736 
737 		m = sc->sc_rxsoft[i].rxs_mbuf;
738 		if (sc->sc_rxdmadesc[i].data_stat & GELIC_RX_IPCSUM) {
739 			m->m_pkthdr.csum_flags |=
740 			    CSUM_IP_CHECKED | CSUM_IP_VALID;
741 		}
742 		if (sc->sc_rxdmadesc[i].data_stat & GELIC_RX_TCPUDPCSUM) {
743 			m->m_pkthdr.csum_flags |=
744 			    CSUM_DATA_VALID | CSUM_PSEUDO_HDR;
745 			m->m_pkthdr.csum_data = 0xffff;
746 		}
747 
748 		if (glc_add_rxbuf(sc, i)) {
749 			ifp->if_ierrors++;
750 			goto requeue;
751 		}
752 
753 		ifp->if_ipackets++;
754 		m->m_pkthdr.rcvif = ifp;
755 		m->m_len = sc->sc_rxdmadesc[i].valid_size;
756 		m->m_pkthdr.len = m->m_len;
757 
758 		/*
759 		 * Remove VLAN tag. Even on early firmwares that do not allow
760 		 * multiple VLANs, the VLAN tag is still in place here.
761 		 */
762 		m_adj(m, 2);
763 
764 		mtx_unlock(&sc->sc_mtx);
765 		(*ifp->if_input)(ifp, m);
766 		mtx_lock(&sc->sc_mtx);
767 
768 	    requeue:
769 		glc_add_rxbuf_dma(sc, i);
770 	}
771 
772 	bus_dmamap_sync(sc->sc_dmadesc_tag, sc->sc_rxdmadesc_map,
773 	    BUS_DMASYNC_PREWRITE);
774 
775 	if (restart_rxdma) {
776 		error = lv1_net_start_rx_dma(sc->sc_bus, sc->sc_dev,
777 		    sc->sc_rxsoft[sc->sc_next_rxdma_slot].rxs_desc, 0);
778 		if (error != 0)
779 			device_printf(sc->sc_self,
780 			    "lv1_net_start_rx_dma error: %d\n", error);
781 	}
782 }
783 
784 static void
785 glc_txintr(struct glc_softc *sc)
786 {
787 	struct ifnet *ifp = sc->sc_ifp;
788 	struct glc_txsoft *txs;
789 	int progress = 0, kickstart = 0, error;
790 
791 	bus_dmamap_sync(sc->sc_dmadesc_tag, sc->sc_txdmadesc_map,
792 	    BUS_DMASYNC_POSTREAD);
793 
794 	while ((txs = STAILQ_FIRST(&sc->sc_txdirtyq)) != NULL) {
795 		if (sc->sc_txdmadesc[txs->txs_lastdesc].cmd_stat
796 		    & GELIC_DESCR_OWNED)
797 			break;
798 
799 		STAILQ_REMOVE_HEAD(&sc->sc_txdirtyq, txs_q);
800 		bus_dmamap_unload(sc->sc_txdma_tag, txs->txs_dmamap);
801 		sc->bsy_txdma_slots -= txs->txs_ndescs;
802 
803 		if (txs->txs_mbuf != NULL) {
804 			m_freem(txs->txs_mbuf);
805 			txs->txs_mbuf = NULL;
806 		}
807 
808 		if ((sc->sc_txdmadesc[txs->txs_lastdesc].cmd_stat & 0xf0000000)
809 		    != 0) {
810 			lv1_net_stop_tx_dma(sc->sc_bus, sc->sc_dev, 0);
811 			kickstart = 1;
812 			ifp->if_oerrors++;
813 		}
814 
815 		if (sc->sc_txdmadesc[txs->txs_lastdesc].cmd_stat &
816 		    GELIC_CMDSTAT_CHAIN_END)
817 			kickstart = 1;
818 
819 		STAILQ_INSERT_TAIL(&sc->sc_txfreeq, txs, txs_q);
820 		ifp->if_opackets++;
821 		progress = 1;
822 	}
823 
824 	if (txs != NULL)
825 		sc->first_used_txdma_slot = txs->txs_firstdesc;
826 	else
827 		sc->first_used_txdma_slot = -1;
828 
829 	if (kickstart || txs != NULL) {
830 		/* Speculatively (or necessarily) start the TX queue again */
831 		error = lv1_net_start_tx_dma(sc->sc_bus, sc->sc_dev,
832 		    sc->sc_txdmadesc_phys +
833 		    txs->txs_firstdesc*sizeof(struct glc_dmadesc), 0);
834 		if (error != 0)
835 			device_printf(sc->sc_self,
836 			    "lv1_net_start_tx_dma error: %d\n", error);
837 	}
838 
839 	if (progress) {
840 		/*
841 		 * We freed some descriptors, so reset IFF_DRV_OACTIVE
842 		 * and restart.
843 		 */
844 		ifp->if_drv_flags &= ~IFF_DRV_OACTIVE;
845 		sc->sc_wdog_timer = STAILQ_EMPTY(&sc->sc_txdirtyq) ? 0 : 5;
846 
847 		if ((ifp->if_drv_flags & IFF_DRV_RUNNING) &&
848 		    !IFQ_DRV_IS_EMPTY(&ifp->if_snd))
849 			glc_start_locked(ifp);
850 	}
851 }
852 
853 static int
854 glc_intr_filter(void *xsc)
855 {
856 	struct glc_softc *sc = xsc;
857 
858 	powerpc_sync();
859 	atomic_set_64(&sc->sc_interrupt_status, *sc->sc_hwirq_status);
860 	return (FILTER_SCHEDULE_THREAD);
861 }
862 
863 static void
864 glc_intr(void *xsc)
865 {
866 	struct glc_softc *sc = xsc;
867 	uint64_t status, linkstat, junk;
868 
869 	mtx_lock(&sc->sc_mtx);
870 
871 	status = atomic_readandclear_64(&sc->sc_interrupt_status);
872 
873 	if (status == 0) {
874 		mtx_unlock(&sc->sc_mtx);
875 		return;
876 	}
877 
878 	if (status & (GELIC_INT_RXDONE | GELIC_INT_RXFRAME))
879 		glc_rxintr(sc);
880 
881 	if (status & (GELIC_INT_TXDONE | GELIC_INT_TX_CHAIN_END))
882 		glc_txintr(sc);
883 
884 	if (status & GELIC_INT_PHY) {
885 		lv1_net_control(sc->sc_bus, sc->sc_dev, GELIC_GET_LINK_STATUS,
886 		    GELIC_VLAN_TX_ETHERNET, 0, 0, &linkstat, &junk);
887 
888 		linkstat = (linkstat & GELIC_LINK_UP) ?
889 		    LINK_STATE_UP : LINK_STATE_DOWN;
890 		if (linkstat != sc->sc_ifp->if_link_state)
891 			if_link_state_change(sc->sc_ifp, linkstat);
892 	}
893 
894 	mtx_unlock(&sc->sc_mtx);
895 }
896 
897 static void
898 glc_media_status(struct ifnet *ifp, struct ifmediareq *ifmr)
899 {
900 	struct glc_softc *sc = ifp->if_softc;
901 	uint64_t status, junk;
902 
903 	ifmr->ifm_status = IFM_AVALID;
904 	ifmr->ifm_active = IFM_ETHER;
905 
906 	lv1_net_control(sc->sc_bus, sc->sc_dev, GELIC_GET_LINK_STATUS,
907 	    GELIC_VLAN_TX_ETHERNET, 0, 0, &status, &junk);
908 
909 	if (status & GELIC_LINK_UP)
910 		ifmr->ifm_status |= IFM_ACTIVE;
911 
912 	if (status & GELIC_SPEED_10)
913 		ifmr->ifm_active |= IFM_10_T;
914 	else if (status & GELIC_SPEED_100)
915 		ifmr->ifm_active |= IFM_100_TX;
916 	else if (status & GELIC_SPEED_1000)
917 		ifmr->ifm_active |= IFM_1000_T;
918 
919 	if (status & GELIC_FULL_DUPLEX)
920 		ifmr->ifm_active |= IFM_FDX;
921 	else
922 		ifmr->ifm_active |= IFM_HDX;
923 }
924 
925 static int
926 glc_media_change(struct ifnet *ifp)
927 {
928 	struct glc_softc *sc = ifp->if_softc;
929 	uint64_t mode, junk;
930 	int result;
931 
932 	if (IFM_TYPE(sc->sc_media.ifm_media) != IFM_ETHER)
933 		return (EINVAL);
934 
935 	switch (IFM_SUBTYPE(sc->sc_media.ifm_media)) {
936 	case IFM_AUTO:
937 		mode = GELIC_AUTO_NEG;
938 		break;
939 	case IFM_10_T:
940 		mode = GELIC_SPEED_10;
941 		break;
942 	case IFM_100_TX:
943 		mode = GELIC_SPEED_100;
944 		break;
945 	case IFM_1000_T:
946 		mode = GELIC_SPEED_1000 | GELIC_FULL_DUPLEX;
947 		break;
948 	default:
949 		return (EINVAL);
950 	}
951 
952 	if (IFM_OPTIONS(sc->sc_media.ifm_media) & IFM_FDX)
953 		mode |= GELIC_FULL_DUPLEX;
954 
955 	result = lv1_net_control(sc->sc_bus, sc->sc_dev, GELIC_SET_LINK_MODE,
956 	    GELIC_VLAN_TX_ETHERNET, mode, 0, &junk, &junk);
957 
958 	return (result ? EIO : 0);
959 }
960 
961