xref: /freebsd/sys/powerpc/ps3/if_glc.c (revision 6574b8ed19b093f0af09501d2c9676c28993cb97)
1 /*-
2  * Copyright (C) 2010 Nathan Whitehorn
3  * All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions
7  * are met:
8  * 1. Redistributions of source code must retain the above copyright
9  *    notice, this list of conditions and the following disclaimer.
10  * 2. Redistributions in binary form must reproduce the above copyright
11  *    notice, this list of conditions and the following disclaimer in the
12  *    documentation and/or other materials provided with the distribution.
13  *
14  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
15  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
16  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
17  * IN NO EVENT SHALL TOOLS GMBH BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
18  * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
19  * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS;
20  * OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
21  * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR
22  * OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF
23  * ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
24  *
25  * $FreeBSD$
26  */
27 
28 #include <sys/param.h>
29 #include <sys/systm.h>
30 #include <sys/sockio.h>
31 #include <sys/endian.h>
32 #include <sys/lock.h>
33 #include <sys/mbuf.h>
34 #include <sys/module.h>
35 #include <sys/malloc.h>
36 #include <sys/mutex.h>
37 #include <sys/kernel.h>
38 #include <sys/socket.h>
39 
40 #include <vm/vm.h>
41 #include <vm/pmap.h>
42 
43 #include <net/bpf.h>
44 #include <net/if.h>
45 #include <net/if_var.h>
46 #include <net/ethernet.h>
47 #include <net/if_media.h>
48 #include <net/if_types.h>
49 #include <net/if_dl.h>
50 
51 #include <machine/pio.h>
52 #include <machine/bus.h>
53 #include <machine/platform.h>
54 #include <machine/pmap.h>
55 #include <machine/resource.h>
56 #include <sys/bus.h>
57 #include <sys/rman.h>
58 
59 #include "ps3bus.h"
60 #include "ps3-hvcall.h"
61 #include "if_glcreg.h"
62 
63 static int	glc_probe(device_t);
64 static int	glc_attach(device_t);
65 static void	glc_init(void *xsc);
66 static void	glc_start(struct ifnet *ifp);
67 static int	glc_ioctl(struct ifnet *ifp, u_long cmd, caddr_t data);
68 static void	glc_set_multicast(struct glc_softc *sc);
69 static int	glc_add_rxbuf(struct glc_softc *sc, int idx);
70 static int	glc_add_rxbuf_dma(struct glc_softc *sc, int idx);
71 static int	glc_encap(struct glc_softc *sc, struct mbuf **m_head,
72 		    bus_addr_t *pktdesc);
73 static int	glc_intr_filter(void *xsc);
74 static void	glc_intr(void *xsc);
75 static void	glc_tick(void *xsc);
76 static void	glc_media_status(struct ifnet *ifp, struct ifmediareq *ifmr);
77 static int	glc_media_change(struct ifnet *ifp);
78 
79 static MALLOC_DEFINE(M_GLC, "gelic", "PS3 GELIC ethernet");
80 
81 static device_method_t glc_methods[] = {
82 	/* Device interface */
83 	DEVMETHOD(device_probe,		glc_probe),
84 	DEVMETHOD(device_attach,	glc_attach),
85 
86 	{ 0, 0 }
87 };
88 
89 static driver_t glc_driver = {
90 	"glc",
91 	glc_methods,
92 	sizeof(struct glc_softc)
93 };
94 
95 static devclass_t glc_devclass;
96 
97 DRIVER_MODULE(glc, ps3bus, glc_driver, glc_devclass, 0, 0);
98 
99 static int
100 glc_probe(device_t dev)
101 {
102 
103 	if (ps3bus_get_bustype(dev) != PS3_BUSTYPE_SYSBUS ||
104 	    ps3bus_get_devtype(dev) != PS3_DEVTYPE_GELIC)
105 		return (ENXIO);
106 
107 	device_set_desc(dev, "Playstation 3 GELIC Network Controller");
108 	return (BUS_PROBE_SPECIFIC);
109 }
110 
111 static void
112 glc_getphys(void *xaddr, bus_dma_segment_t *segs, int nsegs, int error)
113 {
114 	if (error != 0)
115 		return;
116 
117 	*(bus_addr_t *)xaddr = segs[0].ds_addr;
118 }
119 
120 static int
121 glc_attach(device_t dev)
122 {
123 	struct glc_softc *sc;
124 	struct glc_txsoft *txs;
125 	uint64_t mac64, val, junk;
126 	int i, err;
127 
128 	sc = device_get_softc(dev);
129 
130 	sc->sc_bus = ps3bus_get_bus(dev);
131 	sc->sc_dev = ps3bus_get_device(dev);
132 	sc->sc_self = dev;
133 
134 	mtx_init(&sc->sc_mtx, device_get_nameunit(dev), MTX_NETWORK_LOCK,
135 	    MTX_DEF);
136 	callout_init_mtx(&sc->sc_tick_ch, &sc->sc_mtx, 0);
137 	sc->next_txdma_slot = 0;
138 	sc->bsy_txdma_slots = 0;
139 	sc->sc_next_rxdma_slot = 0;
140 	sc->first_used_txdma_slot = -1;
141 
142 	/*
143 	 * Shut down existing tasks.
144 	 */
145 
146 	lv1_net_stop_tx_dma(sc->sc_bus, sc->sc_dev, 0);
147 	lv1_net_stop_rx_dma(sc->sc_bus, sc->sc_dev, 0);
148 
149 	sc->sc_ifp = if_alloc(IFT_ETHER);
150 	sc->sc_ifp->if_softc = sc;
151 
152 	/*
153 	 * Get MAC address and VLAN id
154 	 */
155 
156 	lv1_net_control(sc->sc_bus, sc->sc_dev, GELIC_GET_MAC_ADDRESS,
157 	    0, 0, 0, &mac64, &junk);
158 	memcpy(sc->sc_enaddr, &((uint8_t *)&mac64)[2], sizeof(sc->sc_enaddr));
159 	sc->sc_tx_vlan = sc->sc_rx_vlan = -1;
160 	err = lv1_net_control(sc->sc_bus, sc->sc_dev, GELIC_GET_VLAN_ID,
161 	    GELIC_VLAN_TX_ETHERNET, 0, 0, &val, &junk);
162 	if (err == 0)
163 		sc->sc_tx_vlan = val;
164 	err = lv1_net_control(sc->sc_bus, sc->sc_dev, GELIC_GET_VLAN_ID,
165 	    GELIC_VLAN_RX_ETHERNET, 0, 0, &val, &junk);
166 	if (err == 0)
167 		sc->sc_rx_vlan = val;
168 
169 	/*
170 	 * Set up interrupt handler
171 	 */
172 	sc->sc_irqid = 0;
173 	sc->sc_irq = bus_alloc_resource_any(dev, SYS_RES_IRQ, &sc->sc_irqid,
174 	    RF_ACTIVE);
175 	if (sc->sc_irq == NULL) {
176 		device_printf(dev, "Could not allocate IRQ!\n");
177 		mtx_destroy(&sc->sc_mtx);
178 		return (ENXIO);
179 	}
180 
181 	bus_setup_intr(dev, sc->sc_irq,
182 	    INTR_TYPE_NET | INTR_MPSAFE | INTR_ENTROPY,
183 	    glc_intr_filter, glc_intr, sc, &sc->sc_irqctx);
184 	sc->sc_hwirq_status = (uint64_t *)contigmalloc(8, M_GLC, M_ZERO, 0,
185 	    BUS_SPACE_MAXADDR_32BIT, 8, PAGE_SIZE);
186 	lv1_net_set_interrupt_status_indicator(sc->sc_bus, sc->sc_dev,
187 	    vtophys(sc->sc_hwirq_status), 0);
188 	lv1_net_set_interrupt_mask(sc->sc_bus, sc->sc_dev,
189 	    GELIC_INT_RXDONE | GELIC_INT_RXFRAME | GELIC_INT_PHY |
190 	    GELIC_INT_TX_CHAIN_END, 0);
191 
192 	/*
193 	 * Set up DMA.
194 	 */
195 
196 	err = bus_dma_tag_create(bus_get_dma_tag(dev), 32, 0,
197 	    BUS_SPACE_MAXADDR, BUS_SPACE_MAXADDR, NULL, NULL,
198 	    129*sizeof(struct glc_dmadesc), 1, 128*sizeof(struct glc_dmadesc),
199 	    0, NULL,NULL, &sc->sc_dmadesc_tag);
200 
201 	err = bus_dmamem_alloc(sc->sc_dmadesc_tag, (void **)&sc->sc_txdmadesc,
202 	    BUS_DMA_WAITOK | BUS_DMA_COHERENT | BUS_DMA_ZERO,
203 	    &sc->sc_txdmadesc_map);
204 	err = bus_dmamap_load(sc->sc_dmadesc_tag, sc->sc_txdmadesc_map,
205 	    sc->sc_txdmadesc, 128*sizeof(struct glc_dmadesc), glc_getphys,
206 	    &sc->sc_txdmadesc_phys, 0);
207 	err = bus_dmamem_alloc(sc->sc_dmadesc_tag, (void **)&sc->sc_rxdmadesc,
208 	    BUS_DMA_WAITOK | BUS_DMA_COHERENT | BUS_DMA_ZERO,
209 	    &sc->sc_rxdmadesc_map);
210 	err = bus_dmamap_load(sc->sc_dmadesc_tag, sc->sc_rxdmadesc_map,
211 	    sc->sc_rxdmadesc, 128*sizeof(struct glc_dmadesc), glc_getphys,
212 	    &sc->sc_rxdmadesc_phys, 0);
213 
214 	err = bus_dma_tag_create(bus_get_dma_tag(dev), 128, 0,
215 	    BUS_SPACE_MAXADDR, BUS_SPACE_MAXADDR, NULL, NULL,
216 	    BUS_SPACE_MAXSIZE_32BIT, 0, BUS_SPACE_MAXSIZE_32BIT, 0, NULL,NULL,
217 	    &sc->sc_rxdma_tag);
218 	err = bus_dma_tag_create(bus_get_dma_tag(dev), 1, 0,
219 	    BUS_SPACE_MAXADDR, BUS_SPACE_MAXADDR, NULL, NULL,
220 	    BUS_SPACE_MAXSIZE_32BIT, 16, BUS_SPACE_MAXSIZE_32BIT, 0, NULL,NULL,
221 	    &sc->sc_txdma_tag);
222 
223 	/* init transmit descriptors */
224 	STAILQ_INIT(&sc->sc_txfreeq);
225 	STAILQ_INIT(&sc->sc_txdirtyq);
226 
227 	/* create TX DMA maps */
228 	err = ENOMEM;
229 	for (i = 0; i < GLC_MAX_TX_PACKETS; i++) {
230 		txs = &sc->sc_txsoft[i];
231 		txs->txs_mbuf = NULL;
232 		err = bus_dmamap_create(sc->sc_txdma_tag, 0, &txs->txs_dmamap);
233 		if (err) {
234 			device_printf(dev,
235 			    "unable to create TX DMA map %d, error = %d\n",
236 			    i, err);
237 		}
238 		STAILQ_INSERT_TAIL(&sc->sc_txfreeq, txs, txs_q);
239 	}
240 
241 	/* Create the receive buffer DMA maps. */
242 	for (i = 0; i < GLC_MAX_RX_PACKETS; i++) {
243 		err = bus_dmamap_create(sc->sc_rxdma_tag, 0,
244 		    &sc->sc_rxsoft[i].rxs_dmamap);
245 		if (err) {
246 			device_printf(dev,
247 			    "unable to create RX DMA map %d, error = %d\n",
248 			    i, err);
249 		}
250 		sc->sc_rxsoft[i].rxs_mbuf = NULL;
251 	}
252 
253 	/*
254 	 * Attach to network stack
255 	 */
256 
257 	if_initname(sc->sc_ifp, device_get_name(dev), device_get_unit(dev));
258 	sc->sc_ifp->if_mtu = ETHERMTU;
259 	sc->sc_ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST;
260 	sc->sc_ifp->if_hwassist = CSUM_TCP | CSUM_UDP;
261 	sc->sc_ifp->if_capabilities = IFCAP_HWCSUM | IFCAP_RXCSUM;
262 	sc->sc_ifp->if_capenable = IFCAP_HWCSUM | IFCAP_RXCSUM;
263 	sc->sc_ifp->if_start = glc_start;
264 	sc->sc_ifp->if_ioctl = glc_ioctl;
265 	sc->sc_ifp->if_init = glc_init;
266 
267 	ifmedia_init(&sc->sc_media, IFM_IMASK, glc_media_change,
268 	    glc_media_status);
269 	ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_10_T, 0, NULL);
270 	ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_10_T | IFM_FDX, 0, NULL);
271 	ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_100_TX, 0, NULL);
272 	ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_100_TX | IFM_FDX, 0, NULL);
273 	ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_1000_T | IFM_FDX, 0, NULL);
274 	ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_AUTO, 0, NULL);
275 	ifmedia_set(&sc->sc_media, IFM_ETHER | IFM_AUTO);
276 
277 	IFQ_SET_MAXLEN(&sc->sc_ifp->if_snd, GLC_MAX_TX_PACKETS);
278 	sc->sc_ifp->if_snd.ifq_drv_maxlen = GLC_MAX_TX_PACKETS;
279 	IFQ_SET_READY(&sc->sc_ifp->if_snd);
280 
281 	ether_ifattach(sc->sc_ifp, sc->sc_enaddr);
282 	sc->sc_ifp->if_hwassist = 0;
283 
284 	return (0);
285 
286 	mtx_destroy(&sc->sc_mtx);
287 	if_free(sc->sc_ifp);
288 	return (ENXIO);
289 }
290 
291 static void
292 glc_init_locked(struct glc_softc *sc)
293 {
294 	int i, error;
295 	struct glc_rxsoft *rxs;
296 	struct glc_txsoft *txs;
297 
298 	mtx_assert(&sc->sc_mtx, MA_OWNED);
299 
300 	lv1_net_stop_tx_dma(sc->sc_bus, sc->sc_dev, 0);
301 	lv1_net_stop_rx_dma(sc->sc_bus, sc->sc_dev, 0);
302 
303 	glc_set_multicast(sc);
304 
305 	for (i = 0; i < GLC_MAX_RX_PACKETS; i++) {
306 		rxs = &sc->sc_rxsoft[i];
307 		rxs->rxs_desc_slot = i;
308 
309 		if (rxs->rxs_mbuf == NULL) {
310 			glc_add_rxbuf(sc, i);
311 
312 			if (rxs->rxs_mbuf == NULL) {
313 				rxs->rxs_desc_slot = -1;
314 				break;
315 			}
316 		}
317 
318 		glc_add_rxbuf_dma(sc, i);
319 		bus_dmamap_sync(sc->sc_dmadesc_tag, sc->sc_rxdmadesc_map,
320 		    BUS_DMASYNC_PREREAD);
321 	}
322 
323 	/* Clear TX dirty queue */
324 	while ((txs = STAILQ_FIRST(&sc->sc_txdirtyq)) != NULL) {
325 		STAILQ_REMOVE_HEAD(&sc->sc_txdirtyq, txs_q);
326 		bus_dmamap_unload(sc->sc_txdma_tag, txs->txs_dmamap);
327 
328 		if (txs->txs_mbuf != NULL) {
329 			m_freem(txs->txs_mbuf);
330 			txs->txs_mbuf = NULL;
331 		}
332 
333 		STAILQ_INSERT_TAIL(&sc->sc_txfreeq, txs, txs_q);
334 	}
335 	sc->first_used_txdma_slot = -1;
336 	sc->bsy_txdma_slots = 0;
337 
338 	error = lv1_net_start_rx_dma(sc->sc_bus, sc->sc_dev,
339 	    sc->sc_rxsoft[0].rxs_desc, 0);
340 	if (error != 0)
341 		device_printf(sc->sc_self,
342 		    "lv1_net_start_rx_dma error: %d\n", error);
343 
344 	sc->sc_ifp->if_drv_flags |= IFF_DRV_RUNNING;
345 	sc->sc_ifp->if_drv_flags &= ~IFF_DRV_OACTIVE;
346 	sc->sc_ifpflags = sc->sc_ifp->if_flags;
347 
348 	sc->sc_wdog_timer = 0;
349 	callout_reset(&sc->sc_tick_ch, hz, glc_tick, sc);
350 }
351 
352 static void
353 glc_stop(void *xsc)
354 {
355 	struct glc_softc *sc = xsc;
356 
357 	mtx_assert(&sc->sc_mtx, MA_OWNED);
358 
359 	lv1_net_stop_tx_dma(sc->sc_bus, sc->sc_dev, 0);
360 	lv1_net_stop_rx_dma(sc->sc_bus, sc->sc_dev, 0);
361 }
362 
363 static void
364 glc_init(void *xsc)
365 {
366 	struct glc_softc *sc = xsc;
367 
368 	mtx_lock(&sc->sc_mtx);
369 	glc_init_locked(sc);
370 	mtx_unlock(&sc->sc_mtx);
371 }
372 
373 static void
374 glc_tick(void *xsc)
375 {
376 	struct glc_softc *sc = xsc;
377 
378 	mtx_assert(&sc->sc_mtx, MA_OWNED);
379 
380 	/*
381 	 * XXX: Sometimes the RX queue gets stuck. Poke it periodically until
382 	 * we figure out why. This will fail harmlessly if the RX queue is
383 	 * already running.
384 	 */
385 	lv1_net_start_rx_dma(sc->sc_bus, sc->sc_dev,
386 	    sc->sc_rxsoft[sc->sc_next_rxdma_slot].rxs_desc, 0);
387 
388 	if (sc->sc_wdog_timer == 0 || --sc->sc_wdog_timer != 0) {
389 		callout_reset(&sc->sc_tick_ch, hz, glc_tick, sc);
390 		return;
391 	}
392 
393 	/* Problems */
394 	device_printf(sc->sc_self, "device timeout\n");
395 
396 	glc_init_locked(sc);
397 }
398 
399 static void
400 glc_start_locked(struct ifnet *ifp)
401 {
402 	struct glc_softc *sc = ifp->if_softc;
403 	bus_addr_t first, pktdesc;
404 	int kickstart = 0;
405 	int error;
406 	struct mbuf *mb_head;
407 
408 	mtx_assert(&sc->sc_mtx, MA_OWNED);
409 	first = 0;
410 
411 	if ((ifp->if_drv_flags & (IFF_DRV_RUNNING | IFF_DRV_OACTIVE)) !=
412 	    IFF_DRV_RUNNING)
413 		return;
414 
415 	if (STAILQ_EMPTY(&sc->sc_txdirtyq))
416 		kickstart = 1;
417 
418 	while (!IFQ_DRV_IS_EMPTY(&ifp->if_snd)) {
419 		IFQ_DRV_DEQUEUE(&ifp->if_snd, mb_head);
420 
421 		if (mb_head == NULL)
422 			break;
423 
424 		/* Check if the ring buffer is full */
425 		if (sc->bsy_txdma_slots > 125) {
426 			/* Put the packet back and stop */
427 			ifp->if_drv_flags |= IFF_DRV_OACTIVE;
428 			IFQ_DRV_PREPEND(&ifp->if_snd, mb_head);
429 			break;
430 		}
431 
432 		BPF_MTAP(ifp, mb_head);
433 
434 		if (sc->sc_tx_vlan >= 0)
435 			mb_head = ether_vlanencap(mb_head, sc->sc_tx_vlan);
436 
437 		if (glc_encap(sc, &mb_head, &pktdesc)) {
438 			ifp->if_drv_flags |= IFF_DRV_OACTIVE;
439 			break;
440 		}
441 
442 		if (first == 0)
443 			first = pktdesc;
444 	}
445 
446 	if (kickstart && first != 0) {
447 		error = lv1_net_start_tx_dma(sc->sc_bus, sc->sc_dev, first, 0);
448 		if (error != 0)
449 			device_printf(sc->sc_self,
450 			    "lv1_net_start_tx_dma error: %d\n", error);
451 		sc->sc_wdog_timer = 5;
452 	}
453 }
454 
455 static void
456 glc_start(struct ifnet *ifp)
457 {
458 	struct glc_softc *sc = ifp->if_softc;
459 
460 	mtx_lock(&sc->sc_mtx);
461 	glc_start_locked(ifp);
462 	mtx_unlock(&sc->sc_mtx);
463 }
464 
465 static int
466 glc_ioctl(struct ifnet *ifp, u_long cmd, caddr_t data)
467 {
468 	struct glc_softc *sc = ifp->if_softc;
469 	struct ifreq *ifr = (struct ifreq *)data;
470 	int err = 0;
471 
472 	switch (cmd) {
473 	case SIOCSIFFLAGS:
474                 mtx_lock(&sc->sc_mtx);
475 		if ((ifp->if_flags & IFF_UP) != 0) {
476 			if ((ifp->if_drv_flags & IFF_DRV_RUNNING) != 0 &&
477 			   ((ifp->if_flags ^ sc->sc_ifpflags) &
478 			    (IFF_ALLMULTI | IFF_PROMISC)) != 0)
479 				glc_set_multicast(sc);
480 			else
481 				glc_init_locked(sc);
482 		}
483 		else if ((ifp->if_drv_flags & IFF_DRV_RUNNING) != 0)
484 			glc_stop(sc);
485 		sc->sc_ifpflags = ifp->if_flags;
486 		mtx_unlock(&sc->sc_mtx);
487 		break;
488 	case SIOCADDMULTI:
489 	case SIOCDELMULTI:
490                 mtx_lock(&sc->sc_mtx);
491 		glc_set_multicast(sc);
492                 mtx_unlock(&sc->sc_mtx);
493 		break;
494 	case SIOCGIFMEDIA:
495 	case SIOCSIFMEDIA:
496 		err = ifmedia_ioctl(ifp, ifr, &sc->sc_media, cmd);
497 		break;
498 	default:
499 		err = ether_ioctl(ifp, cmd, data);
500 		break;
501 	}
502 
503 	return (err);
504 }
505 
506 static void
507 glc_set_multicast(struct glc_softc *sc)
508 {
509 	struct ifnet *ifp = sc->sc_ifp;
510 	struct ifmultiaddr *inm;
511 	uint64_t addr;
512 	int naddrs;
513 
514 	/* Clear multicast filter */
515 	lv1_net_remove_multicast_address(sc->sc_bus, sc->sc_dev, 0, 1);
516 
517 	/* Add broadcast */
518 	lv1_net_add_multicast_address(sc->sc_bus, sc->sc_dev,
519 	    0xffffffffffffL, 0);
520 
521 	if ((ifp->if_flags & IFF_ALLMULTI) != 0) {
522 		lv1_net_add_multicast_address(sc->sc_bus, sc->sc_dev, 0, 1);
523 	} else {
524 		if_maddr_rlock(ifp);
525 		naddrs = 1; /* Include broadcast */
526 		TAILQ_FOREACH(inm, &ifp->if_multiaddrs, ifma_link) {
527 			if (inm->ifma_addr->sa_family != AF_LINK)
528 				continue;
529 			addr = 0;
530 			memcpy(&((uint8_t *)(&addr))[2],
531 			    LLADDR((struct sockaddr_dl *)inm->ifma_addr),
532 			    ETHER_ADDR_LEN);
533 
534 			lv1_net_add_multicast_address(sc->sc_bus, sc->sc_dev,
535 			    addr, 0);
536 
537 			/*
538 			 * Filter can only hold 32 addresses, so fall back to
539 			 * the IFF_ALLMULTI case if we have too many.
540 			 */
541 			if (++naddrs >= 32) {
542 				lv1_net_add_multicast_address(sc->sc_bus,
543 				    sc->sc_dev, 0, 1);
544 				break;
545 			}
546 		}
547 		if_maddr_runlock(ifp);
548 	}
549 }
550 
551 static int
552 glc_add_rxbuf(struct glc_softc *sc, int idx)
553 {
554 	struct glc_rxsoft *rxs = &sc->sc_rxsoft[idx];
555 	struct mbuf *m;
556 	bus_dma_segment_t segs[1];
557 	int error, nsegs;
558 
559 	m = m_getcl(M_NOWAIT, MT_DATA, M_PKTHDR);
560 	if (m == NULL)
561 		return (ENOBUFS);
562 	m->m_len = m->m_pkthdr.len = m->m_ext.ext_size;
563 
564 	if (rxs->rxs_mbuf != NULL) {
565 		bus_dmamap_sync(sc->sc_rxdma_tag, rxs->rxs_dmamap,
566 		    BUS_DMASYNC_POSTREAD);
567 		bus_dmamap_unload(sc->sc_rxdma_tag, rxs->rxs_dmamap);
568 	}
569 
570 	error = bus_dmamap_load_mbuf_sg(sc->sc_rxdma_tag, rxs->rxs_dmamap, m,
571 	    segs, &nsegs, BUS_DMA_NOWAIT);
572 	if (error != 0) {
573 		device_printf(sc->sc_self,
574 		    "cannot load RS DMA map %d, error = %d\n", idx, error);
575 		m_freem(m);
576 		return (error);
577 	}
578 	/* If nsegs is wrong then the stack is corrupt. */
579 	KASSERT(nsegs == 1,
580 	    ("%s: too many DMA segments (%d)", __func__, nsegs));
581 	rxs->rxs_mbuf = m;
582 	rxs->segment = segs[0];
583 
584 	bus_dmamap_sync(sc->sc_rxdma_tag, rxs->rxs_dmamap, BUS_DMASYNC_PREREAD);
585 
586 	return (0);
587 }
588 
589 static int
590 glc_add_rxbuf_dma(struct glc_softc *sc, int idx)
591 {
592 	struct glc_rxsoft *rxs = &sc->sc_rxsoft[idx];
593 
594 	bzero(&sc->sc_rxdmadesc[idx], sizeof(sc->sc_rxdmadesc[idx]));
595 	sc->sc_rxdmadesc[idx].paddr = rxs->segment.ds_addr;
596 	sc->sc_rxdmadesc[idx].len = rxs->segment.ds_len;
597 	sc->sc_rxdmadesc[idx].next = sc->sc_rxdmadesc_phys +
598 	    ((idx + 1) % GLC_MAX_RX_PACKETS)*sizeof(sc->sc_rxdmadesc[idx]);
599 	sc->sc_rxdmadesc[idx].cmd_stat = GELIC_DESCR_OWNED;
600 
601 	rxs->rxs_desc_slot = idx;
602 	rxs->rxs_desc = sc->sc_rxdmadesc_phys + idx*sizeof(struct glc_dmadesc);
603 
604         return (0);
605 }
606 
607 static int
608 glc_encap(struct glc_softc *sc, struct mbuf **m_head, bus_addr_t *pktdesc)
609 {
610 	bus_dma_segment_t segs[16];
611 	struct glc_txsoft *txs;
612 	struct mbuf *m;
613 	bus_addr_t firstslotphys;
614 	int i, idx, nsegs, nsegs_max;
615 	int err = 0;
616 
617 	/* Max number of segments is the number of free DMA slots */
618 	nsegs_max = 128 - sc->bsy_txdma_slots;
619 
620 	if (nsegs_max > 16 || sc->first_used_txdma_slot < 0)
621 		nsegs_max = 16;
622 
623 	/* Get a work queue entry. */
624 	if ((txs = STAILQ_FIRST(&sc->sc_txfreeq)) == NULL) {
625 		/* Ran out of descriptors. */
626 		return (ENOBUFS);
627 	}
628 
629 	nsegs = 0;
630 	for (m = *m_head; m != NULL; m = m->m_next)
631 		nsegs++;
632 
633 	if (nsegs > nsegs_max) {
634 		m = m_collapse(*m_head, M_NOWAIT, nsegs_max);
635 		if (m == NULL) {
636 			m_freem(*m_head);
637 			*m_head = NULL;
638 			return (ENOBUFS);
639 		}
640 		*m_head = m;
641 	}
642 
643 	err = bus_dmamap_load_mbuf_sg(sc->sc_txdma_tag, txs->txs_dmamap,
644 	    *m_head, segs, &nsegs, BUS_DMA_NOWAIT);
645 	if (err != 0) {
646 		m_freem(*m_head);
647 		*m_head = NULL;
648 		return (err);
649 	}
650 
651 	KASSERT(nsegs <= 128 - sc->bsy_txdma_slots,
652 	    ("GLC: Mapped too many (%d) DMA segments with %d available",
653 	    nsegs, 128 - sc->bsy_txdma_slots));
654 
655 	if (nsegs == 0) {
656 		m_freem(*m_head);
657 		*m_head = NULL;
658 		return (EIO);
659 	}
660 
661 	txs->txs_ndescs = nsegs;
662 	txs->txs_firstdesc = sc->next_txdma_slot;
663 
664 	idx = txs->txs_firstdesc;
665 	firstslotphys = sc->sc_txdmadesc_phys +
666 	    txs->txs_firstdesc*sizeof(struct glc_dmadesc);
667 
668 	for (i = 0; i < nsegs; i++) {
669 		bzero(&sc->sc_txdmadesc[idx], sizeof(sc->sc_txdmadesc[idx]));
670 		sc->sc_txdmadesc[idx].paddr = segs[i].ds_addr;
671 		sc->sc_txdmadesc[idx].len = segs[i].ds_len;
672 		sc->sc_txdmadesc[idx].next = sc->sc_txdmadesc_phys +
673 		    ((idx + 1) % GLC_MAX_TX_PACKETS)*sizeof(struct glc_dmadesc);
674 		sc->sc_txdmadesc[idx].cmd_stat |= GELIC_CMDSTAT_NOIPSEC;
675 
676 		if (i+1 == nsegs) {
677 			txs->txs_lastdesc = idx;
678 			sc->sc_txdmadesc[idx].next = 0;
679 			sc->sc_txdmadesc[idx].cmd_stat |= GELIC_CMDSTAT_LAST;
680 		}
681 
682 		if ((*m_head)->m_pkthdr.csum_flags & CSUM_TCP)
683 			sc->sc_txdmadesc[idx].cmd_stat |= GELIC_CMDSTAT_CSUM_TCP;
684 		if ((*m_head)->m_pkthdr.csum_flags & CSUM_UDP)
685 			sc->sc_txdmadesc[idx].cmd_stat |= GELIC_CMDSTAT_CSUM_UDP;
686 		sc->sc_txdmadesc[idx].cmd_stat |= GELIC_DESCR_OWNED;
687 
688 		idx = (idx + 1) % GLC_MAX_TX_PACKETS;
689 	}
690 	sc->next_txdma_slot = idx;
691 	sc->bsy_txdma_slots += nsegs;
692 	if (txs->txs_firstdesc != 0)
693 		idx = txs->txs_firstdesc - 1;
694 	else
695 		idx = GLC_MAX_TX_PACKETS - 1;
696 
697 	if (sc->first_used_txdma_slot < 0)
698 		sc->first_used_txdma_slot = txs->txs_firstdesc;
699 
700 	bus_dmamap_sync(sc->sc_txdma_tag, txs->txs_dmamap,
701 	    BUS_DMASYNC_PREWRITE);
702 	sc->sc_txdmadesc[idx].next = firstslotphys;
703 
704 	STAILQ_REMOVE_HEAD(&sc->sc_txfreeq, txs_q);
705 	STAILQ_INSERT_TAIL(&sc->sc_txdirtyq, txs, txs_q);
706 	txs->txs_mbuf = *m_head;
707 	*pktdesc = firstslotphys;
708 
709 	return (0);
710 }
711 
712 static void
713 glc_rxintr(struct glc_softc *sc)
714 {
715 	int i, restart_rxdma, error;
716 	struct mbuf *m;
717 	struct ifnet *ifp = sc->sc_ifp;
718 
719 	bus_dmamap_sync(sc->sc_dmadesc_tag, sc->sc_rxdmadesc_map,
720 	    BUS_DMASYNC_POSTREAD);
721 
722 	restart_rxdma = 0;
723 	while ((sc->sc_rxdmadesc[sc->sc_next_rxdma_slot].cmd_stat &
724 	   GELIC_DESCR_OWNED) == 0) {
725 		i = sc->sc_next_rxdma_slot;
726 		sc->sc_next_rxdma_slot++;
727 		if (sc->sc_next_rxdma_slot >= GLC_MAX_RX_PACKETS)
728 			sc->sc_next_rxdma_slot = 0;
729 
730 		if (sc->sc_rxdmadesc[i].cmd_stat & GELIC_CMDSTAT_CHAIN_END)
731 			restart_rxdma = 1;
732 
733 		if (sc->sc_rxdmadesc[i].rxerror & GELIC_RXERRORS) {
734 			ifp->if_ierrors++;
735 			goto requeue;
736 		}
737 
738 		m = sc->sc_rxsoft[i].rxs_mbuf;
739 		if (sc->sc_rxdmadesc[i].data_stat & GELIC_RX_IPCSUM) {
740 			m->m_pkthdr.csum_flags |=
741 			    CSUM_IP_CHECKED | CSUM_IP_VALID;
742 		}
743 		if (sc->sc_rxdmadesc[i].data_stat & GELIC_RX_TCPUDPCSUM) {
744 			m->m_pkthdr.csum_flags |=
745 			    CSUM_DATA_VALID | CSUM_PSEUDO_HDR;
746 			m->m_pkthdr.csum_data = 0xffff;
747 		}
748 
749 		if (glc_add_rxbuf(sc, i)) {
750 			ifp->if_ierrors++;
751 			goto requeue;
752 		}
753 
754 		ifp->if_ipackets++;
755 		m->m_pkthdr.rcvif = ifp;
756 		m->m_len = sc->sc_rxdmadesc[i].valid_size;
757 		m->m_pkthdr.len = m->m_len;
758 
759 		/*
760 		 * Remove VLAN tag. Even on early firmwares that do not allow
761 		 * multiple VLANs, the VLAN tag is still in place here.
762 		 */
763 		m_adj(m, 2);
764 
765 		mtx_unlock(&sc->sc_mtx);
766 		(*ifp->if_input)(ifp, m);
767 		mtx_lock(&sc->sc_mtx);
768 
769 	    requeue:
770 		glc_add_rxbuf_dma(sc, i);
771 	}
772 
773 	bus_dmamap_sync(sc->sc_dmadesc_tag, sc->sc_rxdmadesc_map,
774 	    BUS_DMASYNC_PREWRITE);
775 
776 	if (restart_rxdma) {
777 		error = lv1_net_start_rx_dma(sc->sc_bus, sc->sc_dev,
778 		    sc->sc_rxsoft[sc->sc_next_rxdma_slot].rxs_desc, 0);
779 		if (error != 0)
780 			device_printf(sc->sc_self,
781 			    "lv1_net_start_rx_dma error: %d\n", error);
782 	}
783 }
784 
785 static void
786 glc_txintr(struct glc_softc *sc)
787 {
788 	struct ifnet *ifp = sc->sc_ifp;
789 	struct glc_txsoft *txs;
790 	int progress = 0, kickstart = 0, error;
791 
792 	bus_dmamap_sync(sc->sc_dmadesc_tag, sc->sc_txdmadesc_map,
793 	    BUS_DMASYNC_POSTREAD);
794 
795 	while ((txs = STAILQ_FIRST(&sc->sc_txdirtyq)) != NULL) {
796 		if (sc->sc_txdmadesc[txs->txs_lastdesc].cmd_stat
797 		    & GELIC_DESCR_OWNED)
798 			break;
799 
800 		STAILQ_REMOVE_HEAD(&sc->sc_txdirtyq, txs_q);
801 		bus_dmamap_unload(sc->sc_txdma_tag, txs->txs_dmamap);
802 		sc->bsy_txdma_slots -= txs->txs_ndescs;
803 
804 		if (txs->txs_mbuf != NULL) {
805 			m_freem(txs->txs_mbuf);
806 			txs->txs_mbuf = NULL;
807 		}
808 
809 		if ((sc->sc_txdmadesc[txs->txs_lastdesc].cmd_stat & 0xf0000000)
810 		    != 0) {
811 			lv1_net_stop_tx_dma(sc->sc_bus, sc->sc_dev, 0);
812 			kickstart = 1;
813 			ifp->if_oerrors++;
814 		}
815 
816 		if (sc->sc_txdmadesc[txs->txs_lastdesc].cmd_stat &
817 		    GELIC_CMDSTAT_CHAIN_END)
818 			kickstart = 1;
819 
820 		STAILQ_INSERT_TAIL(&sc->sc_txfreeq, txs, txs_q);
821 		ifp->if_opackets++;
822 		progress = 1;
823 	}
824 
825 	if (txs != NULL)
826 		sc->first_used_txdma_slot = txs->txs_firstdesc;
827 	else
828 		sc->first_used_txdma_slot = -1;
829 
830 	if (kickstart || txs != NULL) {
831 		/* Speculatively (or necessarily) start the TX queue again */
832 		error = lv1_net_start_tx_dma(sc->sc_bus, sc->sc_dev,
833 		    sc->sc_txdmadesc_phys +
834 		    txs->txs_firstdesc*sizeof(struct glc_dmadesc), 0);
835 		if (error != 0)
836 			device_printf(sc->sc_self,
837 			    "lv1_net_start_tx_dma error: %d\n", error);
838 	}
839 
840 	if (progress) {
841 		/*
842 		 * We freed some descriptors, so reset IFF_DRV_OACTIVE
843 		 * and restart.
844 		 */
845 		ifp->if_drv_flags &= ~IFF_DRV_OACTIVE;
846 		sc->sc_wdog_timer = STAILQ_EMPTY(&sc->sc_txdirtyq) ? 0 : 5;
847 
848 		if ((ifp->if_drv_flags & IFF_DRV_RUNNING) &&
849 		    !IFQ_DRV_IS_EMPTY(&ifp->if_snd))
850 			glc_start_locked(ifp);
851 	}
852 }
853 
854 static int
855 glc_intr_filter(void *xsc)
856 {
857 	struct glc_softc *sc = xsc;
858 
859 	powerpc_sync();
860 	atomic_set_64(&sc->sc_interrupt_status, *sc->sc_hwirq_status);
861 	return (FILTER_SCHEDULE_THREAD);
862 }
863 
864 static void
865 glc_intr(void *xsc)
866 {
867 	struct glc_softc *sc = xsc;
868 	uint64_t status, linkstat, junk;
869 
870 	mtx_lock(&sc->sc_mtx);
871 
872 	status = atomic_readandclear_64(&sc->sc_interrupt_status);
873 
874 	if (status == 0) {
875 		mtx_unlock(&sc->sc_mtx);
876 		return;
877 	}
878 
879 	if (status & (GELIC_INT_RXDONE | GELIC_INT_RXFRAME))
880 		glc_rxintr(sc);
881 
882 	if (status & (GELIC_INT_TXDONE | GELIC_INT_TX_CHAIN_END))
883 		glc_txintr(sc);
884 
885 	if (status & GELIC_INT_PHY) {
886 		lv1_net_control(sc->sc_bus, sc->sc_dev, GELIC_GET_LINK_STATUS,
887 		    GELIC_VLAN_TX_ETHERNET, 0, 0, &linkstat, &junk);
888 
889 		linkstat = (linkstat & GELIC_LINK_UP) ?
890 		    LINK_STATE_UP : LINK_STATE_DOWN;
891 		if (linkstat != sc->sc_ifp->if_link_state)
892 			if_link_state_change(sc->sc_ifp, linkstat);
893 	}
894 
895 	mtx_unlock(&sc->sc_mtx);
896 }
897 
898 static void
899 glc_media_status(struct ifnet *ifp, struct ifmediareq *ifmr)
900 {
901 	struct glc_softc *sc = ifp->if_softc;
902 	uint64_t status, junk;
903 
904 	ifmr->ifm_status = IFM_AVALID;
905 	ifmr->ifm_active = IFM_ETHER;
906 
907 	lv1_net_control(sc->sc_bus, sc->sc_dev, GELIC_GET_LINK_STATUS,
908 	    GELIC_VLAN_TX_ETHERNET, 0, 0, &status, &junk);
909 
910 	if (status & GELIC_LINK_UP)
911 		ifmr->ifm_status |= IFM_ACTIVE;
912 
913 	if (status & GELIC_SPEED_10)
914 		ifmr->ifm_active |= IFM_10_T;
915 	else if (status & GELIC_SPEED_100)
916 		ifmr->ifm_active |= IFM_100_TX;
917 	else if (status & GELIC_SPEED_1000)
918 		ifmr->ifm_active |= IFM_1000_T;
919 
920 	if (status & GELIC_FULL_DUPLEX)
921 		ifmr->ifm_active |= IFM_FDX;
922 	else
923 		ifmr->ifm_active |= IFM_HDX;
924 }
925 
926 static int
927 glc_media_change(struct ifnet *ifp)
928 {
929 	struct glc_softc *sc = ifp->if_softc;
930 	uint64_t mode, junk;
931 	int result;
932 
933 	if (IFM_TYPE(sc->sc_media.ifm_media) != IFM_ETHER)
934 		return (EINVAL);
935 
936 	switch (IFM_SUBTYPE(sc->sc_media.ifm_media)) {
937 	case IFM_AUTO:
938 		mode = GELIC_AUTO_NEG;
939 		break;
940 	case IFM_10_T:
941 		mode = GELIC_SPEED_10;
942 		break;
943 	case IFM_100_TX:
944 		mode = GELIC_SPEED_100;
945 		break;
946 	case IFM_1000_T:
947 		mode = GELIC_SPEED_1000 | GELIC_FULL_DUPLEX;
948 		break;
949 	default:
950 		return (EINVAL);
951 	}
952 
953 	if (IFM_OPTIONS(sc->sc_media.ifm_media) & IFM_FDX)
954 		mode |= GELIC_FULL_DUPLEX;
955 
956 	result = lv1_net_control(sc->sc_bus, sc->sc_dev, GELIC_SET_LINK_MODE,
957 	    GELIC_VLAN_TX_ETHERNET, mode, 0, &junk, &junk);
958 
959 	return (result ? EIO : 0);
960 }
961 
962