1 /*- 2 * SPDX-License-Identifier: BSD-2-Clause-FreeBSD 3 * 4 * Copyright (C) 2010 Nathan Whitehorn 5 * All rights reserved. 6 * 7 * Redistribution and use in source and binary forms, with or without 8 * modification, are permitted provided that the following conditions 9 * are met: 10 * 1. Redistributions of source code must retain the above copyright 11 * notice, this list of conditions and the following disclaimer. 12 * 2. Redistributions in binary form must reproduce the above copyright 13 * notice, this list of conditions and the following disclaimer in the 14 * documentation and/or other materials provided with the distribution. 15 * 16 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR 17 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 18 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. 19 * IN NO EVENT SHALL TOOLS GMBH BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, 20 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, 21 * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; 22 * OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, 23 * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR 24 * OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF 25 * ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 26 * 27 * $FreeBSD$ 28 */ 29 30 #include <sys/param.h> 31 #include <sys/systm.h> 32 #include <sys/sockio.h> 33 #include <sys/endian.h> 34 #include <sys/lock.h> 35 #include <sys/mbuf.h> 36 #include <sys/module.h> 37 #include <sys/malloc.h> 38 #include <sys/mutex.h> 39 #include <sys/kernel.h> 40 #include <sys/socket.h> 41 42 #include <vm/vm.h> 43 #include <vm/pmap.h> 44 45 #include <net/bpf.h> 46 #include <net/if.h> 47 #include <net/if_var.h> 48 #include <net/ethernet.h> 49 #include <net/if_media.h> 50 #include <net/if_types.h> 51 #include <net/if_dl.h> 52 53 #include <machine/pio.h> 54 #include <machine/bus.h> 55 #include <machine/platform.h> 56 #include <machine/resource.h> 57 #include <sys/bus.h> 58 #include <sys/rman.h> 59 60 #include "ps3bus.h" 61 #include "ps3-hvcall.h" 62 #include "if_glcreg.h" 63 64 static int glc_probe(device_t); 65 static int glc_attach(device_t); 66 static void glc_init(void *xsc); 67 static void glc_start(struct ifnet *ifp); 68 static int glc_ioctl(struct ifnet *ifp, u_long cmd, caddr_t data); 69 static void glc_set_multicast(struct glc_softc *sc); 70 static int glc_add_rxbuf(struct glc_softc *sc, int idx); 71 static int glc_add_rxbuf_dma(struct glc_softc *sc, int idx); 72 static int glc_encap(struct glc_softc *sc, struct mbuf **m_head, 73 bus_addr_t *pktdesc); 74 static int glc_intr_filter(void *xsc); 75 static void glc_intr(void *xsc); 76 static void glc_tick(void *xsc); 77 static void glc_media_status(struct ifnet *ifp, struct ifmediareq *ifmr); 78 static int glc_media_change(struct ifnet *ifp); 79 80 static MALLOC_DEFINE(M_GLC, "gelic", "PS3 GELIC ethernet"); 81 82 static device_method_t glc_methods[] = { 83 /* Device interface */ 84 DEVMETHOD(device_probe, glc_probe), 85 DEVMETHOD(device_attach, glc_attach), 86 { 0, 0 } 87 }; 88 89 static driver_t glc_driver = { 90 "glc", 91 glc_methods, 92 sizeof(struct glc_softc) 93 }; 94 95 DRIVER_MODULE(glc, ps3bus, glc_driver, 0, 0); 96 97 static int 98 glc_probe(device_t dev) 99 { 100 101 if (ps3bus_get_bustype(dev) != PS3_BUSTYPE_SYSBUS || 102 ps3bus_get_devtype(dev) != PS3_DEVTYPE_GELIC) 103 return (ENXIO); 104 105 device_set_desc(dev, "Playstation 3 GELIC Network Controller"); 106 return (BUS_PROBE_SPECIFIC); 107 } 108 109 static void 110 glc_getphys(void *xaddr, bus_dma_segment_t *segs, int nsegs, int error) 111 { 112 if (error != 0) 113 return; 114 115 *(bus_addr_t *)xaddr = segs[0].ds_addr; 116 } 117 118 static int 119 glc_attach(device_t dev) 120 { 121 struct glc_softc *sc; 122 struct glc_txsoft *txs; 123 uint64_t mac64, val, junk; 124 int i, err; 125 126 sc = device_get_softc(dev); 127 128 sc->sc_bus = ps3bus_get_bus(dev); 129 sc->sc_dev = ps3bus_get_device(dev); 130 sc->sc_self = dev; 131 132 mtx_init(&sc->sc_mtx, device_get_nameunit(dev), MTX_NETWORK_LOCK, 133 MTX_DEF); 134 callout_init_mtx(&sc->sc_tick_ch, &sc->sc_mtx, 0); 135 sc->next_txdma_slot = 0; 136 sc->bsy_txdma_slots = 0; 137 sc->sc_next_rxdma_slot = 0; 138 sc->first_used_txdma_slot = -1; 139 140 /* 141 * Shut down existing tasks. 142 */ 143 144 lv1_net_stop_tx_dma(sc->sc_bus, sc->sc_dev, 0); 145 lv1_net_stop_rx_dma(sc->sc_bus, sc->sc_dev, 0); 146 147 sc->sc_ifp = if_alloc(IFT_ETHER); 148 sc->sc_ifp->if_softc = sc; 149 150 /* 151 * Get MAC address and VLAN id 152 */ 153 154 lv1_net_control(sc->sc_bus, sc->sc_dev, GELIC_GET_MAC_ADDRESS, 155 0, 0, 0, &mac64, &junk); 156 memcpy(sc->sc_enaddr, &((uint8_t *)&mac64)[2], sizeof(sc->sc_enaddr)); 157 sc->sc_tx_vlan = sc->sc_rx_vlan = -1; 158 err = lv1_net_control(sc->sc_bus, sc->sc_dev, GELIC_GET_VLAN_ID, 159 GELIC_VLAN_TX_ETHERNET, 0, 0, &val, &junk); 160 if (err == 0) 161 sc->sc_tx_vlan = val; 162 err = lv1_net_control(sc->sc_bus, sc->sc_dev, GELIC_GET_VLAN_ID, 163 GELIC_VLAN_RX_ETHERNET, 0, 0, &val, &junk); 164 if (err == 0) 165 sc->sc_rx_vlan = val; 166 167 /* 168 * Set up interrupt handler 169 */ 170 sc->sc_irqid = 0; 171 sc->sc_irq = bus_alloc_resource_any(dev, SYS_RES_IRQ, &sc->sc_irqid, 172 RF_ACTIVE); 173 if (sc->sc_irq == NULL) { 174 device_printf(dev, "Could not allocate IRQ!\n"); 175 mtx_destroy(&sc->sc_mtx); 176 return (ENXIO); 177 } 178 179 bus_setup_intr(dev, sc->sc_irq, 180 INTR_TYPE_NET | INTR_MPSAFE | INTR_ENTROPY, 181 glc_intr_filter, glc_intr, sc, &sc->sc_irqctx); 182 sc->sc_hwirq_status = (uint64_t *)contigmalloc(8, M_GLC, M_ZERO, 0, 183 BUS_SPACE_MAXADDR_32BIT, 8, PAGE_SIZE); 184 lv1_net_set_interrupt_status_indicator(sc->sc_bus, sc->sc_dev, 185 vtophys(sc->sc_hwirq_status), 0); 186 lv1_net_set_interrupt_mask(sc->sc_bus, sc->sc_dev, 187 GELIC_INT_RXDONE | GELIC_INT_RXFRAME | GELIC_INT_PHY | 188 GELIC_INT_TX_CHAIN_END, 0); 189 190 /* 191 * Set up DMA. 192 */ 193 194 err = bus_dma_tag_create(bus_get_dma_tag(dev), 32, 0, 195 BUS_SPACE_MAXADDR, BUS_SPACE_MAXADDR, NULL, NULL, 196 129*sizeof(struct glc_dmadesc), 1, 128*sizeof(struct glc_dmadesc), 197 0, NULL,NULL, &sc->sc_dmadesc_tag); 198 199 err = bus_dmamem_alloc(sc->sc_dmadesc_tag, (void **)&sc->sc_txdmadesc, 200 BUS_DMA_WAITOK | BUS_DMA_COHERENT | BUS_DMA_ZERO, 201 &sc->sc_txdmadesc_map); 202 err = bus_dmamap_load(sc->sc_dmadesc_tag, sc->sc_txdmadesc_map, 203 sc->sc_txdmadesc, 128*sizeof(struct glc_dmadesc), glc_getphys, 204 &sc->sc_txdmadesc_phys, 0); 205 err = bus_dmamem_alloc(sc->sc_dmadesc_tag, (void **)&sc->sc_rxdmadesc, 206 BUS_DMA_WAITOK | BUS_DMA_COHERENT | BUS_DMA_ZERO, 207 &sc->sc_rxdmadesc_map); 208 err = bus_dmamap_load(sc->sc_dmadesc_tag, sc->sc_rxdmadesc_map, 209 sc->sc_rxdmadesc, 128*sizeof(struct glc_dmadesc), glc_getphys, 210 &sc->sc_rxdmadesc_phys, 0); 211 212 err = bus_dma_tag_create(bus_get_dma_tag(dev), 128, 0, 213 BUS_SPACE_MAXADDR, BUS_SPACE_MAXADDR, NULL, NULL, 214 BUS_SPACE_MAXSIZE_32BIT, 0, BUS_SPACE_MAXSIZE_32BIT, 0, NULL,NULL, 215 &sc->sc_rxdma_tag); 216 err = bus_dma_tag_create(bus_get_dma_tag(dev), 1, 0, 217 BUS_SPACE_MAXADDR, BUS_SPACE_MAXADDR, NULL, NULL, 218 BUS_SPACE_MAXSIZE_32BIT, 16, BUS_SPACE_MAXSIZE_32BIT, 0, NULL,NULL, 219 &sc->sc_txdma_tag); 220 221 /* init transmit descriptors */ 222 STAILQ_INIT(&sc->sc_txfreeq); 223 STAILQ_INIT(&sc->sc_txdirtyq); 224 225 /* create TX DMA maps */ 226 err = ENOMEM; 227 for (i = 0; i < GLC_MAX_TX_PACKETS; i++) { 228 txs = &sc->sc_txsoft[i]; 229 txs->txs_mbuf = NULL; 230 err = bus_dmamap_create(sc->sc_txdma_tag, 0, &txs->txs_dmamap); 231 if (err) { 232 device_printf(dev, 233 "unable to create TX DMA map %d, error = %d\n", 234 i, err); 235 } 236 STAILQ_INSERT_TAIL(&sc->sc_txfreeq, txs, txs_q); 237 } 238 239 /* Create the receive buffer DMA maps. */ 240 for (i = 0; i < GLC_MAX_RX_PACKETS; i++) { 241 err = bus_dmamap_create(sc->sc_rxdma_tag, 0, 242 &sc->sc_rxsoft[i].rxs_dmamap); 243 if (err) { 244 device_printf(dev, 245 "unable to create RX DMA map %d, error = %d\n", 246 i, err); 247 } 248 sc->sc_rxsoft[i].rxs_mbuf = NULL; 249 } 250 251 /* 252 * Attach to network stack 253 */ 254 255 if_initname(sc->sc_ifp, device_get_name(dev), device_get_unit(dev)); 256 sc->sc_ifp->if_mtu = ETHERMTU; 257 sc->sc_ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST; 258 sc->sc_ifp->if_hwassist = CSUM_TCP | CSUM_UDP; 259 sc->sc_ifp->if_capabilities = IFCAP_HWCSUM | IFCAP_RXCSUM; 260 sc->sc_ifp->if_capenable = IFCAP_HWCSUM | IFCAP_RXCSUM; 261 sc->sc_ifp->if_start = glc_start; 262 sc->sc_ifp->if_ioctl = glc_ioctl; 263 sc->sc_ifp->if_init = glc_init; 264 265 ifmedia_init(&sc->sc_media, IFM_IMASK, glc_media_change, 266 glc_media_status); 267 ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_10_T, 0, NULL); 268 ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_10_T | IFM_FDX, 0, NULL); 269 ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_100_TX, 0, NULL); 270 ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_100_TX | IFM_FDX, 0, NULL); 271 ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_1000_T | IFM_FDX, 0, NULL); 272 ifmedia_add(&sc->sc_media, IFM_ETHER | IFM_AUTO, 0, NULL); 273 ifmedia_set(&sc->sc_media, IFM_ETHER | IFM_AUTO); 274 275 IFQ_SET_MAXLEN(&sc->sc_ifp->if_snd, GLC_MAX_TX_PACKETS); 276 sc->sc_ifp->if_snd.ifq_drv_maxlen = GLC_MAX_TX_PACKETS; 277 IFQ_SET_READY(&sc->sc_ifp->if_snd); 278 279 ether_ifattach(sc->sc_ifp, sc->sc_enaddr); 280 sc->sc_ifp->if_hwassist = 0; 281 282 return (0); 283 284 mtx_destroy(&sc->sc_mtx); 285 if_free(sc->sc_ifp); 286 return (ENXIO); 287 } 288 289 static void 290 glc_init_locked(struct glc_softc *sc) 291 { 292 int i, error; 293 struct glc_rxsoft *rxs; 294 struct glc_txsoft *txs; 295 296 mtx_assert(&sc->sc_mtx, MA_OWNED); 297 298 lv1_net_stop_tx_dma(sc->sc_bus, sc->sc_dev, 0); 299 lv1_net_stop_rx_dma(sc->sc_bus, sc->sc_dev, 0); 300 301 glc_set_multicast(sc); 302 303 for (i = 0; i < GLC_MAX_RX_PACKETS; i++) { 304 rxs = &sc->sc_rxsoft[i]; 305 rxs->rxs_desc_slot = i; 306 307 if (rxs->rxs_mbuf == NULL) { 308 glc_add_rxbuf(sc, i); 309 310 if (rxs->rxs_mbuf == NULL) { 311 rxs->rxs_desc_slot = -1; 312 break; 313 } 314 } 315 316 glc_add_rxbuf_dma(sc, i); 317 bus_dmamap_sync(sc->sc_dmadesc_tag, sc->sc_rxdmadesc_map, 318 BUS_DMASYNC_PREREAD); 319 } 320 321 /* Clear TX dirty queue */ 322 while ((txs = STAILQ_FIRST(&sc->sc_txdirtyq)) != NULL) { 323 STAILQ_REMOVE_HEAD(&sc->sc_txdirtyq, txs_q); 324 bus_dmamap_unload(sc->sc_txdma_tag, txs->txs_dmamap); 325 326 if (txs->txs_mbuf != NULL) { 327 m_freem(txs->txs_mbuf); 328 txs->txs_mbuf = NULL; 329 } 330 331 STAILQ_INSERT_TAIL(&sc->sc_txfreeq, txs, txs_q); 332 } 333 sc->first_used_txdma_slot = -1; 334 sc->bsy_txdma_slots = 0; 335 336 error = lv1_net_start_rx_dma(sc->sc_bus, sc->sc_dev, 337 sc->sc_rxsoft[0].rxs_desc, 0); 338 if (error != 0) 339 device_printf(sc->sc_self, 340 "lv1_net_start_rx_dma error: %d\n", error); 341 342 sc->sc_ifp->if_drv_flags |= IFF_DRV_RUNNING; 343 sc->sc_ifp->if_drv_flags &= ~IFF_DRV_OACTIVE; 344 sc->sc_ifpflags = sc->sc_ifp->if_flags; 345 346 sc->sc_wdog_timer = 0; 347 callout_reset(&sc->sc_tick_ch, hz, glc_tick, sc); 348 } 349 350 static void 351 glc_stop(void *xsc) 352 { 353 struct glc_softc *sc = xsc; 354 355 mtx_assert(&sc->sc_mtx, MA_OWNED); 356 357 lv1_net_stop_tx_dma(sc->sc_bus, sc->sc_dev, 0); 358 lv1_net_stop_rx_dma(sc->sc_bus, sc->sc_dev, 0); 359 } 360 361 static void 362 glc_init(void *xsc) 363 { 364 struct glc_softc *sc = xsc; 365 366 mtx_lock(&sc->sc_mtx); 367 glc_init_locked(sc); 368 mtx_unlock(&sc->sc_mtx); 369 } 370 371 static void 372 glc_tick(void *xsc) 373 { 374 struct glc_softc *sc = xsc; 375 376 mtx_assert(&sc->sc_mtx, MA_OWNED); 377 378 /* 379 * XXX: Sometimes the RX queue gets stuck. Poke it periodically until 380 * we figure out why. This will fail harmlessly if the RX queue is 381 * already running. 382 */ 383 lv1_net_start_rx_dma(sc->sc_bus, sc->sc_dev, 384 sc->sc_rxsoft[sc->sc_next_rxdma_slot].rxs_desc, 0); 385 386 if (sc->sc_wdog_timer == 0 || --sc->sc_wdog_timer != 0) { 387 callout_reset(&sc->sc_tick_ch, hz, glc_tick, sc); 388 return; 389 } 390 391 /* Problems */ 392 device_printf(sc->sc_self, "device timeout\n"); 393 394 glc_init_locked(sc); 395 } 396 397 static void 398 glc_start_locked(struct ifnet *ifp) 399 { 400 struct glc_softc *sc = ifp->if_softc; 401 bus_addr_t first, pktdesc; 402 int kickstart = 0; 403 int error; 404 struct mbuf *mb_head; 405 406 mtx_assert(&sc->sc_mtx, MA_OWNED); 407 first = 0; 408 409 if ((ifp->if_drv_flags & (IFF_DRV_RUNNING | IFF_DRV_OACTIVE)) != 410 IFF_DRV_RUNNING) 411 return; 412 413 if (STAILQ_EMPTY(&sc->sc_txdirtyq)) 414 kickstart = 1; 415 416 while (!IFQ_DRV_IS_EMPTY(&ifp->if_snd)) { 417 IFQ_DRV_DEQUEUE(&ifp->if_snd, mb_head); 418 419 if (mb_head == NULL) 420 break; 421 422 /* Check if the ring buffer is full */ 423 if (sc->bsy_txdma_slots > 125) { 424 /* Put the packet back and stop */ 425 ifp->if_drv_flags |= IFF_DRV_OACTIVE; 426 IFQ_DRV_PREPEND(&ifp->if_snd, mb_head); 427 break; 428 } 429 430 BPF_MTAP(ifp, mb_head); 431 432 if (sc->sc_tx_vlan >= 0) 433 mb_head = ether_vlanencap(mb_head, sc->sc_tx_vlan); 434 435 if (glc_encap(sc, &mb_head, &pktdesc)) { 436 ifp->if_drv_flags |= IFF_DRV_OACTIVE; 437 break; 438 } 439 440 if (first == 0) 441 first = pktdesc; 442 } 443 444 if (kickstart && first != 0) { 445 error = lv1_net_start_tx_dma(sc->sc_bus, sc->sc_dev, first, 0); 446 if (error != 0) 447 device_printf(sc->sc_self, 448 "lv1_net_start_tx_dma error: %d\n", error); 449 sc->sc_wdog_timer = 5; 450 } 451 } 452 453 static void 454 glc_start(struct ifnet *ifp) 455 { 456 struct glc_softc *sc = ifp->if_softc; 457 458 mtx_lock(&sc->sc_mtx); 459 glc_start_locked(ifp); 460 mtx_unlock(&sc->sc_mtx); 461 } 462 463 static int 464 glc_ioctl(struct ifnet *ifp, u_long cmd, caddr_t data) 465 { 466 struct glc_softc *sc = ifp->if_softc; 467 struct ifreq *ifr = (struct ifreq *)data; 468 int err = 0; 469 470 switch (cmd) { 471 case SIOCSIFFLAGS: 472 mtx_lock(&sc->sc_mtx); 473 if ((ifp->if_flags & IFF_UP) != 0) { 474 if ((ifp->if_drv_flags & IFF_DRV_RUNNING) != 0 && 475 ((ifp->if_flags ^ sc->sc_ifpflags) & 476 (IFF_ALLMULTI | IFF_PROMISC)) != 0) 477 glc_set_multicast(sc); 478 else 479 glc_init_locked(sc); 480 } 481 else if ((ifp->if_drv_flags & IFF_DRV_RUNNING) != 0) 482 glc_stop(sc); 483 sc->sc_ifpflags = ifp->if_flags; 484 mtx_unlock(&sc->sc_mtx); 485 break; 486 case SIOCADDMULTI: 487 case SIOCDELMULTI: 488 mtx_lock(&sc->sc_mtx); 489 glc_set_multicast(sc); 490 mtx_unlock(&sc->sc_mtx); 491 break; 492 case SIOCGIFMEDIA: 493 case SIOCSIFMEDIA: 494 err = ifmedia_ioctl(ifp, ifr, &sc->sc_media, cmd); 495 break; 496 default: 497 err = ether_ioctl(ifp, cmd, data); 498 break; 499 } 500 501 return (err); 502 } 503 504 static u_int 505 glc_add_maddr(void *arg, struct sockaddr_dl *sdl, u_int cnt) 506 { 507 struct glc_softc *sc = arg; 508 uint64_t addr; 509 510 /* 511 * Filter can only hold 32 addresses, so fall back to 512 * the IFF_ALLMULTI case if we have too many. +1 is for 513 * broadcast. 514 */ 515 if (cnt + 1 == 32) 516 return (0); 517 518 addr = 0; 519 memcpy(&((uint8_t *)(&addr))[2], LLADDR(sdl), ETHER_ADDR_LEN); 520 lv1_net_add_multicast_address(sc->sc_bus, sc->sc_dev, addr, 0); 521 522 return (1); 523 } 524 525 static void 526 glc_set_multicast(struct glc_softc *sc) 527 { 528 struct ifnet *ifp = sc->sc_ifp; 529 int naddrs; 530 531 /* Clear multicast filter */ 532 lv1_net_remove_multicast_address(sc->sc_bus, sc->sc_dev, 0, 1); 533 534 /* Add broadcast */ 535 lv1_net_add_multicast_address(sc->sc_bus, sc->sc_dev, 536 0xffffffffffffL, 0); 537 538 if ((ifp->if_flags & IFF_ALLMULTI) != 0) { 539 lv1_net_add_multicast_address(sc->sc_bus, sc->sc_dev, 0, 1); 540 } else { 541 naddrs = if_foreach_llmaddr(ifp, glc_add_maddr, sc); 542 if (naddrs + 1 == 32) 543 lv1_net_add_multicast_address(sc->sc_bus, 544 sc->sc_dev, 0, 1); 545 } 546 } 547 548 static int 549 glc_add_rxbuf(struct glc_softc *sc, int idx) 550 { 551 struct glc_rxsoft *rxs = &sc->sc_rxsoft[idx]; 552 struct mbuf *m; 553 bus_dma_segment_t segs[1]; 554 int error, nsegs; 555 556 m = m_getcl(M_NOWAIT, MT_DATA, M_PKTHDR); 557 if (m == NULL) 558 return (ENOBUFS); 559 m->m_len = m->m_pkthdr.len = m->m_ext.ext_size; 560 561 if (rxs->rxs_mbuf != NULL) { 562 bus_dmamap_sync(sc->sc_rxdma_tag, rxs->rxs_dmamap, 563 BUS_DMASYNC_POSTREAD); 564 bus_dmamap_unload(sc->sc_rxdma_tag, rxs->rxs_dmamap); 565 } 566 567 error = bus_dmamap_load_mbuf_sg(sc->sc_rxdma_tag, rxs->rxs_dmamap, m, 568 segs, &nsegs, BUS_DMA_NOWAIT); 569 if (error != 0) { 570 device_printf(sc->sc_self, 571 "cannot load RS DMA map %d, error = %d\n", idx, error); 572 m_freem(m); 573 return (error); 574 } 575 /* If nsegs is wrong then the stack is corrupt. */ 576 KASSERT(nsegs == 1, 577 ("%s: too many DMA segments (%d)", __func__, nsegs)); 578 rxs->rxs_mbuf = m; 579 rxs->segment = segs[0]; 580 581 bus_dmamap_sync(sc->sc_rxdma_tag, rxs->rxs_dmamap, BUS_DMASYNC_PREREAD); 582 583 return (0); 584 } 585 586 static int 587 glc_add_rxbuf_dma(struct glc_softc *sc, int idx) 588 { 589 struct glc_rxsoft *rxs = &sc->sc_rxsoft[idx]; 590 591 bzero(&sc->sc_rxdmadesc[idx], sizeof(sc->sc_rxdmadesc[idx])); 592 sc->sc_rxdmadesc[idx].paddr = rxs->segment.ds_addr; 593 sc->sc_rxdmadesc[idx].len = rxs->segment.ds_len; 594 sc->sc_rxdmadesc[idx].next = sc->sc_rxdmadesc_phys + 595 ((idx + 1) % GLC_MAX_RX_PACKETS)*sizeof(sc->sc_rxdmadesc[idx]); 596 sc->sc_rxdmadesc[idx].cmd_stat = GELIC_DESCR_OWNED; 597 598 rxs->rxs_desc_slot = idx; 599 rxs->rxs_desc = sc->sc_rxdmadesc_phys + idx*sizeof(struct glc_dmadesc); 600 601 return (0); 602 } 603 604 static int 605 glc_encap(struct glc_softc *sc, struct mbuf **m_head, bus_addr_t *pktdesc) 606 { 607 bus_dma_segment_t segs[16]; 608 struct glc_txsoft *txs; 609 struct mbuf *m; 610 bus_addr_t firstslotphys; 611 int i, idx, nsegs, nsegs_max; 612 int err = 0; 613 614 /* Max number of segments is the number of free DMA slots */ 615 nsegs_max = 128 - sc->bsy_txdma_slots; 616 617 if (nsegs_max > 16 || sc->first_used_txdma_slot < 0) 618 nsegs_max = 16; 619 620 /* Get a work queue entry. */ 621 if ((txs = STAILQ_FIRST(&sc->sc_txfreeq)) == NULL) { 622 /* Ran out of descriptors. */ 623 return (ENOBUFS); 624 } 625 626 nsegs = 0; 627 for (m = *m_head; m != NULL; m = m->m_next) 628 nsegs++; 629 630 if (nsegs > nsegs_max) { 631 m = m_collapse(*m_head, M_NOWAIT, nsegs_max); 632 if (m == NULL) { 633 m_freem(*m_head); 634 *m_head = NULL; 635 return (ENOBUFS); 636 } 637 *m_head = m; 638 } 639 640 err = bus_dmamap_load_mbuf_sg(sc->sc_txdma_tag, txs->txs_dmamap, 641 *m_head, segs, &nsegs, BUS_DMA_NOWAIT); 642 if (err != 0) { 643 m_freem(*m_head); 644 *m_head = NULL; 645 return (err); 646 } 647 648 KASSERT(nsegs <= 128 - sc->bsy_txdma_slots, 649 ("GLC: Mapped too many (%d) DMA segments with %d available", 650 nsegs, 128 - sc->bsy_txdma_slots)); 651 652 if (nsegs == 0) { 653 m_freem(*m_head); 654 *m_head = NULL; 655 return (EIO); 656 } 657 658 txs->txs_ndescs = nsegs; 659 txs->txs_firstdesc = sc->next_txdma_slot; 660 661 idx = txs->txs_firstdesc; 662 firstslotphys = sc->sc_txdmadesc_phys + 663 txs->txs_firstdesc*sizeof(struct glc_dmadesc); 664 665 for (i = 0; i < nsegs; i++) { 666 bzero(&sc->sc_txdmadesc[idx], sizeof(sc->sc_txdmadesc[idx])); 667 sc->sc_txdmadesc[idx].paddr = segs[i].ds_addr; 668 sc->sc_txdmadesc[idx].len = segs[i].ds_len; 669 sc->sc_txdmadesc[idx].next = sc->sc_txdmadesc_phys + 670 ((idx + 1) % GLC_MAX_TX_PACKETS)*sizeof(struct glc_dmadesc); 671 sc->sc_txdmadesc[idx].cmd_stat |= GELIC_CMDSTAT_NOIPSEC; 672 673 if (i+1 == nsegs) { 674 txs->txs_lastdesc = idx; 675 sc->sc_txdmadesc[idx].next = 0; 676 sc->sc_txdmadesc[idx].cmd_stat |= GELIC_CMDSTAT_LAST; 677 } 678 679 if ((*m_head)->m_pkthdr.csum_flags & CSUM_TCP) 680 sc->sc_txdmadesc[idx].cmd_stat |= GELIC_CMDSTAT_CSUM_TCP; 681 if ((*m_head)->m_pkthdr.csum_flags & CSUM_UDP) 682 sc->sc_txdmadesc[idx].cmd_stat |= GELIC_CMDSTAT_CSUM_UDP; 683 sc->sc_txdmadesc[idx].cmd_stat |= GELIC_DESCR_OWNED; 684 685 idx = (idx + 1) % GLC_MAX_TX_PACKETS; 686 } 687 sc->next_txdma_slot = idx; 688 sc->bsy_txdma_slots += nsegs; 689 if (txs->txs_firstdesc != 0) 690 idx = txs->txs_firstdesc - 1; 691 else 692 idx = GLC_MAX_TX_PACKETS - 1; 693 694 if (sc->first_used_txdma_slot < 0) 695 sc->first_used_txdma_slot = txs->txs_firstdesc; 696 697 bus_dmamap_sync(sc->sc_txdma_tag, txs->txs_dmamap, 698 BUS_DMASYNC_PREWRITE); 699 sc->sc_txdmadesc[idx].next = firstslotphys; 700 701 STAILQ_REMOVE_HEAD(&sc->sc_txfreeq, txs_q); 702 STAILQ_INSERT_TAIL(&sc->sc_txdirtyq, txs, txs_q); 703 txs->txs_mbuf = *m_head; 704 *pktdesc = firstslotphys; 705 706 return (0); 707 } 708 709 static void 710 glc_rxintr(struct glc_softc *sc) 711 { 712 int i, restart_rxdma, error; 713 struct mbuf *m; 714 struct ifnet *ifp = sc->sc_ifp; 715 716 bus_dmamap_sync(sc->sc_dmadesc_tag, sc->sc_rxdmadesc_map, 717 BUS_DMASYNC_POSTREAD); 718 719 restart_rxdma = 0; 720 while ((sc->sc_rxdmadesc[sc->sc_next_rxdma_slot].cmd_stat & 721 GELIC_DESCR_OWNED) == 0) { 722 i = sc->sc_next_rxdma_slot; 723 sc->sc_next_rxdma_slot++; 724 if (sc->sc_next_rxdma_slot >= GLC_MAX_RX_PACKETS) 725 sc->sc_next_rxdma_slot = 0; 726 727 if (sc->sc_rxdmadesc[i].cmd_stat & GELIC_CMDSTAT_CHAIN_END) 728 restart_rxdma = 1; 729 730 if (sc->sc_rxdmadesc[i].rxerror & GELIC_RXERRORS) { 731 if_inc_counter(ifp, IFCOUNTER_IERRORS, 1); 732 goto requeue; 733 } 734 735 m = sc->sc_rxsoft[i].rxs_mbuf; 736 if (sc->sc_rxdmadesc[i].data_stat & GELIC_RX_IPCSUM) { 737 m->m_pkthdr.csum_flags |= 738 CSUM_IP_CHECKED | CSUM_IP_VALID; 739 } 740 if (sc->sc_rxdmadesc[i].data_stat & GELIC_RX_TCPUDPCSUM) { 741 m->m_pkthdr.csum_flags |= 742 CSUM_DATA_VALID | CSUM_PSEUDO_HDR; 743 m->m_pkthdr.csum_data = 0xffff; 744 } 745 746 if (glc_add_rxbuf(sc, i)) { 747 if_inc_counter(ifp, IFCOUNTER_IERRORS, 1); 748 goto requeue; 749 } 750 751 if_inc_counter(ifp, IFCOUNTER_IPACKETS, 1); 752 m->m_pkthdr.rcvif = ifp; 753 m->m_len = sc->sc_rxdmadesc[i].valid_size; 754 m->m_pkthdr.len = m->m_len; 755 756 /* 757 * Remove VLAN tag. Even on early firmwares that do not allow 758 * multiple VLANs, the VLAN tag is still in place here. 759 */ 760 m_adj(m, 2); 761 762 mtx_unlock(&sc->sc_mtx); 763 (*ifp->if_input)(ifp, m); 764 mtx_lock(&sc->sc_mtx); 765 766 requeue: 767 glc_add_rxbuf_dma(sc, i); 768 } 769 770 bus_dmamap_sync(sc->sc_dmadesc_tag, sc->sc_rxdmadesc_map, 771 BUS_DMASYNC_PREWRITE); 772 773 if (restart_rxdma) { 774 error = lv1_net_start_rx_dma(sc->sc_bus, sc->sc_dev, 775 sc->sc_rxsoft[sc->sc_next_rxdma_slot].rxs_desc, 0); 776 if (error != 0) 777 device_printf(sc->sc_self, 778 "lv1_net_start_rx_dma error: %d\n", error); 779 } 780 } 781 782 static void 783 glc_txintr(struct glc_softc *sc) 784 { 785 struct ifnet *ifp = sc->sc_ifp; 786 struct glc_txsoft *txs; 787 int progress = 0, kickstart = 0, error; 788 789 bus_dmamap_sync(sc->sc_dmadesc_tag, sc->sc_txdmadesc_map, 790 BUS_DMASYNC_POSTREAD); 791 792 while ((txs = STAILQ_FIRST(&sc->sc_txdirtyq)) != NULL) { 793 if (sc->sc_txdmadesc[txs->txs_lastdesc].cmd_stat 794 & GELIC_DESCR_OWNED) 795 break; 796 797 STAILQ_REMOVE_HEAD(&sc->sc_txdirtyq, txs_q); 798 bus_dmamap_unload(sc->sc_txdma_tag, txs->txs_dmamap); 799 sc->bsy_txdma_slots -= txs->txs_ndescs; 800 801 if (txs->txs_mbuf != NULL) { 802 m_freem(txs->txs_mbuf); 803 txs->txs_mbuf = NULL; 804 } 805 806 if ((sc->sc_txdmadesc[txs->txs_lastdesc].cmd_stat & 0xf0000000) 807 != 0) { 808 lv1_net_stop_tx_dma(sc->sc_bus, sc->sc_dev, 0); 809 kickstart = 1; 810 if_inc_counter(ifp, IFCOUNTER_OERRORS, 1); 811 } 812 813 if (sc->sc_txdmadesc[txs->txs_lastdesc].cmd_stat & 814 GELIC_CMDSTAT_CHAIN_END) 815 kickstart = 1; 816 817 STAILQ_INSERT_TAIL(&sc->sc_txfreeq, txs, txs_q); 818 if_inc_counter(ifp, IFCOUNTER_OPACKETS, 1); 819 progress = 1; 820 } 821 822 if (txs != NULL) 823 sc->first_used_txdma_slot = txs->txs_firstdesc; 824 else 825 sc->first_used_txdma_slot = -1; 826 827 if (kickstart || txs != NULL) { 828 /* Speculatively (or necessarily) start the TX queue again */ 829 error = lv1_net_start_tx_dma(sc->sc_bus, sc->sc_dev, 830 sc->sc_txdmadesc_phys + 831 ((txs == NULL) ? 0 : txs->txs_firstdesc)* 832 sizeof(struct glc_dmadesc), 0); 833 if (error != 0) 834 device_printf(sc->sc_self, 835 "lv1_net_start_tx_dma error: %d\n", error); 836 } 837 838 if (progress) { 839 /* 840 * We freed some descriptors, so reset IFF_DRV_OACTIVE 841 * and restart. 842 */ 843 ifp->if_drv_flags &= ~IFF_DRV_OACTIVE; 844 sc->sc_wdog_timer = STAILQ_EMPTY(&sc->sc_txdirtyq) ? 0 : 5; 845 846 if ((ifp->if_drv_flags & IFF_DRV_RUNNING) && 847 !IFQ_DRV_IS_EMPTY(&ifp->if_snd)) 848 glc_start_locked(ifp); 849 } 850 } 851 852 static int 853 glc_intr_filter(void *xsc) 854 { 855 struct glc_softc *sc = xsc; 856 857 powerpc_sync(); 858 atomic_set_64(&sc->sc_interrupt_status, *sc->sc_hwirq_status); 859 return (FILTER_SCHEDULE_THREAD); 860 } 861 862 static void 863 glc_intr(void *xsc) 864 { 865 struct glc_softc *sc = xsc; 866 uint64_t status, linkstat, junk; 867 868 mtx_lock(&sc->sc_mtx); 869 870 status = atomic_readandclear_64(&sc->sc_interrupt_status); 871 872 if (status == 0) { 873 mtx_unlock(&sc->sc_mtx); 874 return; 875 } 876 877 if (status & (GELIC_INT_RXDONE | GELIC_INT_RXFRAME)) 878 glc_rxintr(sc); 879 880 if (status & (GELIC_INT_TXDONE | GELIC_INT_TX_CHAIN_END)) 881 glc_txintr(sc); 882 883 if (status & GELIC_INT_PHY) { 884 lv1_net_control(sc->sc_bus, sc->sc_dev, GELIC_GET_LINK_STATUS, 885 GELIC_VLAN_TX_ETHERNET, 0, 0, &linkstat, &junk); 886 887 linkstat = (linkstat & GELIC_LINK_UP) ? 888 LINK_STATE_UP : LINK_STATE_DOWN; 889 if (linkstat != sc->sc_ifp->if_link_state) 890 if_link_state_change(sc->sc_ifp, linkstat); 891 } 892 893 mtx_unlock(&sc->sc_mtx); 894 } 895 896 static void 897 glc_media_status(struct ifnet *ifp, struct ifmediareq *ifmr) 898 { 899 struct glc_softc *sc = ifp->if_softc; 900 uint64_t status, junk; 901 902 ifmr->ifm_status = IFM_AVALID; 903 ifmr->ifm_active = IFM_ETHER; 904 905 lv1_net_control(sc->sc_bus, sc->sc_dev, GELIC_GET_LINK_STATUS, 906 GELIC_VLAN_TX_ETHERNET, 0, 0, &status, &junk); 907 908 if (status & GELIC_LINK_UP) 909 ifmr->ifm_status |= IFM_ACTIVE; 910 911 if (status & GELIC_SPEED_10) 912 ifmr->ifm_active |= IFM_10_T; 913 else if (status & GELIC_SPEED_100) 914 ifmr->ifm_active |= IFM_100_TX; 915 else if (status & GELIC_SPEED_1000) 916 ifmr->ifm_active |= IFM_1000_T; 917 918 if (status & GELIC_FULL_DUPLEX) 919 ifmr->ifm_active |= IFM_FDX; 920 else 921 ifmr->ifm_active |= IFM_HDX; 922 } 923 924 static int 925 glc_media_change(struct ifnet *ifp) 926 { 927 struct glc_softc *sc = ifp->if_softc; 928 uint64_t mode, junk; 929 int result; 930 931 if (IFM_TYPE(sc->sc_media.ifm_media) != IFM_ETHER) 932 return (EINVAL); 933 934 switch (IFM_SUBTYPE(sc->sc_media.ifm_media)) { 935 case IFM_AUTO: 936 mode = GELIC_AUTO_NEG; 937 break; 938 case IFM_10_T: 939 mode = GELIC_SPEED_10; 940 break; 941 case IFM_100_TX: 942 mode = GELIC_SPEED_100; 943 break; 944 case IFM_1000_T: 945 mode = GELIC_SPEED_1000 | GELIC_FULL_DUPLEX; 946 break; 947 default: 948 return (EINVAL); 949 } 950 951 if (IFM_OPTIONS(sc->sc_media.ifm_media) & IFM_FDX) 952 mode |= GELIC_FULL_DUPLEX; 953 954 result = lv1_net_control(sc->sc_bus, sc->sc_dev, GELIC_SET_LINK_MODE, 955 GELIC_VLAN_TX_ETHERNET, mode, 0, &junk, &junk); 956 957 return (result ? EIO : 0); 958 } 959