1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright 2008 Sun Microsystems, Inc. All rights reserved. 23 * Use is subject to license terms. 24 */ 25 26 #include "rge.h" 27 28 /* 29 * This is the string displayed by modinfo, etc. 30 * Make sure you keep the version ID up to date! 31 */ 32 static char rge_ident[] = "Realtek 1Gb Ethernet v1.11"; 33 34 /* 35 * Used for buffers allocated by ddi_dma_mem_alloc() 36 */ 37 static ddi_dma_attr_t dma_attr_buf = { 38 DMA_ATTR_V0, /* dma_attr version */ 39 (uint32_t)0, /* dma_attr_addr_lo */ 40 (uint32_t)0xFFFFFFFF, /* dma_attr_addr_hi */ 41 (uint32_t)0xFFFFFFFF, /* dma_attr_count_max */ 42 (uint32_t)16, /* dma_attr_align */ 43 0xFFFFFFFF, /* dma_attr_burstsizes */ 44 1, /* dma_attr_minxfer */ 45 (uint32_t)0xFFFFFFFF, /* dma_attr_maxxfer */ 46 (uint32_t)0xFFFFFFFF, /* dma_attr_seg */ 47 1, /* dma_attr_sgllen */ 48 1, /* dma_attr_granular */ 49 0, /* dma_attr_flags */ 50 }; 51 52 /* 53 * Used for BDs allocated by ddi_dma_mem_alloc() 54 */ 55 static ddi_dma_attr_t dma_attr_desc = { 56 DMA_ATTR_V0, /* dma_attr version */ 57 (uint32_t)0, /* dma_attr_addr_lo */ 58 (uint32_t)0xFFFFFFFF, /* dma_attr_addr_hi */ 59 (uint32_t)0xFFFFFFFF, /* dma_attr_count_max */ 60 (uint32_t)256, /* dma_attr_align */ 61 0xFFFFFFFF, /* dma_attr_burstsizes */ 62 1, /* dma_attr_minxfer */ 63 (uint32_t)0xFFFFFFFF, /* dma_attr_maxxfer */ 64 (uint32_t)0xFFFFFFFF, /* dma_attr_seg */ 65 1, /* dma_attr_sgllen */ 66 1, /* dma_attr_granular */ 67 0, /* dma_attr_flags */ 68 }; 69 70 /* 71 * PIO access attributes for registers 72 */ 73 static ddi_device_acc_attr_t rge_reg_accattr = { 74 DDI_DEVICE_ATTR_V0, 75 DDI_STRUCTURE_LE_ACC, 76 DDI_STRICTORDER_ACC, 77 DDI_DEFAULT_ACC 78 }; 79 80 /* 81 * DMA access attributes for descriptors 82 */ 83 static ddi_device_acc_attr_t rge_desc_accattr = { 84 DDI_DEVICE_ATTR_V0, 85 DDI_NEVERSWAP_ACC, 86 DDI_STRICTORDER_ACC, 87 DDI_DEFAULT_ACC 88 }; 89 90 /* 91 * DMA access attributes for data 92 */ 93 static ddi_device_acc_attr_t rge_buf_accattr = { 94 DDI_DEVICE_ATTR_V0, 95 DDI_NEVERSWAP_ACC, 96 DDI_STRICTORDER_ACC, 97 DDI_DEFAULT_ACC 98 }; 99 100 /* 101 * Property names 102 */ 103 static char debug_propname[] = "rge_debug_flags"; 104 static char mtu_propname[] = "default_mtu"; 105 static char msi_propname[] = "msi_enable"; 106 107 static int rge_m_start(void *); 108 static void rge_m_stop(void *); 109 static int rge_m_promisc(void *, boolean_t); 110 static int rge_m_multicst(void *, boolean_t, const uint8_t *); 111 static int rge_m_unicst(void *, const uint8_t *); 112 static void rge_m_resources(void *); 113 static void rge_m_ioctl(void *, queue_t *, mblk_t *); 114 static boolean_t rge_m_getcapab(void *, mac_capab_t, void *); 115 116 #define RGE_M_CALLBACK_FLAGS (MC_RESOURCES | MC_IOCTL | MC_GETCAPAB) 117 118 static mac_callbacks_t rge_m_callbacks = { 119 RGE_M_CALLBACK_FLAGS, 120 rge_m_stat, 121 rge_m_start, 122 rge_m_stop, 123 rge_m_promisc, 124 rge_m_multicst, 125 rge_m_unicst, 126 rge_m_tx, 127 rge_m_resources, 128 rge_m_ioctl, 129 rge_m_getcapab 130 }; 131 132 /* 133 * Allocate an area of memory and a DMA handle for accessing it 134 */ 135 static int 136 rge_alloc_dma_mem(rge_t *rgep, size_t memsize, ddi_dma_attr_t *dma_attr_p, 137 ddi_device_acc_attr_t *acc_attr_p, uint_t dma_flags, dma_area_t *dma_p) 138 { 139 caddr_t vaddr; 140 int err; 141 142 /* 143 * Allocate handle 144 */ 145 err = ddi_dma_alloc_handle(rgep->devinfo, dma_attr_p, 146 DDI_DMA_SLEEP, NULL, &dma_p->dma_hdl); 147 if (err != DDI_SUCCESS) { 148 dma_p->dma_hdl = NULL; 149 return (DDI_FAILURE); 150 } 151 152 /* 153 * Allocate memory 154 */ 155 err = ddi_dma_mem_alloc(dma_p->dma_hdl, memsize, acc_attr_p, 156 dma_flags & (DDI_DMA_CONSISTENT | DDI_DMA_STREAMING), 157 DDI_DMA_SLEEP, NULL, &vaddr, &dma_p->alength, &dma_p->acc_hdl); 158 if (err != DDI_SUCCESS) { 159 ddi_dma_free_handle(&dma_p->dma_hdl); 160 dma_p->dma_hdl = NULL; 161 dma_p->acc_hdl = NULL; 162 return (DDI_FAILURE); 163 } 164 165 /* 166 * Bind the two together 167 */ 168 dma_p->mem_va = vaddr; 169 err = ddi_dma_addr_bind_handle(dma_p->dma_hdl, NULL, 170 vaddr, dma_p->alength, dma_flags, DDI_DMA_SLEEP, NULL, 171 &dma_p->cookie, &dma_p->ncookies); 172 if (err != DDI_DMA_MAPPED || dma_p->ncookies != 1) { 173 ddi_dma_mem_free(&dma_p->acc_hdl); 174 ddi_dma_free_handle(&dma_p->dma_hdl); 175 dma_p->acc_hdl = NULL; 176 dma_p->dma_hdl = NULL; 177 return (DDI_FAILURE); 178 } 179 180 dma_p->nslots = ~0U; 181 dma_p->size = ~0U; 182 dma_p->token = ~0U; 183 dma_p->offset = 0; 184 return (DDI_SUCCESS); 185 } 186 187 /* 188 * Free one allocated area of DMAable memory 189 */ 190 static void 191 rge_free_dma_mem(dma_area_t *dma_p) 192 { 193 if (dma_p->dma_hdl != NULL) { 194 if (dma_p->ncookies) { 195 (void) ddi_dma_unbind_handle(dma_p->dma_hdl); 196 dma_p->ncookies = 0; 197 } 198 ddi_dma_free_handle(&dma_p->dma_hdl); 199 dma_p->dma_hdl = NULL; 200 } 201 202 if (dma_p->acc_hdl != NULL) { 203 ddi_dma_mem_free(&dma_p->acc_hdl); 204 dma_p->acc_hdl = NULL; 205 } 206 } 207 208 /* 209 * Utility routine to carve a slice off a chunk of allocated memory, 210 * updating the chunk descriptor accordingly. The size of the slice 211 * is given by the product of the <qty> and <size> parameters. 212 */ 213 static void 214 rge_slice_chunk(dma_area_t *slice, dma_area_t *chunk, 215 uint32_t qty, uint32_t size) 216 { 217 static uint32_t sequence = 0xbcd5704a; 218 size_t totsize; 219 220 totsize = qty*size; 221 ASSERT(totsize <= chunk->alength); 222 223 *slice = *chunk; 224 slice->nslots = qty; 225 slice->size = size; 226 slice->alength = totsize; 227 slice->token = ++sequence; 228 229 chunk->mem_va = (caddr_t)chunk->mem_va + totsize; 230 chunk->alength -= totsize; 231 chunk->offset += totsize; 232 chunk->cookie.dmac_laddress += totsize; 233 chunk->cookie.dmac_size -= totsize; 234 } 235 236 static int 237 rge_alloc_bufs(rge_t *rgep) 238 { 239 size_t txdescsize; 240 size_t rxdescsize; 241 int err; 242 243 /* 244 * Allocate memory & handle for packet statistics 245 */ 246 err = rge_alloc_dma_mem(rgep, 247 RGE_STATS_DUMP_SIZE, 248 &dma_attr_desc, 249 &rge_desc_accattr, 250 DDI_DMA_RDWR | DDI_DMA_CONSISTENT, 251 &rgep->dma_area_stats); 252 if (err != DDI_SUCCESS) 253 return (DDI_FAILURE); 254 rgep->hw_stats = DMA_VPTR(rgep->dma_area_stats); 255 256 /* 257 * Allocate memory & handle for Tx descriptor ring 258 */ 259 txdescsize = RGE_SEND_SLOTS * sizeof (rge_bd_t); 260 err = rge_alloc_dma_mem(rgep, 261 txdescsize, 262 &dma_attr_desc, 263 &rge_desc_accattr, 264 DDI_DMA_RDWR | DDI_DMA_CONSISTENT, 265 &rgep->dma_area_txdesc); 266 if (err != DDI_SUCCESS) 267 return (DDI_FAILURE); 268 269 /* 270 * Allocate memory & handle for Rx descriptor ring 271 */ 272 rxdescsize = RGE_RECV_SLOTS * sizeof (rge_bd_t); 273 err = rge_alloc_dma_mem(rgep, 274 rxdescsize, 275 &dma_attr_desc, 276 &rge_desc_accattr, 277 DDI_DMA_RDWR | DDI_DMA_CONSISTENT, 278 &rgep->dma_area_rxdesc); 279 if (err != DDI_SUCCESS) 280 return (DDI_FAILURE); 281 282 return (DDI_SUCCESS); 283 } 284 285 /* 286 * rge_free_bufs() -- free descriptors/buffers allocated for this 287 * device instance. 288 */ 289 static void 290 rge_free_bufs(rge_t *rgep) 291 { 292 rge_free_dma_mem(&rgep->dma_area_stats); 293 rge_free_dma_mem(&rgep->dma_area_txdesc); 294 rge_free_dma_mem(&rgep->dma_area_rxdesc); 295 } 296 297 /* 298 * ========== Transmit and receive ring reinitialisation ========== 299 */ 300 301 /* 302 * These <reinit> routines each reset the rx/tx rings to an initial 303 * state, assuming that the corresponding <init> routine has already 304 * been called exactly once. 305 */ 306 static void 307 rge_reinit_send_ring(rge_t *rgep) 308 { 309 sw_sbd_t *ssbdp; 310 rge_bd_t *bdp; 311 uint32_t slot; 312 313 /* 314 * re-init send ring 315 */ 316 DMA_ZERO(rgep->tx_desc); 317 ssbdp = rgep->sw_sbds; 318 bdp = rgep->tx_ring; 319 for (slot = 0; slot < RGE_SEND_SLOTS; slot++) { 320 bdp->host_buf_addr = 321 RGE_BSWAP_32(ssbdp->pbuf.cookie.dmac_laddress); 322 bdp->host_buf_addr_hi = 323 RGE_BSWAP_32(ssbdp->pbuf.cookie.dmac_laddress >> 32); 324 /* last BD in Tx ring */ 325 if (slot == (RGE_SEND_SLOTS - 1)) 326 bdp->flags_len = RGE_BSWAP_32(BD_FLAG_EOR); 327 ssbdp++; 328 bdp++; 329 } 330 DMA_SYNC(rgep->tx_desc, DDI_DMA_SYNC_FORDEV); 331 rgep->tx_next = 0; 332 rgep->tc_next = 0; 333 rgep->tc_tail = 0; 334 rgep->tx_flow = 0; 335 rgep->tx_free = RGE_SEND_SLOTS; 336 } 337 338 static void 339 rge_reinit_recv_ring(rge_t *rgep) 340 { 341 rge_bd_t *bdp; 342 sw_rbd_t *srbdp; 343 dma_area_t *pbuf; 344 uint32_t slot; 345 346 /* 347 * re-init receive ring 348 */ 349 DMA_ZERO(rgep->rx_desc); 350 srbdp = rgep->sw_rbds; 351 bdp = rgep->rx_ring; 352 for (slot = 0; slot < RGE_RECV_SLOTS; slot++) { 353 pbuf = &srbdp->rx_buf->pbuf; 354 bdp->host_buf_addr = 355 RGE_BSWAP_32(pbuf->cookie.dmac_laddress + rgep->head_room); 356 bdp->host_buf_addr_hi = 357 RGE_BSWAP_32(pbuf->cookie.dmac_laddress >> 32); 358 bdp->flags_len = RGE_BSWAP_32(BD_FLAG_HW_OWN | 359 (rgep->rxbuf_size - rgep->head_room)); 360 /* last BD in Tx ring */ 361 if (slot == (RGE_RECV_SLOTS - 1)) 362 bdp->flags_len |= RGE_BSWAP_32(BD_FLAG_EOR); 363 srbdp++; 364 bdp++; 365 } 366 DMA_SYNC(rgep->rx_desc, DDI_DMA_SYNC_FORDEV); 367 rgep->watchdog = 0; 368 rgep->rx_next = 0; 369 } 370 371 static void 372 rge_reinit_buf_ring(rge_t *rgep) 373 { 374 375 if (rgep->chip_flags & CHIP_FLAG_FORCE_BCOPY) 376 return; 377 378 /* 379 * If all the up-sending buffers haven't been returned to driver, 380 * use bcopy() only in rx process. 381 */ 382 if (rgep->rx_free != RGE_BUF_SLOTS) 383 rgep->rx_bcopy = B_TRUE; 384 } 385 386 static void 387 rge_reinit_rings(rge_t *rgep) 388 { 389 rge_reinit_send_ring(rgep); 390 rge_reinit_recv_ring(rgep); 391 rge_reinit_buf_ring(rgep); 392 } 393 394 static void 395 rge_fini_send_ring(rge_t *rgep) 396 { 397 sw_sbd_t *ssbdp; 398 uint32_t slot; 399 400 ssbdp = rgep->sw_sbds; 401 for (slot = 0; slot < RGE_SEND_SLOTS; ++slot) { 402 rge_free_dma_mem(&ssbdp->pbuf); 403 ssbdp++; 404 } 405 406 kmem_free(rgep->sw_sbds, RGE_SEND_SLOTS * sizeof (sw_sbd_t)); 407 rgep->sw_sbds = NULL; 408 } 409 410 static void 411 rge_fini_recv_ring(rge_t *rgep) 412 { 413 sw_rbd_t *srbdp; 414 uint32_t slot; 415 416 srbdp = rgep->sw_rbds; 417 for (slot = 0; slot < RGE_RECV_SLOTS; ++srbdp, ++slot) { 418 if (srbdp->rx_buf) { 419 if (srbdp->rx_buf->mp != NULL) { 420 freemsg(srbdp->rx_buf->mp); 421 srbdp->rx_buf->mp = NULL; 422 } 423 rge_free_dma_mem(&srbdp->rx_buf->pbuf); 424 kmem_free(srbdp->rx_buf, sizeof (dma_buf_t)); 425 srbdp->rx_buf = NULL; 426 } 427 } 428 429 kmem_free(rgep->sw_rbds, RGE_RECV_SLOTS * sizeof (sw_rbd_t)); 430 rgep->sw_rbds = NULL; 431 } 432 433 static void 434 rge_fini_buf_ring(rge_t *rgep) 435 { 436 sw_rbd_t *srbdp; 437 uint32_t slot; 438 439 if (rgep->chip_flags & CHIP_FLAG_FORCE_BCOPY) 440 return; 441 442 ASSERT(rgep->rx_free == RGE_BUF_SLOTS); 443 444 srbdp = rgep->free_srbds; 445 for (slot = 0; slot < RGE_BUF_SLOTS; ++srbdp, ++slot) { 446 if (srbdp->rx_buf != NULL) { 447 if (srbdp->rx_buf->mp != NULL) { 448 freemsg(srbdp->rx_buf->mp); 449 srbdp->rx_buf->mp = NULL; 450 } 451 rge_free_dma_mem(&srbdp->rx_buf->pbuf); 452 kmem_free(srbdp->rx_buf, sizeof (dma_buf_t)); 453 srbdp->rx_buf = NULL; 454 } 455 } 456 457 kmem_free(rgep->free_srbds, RGE_BUF_SLOTS * sizeof (sw_rbd_t)); 458 rgep->free_srbds = NULL; 459 } 460 461 static void 462 rge_fini_rings(rge_t *rgep) 463 { 464 rge_fini_send_ring(rgep); 465 rge_fini_recv_ring(rgep); 466 rge_fini_buf_ring(rgep); 467 } 468 469 static int 470 rge_init_send_ring(rge_t *rgep) 471 { 472 uint32_t slot; 473 sw_sbd_t *ssbdp; 474 dma_area_t *pbuf; 475 dma_area_t desc; 476 int err; 477 478 /* 479 * Allocate the array of s/w Tx Buffer Descriptors 480 */ 481 ssbdp = kmem_zalloc(RGE_SEND_SLOTS*sizeof (*ssbdp), KM_SLEEP); 482 rgep->sw_sbds = ssbdp; 483 484 /* 485 * Init send ring 486 */ 487 rgep->tx_desc = rgep->dma_area_txdesc; 488 DMA_ZERO(rgep->tx_desc); 489 rgep->tx_ring = rgep->tx_desc.mem_va; 490 491 desc = rgep->tx_desc; 492 for (slot = 0; slot < RGE_SEND_SLOTS; slot++) { 493 rge_slice_chunk(&ssbdp->desc, &desc, 1, sizeof (rge_bd_t)); 494 495 /* 496 * Allocate memory & handle for Tx buffers 497 */ 498 pbuf = &ssbdp->pbuf; 499 err = rge_alloc_dma_mem(rgep, rgep->txbuf_size, 500 &dma_attr_buf, &rge_buf_accattr, 501 DDI_DMA_WRITE | DDI_DMA_STREAMING, pbuf); 502 if (err != DDI_SUCCESS) { 503 rge_error(rgep, 504 "rge_init_send_ring: alloc tx buffer failed"); 505 rge_fini_send_ring(rgep); 506 return (DDI_FAILURE); 507 } 508 ssbdp++; 509 } 510 ASSERT(desc.alength == 0); 511 512 DMA_SYNC(rgep->tx_desc, DDI_DMA_SYNC_FORDEV); 513 return (DDI_SUCCESS); 514 } 515 516 static int 517 rge_init_recv_ring(rge_t *rgep) 518 { 519 uint32_t slot; 520 sw_rbd_t *srbdp; 521 dma_buf_t *rx_buf; 522 dma_area_t *pbuf; 523 int err; 524 525 /* 526 * Allocate the array of s/w Rx Buffer Descriptors 527 */ 528 srbdp = kmem_zalloc(RGE_RECV_SLOTS*sizeof (*srbdp), KM_SLEEP); 529 rgep->sw_rbds = srbdp; 530 531 /* 532 * Init receive ring 533 */ 534 rgep->rx_next = 0; 535 rgep->rx_desc = rgep->dma_area_rxdesc; 536 DMA_ZERO(rgep->rx_desc); 537 rgep->rx_ring = rgep->rx_desc.mem_va; 538 539 for (slot = 0; slot < RGE_RECV_SLOTS; slot++) { 540 srbdp->rx_buf = rx_buf = 541 kmem_zalloc(sizeof (dma_buf_t), KM_SLEEP); 542 543 /* 544 * Allocate memory & handle for Rx buffers 545 */ 546 pbuf = &rx_buf->pbuf; 547 err = rge_alloc_dma_mem(rgep, rgep->rxbuf_size, 548 &dma_attr_buf, &rge_buf_accattr, 549 DDI_DMA_READ | DDI_DMA_STREAMING, pbuf); 550 if (err != DDI_SUCCESS) { 551 rge_fini_recv_ring(rgep); 552 rge_error(rgep, 553 "rge_init_recv_ring: alloc rx buffer failed"); 554 return (DDI_FAILURE); 555 } 556 557 pbuf->alength -= rgep->head_room; 558 pbuf->offset += rgep->head_room; 559 if (!(rgep->chip_flags & CHIP_FLAG_FORCE_BCOPY)) { 560 rx_buf->rx_recycle.free_func = rge_rx_recycle; 561 rx_buf->rx_recycle.free_arg = (caddr_t)rx_buf; 562 rx_buf->private = (caddr_t)rgep; 563 rx_buf->mp = desballoc(DMA_VPTR(rx_buf->pbuf), 564 rgep->rxbuf_size, 0, &rx_buf->rx_recycle); 565 if (rx_buf->mp == NULL) { 566 rge_fini_recv_ring(rgep); 567 rge_problem(rgep, 568 "rge_init_recv_ring: desballoc() failed"); 569 return (DDI_FAILURE); 570 } 571 } 572 srbdp++; 573 } 574 DMA_SYNC(rgep->rx_desc, DDI_DMA_SYNC_FORDEV); 575 return (DDI_SUCCESS); 576 } 577 578 static int 579 rge_init_buf_ring(rge_t *rgep) 580 { 581 uint32_t slot; 582 sw_rbd_t *free_srbdp; 583 dma_buf_t *rx_buf; 584 dma_area_t *pbuf; 585 int err; 586 587 if (rgep->chip_flags & CHIP_FLAG_FORCE_BCOPY) { 588 rgep->rx_bcopy = B_TRUE; 589 return (DDI_SUCCESS); 590 } 591 592 /* 593 * Allocate the array of s/w free Buffer Descriptors 594 */ 595 free_srbdp = kmem_zalloc(RGE_BUF_SLOTS*sizeof (*free_srbdp), KM_SLEEP); 596 rgep->free_srbds = free_srbdp; 597 598 /* 599 * Init free buffer ring 600 */ 601 rgep->rc_next = 0; 602 rgep->rf_next = 0; 603 rgep->rx_bcopy = B_FALSE; 604 rgep->rx_free = RGE_BUF_SLOTS; 605 for (slot = 0; slot < RGE_BUF_SLOTS; slot++) { 606 free_srbdp->rx_buf = rx_buf = 607 kmem_zalloc(sizeof (dma_buf_t), KM_SLEEP); 608 609 /* 610 * Allocate memory & handle for free Rx buffers 611 */ 612 pbuf = &rx_buf->pbuf; 613 err = rge_alloc_dma_mem(rgep, rgep->rxbuf_size, 614 &dma_attr_buf, &rge_buf_accattr, 615 DDI_DMA_READ | DDI_DMA_STREAMING, pbuf); 616 if (err != DDI_SUCCESS) { 617 rge_fini_buf_ring(rgep); 618 rge_error(rgep, 619 "rge_init_buf_ring: alloc rx free buffer failed"); 620 return (DDI_FAILURE); 621 } 622 pbuf->alength -= rgep->head_room; 623 pbuf->offset += rgep->head_room; 624 rx_buf->rx_recycle.free_func = rge_rx_recycle; 625 rx_buf->rx_recycle.free_arg = (caddr_t)rx_buf; 626 rx_buf->private = (caddr_t)rgep; 627 rx_buf->mp = desballoc(DMA_VPTR(rx_buf->pbuf), 628 rgep->rxbuf_size, 0, &rx_buf->rx_recycle); 629 if (rx_buf->mp == NULL) { 630 rge_fini_buf_ring(rgep); 631 rge_problem(rgep, 632 "rge_init_buf_ring: desballoc() failed"); 633 return (DDI_FAILURE); 634 } 635 free_srbdp++; 636 } 637 return (DDI_SUCCESS); 638 } 639 640 static int 641 rge_init_rings(rge_t *rgep) 642 { 643 int err; 644 645 err = rge_init_send_ring(rgep); 646 if (err != DDI_SUCCESS) 647 return (DDI_FAILURE); 648 649 err = rge_init_recv_ring(rgep); 650 if (err != DDI_SUCCESS) { 651 rge_fini_send_ring(rgep); 652 return (DDI_FAILURE); 653 } 654 655 err = rge_init_buf_ring(rgep); 656 if (err != DDI_SUCCESS) { 657 rge_fini_send_ring(rgep); 658 rge_fini_recv_ring(rgep); 659 return (DDI_FAILURE); 660 } 661 662 return (DDI_SUCCESS); 663 } 664 665 /* 666 * ========== Internal state management entry points ========== 667 */ 668 669 #undef RGE_DBG 670 #define RGE_DBG RGE_DBG_NEMO /* debug flag for this code */ 671 672 /* 673 * These routines provide all the functionality required by the 674 * corresponding MAC layer entry points, but don't update the 675 * MAC state so they can be called internally without disturbing 676 * our record of what NEMO thinks we should be doing ... 677 */ 678 679 /* 680 * rge_reset() -- reset h/w & rings to initial state 681 */ 682 static void 683 rge_reset(rge_t *rgep) 684 { 685 ASSERT(mutex_owned(rgep->genlock)); 686 687 /* 688 * Grab all the other mutexes in the world (this should 689 * ensure no other threads are manipulating driver state) 690 */ 691 mutex_enter(rgep->rx_lock); 692 mutex_enter(rgep->rc_lock); 693 rw_enter(rgep->errlock, RW_WRITER); 694 695 (void) rge_chip_reset(rgep); 696 rge_reinit_rings(rgep); 697 rge_chip_init(rgep); 698 699 /* 700 * Free the world ... 701 */ 702 rw_exit(rgep->errlock); 703 mutex_exit(rgep->rc_lock); 704 mutex_exit(rgep->rx_lock); 705 706 rgep->stats.rpackets = 0; 707 rgep->stats.rbytes = 0; 708 rgep->stats.opackets = 0; 709 rgep->stats.obytes = 0; 710 rgep->stats.tx_pre_ismax = B_FALSE; 711 rgep->stats.tx_cur_ismax = B_FALSE; 712 713 RGE_DEBUG(("rge_reset($%p) done", (void *)rgep)); 714 } 715 716 /* 717 * rge_stop() -- stop processing, don't reset h/w or rings 718 */ 719 static void 720 rge_stop(rge_t *rgep) 721 { 722 ASSERT(mutex_owned(rgep->genlock)); 723 724 rge_chip_stop(rgep, B_FALSE); 725 726 RGE_DEBUG(("rge_stop($%p) done", (void *)rgep)); 727 } 728 729 /* 730 * rge_start() -- start transmitting/receiving 731 */ 732 static void 733 rge_start(rge_t *rgep) 734 { 735 ASSERT(mutex_owned(rgep->genlock)); 736 737 /* 738 * Start chip processing, including enabling interrupts 739 */ 740 rge_chip_start(rgep); 741 rgep->watchdog = 0; 742 } 743 744 /* 745 * rge_restart - restart transmitting/receiving after error or suspend 746 */ 747 void 748 rge_restart(rge_t *rgep) 749 { 750 uint32_t i; 751 752 ASSERT(mutex_owned(rgep->genlock)); 753 /* 754 * Wait for posted buffer to be freed... 755 */ 756 if (!rgep->rx_bcopy) { 757 for (i = 0; i < RXBUFF_FREE_LOOP; i++) { 758 if (rgep->rx_free == RGE_BUF_SLOTS) 759 break; 760 drv_usecwait(1000); 761 RGE_DEBUG(("rge_restart: waiting for rx buf free...")); 762 } 763 } 764 rge_reset(rgep); 765 rgep->stats.chip_reset++; 766 if (rgep->rge_mac_state == RGE_MAC_STARTED) { 767 rge_start(rgep); 768 rgep->resched_needed = B_TRUE; 769 (void) ddi_intr_trigger_softint(rgep->resched_hdl, NULL); 770 } 771 } 772 773 774 /* 775 * ========== Nemo-required management entry points ========== 776 */ 777 778 #undef RGE_DBG 779 #define RGE_DBG RGE_DBG_NEMO /* debug flag for this code */ 780 781 /* 782 * rge_m_stop() -- stop transmitting/receiving 783 */ 784 static void 785 rge_m_stop(void *arg) 786 { 787 rge_t *rgep = arg; /* private device info */ 788 uint32_t i; 789 790 /* 791 * Just stop processing, then record new MAC state 792 */ 793 mutex_enter(rgep->genlock); 794 if (rgep->suspended) { 795 ASSERT(rgep->rge_mac_state == RGE_MAC_STOPPED); 796 mutex_exit(rgep->genlock); 797 return; 798 } 799 rge_stop(rgep); 800 /* 801 * Wait for posted buffer to be freed... 802 */ 803 if (!rgep->rx_bcopy) { 804 for (i = 0; i < RXBUFF_FREE_LOOP; i++) { 805 if (rgep->rx_free == RGE_BUF_SLOTS) 806 break; 807 drv_usecwait(1000); 808 RGE_DEBUG(("rge_m_stop: waiting for rx buf free...")); 809 } 810 } 811 rgep->rge_mac_state = RGE_MAC_STOPPED; 812 RGE_DEBUG(("rge_m_stop($%p) done", arg)); 813 mutex_exit(rgep->genlock); 814 } 815 816 /* 817 * rge_m_start() -- start transmitting/receiving 818 */ 819 static int 820 rge_m_start(void *arg) 821 { 822 rge_t *rgep = arg; /* private device info */ 823 824 mutex_enter(rgep->genlock); 825 if (rgep->suspended) { 826 mutex_exit(rgep->genlock); 827 return (DDI_FAILURE); 828 } 829 /* 830 * Clear hw/sw statistics 831 */ 832 DMA_ZERO(rgep->dma_area_stats); 833 bzero(&rgep->stats, sizeof (rge_stats_t)); 834 835 /* 836 * Start processing and record new MAC state 837 */ 838 rge_reset(rgep); 839 rge_start(rgep); 840 rgep->rge_mac_state = RGE_MAC_STARTED; 841 RGE_DEBUG(("rge_m_start($%p) done", arg)); 842 843 mutex_exit(rgep->genlock); 844 845 return (0); 846 } 847 848 /* 849 * rge_m_unicst_set() -- set the physical network address 850 */ 851 static int 852 rge_m_unicst(void *arg, const uint8_t *macaddr) 853 { 854 rge_t *rgep = arg; /* private device info */ 855 856 /* 857 * Remember the new current address in the driver state 858 * Sync the chip's idea of the address too ... 859 */ 860 mutex_enter(rgep->genlock); 861 bcopy(macaddr, rgep->netaddr, ETHERADDRL); 862 863 if (rgep->suspended) { 864 mutex_exit(rgep->genlock); 865 return (DDI_SUCCESS); 866 } 867 868 rge_chip_sync(rgep, RGE_SET_MAC); 869 mutex_exit(rgep->genlock); 870 871 return (0); 872 } 873 874 /* 875 * Compute the index of the required bit in the multicast hash map. 876 * This must mirror the way the hardware actually does it! 877 */ 878 static uint32_t 879 rge_hash_index(const uint8_t *mca) 880 { 881 uint32_t crc = (uint32_t)RGE_HASH_CRC; 882 uint32_t const POLY = RGE_HASH_POLY; 883 uint32_t msb; 884 int bytes; 885 uchar_t currentbyte; 886 uint32_t index; 887 int bit; 888 889 for (bytes = 0; bytes < ETHERADDRL; bytes++) { 890 currentbyte = mca[bytes]; 891 for (bit = 0; bit < 8; bit++) { 892 msb = crc >> 31; 893 crc <<= 1; 894 if (msb ^ (currentbyte & 1)) 895 crc ^= POLY; 896 currentbyte >>= 1; 897 } 898 } 899 index = crc >> 26; 900 /* the index value is between 0 and 63(0x3f) */ 901 902 return (index); 903 } 904 905 /* 906 * rge_m_multicst_add() -- enable/disable a multicast address 907 */ 908 static int 909 rge_m_multicst(void *arg, boolean_t add, const uint8_t *mca) 910 { 911 rge_t *rgep = arg; /* private device info */ 912 struct ether_addr *addr; 913 uint32_t index; 914 uint32_t reg; 915 uint8_t *hashp; 916 917 mutex_enter(rgep->genlock); 918 hashp = rgep->mcast_hash; 919 addr = (struct ether_addr *)mca; 920 /* 921 * Calculate the Multicast address hash index value 922 * Normally, the position of MAR0-MAR7 is 923 * MAR0: offset 0x08, ..., MAR7: offset 0x0F. 924 * 925 * For pcie chipset, the position of MAR0-MAR7 is 926 * different from others: 927 * MAR0: offset 0x0F, ..., MAR7: offset 0x08. 928 */ 929 index = rge_hash_index(addr->ether_addr_octet); 930 if (rgep->chipid.is_pcie) 931 reg = (~(index / RGE_MCAST_NUM)) & 0x7; 932 else 933 reg = index / RGE_MCAST_NUM; 934 935 if (add) { 936 if (rgep->mcast_refs[index]++) { 937 mutex_exit(rgep->genlock); 938 return (0); 939 } 940 hashp[reg] |= 1 << (index % RGE_MCAST_NUM); 941 } else { 942 if (--rgep->mcast_refs[index]) { 943 mutex_exit(rgep->genlock); 944 return (0); 945 } 946 hashp[reg] &= ~ (1 << (index % RGE_MCAST_NUM)); 947 } 948 949 if (rgep->suspended) { 950 mutex_exit(rgep->genlock); 951 return (DDI_SUCCESS); 952 } 953 954 /* 955 * Set multicast register 956 */ 957 rge_chip_sync(rgep, RGE_SET_MUL); 958 959 mutex_exit(rgep->genlock); 960 return (0); 961 } 962 963 /* 964 * rge_m_promisc() -- set or reset promiscuous mode on the board 965 * 966 * Program the hardware to enable/disable promiscuous and/or 967 * receive-all-multicast modes. 968 */ 969 static int 970 rge_m_promisc(void *arg, boolean_t on) 971 { 972 rge_t *rgep = arg; 973 974 /* 975 * Store MAC layer specified mode and pass to chip layer to update h/w 976 */ 977 mutex_enter(rgep->genlock); 978 979 if (rgep->promisc == on) { 980 mutex_exit(rgep->genlock); 981 return (0); 982 } 983 rgep->promisc = on; 984 985 if (rgep->suspended) { 986 mutex_exit(rgep->genlock); 987 return (DDI_SUCCESS); 988 } 989 990 rge_chip_sync(rgep, RGE_SET_PROMISC); 991 RGE_DEBUG(("rge_m_promisc_set($%p) done", arg)); 992 mutex_exit(rgep->genlock); 993 return (0); 994 } 995 996 /* 997 * Loopback ioctl code 998 */ 999 1000 static lb_property_t loopmodes[] = { 1001 { normal, "normal", RGE_LOOP_NONE }, 1002 { internal, "PHY", RGE_LOOP_INTERNAL_PHY }, 1003 { internal, "MAC", RGE_LOOP_INTERNAL_MAC } 1004 }; 1005 1006 static enum ioc_reply 1007 rge_set_loop_mode(rge_t *rgep, uint32_t mode) 1008 { 1009 /* 1010 * If the mode isn't being changed, there's nothing to do ... 1011 */ 1012 if (mode == rgep->param_loop_mode) 1013 return (IOC_ACK); 1014 1015 /* 1016 * Validate the requested mode and prepare a suitable message 1017 * to explain the link down/up cycle that the change will 1018 * probably induce ... 1019 */ 1020 switch (mode) { 1021 default: 1022 return (IOC_INVAL); 1023 1024 case RGE_LOOP_NONE: 1025 case RGE_LOOP_INTERNAL_PHY: 1026 case RGE_LOOP_INTERNAL_MAC: 1027 break; 1028 } 1029 1030 /* 1031 * All OK; tell the caller to reprogram 1032 * the PHY and/or MAC for the new mode ... 1033 */ 1034 rgep->param_loop_mode = mode; 1035 return (IOC_RESTART_ACK); 1036 } 1037 1038 static enum ioc_reply 1039 rge_loop_ioctl(rge_t *rgep, queue_t *wq, mblk_t *mp, struct iocblk *iocp) 1040 { 1041 lb_info_sz_t *lbsp; 1042 lb_property_t *lbpp; 1043 uint32_t *lbmp; 1044 int cmd; 1045 1046 _NOTE(ARGUNUSED(wq)) 1047 1048 /* 1049 * Validate format of ioctl 1050 */ 1051 if (mp->b_cont == NULL) 1052 return (IOC_INVAL); 1053 1054 cmd = iocp->ioc_cmd; 1055 switch (cmd) { 1056 default: 1057 /* NOTREACHED */ 1058 rge_error(rgep, "rge_loop_ioctl: invalid cmd 0x%x", cmd); 1059 return (IOC_INVAL); 1060 1061 case LB_GET_INFO_SIZE: 1062 if (iocp->ioc_count != sizeof (lb_info_sz_t)) 1063 return (IOC_INVAL); 1064 lbsp = (lb_info_sz_t *)mp->b_cont->b_rptr; 1065 *lbsp = sizeof (loopmodes); 1066 return (IOC_REPLY); 1067 1068 case LB_GET_INFO: 1069 if (iocp->ioc_count != sizeof (loopmodes)) 1070 return (IOC_INVAL); 1071 lbpp = (lb_property_t *)mp->b_cont->b_rptr; 1072 bcopy(loopmodes, lbpp, sizeof (loopmodes)); 1073 return (IOC_REPLY); 1074 1075 case LB_GET_MODE: 1076 if (iocp->ioc_count != sizeof (uint32_t)) 1077 return (IOC_INVAL); 1078 lbmp = (uint32_t *)mp->b_cont->b_rptr; 1079 *lbmp = rgep->param_loop_mode; 1080 return (IOC_REPLY); 1081 1082 case LB_SET_MODE: 1083 if (iocp->ioc_count != sizeof (uint32_t)) 1084 return (IOC_INVAL); 1085 lbmp = (uint32_t *)mp->b_cont->b_rptr; 1086 return (rge_set_loop_mode(rgep, *lbmp)); 1087 } 1088 } 1089 1090 /* 1091 * Specific rge IOCTLs, the MAC layer handles the generic ones. 1092 */ 1093 static void 1094 rge_m_ioctl(void *arg, queue_t *wq, mblk_t *mp) 1095 { 1096 rge_t *rgep = arg; 1097 struct iocblk *iocp; 1098 enum ioc_reply status; 1099 boolean_t need_privilege; 1100 int err; 1101 int cmd; 1102 1103 /* 1104 * If suspended, we might actually be able to do some of 1105 * these ioctls, but it is harder to make sure they occur 1106 * without actually putting the hardware in an undesireable 1107 * state. So just NAK it. 1108 */ 1109 mutex_enter(rgep->genlock); 1110 if (rgep->suspended) { 1111 miocnak(wq, mp, 0, EINVAL); 1112 mutex_exit(rgep->genlock); 1113 return; 1114 } 1115 mutex_exit(rgep->genlock); 1116 1117 /* 1118 * Validate the command before bothering with the mutex ... 1119 */ 1120 iocp = (struct iocblk *)mp->b_rptr; 1121 iocp->ioc_error = 0; 1122 need_privilege = B_TRUE; 1123 cmd = iocp->ioc_cmd; 1124 switch (cmd) { 1125 default: 1126 miocnak(wq, mp, 0, EINVAL); 1127 return; 1128 1129 case RGE_MII_READ: 1130 case RGE_MII_WRITE: 1131 case RGE_DIAG: 1132 case RGE_PEEK: 1133 case RGE_POKE: 1134 case RGE_PHY_RESET: 1135 case RGE_SOFT_RESET: 1136 case RGE_HARD_RESET: 1137 break; 1138 1139 case LB_GET_INFO_SIZE: 1140 case LB_GET_INFO: 1141 case LB_GET_MODE: 1142 need_privilege = B_FALSE; 1143 /* FALLTHRU */ 1144 case LB_SET_MODE: 1145 break; 1146 1147 case ND_GET: 1148 need_privilege = B_FALSE; 1149 /* FALLTHRU */ 1150 case ND_SET: 1151 break; 1152 } 1153 1154 if (need_privilege) { 1155 /* 1156 * Check for specific net_config privilege 1157 */ 1158 err = secpolicy_net_config(iocp->ioc_cr, B_FALSE); 1159 if (err != 0) { 1160 miocnak(wq, mp, 0, err); 1161 return; 1162 } 1163 } 1164 1165 mutex_enter(rgep->genlock); 1166 1167 switch (cmd) { 1168 default: 1169 _NOTE(NOTREACHED) 1170 status = IOC_INVAL; 1171 break; 1172 1173 case RGE_MII_READ: 1174 case RGE_MII_WRITE: 1175 case RGE_DIAG: 1176 case RGE_PEEK: 1177 case RGE_POKE: 1178 case RGE_PHY_RESET: 1179 case RGE_SOFT_RESET: 1180 case RGE_HARD_RESET: 1181 status = rge_chip_ioctl(rgep, wq, mp, iocp); 1182 break; 1183 1184 case LB_GET_INFO_SIZE: 1185 case LB_GET_INFO: 1186 case LB_GET_MODE: 1187 case LB_SET_MODE: 1188 status = rge_loop_ioctl(rgep, wq, mp, iocp); 1189 break; 1190 1191 case ND_GET: 1192 case ND_SET: 1193 status = rge_nd_ioctl(rgep, wq, mp, iocp); 1194 break; 1195 } 1196 1197 /* 1198 * Do we need to reprogram the PHY and/or the MAC? 1199 * Do it now, while we still have the mutex. 1200 * 1201 * Note: update the PHY first, 'cos it controls the 1202 * speed/duplex parameters that the MAC code uses. 1203 */ 1204 switch (status) { 1205 case IOC_RESTART_REPLY: 1206 case IOC_RESTART_ACK: 1207 rge_phy_update(rgep); 1208 break; 1209 } 1210 1211 mutex_exit(rgep->genlock); 1212 1213 /* 1214 * Finally, decide how to reply 1215 */ 1216 switch (status) { 1217 default: 1218 case IOC_INVAL: 1219 /* 1220 * Error, reply with a NAK and EINVAL or the specified error 1221 */ 1222 miocnak(wq, mp, 0, iocp->ioc_error == 0 ? 1223 EINVAL : iocp->ioc_error); 1224 break; 1225 1226 case IOC_DONE: 1227 /* 1228 * OK, reply already sent 1229 */ 1230 break; 1231 1232 case IOC_RESTART_ACK: 1233 case IOC_ACK: 1234 /* 1235 * OK, reply with an ACK 1236 */ 1237 miocack(wq, mp, 0, 0); 1238 break; 1239 1240 case IOC_RESTART_REPLY: 1241 case IOC_REPLY: 1242 /* 1243 * OK, send prepared reply as ACK or NAK 1244 */ 1245 mp->b_datap->db_type = iocp->ioc_error == 0 ? 1246 M_IOCACK : M_IOCNAK; 1247 qreply(wq, mp); 1248 break; 1249 } 1250 } 1251 1252 static void 1253 rge_m_resources(void *arg) 1254 { 1255 rge_t *rgep = arg; 1256 mac_rx_fifo_t mrf; 1257 1258 mutex_enter(rgep->genlock); 1259 1260 /* 1261 * Register Rx rings as resources and save mac 1262 * resource id for future reference 1263 */ 1264 mrf.mrf_type = MAC_RX_FIFO; 1265 mrf.mrf_blank = rge_chip_blank; 1266 mrf.mrf_arg = (void *)rgep; 1267 mrf.mrf_normal_blank_time = RGE_RX_INT_TIME; 1268 mrf.mrf_normal_pkt_count = RGE_RX_INT_PKTS; 1269 rgep->handle = mac_resource_add(rgep->mh, (mac_resource_t *)&mrf); 1270 1271 mutex_exit(rgep->genlock); 1272 } 1273 1274 /* ARGSUSED */ 1275 static boolean_t 1276 rge_m_getcapab(void *arg, mac_capab_t cap, void *cap_data) 1277 { 1278 rge_t *rgep = arg; 1279 1280 switch (cap) { 1281 case MAC_CAPAB_HCKSUM: { 1282 uint32_t *hcksum_txflags = cap_data; 1283 switch (rgep->chipid.mac_ver) { 1284 case MAC_VER_8169: 1285 case MAC_VER_8169S_D: 1286 case MAC_VER_8169S_E: 1287 case MAC_VER_8169SB: 1288 case MAC_VER_8169SC: 1289 case MAC_VER_8168: 1290 case MAC_VER_8168B_B: 1291 case MAC_VER_8168B_C: 1292 case MAC_VER_8101E: 1293 *hcksum_txflags = HCKSUM_INET_FULL_V4 | 1294 HCKSUM_IPHDRCKSUM; 1295 break; 1296 case MAC_VER_8168B_D: 1297 case MAC_VER_8101E_B: 1298 case MAC_VER_8101E_C: 1299 default: 1300 *hcksum_txflags = 0; 1301 break; 1302 } 1303 break; 1304 } 1305 case MAC_CAPAB_POLL: 1306 /* 1307 * There's nothing for us to fill in, simply returning 1308 * B_TRUE stating that we support polling is sufficient. 1309 */ 1310 break; 1311 default: 1312 return (B_FALSE); 1313 } 1314 return (B_TRUE); 1315 } 1316 1317 /* 1318 * ============ Init MSI/Fixed Interrupt routines ============== 1319 */ 1320 1321 /* 1322 * rge_add_intrs: 1323 * 1324 * Register FIXED or MSI interrupts. 1325 */ 1326 static int 1327 rge_add_intrs(rge_t *rgep, int intr_type) 1328 { 1329 dev_info_t *dip = rgep->devinfo; 1330 int avail; 1331 int actual; 1332 int intr_size; 1333 int count; 1334 int i, j; 1335 int ret; 1336 1337 /* Get number of interrupts */ 1338 ret = ddi_intr_get_nintrs(dip, intr_type, &count); 1339 if ((ret != DDI_SUCCESS) || (count == 0)) { 1340 rge_error(rgep, "ddi_intr_get_nintrs() failure, ret: %d, " 1341 "count: %d", ret, count); 1342 return (DDI_FAILURE); 1343 } 1344 1345 /* Get number of available interrupts */ 1346 ret = ddi_intr_get_navail(dip, intr_type, &avail); 1347 if ((ret != DDI_SUCCESS) || (avail == 0)) { 1348 rge_error(rgep, "ddi_intr_get_navail() failure, " 1349 "ret: %d, avail: %d\n", ret, avail); 1350 return (DDI_FAILURE); 1351 } 1352 1353 /* Allocate an array of interrupt handles */ 1354 intr_size = count * sizeof (ddi_intr_handle_t); 1355 rgep->htable = kmem_alloc(intr_size, KM_SLEEP); 1356 rgep->intr_rqst = count; 1357 1358 /* Call ddi_intr_alloc() */ 1359 ret = ddi_intr_alloc(dip, rgep->htable, intr_type, 0, 1360 count, &actual, DDI_INTR_ALLOC_NORMAL); 1361 if (ret != DDI_SUCCESS || actual == 0) { 1362 rge_error(rgep, "ddi_intr_alloc() failed %d\n", ret); 1363 kmem_free(rgep->htable, intr_size); 1364 return (DDI_FAILURE); 1365 } 1366 if (actual < count) { 1367 rge_log(rgep, "ddi_intr_alloc() Requested: %d, Received: %d\n", 1368 count, actual); 1369 } 1370 rgep->intr_cnt = actual; 1371 1372 /* 1373 * Get priority for first msi, assume remaining are all the same 1374 */ 1375 if ((ret = ddi_intr_get_pri(rgep->htable[0], &rgep->intr_pri)) != 1376 DDI_SUCCESS) { 1377 rge_error(rgep, "ddi_intr_get_pri() failed %d\n", ret); 1378 /* Free already allocated intr */ 1379 for (i = 0; i < actual; i++) { 1380 (void) ddi_intr_free(rgep->htable[i]); 1381 } 1382 kmem_free(rgep->htable, intr_size); 1383 return (DDI_FAILURE); 1384 } 1385 1386 /* Test for high level mutex */ 1387 if (rgep->intr_pri >= ddi_intr_get_hilevel_pri()) { 1388 rge_error(rgep, "rge_add_intrs:" 1389 "Hi level interrupt not supported"); 1390 for (i = 0; i < actual; i++) 1391 (void) ddi_intr_free(rgep->htable[i]); 1392 kmem_free(rgep->htable, intr_size); 1393 return (DDI_FAILURE); 1394 } 1395 1396 /* Call ddi_intr_add_handler() */ 1397 for (i = 0; i < actual; i++) { 1398 if ((ret = ddi_intr_add_handler(rgep->htable[i], rge_intr, 1399 (caddr_t)rgep, (caddr_t)(uintptr_t)i)) != DDI_SUCCESS) { 1400 rge_error(rgep, "ddi_intr_add_handler() " 1401 "failed %d\n", ret); 1402 /* Remove already added intr */ 1403 for (j = 0; j < i; j++) 1404 (void) ddi_intr_remove_handler(rgep->htable[j]); 1405 /* Free already allocated intr */ 1406 for (i = 0; i < actual; i++) { 1407 (void) ddi_intr_free(rgep->htable[i]); 1408 } 1409 kmem_free(rgep->htable, intr_size); 1410 return (DDI_FAILURE); 1411 } 1412 } 1413 1414 if ((ret = ddi_intr_get_cap(rgep->htable[0], &rgep->intr_cap)) 1415 != DDI_SUCCESS) { 1416 rge_error(rgep, "ddi_intr_get_cap() failed %d\n", ret); 1417 for (i = 0; i < actual; i++) { 1418 (void) ddi_intr_remove_handler(rgep->htable[i]); 1419 (void) ddi_intr_free(rgep->htable[i]); 1420 } 1421 kmem_free(rgep->htable, intr_size); 1422 return (DDI_FAILURE); 1423 } 1424 1425 return (DDI_SUCCESS); 1426 } 1427 1428 /* 1429 * rge_rem_intrs: 1430 * 1431 * Unregister FIXED or MSI interrupts 1432 */ 1433 static void 1434 rge_rem_intrs(rge_t *rgep) 1435 { 1436 int i; 1437 1438 /* Disable all interrupts */ 1439 if (rgep->intr_cap & DDI_INTR_FLAG_BLOCK) { 1440 /* Call ddi_intr_block_disable() */ 1441 (void) ddi_intr_block_disable(rgep->htable, rgep->intr_cnt); 1442 } else { 1443 for (i = 0; i < rgep->intr_cnt; i++) { 1444 (void) ddi_intr_disable(rgep->htable[i]); 1445 } 1446 } 1447 1448 /* Call ddi_intr_remove_handler() */ 1449 for (i = 0; i < rgep->intr_cnt; i++) { 1450 (void) ddi_intr_remove_handler(rgep->htable[i]); 1451 (void) ddi_intr_free(rgep->htable[i]); 1452 } 1453 1454 kmem_free(rgep->htable, rgep->intr_rqst * sizeof (ddi_intr_handle_t)); 1455 } 1456 1457 /* 1458 * ========== Per-instance setup/teardown code ========== 1459 */ 1460 1461 #undef RGE_DBG 1462 #define RGE_DBG RGE_DBG_INIT /* debug flag for this code */ 1463 1464 static void 1465 rge_unattach(rge_t *rgep) 1466 { 1467 /* 1468 * Flag that no more activity may be initiated 1469 */ 1470 rgep->progress &= ~PROGRESS_READY; 1471 rgep->rge_mac_state = RGE_MAC_UNATTACH; 1472 1473 /* 1474 * Quiesce the PHY and MAC (leave it reset but still powered). 1475 * Clean up and free all RGE data structures 1476 */ 1477 if (rgep->periodic_id != NULL) { 1478 ddi_periodic_delete(rgep->periodic_id); 1479 rgep->periodic_id = NULL; 1480 } 1481 1482 if (rgep->progress & PROGRESS_KSTATS) 1483 rge_fini_kstats(rgep); 1484 1485 if (rgep->progress & PROGRESS_PHY) 1486 (void) rge_phy_reset(rgep); 1487 1488 if (rgep->progress & PROGRESS_INIT) { 1489 mutex_enter(rgep->genlock); 1490 (void) rge_chip_reset(rgep); 1491 mutex_exit(rgep->genlock); 1492 rge_fini_rings(rgep); 1493 } 1494 1495 if (rgep->progress & PROGRESS_INTR) { 1496 rge_rem_intrs(rgep); 1497 mutex_destroy(rgep->rc_lock); 1498 mutex_destroy(rgep->rx_lock); 1499 mutex_destroy(rgep->tc_lock); 1500 mutex_destroy(rgep->tx_lock); 1501 rw_destroy(rgep->errlock); 1502 mutex_destroy(rgep->genlock); 1503 } 1504 1505 if (rgep->progress & PROGRESS_FACTOTUM) 1506 (void) ddi_intr_remove_softint(rgep->factotum_hdl); 1507 1508 if (rgep->progress & PROGRESS_RESCHED) 1509 (void) ddi_intr_remove_softint(rgep->resched_hdl); 1510 1511 if (rgep->progress & PROGRESS_NDD) 1512 rge_nd_cleanup(rgep); 1513 1514 rge_free_bufs(rgep); 1515 1516 if (rgep->progress & PROGRESS_REGS) 1517 ddi_regs_map_free(&rgep->io_handle); 1518 1519 if (rgep->progress & PROGRESS_CFG) 1520 pci_config_teardown(&rgep->cfg_handle); 1521 1522 ddi_remove_minor_node(rgep->devinfo, NULL); 1523 kmem_free(rgep, sizeof (*rgep)); 1524 } 1525 1526 static int 1527 rge_resume(dev_info_t *devinfo) 1528 { 1529 rge_t *rgep; /* Our private data */ 1530 chip_id_t *cidp; 1531 chip_id_t chipid; 1532 1533 rgep = ddi_get_driver_private(devinfo); 1534 1535 /* 1536 * If there are state inconsistancies, this is bad. Returning 1537 * DDI_FAILURE here will eventually cause the machine to panic, 1538 * so it is best done here so that there is a possibility of 1539 * debugging the problem. 1540 */ 1541 if (rgep == NULL) 1542 cmn_err(CE_PANIC, 1543 "rge: ngep returned from ddi_get_driver_private was NULL"); 1544 1545 /* 1546 * Refuse to resume if the data structures aren't consistent 1547 */ 1548 if (rgep->devinfo != devinfo) 1549 cmn_err(CE_PANIC, 1550 "rge: passed devinfo not the same as saved devinfo"); 1551 1552 /* 1553 * Read chip ID & set up config space command register(s) 1554 * Refuse to resume if the chip has changed its identity! 1555 */ 1556 cidp = &rgep->chipid; 1557 rge_chip_cfg_init(rgep, &chipid); 1558 if (chipid.vendor != cidp->vendor) 1559 return (DDI_FAILURE); 1560 if (chipid.device != cidp->device) 1561 return (DDI_FAILURE); 1562 if (chipid.revision != cidp->revision) 1563 return (DDI_FAILURE); 1564 1565 mutex_enter(rgep->genlock); 1566 1567 /* 1568 * Only in one case, this conditional branch can be executed: the port 1569 * hasn't been plumbed. 1570 */ 1571 if (rgep->suspended == B_FALSE) { 1572 mutex_exit(rgep->genlock); 1573 return (DDI_SUCCESS); 1574 } 1575 rgep->rge_mac_state = RGE_MAC_STARTED; 1576 /* 1577 * All OK, reinitialise h/w & kick off NEMO scheduling 1578 */ 1579 rge_restart(rgep); 1580 rgep->suspended = B_FALSE; 1581 1582 mutex_exit(rgep->genlock); 1583 1584 return (DDI_SUCCESS); 1585 } 1586 1587 1588 /* 1589 * attach(9E) -- Attach a device to the system 1590 * 1591 * Called once for each board successfully probed. 1592 */ 1593 static int 1594 rge_attach(dev_info_t *devinfo, ddi_attach_cmd_t cmd) 1595 { 1596 rge_t *rgep; /* Our private data */ 1597 mac_register_t *macp; 1598 chip_id_t *cidp; 1599 int intr_types; 1600 caddr_t regs; 1601 int instance; 1602 int i; 1603 int err; 1604 1605 /* 1606 * we don't support high level interrupts in the driver 1607 */ 1608 if (ddi_intr_hilevel(devinfo, 0) != 0) { 1609 cmn_err(CE_WARN, 1610 "rge_attach -- unsupported high level interrupt"); 1611 return (DDI_FAILURE); 1612 } 1613 1614 instance = ddi_get_instance(devinfo); 1615 RGE_GTRACE(("rge_attach($%p, %d) instance %d", 1616 (void *)devinfo, cmd, instance)); 1617 RGE_BRKPT(NULL, "rge_attach"); 1618 1619 switch (cmd) { 1620 default: 1621 return (DDI_FAILURE); 1622 1623 case DDI_RESUME: 1624 return (rge_resume(devinfo)); 1625 1626 case DDI_ATTACH: 1627 break; 1628 } 1629 1630 rgep = kmem_zalloc(sizeof (*rgep), KM_SLEEP); 1631 ddi_set_driver_private(devinfo, rgep); 1632 rgep->devinfo = devinfo; 1633 1634 /* 1635 * Initialize more fields in RGE private data 1636 */ 1637 rgep->rge_mac_state = RGE_MAC_ATTACH; 1638 rgep->debug = ddi_prop_get_int(DDI_DEV_T_ANY, devinfo, 1639 DDI_PROP_DONTPASS, debug_propname, rge_debug); 1640 rgep->default_mtu = ddi_prop_get_int(DDI_DEV_T_ANY, devinfo, 1641 DDI_PROP_DONTPASS, mtu_propname, ETHERMTU); 1642 rgep->msi_enable = ddi_prop_get_int(DDI_DEV_T_ANY, devinfo, 1643 DDI_PROP_DONTPASS, msi_propname, B_TRUE); 1644 (void) snprintf(rgep->ifname, sizeof (rgep->ifname), "%s%d", 1645 RGE_DRIVER_NAME, instance); 1646 1647 /* 1648 * Map config space registers 1649 * Read chip ID & set up config space command register(s) 1650 * 1651 * Note: this leaves the chip accessible by Memory Space 1652 * accesses, but with interrupts and Bus Mastering off. 1653 * This should ensure that nothing untoward will happen 1654 * if it has been left active by the (net-)bootloader. 1655 * We'll re-enable Bus Mastering once we've reset the chip, 1656 * and allow interrupts only when everything else is set up. 1657 */ 1658 err = pci_config_setup(devinfo, &rgep->cfg_handle); 1659 if (err != DDI_SUCCESS) { 1660 rge_problem(rgep, "pci_config_setup() failed"); 1661 goto attach_fail; 1662 } 1663 rgep->progress |= PROGRESS_CFG; 1664 cidp = &rgep->chipid; 1665 bzero(cidp, sizeof (*cidp)); 1666 rge_chip_cfg_init(rgep, cidp); 1667 1668 /* 1669 * Map operating registers 1670 */ 1671 err = ddi_regs_map_setup(devinfo, 1, ®s, 1672 0, 0, &rge_reg_accattr, &rgep->io_handle); 1673 if (err != DDI_SUCCESS) { 1674 rge_problem(rgep, "ddi_regs_map_setup() failed"); 1675 goto attach_fail; 1676 } 1677 rgep->io_regs = regs; 1678 rgep->progress |= PROGRESS_REGS; 1679 1680 /* 1681 * Characterise the device, so we know its requirements. 1682 * Then allocate the appropriate TX and RX descriptors & buffers. 1683 */ 1684 rge_chip_ident(rgep); 1685 err = rge_alloc_bufs(rgep); 1686 if (err != DDI_SUCCESS) { 1687 rge_problem(rgep, "DMA buffer allocation failed"); 1688 goto attach_fail; 1689 } 1690 1691 /* 1692 * Register NDD-tweakable parameters 1693 */ 1694 if (rge_nd_init(rgep)) { 1695 rge_problem(rgep, "rge_nd_init() failed"); 1696 goto attach_fail; 1697 } 1698 rgep->progress |= PROGRESS_NDD; 1699 1700 /* 1701 * Add the softint handlers: 1702 * 1703 * Both of these handlers are used to avoid restrictions on the 1704 * context and/or mutexes required for some operations. In 1705 * particular, the hardware interrupt handler and its subfunctions 1706 * can detect a number of conditions that we don't want to handle 1707 * in that context or with that set of mutexes held. So, these 1708 * softints are triggered instead: 1709 * 1710 * the <resched> softint is triggered if if we have previously 1711 * had to refuse to send a packet because of resource shortage 1712 * (we've run out of transmit buffers), but the send completion 1713 * interrupt handler has now detected that more buffers have 1714 * become available. 1715 * 1716 * the <factotum> is triggered if the h/w interrupt handler 1717 * sees the <link state changed> or <error> bits in the status 1718 * block. It's also triggered periodically to poll the link 1719 * state, just in case we aren't getting link status change 1720 * interrupts ... 1721 */ 1722 err = ddi_intr_add_softint(devinfo, &rgep->resched_hdl, 1723 DDI_INTR_SOFTPRI_MIN, rge_reschedule, (caddr_t)rgep); 1724 if (err != DDI_SUCCESS) { 1725 rge_problem(rgep, "ddi_intr_add_softint() failed"); 1726 goto attach_fail; 1727 } 1728 rgep->progress |= PROGRESS_RESCHED; 1729 err = ddi_intr_add_softint(devinfo, &rgep->factotum_hdl, 1730 DDI_INTR_SOFTPRI_MIN, rge_chip_factotum, (caddr_t)rgep); 1731 if (err != DDI_SUCCESS) { 1732 rge_problem(rgep, "ddi_intr_add_softint() failed"); 1733 goto attach_fail; 1734 } 1735 rgep->progress |= PROGRESS_FACTOTUM; 1736 1737 /* 1738 * Get supported interrupt types 1739 */ 1740 if (ddi_intr_get_supported_types(devinfo, &intr_types) 1741 != DDI_SUCCESS) { 1742 rge_error(rgep, "ddi_intr_get_supported_types failed\n"); 1743 goto attach_fail; 1744 } 1745 1746 /* 1747 * Add the h/w interrupt handler and initialise mutexes 1748 * RTL8101E is observed to have MSI invalidation issue after S/R. 1749 * So the FIXED interrupt is used instead. 1750 */ 1751 if (rgep->chipid.mac_ver == MAC_VER_8101E) 1752 rgep->msi_enable = B_FALSE; 1753 if ((intr_types & DDI_INTR_TYPE_MSI) && rgep->msi_enable) { 1754 if (rge_add_intrs(rgep, DDI_INTR_TYPE_MSI) != DDI_SUCCESS) { 1755 rge_error(rgep, "MSI registration failed, " 1756 "trying FIXED interrupt type\n"); 1757 } else { 1758 rge_log(rgep, "Using MSI interrupt type\n"); 1759 rgep->intr_type = DDI_INTR_TYPE_MSI; 1760 rgep->progress |= PROGRESS_INTR; 1761 } 1762 } 1763 if (!(rgep->progress & PROGRESS_INTR) && 1764 (intr_types & DDI_INTR_TYPE_FIXED)) { 1765 if (rge_add_intrs(rgep, DDI_INTR_TYPE_FIXED) != DDI_SUCCESS) { 1766 rge_error(rgep, "FIXED interrupt " 1767 "registration failed\n"); 1768 goto attach_fail; 1769 } 1770 rge_log(rgep, "Using FIXED interrupt type\n"); 1771 rgep->intr_type = DDI_INTR_TYPE_FIXED; 1772 rgep->progress |= PROGRESS_INTR; 1773 } 1774 if (!(rgep->progress & PROGRESS_INTR)) { 1775 rge_error(rgep, "No interrupts registered\n"); 1776 goto attach_fail; 1777 } 1778 mutex_init(rgep->genlock, NULL, MUTEX_DRIVER, 1779 DDI_INTR_PRI(rgep->intr_pri)); 1780 rw_init(rgep->errlock, NULL, RW_DRIVER, 1781 DDI_INTR_PRI(rgep->intr_pri)); 1782 mutex_init(rgep->tx_lock, NULL, MUTEX_DRIVER, 1783 DDI_INTR_PRI(rgep->intr_pri)); 1784 mutex_init(rgep->tc_lock, NULL, MUTEX_DRIVER, 1785 DDI_INTR_PRI(rgep->intr_pri)); 1786 mutex_init(rgep->rx_lock, NULL, MUTEX_DRIVER, 1787 DDI_INTR_PRI(rgep->intr_pri)); 1788 mutex_init(rgep->rc_lock, NULL, MUTEX_DRIVER, 1789 DDI_INTR_PRI(rgep->intr_pri)); 1790 1791 /* 1792 * Initialize rings 1793 */ 1794 err = rge_init_rings(rgep); 1795 if (err != DDI_SUCCESS) { 1796 rge_problem(rgep, "rge_init_rings() failed"); 1797 goto attach_fail; 1798 } 1799 rgep->progress |= PROGRESS_INIT; 1800 1801 /* 1802 * Now that mutex locks are initialized, enable interrupts. 1803 */ 1804 if (rgep->intr_cap & DDI_INTR_FLAG_BLOCK) { 1805 /* Call ddi_intr_block_enable() for MSI interrupts */ 1806 (void) ddi_intr_block_enable(rgep->htable, rgep->intr_cnt); 1807 } else { 1808 /* Call ddi_intr_enable for MSI or FIXED interrupts */ 1809 for (i = 0; i < rgep->intr_cnt; i++) { 1810 (void) ddi_intr_enable(rgep->htable[i]); 1811 } 1812 } 1813 1814 /* 1815 * Initialise link state variables 1816 * Stop, reset & reinitialise the chip. 1817 * Initialise the (internal) PHY. 1818 */ 1819 rgep->param_link_up = LINK_STATE_UNKNOWN; 1820 1821 /* 1822 * Reset chip & rings to initial state; also reset address 1823 * filtering, promiscuity, loopback mode. 1824 */ 1825 mutex_enter(rgep->genlock); 1826 (void) rge_chip_reset(rgep); 1827 rge_chip_sync(rgep, RGE_GET_MAC); 1828 bzero(rgep->mcast_hash, sizeof (rgep->mcast_hash)); 1829 bzero(rgep->mcast_refs, sizeof (rgep->mcast_refs)); 1830 rgep->promisc = B_FALSE; 1831 rgep->param_loop_mode = RGE_LOOP_NONE; 1832 mutex_exit(rgep->genlock); 1833 rge_phy_init(rgep); 1834 rgep->progress |= PROGRESS_PHY; 1835 1836 /* 1837 * Create & initialise named kstats 1838 */ 1839 rge_init_kstats(rgep, instance); 1840 rgep->progress |= PROGRESS_KSTATS; 1841 1842 if ((macp = mac_alloc(MAC_VERSION)) == NULL) 1843 goto attach_fail; 1844 macp->m_type_ident = MAC_PLUGIN_IDENT_ETHER; 1845 macp->m_driver = rgep; 1846 macp->m_dip = devinfo; 1847 macp->m_src_addr = rgep->netaddr; 1848 macp->m_callbacks = &rge_m_callbacks; 1849 macp->m_min_sdu = 0; 1850 macp->m_max_sdu = rgep->default_mtu; 1851 macp->m_margin = VLAN_TAGSZ; 1852 1853 /* 1854 * Finally, we're ready to register ourselves with the MAC layer 1855 * interface; if this succeeds, we're all ready to start() 1856 */ 1857 err = mac_register(macp, &rgep->mh); 1858 mac_free(macp); 1859 if (err != 0) 1860 goto attach_fail; 1861 1862 /* 1863 * Register a periodical handler. 1864 * reg_chip_cyclic() is invoked in kernel context. 1865 */ 1866 rgep->periodic_id = ddi_periodic_add(rge_chip_cyclic, rgep, 1867 RGE_CYCLIC_PERIOD, DDI_IPL_0); 1868 1869 rgep->progress |= PROGRESS_READY; 1870 return (DDI_SUCCESS); 1871 1872 attach_fail: 1873 rge_unattach(rgep); 1874 return (DDI_FAILURE); 1875 } 1876 1877 /* 1878 * rge_suspend() -- suspend transmit/receive for powerdown 1879 */ 1880 static int 1881 rge_suspend(rge_t *rgep) 1882 { 1883 /* 1884 * Stop processing and idle (powerdown) the PHY ... 1885 */ 1886 mutex_enter(rgep->genlock); 1887 rw_enter(rgep->errlock, RW_READER); 1888 1889 if (rgep->rge_mac_state != RGE_MAC_STARTED) { 1890 mutex_exit(rgep->genlock); 1891 return (DDI_SUCCESS); 1892 } 1893 1894 rgep->suspended = B_TRUE; 1895 rge_stop(rgep); 1896 rgep->rge_mac_state = RGE_MAC_STOPPED; 1897 1898 rw_exit(rgep->errlock); 1899 mutex_exit(rgep->genlock); 1900 1901 return (DDI_SUCCESS); 1902 } 1903 1904 /* 1905 * detach(9E) -- Detach a device from the system 1906 */ 1907 static int 1908 rge_detach(dev_info_t *devinfo, ddi_detach_cmd_t cmd) 1909 { 1910 rge_t *rgep; 1911 1912 RGE_GTRACE(("rge_detach($%p, %d)", (void *)devinfo, cmd)); 1913 1914 rgep = ddi_get_driver_private(devinfo); 1915 1916 switch (cmd) { 1917 default: 1918 return (DDI_FAILURE); 1919 1920 case DDI_SUSPEND: 1921 return (rge_suspend(rgep)); 1922 1923 case DDI_DETACH: 1924 break; 1925 } 1926 1927 /* 1928 * If there is any posted buffer, the driver should reject to be 1929 * detached. Need notice upper layer to release them. 1930 */ 1931 if (!(rgep->chip_flags & CHIP_FLAG_FORCE_BCOPY) && 1932 rgep->rx_free != RGE_BUF_SLOTS) 1933 return (DDI_FAILURE); 1934 1935 /* 1936 * Unregister from the MAC layer subsystem. This can fail, in 1937 * particular if there are DLPI style-2 streams still open - 1938 * in which case we just return failure without shutting 1939 * down chip operations. 1940 */ 1941 if (mac_unregister(rgep->mh) != 0) 1942 return (DDI_FAILURE); 1943 1944 /* 1945 * All activity stopped, so we can clean up & exit 1946 */ 1947 rge_unattach(rgep); 1948 return (DDI_SUCCESS); 1949 } 1950 1951 1952 /* 1953 * ========== Module Loading Data & Entry Points ========== 1954 */ 1955 1956 #undef RGE_DBG 1957 #define RGE_DBG RGE_DBG_INIT /* debug flag for this code */ 1958 DDI_DEFINE_STREAM_OPS(rge_dev_ops, nulldev, nulldev, rge_attach, rge_detach, 1959 nodev, NULL, D_MP, NULL); 1960 1961 static struct modldrv rge_modldrv = { 1962 &mod_driverops, /* Type of module. This one is a driver */ 1963 rge_ident, /* short description */ 1964 &rge_dev_ops /* driver specific ops */ 1965 }; 1966 1967 static struct modlinkage modlinkage = { 1968 MODREV_1, (void *)&rge_modldrv, NULL 1969 }; 1970 1971 1972 int 1973 _info(struct modinfo *modinfop) 1974 { 1975 return (mod_info(&modlinkage, modinfop)); 1976 } 1977 1978 int 1979 _init(void) 1980 { 1981 int status; 1982 1983 mac_init_ops(&rge_dev_ops, "rge"); 1984 status = mod_install(&modlinkage); 1985 if (status == DDI_SUCCESS) 1986 mutex_init(rge_log_mutex, NULL, MUTEX_DRIVER, NULL); 1987 else 1988 mac_fini_ops(&rge_dev_ops); 1989 1990 return (status); 1991 } 1992 1993 int 1994 _fini(void) 1995 { 1996 int status; 1997 1998 status = mod_remove(&modlinkage); 1999 if (status == DDI_SUCCESS) { 2000 mac_fini_ops(&rge_dev_ops); 2001 mutex_destroy(rge_log_mutex); 2002 } 2003 return (status); 2004 } 2005