1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright 2010 Sun Microsystems, Inc. All rights reserved. 23 * Use is subject to license terms. 24 */ 25 26 #include "rge.h" 27 28 /* 29 * This is the string displayed by modinfo, etc. 30 * Make sure you keep the version ID up to date! 31 */ 32 static char rge_ident[] = "Realtek 1Gb Ethernet"; 33 34 /* 35 * Used for buffers allocated by ddi_dma_mem_alloc() 36 */ 37 static ddi_dma_attr_t dma_attr_buf = { 38 DMA_ATTR_V0, /* dma_attr version */ 39 (uint32_t)0, /* dma_attr_addr_lo */ 40 (uint32_t)0xFFFFFFFF, /* dma_attr_addr_hi */ 41 (uint32_t)0xFFFFFFFF, /* dma_attr_count_max */ 42 (uint32_t)16, /* dma_attr_align */ 43 0xFFFFFFFF, /* dma_attr_burstsizes */ 44 1, /* dma_attr_minxfer */ 45 (uint32_t)0xFFFFFFFF, /* dma_attr_maxxfer */ 46 (uint32_t)0xFFFFFFFF, /* dma_attr_seg */ 47 1, /* dma_attr_sgllen */ 48 1, /* dma_attr_granular */ 49 0, /* dma_attr_flags */ 50 }; 51 52 /* 53 * Used for BDs allocated by ddi_dma_mem_alloc() 54 */ 55 static ddi_dma_attr_t dma_attr_desc = { 56 DMA_ATTR_V0, /* dma_attr version */ 57 (uint32_t)0, /* dma_attr_addr_lo */ 58 (uint32_t)0xFFFFFFFF, /* dma_attr_addr_hi */ 59 (uint32_t)0xFFFFFFFF, /* dma_attr_count_max */ 60 (uint32_t)256, /* dma_attr_align */ 61 0xFFFFFFFF, /* dma_attr_burstsizes */ 62 1, /* dma_attr_minxfer */ 63 (uint32_t)0xFFFFFFFF, /* dma_attr_maxxfer */ 64 (uint32_t)0xFFFFFFFF, /* dma_attr_seg */ 65 1, /* dma_attr_sgllen */ 66 1, /* dma_attr_granular */ 67 0, /* dma_attr_flags */ 68 }; 69 70 /* 71 * PIO access attributes for registers 72 */ 73 static ddi_device_acc_attr_t rge_reg_accattr = { 74 DDI_DEVICE_ATTR_V0, 75 DDI_STRUCTURE_LE_ACC, 76 DDI_STRICTORDER_ACC, 77 DDI_DEFAULT_ACC 78 }; 79 80 /* 81 * DMA access attributes for descriptors 82 */ 83 static ddi_device_acc_attr_t rge_desc_accattr = { 84 DDI_DEVICE_ATTR_V0, 85 DDI_NEVERSWAP_ACC, 86 DDI_STRICTORDER_ACC, 87 DDI_DEFAULT_ACC 88 }; 89 90 /* 91 * DMA access attributes for data 92 */ 93 static ddi_device_acc_attr_t rge_buf_accattr = { 94 DDI_DEVICE_ATTR_V0, 95 DDI_NEVERSWAP_ACC, 96 DDI_STRICTORDER_ACC, 97 DDI_DEFAULT_ACC 98 }; 99 100 /* 101 * Property names 102 */ 103 static char debug_propname[] = "rge_debug_flags"; 104 static char mtu_propname[] = "default_mtu"; 105 static char msi_propname[] = "msi_enable"; 106 107 static int rge_m_start(void *); 108 static void rge_m_stop(void *); 109 static int rge_m_promisc(void *, boolean_t); 110 static int rge_m_multicst(void *, boolean_t, const uint8_t *); 111 static int rge_m_unicst(void *, const uint8_t *); 112 static void rge_m_ioctl(void *, queue_t *, mblk_t *); 113 static boolean_t rge_m_getcapab(void *, mac_capab_t, void *); 114 115 #define RGE_M_CALLBACK_FLAGS (MC_IOCTL | MC_GETCAPAB) 116 117 static mac_callbacks_t rge_m_callbacks = { 118 RGE_M_CALLBACK_FLAGS, 119 rge_m_stat, 120 rge_m_start, 121 rge_m_stop, 122 rge_m_promisc, 123 rge_m_multicst, 124 rge_m_unicst, 125 rge_m_tx, 126 rge_m_ioctl, 127 rge_m_getcapab 128 }; 129 130 /* 131 * Allocate an area of memory and a DMA handle for accessing it 132 */ 133 static int 134 rge_alloc_dma_mem(rge_t *rgep, size_t memsize, ddi_dma_attr_t *dma_attr_p, 135 ddi_device_acc_attr_t *acc_attr_p, uint_t dma_flags, dma_area_t *dma_p) 136 { 137 caddr_t vaddr; 138 int err; 139 140 /* 141 * Allocate handle 142 */ 143 err = ddi_dma_alloc_handle(rgep->devinfo, dma_attr_p, 144 DDI_DMA_SLEEP, NULL, &dma_p->dma_hdl); 145 if (err != DDI_SUCCESS) { 146 dma_p->dma_hdl = NULL; 147 return (DDI_FAILURE); 148 } 149 150 /* 151 * Allocate memory 152 */ 153 err = ddi_dma_mem_alloc(dma_p->dma_hdl, memsize, acc_attr_p, 154 dma_flags & (DDI_DMA_CONSISTENT | DDI_DMA_STREAMING), 155 DDI_DMA_SLEEP, NULL, &vaddr, &dma_p->alength, &dma_p->acc_hdl); 156 if (err != DDI_SUCCESS) { 157 ddi_dma_free_handle(&dma_p->dma_hdl); 158 dma_p->dma_hdl = NULL; 159 dma_p->acc_hdl = NULL; 160 return (DDI_FAILURE); 161 } 162 163 /* 164 * Bind the two together 165 */ 166 dma_p->mem_va = vaddr; 167 err = ddi_dma_addr_bind_handle(dma_p->dma_hdl, NULL, 168 vaddr, dma_p->alength, dma_flags, DDI_DMA_SLEEP, NULL, 169 &dma_p->cookie, &dma_p->ncookies); 170 if (err != DDI_DMA_MAPPED || dma_p->ncookies != 1) { 171 ddi_dma_mem_free(&dma_p->acc_hdl); 172 ddi_dma_free_handle(&dma_p->dma_hdl); 173 dma_p->acc_hdl = NULL; 174 dma_p->dma_hdl = NULL; 175 return (DDI_FAILURE); 176 } 177 178 dma_p->nslots = ~0U; 179 dma_p->size = ~0U; 180 dma_p->token = ~0U; 181 dma_p->offset = 0; 182 return (DDI_SUCCESS); 183 } 184 185 /* 186 * Free one allocated area of DMAable memory 187 */ 188 static void 189 rge_free_dma_mem(dma_area_t *dma_p) 190 { 191 if (dma_p->dma_hdl != NULL) { 192 if (dma_p->ncookies) { 193 (void) ddi_dma_unbind_handle(dma_p->dma_hdl); 194 dma_p->ncookies = 0; 195 } 196 ddi_dma_free_handle(&dma_p->dma_hdl); 197 dma_p->dma_hdl = NULL; 198 } 199 200 if (dma_p->acc_hdl != NULL) { 201 ddi_dma_mem_free(&dma_p->acc_hdl); 202 dma_p->acc_hdl = NULL; 203 } 204 } 205 206 /* 207 * Utility routine to carve a slice off a chunk of allocated memory, 208 * updating the chunk descriptor accordingly. The size of the slice 209 * is given by the product of the <qty> and <size> parameters. 210 */ 211 static void 212 rge_slice_chunk(dma_area_t *slice, dma_area_t *chunk, 213 uint32_t qty, uint32_t size) 214 { 215 static uint32_t sequence = 0xbcd5704a; 216 size_t totsize; 217 218 totsize = qty*size; 219 ASSERT(totsize <= chunk->alength); 220 221 *slice = *chunk; 222 slice->nslots = qty; 223 slice->size = size; 224 slice->alength = totsize; 225 slice->token = ++sequence; 226 227 chunk->mem_va = (caddr_t)chunk->mem_va + totsize; 228 chunk->alength -= totsize; 229 chunk->offset += totsize; 230 chunk->cookie.dmac_laddress += totsize; 231 chunk->cookie.dmac_size -= totsize; 232 } 233 234 static int 235 rge_alloc_bufs(rge_t *rgep) 236 { 237 size_t txdescsize; 238 size_t rxdescsize; 239 int err; 240 241 /* 242 * Allocate memory & handle for packet statistics 243 */ 244 err = rge_alloc_dma_mem(rgep, 245 RGE_STATS_DUMP_SIZE, 246 &dma_attr_desc, 247 &rge_desc_accattr, 248 DDI_DMA_RDWR | DDI_DMA_CONSISTENT, 249 &rgep->dma_area_stats); 250 if (err != DDI_SUCCESS) 251 return (DDI_FAILURE); 252 rgep->hw_stats = DMA_VPTR(rgep->dma_area_stats); 253 254 /* 255 * Allocate memory & handle for Tx descriptor ring 256 */ 257 txdescsize = RGE_SEND_SLOTS * sizeof (rge_bd_t); 258 err = rge_alloc_dma_mem(rgep, 259 txdescsize, 260 &dma_attr_desc, 261 &rge_desc_accattr, 262 DDI_DMA_RDWR | DDI_DMA_CONSISTENT, 263 &rgep->dma_area_txdesc); 264 if (err != DDI_SUCCESS) 265 return (DDI_FAILURE); 266 267 /* 268 * Allocate memory & handle for Rx descriptor ring 269 */ 270 rxdescsize = RGE_RECV_SLOTS * sizeof (rge_bd_t); 271 err = rge_alloc_dma_mem(rgep, 272 rxdescsize, 273 &dma_attr_desc, 274 &rge_desc_accattr, 275 DDI_DMA_RDWR | DDI_DMA_CONSISTENT, 276 &rgep->dma_area_rxdesc); 277 if (err != DDI_SUCCESS) 278 return (DDI_FAILURE); 279 280 return (DDI_SUCCESS); 281 } 282 283 /* 284 * rge_free_bufs() -- free descriptors/buffers allocated for this 285 * device instance. 286 */ 287 static void 288 rge_free_bufs(rge_t *rgep) 289 { 290 rge_free_dma_mem(&rgep->dma_area_stats); 291 rge_free_dma_mem(&rgep->dma_area_txdesc); 292 rge_free_dma_mem(&rgep->dma_area_rxdesc); 293 } 294 295 /* 296 * ========== Transmit and receive ring reinitialisation ========== 297 */ 298 299 /* 300 * These <reinit> routines each reset the rx/tx rings to an initial 301 * state, assuming that the corresponding <init> routine has already 302 * been called exactly once. 303 */ 304 static void 305 rge_reinit_send_ring(rge_t *rgep) 306 { 307 sw_sbd_t *ssbdp; 308 rge_bd_t *bdp; 309 uint32_t slot; 310 311 /* 312 * re-init send ring 313 */ 314 DMA_ZERO(rgep->tx_desc); 315 ssbdp = rgep->sw_sbds; 316 bdp = rgep->tx_ring; 317 for (slot = 0; slot < RGE_SEND_SLOTS; slot++) { 318 bdp->host_buf_addr = 319 RGE_BSWAP_32(ssbdp->pbuf.cookie.dmac_laddress); 320 bdp->host_buf_addr_hi = 321 RGE_BSWAP_32(ssbdp->pbuf.cookie.dmac_laddress >> 32); 322 /* last BD in Tx ring */ 323 if (slot == (RGE_SEND_SLOTS - 1)) 324 bdp->flags_len = RGE_BSWAP_32(BD_FLAG_EOR); 325 ssbdp++; 326 bdp++; 327 } 328 DMA_SYNC(rgep->tx_desc, DDI_DMA_SYNC_FORDEV); 329 rgep->tx_next = 0; 330 rgep->tc_next = 0; 331 rgep->tc_tail = 0; 332 rgep->tx_flow = 0; 333 rgep->tx_free = RGE_SEND_SLOTS; 334 } 335 336 static void 337 rge_reinit_recv_ring(rge_t *rgep) 338 { 339 rge_bd_t *bdp; 340 sw_rbd_t *srbdp; 341 dma_area_t *pbuf; 342 uint32_t slot; 343 344 /* 345 * re-init receive ring 346 */ 347 DMA_ZERO(rgep->rx_desc); 348 srbdp = rgep->sw_rbds; 349 bdp = rgep->rx_ring; 350 for (slot = 0; slot < RGE_RECV_SLOTS; slot++) { 351 pbuf = &srbdp->rx_buf->pbuf; 352 bdp->host_buf_addr = 353 RGE_BSWAP_32(pbuf->cookie.dmac_laddress + rgep->head_room); 354 bdp->host_buf_addr_hi = 355 RGE_BSWAP_32(pbuf->cookie.dmac_laddress >> 32); 356 bdp->flags_len = RGE_BSWAP_32(BD_FLAG_HW_OWN | 357 (rgep->rxbuf_size - rgep->head_room)); 358 /* last BD in Tx ring */ 359 if (slot == (RGE_RECV_SLOTS - 1)) 360 bdp->flags_len |= RGE_BSWAP_32(BD_FLAG_EOR); 361 srbdp++; 362 bdp++; 363 } 364 DMA_SYNC(rgep->rx_desc, DDI_DMA_SYNC_FORDEV); 365 rgep->watchdog = 0; 366 rgep->rx_next = 0; 367 } 368 369 static void 370 rge_reinit_buf_ring(rge_t *rgep) 371 { 372 373 if (rgep->chip_flags & CHIP_FLAG_FORCE_BCOPY) 374 return; 375 376 /* 377 * If all the up-sending buffers haven't been returned to driver, 378 * use bcopy() only in rx process. 379 */ 380 if (rgep->rx_free != RGE_BUF_SLOTS) 381 rgep->rx_bcopy = B_TRUE; 382 } 383 384 static void 385 rge_reinit_rings(rge_t *rgep) 386 { 387 rge_reinit_send_ring(rgep); 388 rge_reinit_recv_ring(rgep); 389 rge_reinit_buf_ring(rgep); 390 } 391 392 static void 393 rge_fini_send_ring(rge_t *rgep) 394 { 395 sw_sbd_t *ssbdp; 396 uint32_t slot; 397 398 ssbdp = rgep->sw_sbds; 399 for (slot = 0; slot < RGE_SEND_SLOTS; ++slot) { 400 rge_free_dma_mem(&ssbdp->pbuf); 401 ssbdp++; 402 } 403 404 kmem_free(rgep->sw_sbds, RGE_SEND_SLOTS * sizeof (sw_sbd_t)); 405 rgep->sw_sbds = NULL; 406 } 407 408 static void 409 rge_fini_recv_ring(rge_t *rgep) 410 { 411 sw_rbd_t *srbdp; 412 uint32_t slot; 413 414 srbdp = rgep->sw_rbds; 415 for (slot = 0; slot < RGE_RECV_SLOTS; ++srbdp, ++slot) { 416 if (srbdp->rx_buf) { 417 if (srbdp->rx_buf->mp != NULL) { 418 freemsg(srbdp->rx_buf->mp); 419 srbdp->rx_buf->mp = NULL; 420 } 421 rge_free_dma_mem(&srbdp->rx_buf->pbuf); 422 kmem_free(srbdp->rx_buf, sizeof (dma_buf_t)); 423 srbdp->rx_buf = NULL; 424 } 425 } 426 427 kmem_free(rgep->sw_rbds, RGE_RECV_SLOTS * sizeof (sw_rbd_t)); 428 rgep->sw_rbds = NULL; 429 } 430 431 static void 432 rge_fini_buf_ring(rge_t *rgep) 433 { 434 sw_rbd_t *srbdp; 435 uint32_t slot; 436 437 if (rgep->chip_flags & CHIP_FLAG_FORCE_BCOPY) 438 return; 439 440 ASSERT(rgep->rx_free == RGE_BUF_SLOTS); 441 442 srbdp = rgep->free_srbds; 443 for (slot = 0; slot < RGE_BUF_SLOTS; ++srbdp, ++slot) { 444 if (srbdp->rx_buf != NULL) { 445 if (srbdp->rx_buf->mp != NULL) { 446 freemsg(srbdp->rx_buf->mp); 447 srbdp->rx_buf->mp = NULL; 448 } 449 rge_free_dma_mem(&srbdp->rx_buf->pbuf); 450 kmem_free(srbdp->rx_buf, sizeof (dma_buf_t)); 451 srbdp->rx_buf = NULL; 452 } 453 } 454 455 kmem_free(rgep->free_srbds, RGE_BUF_SLOTS * sizeof (sw_rbd_t)); 456 rgep->free_srbds = NULL; 457 } 458 459 static void 460 rge_fini_rings(rge_t *rgep) 461 { 462 rge_fini_send_ring(rgep); 463 rge_fini_recv_ring(rgep); 464 rge_fini_buf_ring(rgep); 465 } 466 467 static int 468 rge_init_send_ring(rge_t *rgep) 469 { 470 uint32_t slot; 471 sw_sbd_t *ssbdp; 472 dma_area_t *pbuf; 473 dma_area_t desc; 474 int err; 475 476 /* 477 * Allocate the array of s/w Tx Buffer Descriptors 478 */ 479 ssbdp = kmem_zalloc(RGE_SEND_SLOTS*sizeof (*ssbdp), KM_SLEEP); 480 rgep->sw_sbds = ssbdp; 481 482 /* 483 * Init send ring 484 */ 485 rgep->tx_desc = rgep->dma_area_txdesc; 486 DMA_ZERO(rgep->tx_desc); 487 rgep->tx_ring = rgep->tx_desc.mem_va; 488 489 desc = rgep->tx_desc; 490 for (slot = 0; slot < RGE_SEND_SLOTS; slot++) { 491 rge_slice_chunk(&ssbdp->desc, &desc, 1, sizeof (rge_bd_t)); 492 493 /* 494 * Allocate memory & handle for Tx buffers 495 */ 496 pbuf = &ssbdp->pbuf; 497 err = rge_alloc_dma_mem(rgep, rgep->txbuf_size, 498 &dma_attr_buf, &rge_buf_accattr, 499 DDI_DMA_WRITE | DDI_DMA_STREAMING, pbuf); 500 if (err != DDI_SUCCESS) { 501 rge_error(rgep, 502 "rge_init_send_ring: alloc tx buffer failed"); 503 rge_fini_send_ring(rgep); 504 return (DDI_FAILURE); 505 } 506 ssbdp++; 507 } 508 ASSERT(desc.alength == 0); 509 510 DMA_SYNC(rgep->tx_desc, DDI_DMA_SYNC_FORDEV); 511 return (DDI_SUCCESS); 512 } 513 514 static int 515 rge_init_recv_ring(rge_t *rgep) 516 { 517 uint32_t slot; 518 sw_rbd_t *srbdp; 519 dma_buf_t *rx_buf; 520 dma_area_t *pbuf; 521 int err; 522 523 /* 524 * Allocate the array of s/w Rx Buffer Descriptors 525 */ 526 srbdp = kmem_zalloc(RGE_RECV_SLOTS*sizeof (*srbdp), KM_SLEEP); 527 rgep->sw_rbds = srbdp; 528 529 /* 530 * Init receive ring 531 */ 532 rgep->rx_next = 0; 533 rgep->rx_desc = rgep->dma_area_rxdesc; 534 DMA_ZERO(rgep->rx_desc); 535 rgep->rx_ring = rgep->rx_desc.mem_va; 536 537 for (slot = 0; slot < RGE_RECV_SLOTS; slot++) { 538 srbdp->rx_buf = rx_buf = 539 kmem_zalloc(sizeof (dma_buf_t), KM_SLEEP); 540 541 /* 542 * Allocate memory & handle for Rx buffers 543 */ 544 pbuf = &rx_buf->pbuf; 545 err = rge_alloc_dma_mem(rgep, rgep->rxbuf_size, 546 &dma_attr_buf, &rge_buf_accattr, 547 DDI_DMA_READ | DDI_DMA_STREAMING, pbuf); 548 if (err != DDI_SUCCESS) { 549 rge_fini_recv_ring(rgep); 550 rge_error(rgep, 551 "rge_init_recv_ring: alloc rx buffer failed"); 552 return (DDI_FAILURE); 553 } 554 555 pbuf->alength -= rgep->head_room; 556 pbuf->offset += rgep->head_room; 557 if (!(rgep->chip_flags & CHIP_FLAG_FORCE_BCOPY)) { 558 rx_buf->rx_recycle.free_func = rge_rx_recycle; 559 rx_buf->rx_recycle.free_arg = (caddr_t)rx_buf; 560 rx_buf->private = (caddr_t)rgep; 561 rx_buf->mp = desballoc(DMA_VPTR(rx_buf->pbuf), 562 rgep->rxbuf_size, 0, &rx_buf->rx_recycle); 563 if (rx_buf->mp == NULL) { 564 rge_fini_recv_ring(rgep); 565 rge_problem(rgep, 566 "rge_init_recv_ring: desballoc() failed"); 567 return (DDI_FAILURE); 568 } 569 } 570 srbdp++; 571 } 572 DMA_SYNC(rgep->rx_desc, DDI_DMA_SYNC_FORDEV); 573 return (DDI_SUCCESS); 574 } 575 576 static int 577 rge_init_buf_ring(rge_t *rgep) 578 { 579 uint32_t slot; 580 sw_rbd_t *free_srbdp; 581 dma_buf_t *rx_buf; 582 dma_area_t *pbuf; 583 int err; 584 585 if (rgep->chip_flags & CHIP_FLAG_FORCE_BCOPY) { 586 rgep->rx_bcopy = B_TRUE; 587 return (DDI_SUCCESS); 588 } 589 590 /* 591 * Allocate the array of s/w free Buffer Descriptors 592 */ 593 free_srbdp = kmem_zalloc(RGE_BUF_SLOTS*sizeof (*free_srbdp), KM_SLEEP); 594 rgep->free_srbds = free_srbdp; 595 596 /* 597 * Init free buffer ring 598 */ 599 rgep->rc_next = 0; 600 rgep->rf_next = 0; 601 rgep->rx_bcopy = B_FALSE; 602 rgep->rx_free = RGE_BUF_SLOTS; 603 for (slot = 0; slot < RGE_BUF_SLOTS; slot++) { 604 free_srbdp->rx_buf = rx_buf = 605 kmem_zalloc(sizeof (dma_buf_t), KM_SLEEP); 606 607 /* 608 * Allocate memory & handle for free Rx buffers 609 */ 610 pbuf = &rx_buf->pbuf; 611 err = rge_alloc_dma_mem(rgep, rgep->rxbuf_size, 612 &dma_attr_buf, &rge_buf_accattr, 613 DDI_DMA_READ | DDI_DMA_STREAMING, pbuf); 614 if (err != DDI_SUCCESS) { 615 rge_fini_buf_ring(rgep); 616 rge_error(rgep, 617 "rge_init_buf_ring: alloc rx free buffer failed"); 618 return (DDI_FAILURE); 619 } 620 pbuf->alength -= rgep->head_room; 621 pbuf->offset += rgep->head_room; 622 rx_buf->rx_recycle.free_func = rge_rx_recycle; 623 rx_buf->rx_recycle.free_arg = (caddr_t)rx_buf; 624 rx_buf->private = (caddr_t)rgep; 625 rx_buf->mp = desballoc(DMA_VPTR(rx_buf->pbuf), 626 rgep->rxbuf_size, 0, &rx_buf->rx_recycle); 627 if (rx_buf->mp == NULL) { 628 rge_fini_buf_ring(rgep); 629 rge_problem(rgep, 630 "rge_init_buf_ring: desballoc() failed"); 631 return (DDI_FAILURE); 632 } 633 free_srbdp++; 634 } 635 return (DDI_SUCCESS); 636 } 637 638 static int 639 rge_init_rings(rge_t *rgep) 640 { 641 int err; 642 643 err = rge_init_send_ring(rgep); 644 if (err != DDI_SUCCESS) 645 return (DDI_FAILURE); 646 647 err = rge_init_recv_ring(rgep); 648 if (err != DDI_SUCCESS) { 649 rge_fini_send_ring(rgep); 650 return (DDI_FAILURE); 651 } 652 653 err = rge_init_buf_ring(rgep); 654 if (err != DDI_SUCCESS) { 655 rge_fini_send_ring(rgep); 656 rge_fini_recv_ring(rgep); 657 return (DDI_FAILURE); 658 } 659 660 return (DDI_SUCCESS); 661 } 662 663 /* 664 * ========== Internal state management entry points ========== 665 */ 666 667 #undef RGE_DBG 668 #define RGE_DBG RGE_DBG_NEMO /* debug flag for this code */ 669 670 /* 671 * These routines provide all the functionality required by the 672 * corresponding MAC layer entry points, but don't update the 673 * MAC state so they can be called internally without disturbing 674 * our record of what NEMO thinks we should be doing ... 675 */ 676 677 /* 678 * rge_reset() -- reset h/w & rings to initial state 679 */ 680 static void 681 rge_reset(rge_t *rgep) 682 { 683 ASSERT(mutex_owned(rgep->genlock)); 684 685 /* 686 * Grab all the other mutexes in the world (this should 687 * ensure no other threads are manipulating driver state) 688 */ 689 mutex_enter(rgep->rx_lock); 690 mutex_enter(rgep->rc_lock); 691 rw_enter(rgep->errlock, RW_WRITER); 692 693 (void) rge_chip_reset(rgep); 694 rge_reinit_rings(rgep); 695 rge_chip_init(rgep); 696 697 /* 698 * Free the world ... 699 */ 700 rw_exit(rgep->errlock); 701 mutex_exit(rgep->rc_lock); 702 mutex_exit(rgep->rx_lock); 703 704 rgep->stats.rpackets = 0; 705 rgep->stats.rbytes = 0; 706 rgep->stats.opackets = 0; 707 rgep->stats.obytes = 0; 708 rgep->stats.tx_pre_ismax = B_FALSE; 709 rgep->stats.tx_cur_ismax = B_FALSE; 710 711 RGE_DEBUG(("rge_reset($%p) done", (void *)rgep)); 712 } 713 714 /* 715 * rge_stop() -- stop processing, don't reset h/w or rings 716 */ 717 static void 718 rge_stop(rge_t *rgep) 719 { 720 ASSERT(mutex_owned(rgep->genlock)); 721 722 rge_chip_stop(rgep, B_FALSE); 723 724 RGE_DEBUG(("rge_stop($%p) done", (void *)rgep)); 725 } 726 727 /* 728 * rge_start() -- start transmitting/receiving 729 */ 730 static void 731 rge_start(rge_t *rgep) 732 { 733 ASSERT(mutex_owned(rgep->genlock)); 734 735 /* 736 * Start chip processing, including enabling interrupts 737 */ 738 rge_chip_start(rgep); 739 rgep->watchdog = 0; 740 } 741 742 /* 743 * rge_restart - restart transmitting/receiving after error or suspend 744 */ 745 void 746 rge_restart(rge_t *rgep) 747 { 748 uint32_t i; 749 750 ASSERT(mutex_owned(rgep->genlock)); 751 /* 752 * Wait for posted buffer to be freed... 753 */ 754 if (!rgep->rx_bcopy) { 755 for (i = 0; i < RXBUFF_FREE_LOOP; i++) { 756 if (rgep->rx_free == RGE_BUF_SLOTS) 757 break; 758 drv_usecwait(1000); 759 RGE_DEBUG(("rge_restart: waiting for rx buf free...")); 760 } 761 } 762 rge_reset(rgep); 763 rgep->stats.chip_reset++; 764 if (rgep->rge_mac_state == RGE_MAC_STARTED) { 765 rge_start(rgep); 766 rgep->resched_needed = B_TRUE; 767 (void) ddi_intr_trigger_softint(rgep->resched_hdl, NULL); 768 } 769 } 770 771 772 /* 773 * ========== Nemo-required management entry points ========== 774 */ 775 776 #undef RGE_DBG 777 #define RGE_DBG RGE_DBG_NEMO /* debug flag for this code */ 778 779 /* 780 * rge_m_stop() -- stop transmitting/receiving 781 */ 782 static void 783 rge_m_stop(void *arg) 784 { 785 rge_t *rgep = arg; /* private device info */ 786 uint32_t i; 787 788 /* 789 * Just stop processing, then record new MAC state 790 */ 791 mutex_enter(rgep->genlock); 792 if (rgep->suspended) { 793 ASSERT(rgep->rge_mac_state == RGE_MAC_STOPPED); 794 mutex_exit(rgep->genlock); 795 return; 796 } 797 rge_stop(rgep); 798 /* 799 * Wait for posted buffer to be freed... 800 */ 801 if (!rgep->rx_bcopy) { 802 for (i = 0; i < RXBUFF_FREE_LOOP; i++) { 803 if (rgep->rx_free == RGE_BUF_SLOTS) 804 break; 805 drv_usecwait(1000); 806 RGE_DEBUG(("rge_m_stop: waiting for rx buf free...")); 807 } 808 } 809 rgep->rge_mac_state = RGE_MAC_STOPPED; 810 RGE_DEBUG(("rge_m_stop($%p) done", arg)); 811 mutex_exit(rgep->genlock); 812 } 813 814 /* 815 * rge_m_start() -- start transmitting/receiving 816 */ 817 static int 818 rge_m_start(void *arg) 819 { 820 rge_t *rgep = arg; /* private device info */ 821 822 mutex_enter(rgep->genlock); 823 if (rgep->suspended) { 824 mutex_exit(rgep->genlock); 825 return (DDI_FAILURE); 826 } 827 /* 828 * Clear hw/sw statistics 829 */ 830 DMA_ZERO(rgep->dma_area_stats); 831 bzero(&rgep->stats, sizeof (rge_stats_t)); 832 833 /* 834 * Start processing and record new MAC state 835 */ 836 rge_reset(rgep); 837 rge_start(rgep); 838 rgep->rge_mac_state = RGE_MAC_STARTED; 839 RGE_DEBUG(("rge_m_start($%p) done", arg)); 840 841 mutex_exit(rgep->genlock); 842 843 return (0); 844 } 845 846 /* 847 * rge_m_unicst_set() -- set the physical network address 848 */ 849 static int 850 rge_m_unicst(void *arg, const uint8_t *macaddr) 851 { 852 rge_t *rgep = arg; /* private device info */ 853 854 /* 855 * Remember the new current address in the driver state 856 * Sync the chip's idea of the address too ... 857 */ 858 mutex_enter(rgep->genlock); 859 bcopy(macaddr, rgep->netaddr, ETHERADDRL); 860 861 if (rgep->suspended) { 862 mutex_exit(rgep->genlock); 863 return (DDI_SUCCESS); 864 } 865 866 rge_chip_sync(rgep, RGE_SET_MAC); 867 mutex_exit(rgep->genlock); 868 869 return (0); 870 } 871 872 /* 873 * Compute the index of the required bit in the multicast hash map. 874 * This must mirror the way the hardware actually does it! 875 */ 876 static uint32_t 877 rge_hash_index(const uint8_t *mca) 878 { 879 uint32_t crc = (uint32_t)RGE_HASH_CRC; 880 uint32_t const POLY = RGE_HASH_POLY; 881 uint32_t msb; 882 int bytes; 883 uchar_t currentbyte; 884 uint32_t index; 885 int bit; 886 887 for (bytes = 0; bytes < ETHERADDRL; bytes++) { 888 currentbyte = mca[bytes]; 889 for (bit = 0; bit < 8; bit++) { 890 msb = crc >> 31; 891 crc <<= 1; 892 if (msb ^ (currentbyte & 1)) 893 crc ^= POLY; 894 currentbyte >>= 1; 895 } 896 } 897 index = crc >> 26; 898 /* the index value is between 0 and 63(0x3f) */ 899 900 return (index); 901 } 902 903 /* 904 * rge_m_multicst_add() -- enable/disable a multicast address 905 */ 906 static int 907 rge_m_multicst(void *arg, boolean_t add, const uint8_t *mca) 908 { 909 rge_t *rgep = arg; /* private device info */ 910 struct ether_addr *addr; 911 uint32_t index; 912 uint32_t reg; 913 uint8_t *hashp; 914 915 mutex_enter(rgep->genlock); 916 hashp = rgep->mcast_hash; 917 addr = (struct ether_addr *)mca; 918 /* 919 * Calculate the Multicast address hash index value 920 * Normally, the position of MAR0-MAR7 is 921 * MAR0: offset 0x08, ..., MAR7: offset 0x0F. 922 * 923 * For pcie chipset, the position of MAR0-MAR7 is 924 * different from others: 925 * MAR0: offset 0x0F, ..., MAR7: offset 0x08. 926 */ 927 index = rge_hash_index(addr->ether_addr_octet); 928 if (rgep->chipid.is_pcie) 929 reg = (~(index / RGE_MCAST_NUM)) & 0x7; 930 else 931 reg = index / RGE_MCAST_NUM; 932 933 if (add) { 934 if (rgep->mcast_refs[index]++) { 935 mutex_exit(rgep->genlock); 936 return (0); 937 } 938 hashp[reg] |= 1 << (index % RGE_MCAST_NUM); 939 } else { 940 if (--rgep->mcast_refs[index]) { 941 mutex_exit(rgep->genlock); 942 return (0); 943 } 944 hashp[reg] &= ~ (1 << (index % RGE_MCAST_NUM)); 945 } 946 947 if (rgep->suspended) { 948 mutex_exit(rgep->genlock); 949 return (DDI_SUCCESS); 950 } 951 952 /* 953 * Set multicast register 954 */ 955 rge_chip_sync(rgep, RGE_SET_MUL); 956 957 mutex_exit(rgep->genlock); 958 return (0); 959 } 960 961 /* 962 * rge_m_promisc() -- set or reset promiscuous mode on the board 963 * 964 * Program the hardware to enable/disable promiscuous and/or 965 * receive-all-multicast modes. 966 */ 967 static int 968 rge_m_promisc(void *arg, boolean_t on) 969 { 970 rge_t *rgep = arg; 971 972 /* 973 * Store MAC layer specified mode and pass to chip layer to update h/w 974 */ 975 mutex_enter(rgep->genlock); 976 977 if (rgep->promisc == on) { 978 mutex_exit(rgep->genlock); 979 return (0); 980 } 981 rgep->promisc = on; 982 983 if (rgep->suspended) { 984 mutex_exit(rgep->genlock); 985 return (DDI_SUCCESS); 986 } 987 988 rge_chip_sync(rgep, RGE_SET_PROMISC); 989 RGE_DEBUG(("rge_m_promisc_set($%p) done", arg)); 990 mutex_exit(rgep->genlock); 991 return (0); 992 } 993 994 /* 995 * Loopback ioctl code 996 */ 997 998 static lb_property_t loopmodes[] = { 999 { normal, "normal", RGE_LOOP_NONE }, 1000 { internal, "PHY", RGE_LOOP_INTERNAL_PHY }, 1001 { internal, "MAC", RGE_LOOP_INTERNAL_MAC } 1002 }; 1003 1004 static enum ioc_reply 1005 rge_set_loop_mode(rge_t *rgep, uint32_t mode) 1006 { 1007 /* 1008 * If the mode isn't being changed, there's nothing to do ... 1009 */ 1010 if (mode == rgep->param_loop_mode) 1011 return (IOC_ACK); 1012 1013 /* 1014 * Validate the requested mode and prepare a suitable message 1015 * to explain the link down/up cycle that the change will 1016 * probably induce ... 1017 */ 1018 switch (mode) { 1019 default: 1020 return (IOC_INVAL); 1021 1022 case RGE_LOOP_NONE: 1023 case RGE_LOOP_INTERNAL_PHY: 1024 case RGE_LOOP_INTERNAL_MAC: 1025 break; 1026 } 1027 1028 /* 1029 * All OK; tell the caller to reprogram 1030 * the PHY and/or MAC for the new mode ... 1031 */ 1032 rgep->param_loop_mode = mode; 1033 return (IOC_RESTART_ACK); 1034 } 1035 1036 static enum ioc_reply 1037 rge_loop_ioctl(rge_t *rgep, queue_t *wq, mblk_t *mp, struct iocblk *iocp) 1038 { 1039 lb_info_sz_t *lbsp; 1040 lb_property_t *lbpp; 1041 uint32_t *lbmp; 1042 int cmd; 1043 1044 _NOTE(ARGUNUSED(wq)) 1045 1046 /* 1047 * Validate format of ioctl 1048 */ 1049 if (mp->b_cont == NULL) 1050 return (IOC_INVAL); 1051 1052 cmd = iocp->ioc_cmd; 1053 switch (cmd) { 1054 default: 1055 /* NOTREACHED */ 1056 rge_error(rgep, "rge_loop_ioctl: invalid cmd 0x%x", cmd); 1057 return (IOC_INVAL); 1058 1059 case LB_GET_INFO_SIZE: 1060 if (iocp->ioc_count != sizeof (lb_info_sz_t)) 1061 return (IOC_INVAL); 1062 lbsp = (lb_info_sz_t *)mp->b_cont->b_rptr; 1063 *lbsp = sizeof (loopmodes); 1064 return (IOC_REPLY); 1065 1066 case LB_GET_INFO: 1067 if (iocp->ioc_count != sizeof (loopmodes)) 1068 return (IOC_INVAL); 1069 lbpp = (lb_property_t *)mp->b_cont->b_rptr; 1070 bcopy(loopmodes, lbpp, sizeof (loopmodes)); 1071 return (IOC_REPLY); 1072 1073 case LB_GET_MODE: 1074 if (iocp->ioc_count != sizeof (uint32_t)) 1075 return (IOC_INVAL); 1076 lbmp = (uint32_t *)mp->b_cont->b_rptr; 1077 *lbmp = rgep->param_loop_mode; 1078 return (IOC_REPLY); 1079 1080 case LB_SET_MODE: 1081 if (iocp->ioc_count != sizeof (uint32_t)) 1082 return (IOC_INVAL); 1083 lbmp = (uint32_t *)mp->b_cont->b_rptr; 1084 return (rge_set_loop_mode(rgep, *lbmp)); 1085 } 1086 } 1087 1088 /* 1089 * Specific rge IOCTLs, the MAC layer handles the generic ones. 1090 */ 1091 static void 1092 rge_m_ioctl(void *arg, queue_t *wq, mblk_t *mp) 1093 { 1094 rge_t *rgep = arg; 1095 struct iocblk *iocp; 1096 enum ioc_reply status; 1097 boolean_t need_privilege; 1098 int err; 1099 int cmd; 1100 1101 /* 1102 * If suspended, we might actually be able to do some of 1103 * these ioctls, but it is harder to make sure they occur 1104 * without actually putting the hardware in an undesireable 1105 * state. So just NAK it. 1106 */ 1107 mutex_enter(rgep->genlock); 1108 if (rgep->suspended) { 1109 miocnak(wq, mp, 0, EINVAL); 1110 mutex_exit(rgep->genlock); 1111 return; 1112 } 1113 mutex_exit(rgep->genlock); 1114 1115 /* 1116 * Validate the command before bothering with the mutex ... 1117 */ 1118 iocp = (struct iocblk *)mp->b_rptr; 1119 iocp->ioc_error = 0; 1120 need_privilege = B_TRUE; 1121 cmd = iocp->ioc_cmd; 1122 switch (cmd) { 1123 default: 1124 miocnak(wq, mp, 0, EINVAL); 1125 return; 1126 1127 case RGE_MII_READ: 1128 case RGE_MII_WRITE: 1129 case RGE_DIAG: 1130 case RGE_PEEK: 1131 case RGE_POKE: 1132 case RGE_PHY_RESET: 1133 case RGE_SOFT_RESET: 1134 case RGE_HARD_RESET: 1135 break; 1136 1137 case LB_GET_INFO_SIZE: 1138 case LB_GET_INFO: 1139 case LB_GET_MODE: 1140 need_privilege = B_FALSE; 1141 /* FALLTHRU */ 1142 case LB_SET_MODE: 1143 break; 1144 1145 case ND_GET: 1146 need_privilege = B_FALSE; 1147 /* FALLTHRU */ 1148 case ND_SET: 1149 break; 1150 } 1151 1152 if (need_privilege) { 1153 /* 1154 * Check for specific net_config privilege 1155 */ 1156 err = secpolicy_net_config(iocp->ioc_cr, B_FALSE); 1157 if (err != 0) { 1158 miocnak(wq, mp, 0, err); 1159 return; 1160 } 1161 } 1162 1163 mutex_enter(rgep->genlock); 1164 1165 switch (cmd) { 1166 default: 1167 _NOTE(NOTREACHED) 1168 status = IOC_INVAL; 1169 break; 1170 1171 case RGE_MII_READ: 1172 case RGE_MII_WRITE: 1173 case RGE_DIAG: 1174 case RGE_PEEK: 1175 case RGE_POKE: 1176 case RGE_PHY_RESET: 1177 case RGE_SOFT_RESET: 1178 case RGE_HARD_RESET: 1179 status = rge_chip_ioctl(rgep, wq, mp, iocp); 1180 break; 1181 1182 case LB_GET_INFO_SIZE: 1183 case LB_GET_INFO: 1184 case LB_GET_MODE: 1185 case LB_SET_MODE: 1186 status = rge_loop_ioctl(rgep, wq, mp, iocp); 1187 break; 1188 1189 case ND_GET: 1190 case ND_SET: 1191 status = rge_nd_ioctl(rgep, wq, mp, iocp); 1192 break; 1193 } 1194 1195 /* 1196 * Do we need to reprogram the PHY and/or the MAC? 1197 * Do it now, while we still have the mutex. 1198 * 1199 * Note: update the PHY first, 'cos it controls the 1200 * speed/duplex parameters that the MAC code uses. 1201 */ 1202 switch (status) { 1203 case IOC_RESTART_REPLY: 1204 case IOC_RESTART_ACK: 1205 rge_phy_update(rgep); 1206 break; 1207 } 1208 1209 mutex_exit(rgep->genlock); 1210 1211 /* 1212 * Finally, decide how to reply 1213 */ 1214 switch (status) { 1215 default: 1216 case IOC_INVAL: 1217 /* 1218 * Error, reply with a NAK and EINVAL or the specified error 1219 */ 1220 miocnak(wq, mp, 0, iocp->ioc_error == 0 ? 1221 EINVAL : iocp->ioc_error); 1222 break; 1223 1224 case IOC_DONE: 1225 /* 1226 * OK, reply already sent 1227 */ 1228 break; 1229 1230 case IOC_RESTART_ACK: 1231 case IOC_ACK: 1232 /* 1233 * OK, reply with an ACK 1234 */ 1235 miocack(wq, mp, 0, 0); 1236 break; 1237 1238 case IOC_RESTART_REPLY: 1239 case IOC_REPLY: 1240 /* 1241 * OK, send prepared reply as ACK or NAK 1242 */ 1243 mp->b_datap->db_type = iocp->ioc_error == 0 ? 1244 M_IOCACK : M_IOCNAK; 1245 qreply(wq, mp); 1246 break; 1247 } 1248 } 1249 1250 /* ARGSUSED */ 1251 static boolean_t 1252 rge_m_getcapab(void *arg, mac_capab_t cap, void *cap_data) 1253 { 1254 rge_t *rgep = arg; 1255 1256 switch (cap) { 1257 case MAC_CAPAB_HCKSUM: { 1258 uint32_t *hcksum_txflags = cap_data; 1259 switch (rgep->chipid.mac_ver) { 1260 case MAC_VER_8169: 1261 case MAC_VER_8169S_D: 1262 case MAC_VER_8169S_E: 1263 case MAC_VER_8169SB: 1264 case MAC_VER_8169SC: 1265 case MAC_VER_8168: 1266 case MAC_VER_8168B_B: 1267 case MAC_VER_8168B_C: 1268 case MAC_VER_8101E: 1269 *hcksum_txflags = HCKSUM_INET_FULL_V4 | 1270 HCKSUM_IPHDRCKSUM; 1271 break; 1272 case MAC_VER_8168C: 1273 case MAC_VER_8101E_B: 1274 case MAC_VER_8101E_C: 1275 default: 1276 *hcksum_txflags = 0; 1277 break; 1278 } 1279 break; 1280 } 1281 default: 1282 return (B_FALSE); 1283 } 1284 return (B_TRUE); 1285 } 1286 1287 /* 1288 * ============ Init MSI/Fixed Interrupt routines ============== 1289 */ 1290 1291 /* 1292 * rge_add_intrs: 1293 * 1294 * Register FIXED or MSI interrupts. 1295 */ 1296 static int 1297 rge_add_intrs(rge_t *rgep, int intr_type) 1298 { 1299 dev_info_t *dip = rgep->devinfo; 1300 int avail; 1301 int actual; 1302 int intr_size; 1303 int count; 1304 int i, j; 1305 int ret; 1306 1307 /* Get number of interrupts */ 1308 ret = ddi_intr_get_nintrs(dip, intr_type, &count); 1309 if ((ret != DDI_SUCCESS) || (count == 0)) { 1310 rge_error(rgep, "ddi_intr_get_nintrs() failure, ret: %d, " 1311 "count: %d", ret, count); 1312 return (DDI_FAILURE); 1313 } 1314 1315 /* Get number of available interrupts */ 1316 ret = ddi_intr_get_navail(dip, intr_type, &avail); 1317 if ((ret != DDI_SUCCESS) || (avail == 0)) { 1318 rge_error(rgep, "ddi_intr_get_navail() failure, " 1319 "ret: %d, avail: %d\n", ret, avail); 1320 return (DDI_FAILURE); 1321 } 1322 1323 /* Allocate an array of interrupt handles */ 1324 intr_size = count * sizeof (ddi_intr_handle_t); 1325 rgep->htable = kmem_alloc(intr_size, KM_SLEEP); 1326 rgep->intr_rqst = count; 1327 1328 /* Call ddi_intr_alloc() */ 1329 ret = ddi_intr_alloc(dip, rgep->htable, intr_type, 0, 1330 count, &actual, DDI_INTR_ALLOC_NORMAL); 1331 if (ret != DDI_SUCCESS || actual == 0) { 1332 rge_error(rgep, "ddi_intr_alloc() failed %d\n", ret); 1333 kmem_free(rgep->htable, intr_size); 1334 return (DDI_FAILURE); 1335 } 1336 if (actual < count) { 1337 rge_log(rgep, "ddi_intr_alloc() Requested: %d, Received: %d\n", 1338 count, actual); 1339 } 1340 rgep->intr_cnt = actual; 1341 1342 /* 1343 * Get priority for first msi, assume remaining are all the same 1344 */ 1345 if ((ret = ddi_intr_get_pri(rgep->htable[0], &rgep->intr_pri)) != 1346 DDI_SUCCESS) { 1347 rge_error(rgep, "ddi_intr_get_pri() failed %d\n", ret); 1348 /* Free already allocated intr */ 1349 for (i = 0; i < actual; i++) { 1350 (void) ddi_intr_free(rgep->htable[i]); 1351 } 1352 kmem_free(rgep->htable, intr_size); 1353 return (DDI_FAILURE); 1354 } 1355 1356 /* Test for high level mutex */ 1357 if (rgep->intr_pri >= ddi_intr_get_hilevel_pri()) { 1358 rge_error(rgep, "rge_add_intrs:" 1359 "Hi level interrupt not supported"); 1360 for (i = 0; i < actual; i++) 1361 (void) ddi_intr_free(rgep->htable[i]); 1362 kmem_free(rgep->htable, intr_size); 1363 return (DDI_FAILURE); 1364 } 1365 1366 /* Call ddi_intr_add_handler() */ 1367 for (i = 0; i < actual; i++) { 1368 if ((ret = ddi_intr_add_handler(rgep->htable[i], rge_intr, 1369 (caddr_t)rgep, (caddr_t)(uintptr_t)i)) != DDI_SUCCESS) { 1370 rge_error(rgep, "ddi_intr_add_handler() " 1371 "failed %d\n", ret); 1372 /* Remove already added intr */ 1373 for (j = 0; j < i; j++) 1374 (void) ddi_intr_remove_handler(rgep->htable[j]); 1375 /* Free already allocated intr */ 1376 for (i = 0; i < actual; i++) { 1377 (void) ddi_intr_free(rgep->htable[i]); 1378 } 1379 kmem_free(rgep->htable, intr_size); 1380 return (DDI_FAILURE); 1381 } 1382 } 1383 1384 if ((ret = ddi_intr_get_cap(rgep->htable[0], &rgep->intr_cap)) 1385 != DDI_SUCCESS) { 1386 rge_error(rgep, "ddi_intr_get_cap() failed %d\n", ret); 1387 for (i = 0; i < actual; i++) { 1388 (void) ddi_intr_remove_handler(rgep->htable[i]); 1389 (void) ddi_intr_free(rgep->htable[i]); 1390 } 1391 kmem_free(rgep->htable, intr_size); 1392 return (DDI_FAILURE); 1393 } 1394 1395 return (DDI_SUCCESS); 1396 } 1397 1398 /* 1399 * rge_rem_intrs: 1400 * 1401 * Unregister FIXED or MSI interrupts 1402 */ 1403 static void 1404 rge_rem_intrs(rge_t *rgep) 1405 { 1406 int i; 1407 1408 /* Disable all interrupts */ 1409 if (rgep->intr_cap & DDI_INTR_FLAG_BLOCK) { 1410 /* Call ddi_intr_block_disable() */ 1411 (void) ddi_intr_block_disable(rgep->htable, rgep->intr_cnt); 1412 } else { 1413 for (i = 0; i < rgep->intr_cnt; i++) { 1414 (void) ddi_intr_disable(rgep->htable[i]); 1415 } 1416 } 1417 1418 /* Call ddi_intr_remove_handler() */ 1419 for (i = 0; i < rgep->intr_cnt; i++) { 1420 (void) ddi_intr_remove_handler(rgep->htable[i]); 1421 (void) ddi_intr_free(rgep->htable[i]); 1422 } 1423 1424 kmem_free(rgep->htable, rgep->intr_rqst * sizeof (ddi_intr_handle_t)); 1425 } 1426 1427 /* 1428 * ========== Per-instance setup/teardown code ========== 1429 */ 1430 1431 #undef RGE_DBG 1432 #define RGE_DBG RGE_DBG_INIT /* debug flag for this code */ 1433 1434 static void 1435 rge_unattach(rge_t *rgep) 1436 { 1437 /* 1438 * Flag that no more activity may be initiated 1439 */ 1440 rgep->progress &= ~PROGRESS_READY; 1441 rgep->rge_mac_state = RGE_MAC_UNATTACH; 1442 1443 /* 1444 * Quiesce the PHY and MAC (leave it reset but still powered). 1445 * Clean up and free all RGE data structures 1446 */ 1447 if (rgep->periodic_id != NULL) { 1448 ddi_periodic_delete(rgep->periodic_id); 1449 rgep->periodic_id = NULL; 1450 } 1451 1452 if (rgep->progress & PROGRESS_KSTATS) 1453 rge_fini_kstats(rgep); 1454 1455 if (rgep->progress & PROGRESS_PHY) 1456 (void) rge_phy_reset(rgep); 1457 1458 if (rgep->progress & PROGRESS_INIT) { 1459 mutex_enter(rgep->genlock); 1460 (void) rge_chip_reset(rgep); 1461 mutex_exit(rgep->genlock); 1462 rge_fini_rings(rgep); 1463 } 1464 1465 if (rgep->progress & PROGRESS_INTR) { 1466 rge_rem_intrs(rgep); 1467 mutex_destroy(rgep->rc_lock); 1468 mutex_destroy(rgep->rx_lock); 1469 mutex_destroy(rgep->tc_lock); 1470 mutex_destroy(rgep->tx_lock); 1471 rw_destroy(rgep->errlock); 1472 mutex_destroy(rgep->genlock); 1473 } 1474 1475 if (rgep->progress & PROGRESS_FACTOTUM) 1476 (void) ddi_intr_remove_softint(rgep->factotum_hdl); 1477 1478 if (rgep->progress & PROGRESS_RESCHED) 1479 (void) ddi_intr_remove_softint(rgep->resched_hdl); 1480 1481 if (rgep->progress & PROGRESS_NDD) 1482 rge_nd_cleanup(rgep); 1483 1484 rge_free_bufs(rgep); 1485 1486 if (rgep->progress & PROGRESS_REGS) 1487 ddi_regs_map_free(&rgep->io_handle); 1488 1489 if (rgep->progress & PROGRESS_CFG) 1490 pci_config_teardown(&rgep->cfg_handle); 1491 1492 ddi_remove_minor_node(rgep->devinfo, NULL); 1493 kmem_free(rgep, sizeof (*rgep)); 1494 } 1495 1496 static int 1497 rge_resume(dev_info_t *devinfo) 1498 { 1499 rge_t *rgep; /* Our private data */ 1500 chip_id_t *cidp; 1501 chip_id_t chipid; 1502 1503 rgep = ddi_get_driver_private(devinfo); 1504 1505 /* 1506 * If there are state inconsistancies, this is bad. Returning 1507 * DDI_FAILURE here will eventually cause the machine to panic, 1508 * so it is best done here so that there is a possibility of 1509 * debugging the problem. 1510 */ 1511 if (rgep == NULL) 1512 cmn_err(CE_PANIC, 1513 "rge: ngep returned from ddi_get_driver_private was NULL"); 1514 1515 /* 1516 * Refuse to resume if the data structures aren't consistent 1517 */ 1518 if (rgep->devinfo != devinfo) 1519 cmn_err(CE_PANIC, 1520 "rge: passed devinfo not the same as saved devinfo"); 1521 1522 /* 1523 * Read chip ID & set up config space command register(s) 1524 * Refuse to resume if the chip has changed its identity! 1525 */ 1526 cidp = &rgep->chipid; 1527 rge_chip_cfg_init(rgep, &chipid); 1528 if (chipid.vendor != cidp->vendor) 1529 return (DDI_FAILURE); 1530 if (chipid.device != cidp->device) 1531 return (DDI_FAILURE); 1532 if (chipid.revision != cidp->revision) 1533 return (DDI_FAILURE); 1534 1535 mutex_enter(rgep->genlock); 1536 1537 /* 1538 * Only in one case, this conditional branch can be executed: the port 1539 * hasn't been plumbed. 1540 */ 1541 if (rgep->suspended == B_FALSE) { 1542 mutex_exit(rgep->genlock); 1543 return (DDI_SUCCESS); 1544 } 1545 rgep->rge_mac_state = RGE_MAC_STARTED; 1546 /* 1547 * All OK, reinitialise h/w & kick off NEMO scheduling 1548 */ 1549 rge_restart(rgep); 1550 rgep->suspended = B_FALSE; 1551 1552 mutex_exit(rgep->genlock); 1553 1554 return (DDI_SUCCESS); 1555 } 1556 1557 1558 /* 1559 * attach(9E) -- Attach a device to the system 1560 * 1561 * Called once for each board successfully probed. 1562 */ 1563 static int 1564 rge_attach(dev_info_t *devinfo, ddi_attach_cmd_t cmd) 1565 { 1566 rge_t *rgep; /* Our private data */ 1567 mac_register_t *macp; 1568 chip_id_t *cidp; 1569 int intr_types; 1570 caddr_t regs; 1571 int instance; 1572 int i; 1573 int err; 1574 1575 /* 1576 * we don't support high level interrupts in the driver 1577 */ 1578 if (ddi_intr_hilevel(devinfo, 0) != 0) { 1579 cmn_err(CE_WARN, 1580 "rge_attach -- unsupported high level interrupt"); 1581 return (DDI_FAILURE); 1582 } 1583 1584 instance = ddi_get_instance(devinfo); 1585 RGE_GTRACE(("rge_attach($%p, %d) instance %d", 1586 (void *)devinfo, cmd, instance)); 1587 RGE_BRKPT(NULL, "rge_attach"); 1588 1589 switch (cmd) { 1590 default: 1591 return (DDI_FAILURE); 1592 1593 case DDI_RESUME: 1594 return (rge_resume(devinfo)); 1595 1596 case DDI_ATTACH: 1597 break; 1598 } 1599 1600 rgep = kmem_zalloc(sizeof (*rgep), KM_SLEEP); 1601 ddi_set_driver_private(devinfo, rgep); 1602 rgep->devinfo = devinfo; 1603 1604 /* 1605 * Initialize more fields in RGE private data 1606 */ 1607 rgep->rge_mac_state = RGE_MAC_ATTACH; 1608 rgep->debug = ddi_prop_get_int(DDI_DEV_T_ANY, devinfo, 1609 DDI_PROP_DONTPASS, debug_propname, rge_debug); 1610 rgep->default_mtu = ddi_prop_get_int(DDI_DEV_T_ANY, devinfo, 1611 DDI_PROP_DONTPASS, mtu_propname, ETHERMTU); 1612 rgep->msi_enable = ddi_prop_get_int(DDI_DEV_T_ANY, devinfo, 1613 DDI_PROP_DONTPASS, msi_propname, B_TRUE); 1614 (void) snprintf(rgep->ifname, sizeof (rgep->ifname), "%s%d", 1615 RGE_DRIVER_NAME, instance); 1616 1617 /* 1618 * Map config space registers 1619 * Read chip ID & set up config space command register(s) 1620 * 1621 * Note: this leaves the chip accessible by Memory Space 1622 * accesses, but with interrupts and Bus Mastering off. 1623 * This should ensure that nothing untoward will happen 1624 * if it has been left active by the (net-)bootloader. 1625 * We'll re-enable Bus Mastering once we've reset the chip, 1626 * and allow interrupts only when everything else is set up. 1627 */ 1628 err = pci_config_setup(devinfo, &rgep->cfg_handle); 1629 if (err != DDI_SUCCESS) { 1630 rge_problem(rgep, "pci_config_setup() failed"); 1631 goto attach_fail; 1632 } 1633 rgep->progress |= PROGRESS_CFG; 1634 cidp = &rgep->chipid; 1635 bzero(cidp, sizeof (*cidp)); 1636 rge_chip_cfg_init(rgep, cidp); 1637 1638 /* 1639 * Map operating registers 1640 */ 1641 err = ddi_regs_map_setup(devinfo, 2, ®s, 1642 0, 0, &rge_reg_accattr, &rgep->io_handle); 1643 1644 /* 1645 * MMIO map will fail if the assigned address is bigger than 4G 1646 * then choose I/O map 1647 */ 1648 if (err != DDI_SUCCESS) { 1649 err = ddi_regs_map_setup(devinfo, 1, ®s, 1650 0, 0, &rge_reg_accattr, &rgep->io_handle); 1651 } 1652 if (err != DDI_SUCCESS) { 1653 rge_problem(rgep, "ddi_regs_map_setup() failed"); 1654 goto attach_fail; 1655 } 1656 rgep->io_regs = regs; 1657 rgep->progress |= PROGRESS_REGS; 1658 1659 /* 1660 * Characterise the device, so we know its requirements. 1661 * Then allocate the appropriate TX and RX descriptors & buffers. 1662 */ 1663 rge_chip_ident(rgep); 1664 err = rge_alloc_bufs(rgep); 1665 if (err != DDI_SUCCESS) { 1666 rge_problem(rgep, "DMA buffer allocation failed"); 1667 goto attach_fail; 1668 } 1669 1670 /* 1671 * Register NDD-tweakable parameters 1672 */ 1673 if (rge_nd_init(rgep)) { 1674 rge_problem(rgep, "rge_nd_init() failed"); 1675 goto attach_fail; 1676 } 1677 rgep->progress |= PROGRESS_NDD; 1678 1679 /* 1680 * Add the softint handlers: 1681 * 1682 * Both of these handlers are used to avoid restrictions on the 1683 * context and/or mutexes required for some operations. In 1684 * particular, the hardware interrupt handler and its subfunctions 1685 * can detect a number of conditions that we don't want to handle 1686 * in that context or with that set of mutexes held. So, these 1687 * softints are triggered instead: 1688 * 1689 * the <resched> softint is triggered if if we have previously 1690 * had to refuse to send a packet because of resource shortage 1691 * (we've run out of transmit buffers), but the send completion 1692 * interrupt handler has now detected that more buffers have 1693 * become available. 1694 * 1695 * the <factotum> is triggered if the h/w interrupt handler 1696 * sees the <link state changed> or <error> bits in the status 1697 * block. It's also triggered periodically to poll the link 1698 * state, just in case we aren't getting link status change 1699 * interrupts ... 1700 */ 1701 err = ddi_intr_add_softint(devinfo, &rgep->resched_hdl, 1702 DDI_INTR_SOFTPRI_MIN, rge_reschedule, (caddr_t)rgep); 1703 if (err != DDI_SUCCESS) { 1704 rge_problem(rgep, "ddi_intr_add_softint() failed"); 1705 goto attach_fail; 1706 } 1707 rgep->progress |= PROGRESS_RESCHED; 1708 err = ddi_intr_add_softint(devinfo, &rgep->factotum_hdl, 1709 DDI_INTR_SOFTPRI_MIN, rge_chip_factotum, (caddr_t)rgep); 1710 if (err != DDI_SUCCESS) { 1711 rge_problem(rgep, "ddi_intr_add_softint() failed"); 1712 goto attach_fail; 1713 } 1714 rgep->progress |= PROGRESS_FACTOTUM; 1715 1716 /* 1717 * Get supported interrupt types 1718 */ 1719 if (ddi_intr_get_supported_types(devinfo, &intr_types) 1720 != DDI_SUCCESS) { 1721 rge_error(rgep, "ddi_intr_get_supported_types failed\n"); 1722 goto attach_fail; 1723 } 1724 1725 /* 1726 * Add the h/w interrupt handler and initialise mutexes 1727 * RTL8101E is observed to have MSI invalidation issue after S/R. 1728 * So the FIXED interrupt is used instead. 1729 */ 1730 if (rgep->chipid.mac_ver == MAC_VER_8101E) 1731 rgep->msi_enable = B_FALSE; 1732 if ((intr_types & DDI_INTR_TYPE_MSI) && rgep->msi_enable) { 1733 if (rge_add_intrs(rgep, DDI_INTR_TYPE_MSI) != DDI_SUCCESS) { 1734 rge_error(rgep, "MSI registration failed, " 1735 "trying FIXED interrupt type\n"); 1736 } else { 1737 rge_log(rgep, "Using MSI interrupt type\n"); 1738 rgep->intr_type = DDI_INTR_TYPE_MSI; 1739 rgep->progress |= PROGRESS_INTR; 1740 } 1741 } 1742 if (!(rgep->progress & PROGRESS_INTR) && 1743 (intr_types & DDI_INTR_TYPE_FIXED)) { 1744 if (rge_add_intrs(rgep, DDI_INTR_TYPE_FIXED) != DDI_SUCCESS) { 1745 rge_error(rgep, "FIXED interrupt " 1746 "registration failed\n"); 1747 goto attach_fail; 1748 } 1749 rge_log(rgep, "Using FIXED interrupt type\n"); 1750 rgep->intr_type = DDI_INTR_TYPE_FIXED; 1751 rgep->progress |= PROGRESS_INTR; 1752 } 1753 if (!(rgep->progress & PROGRESS_INTR)) { 1754 rge_error(rgep, "No interrupts registered\n"); 1755 goto attach_fail; 1756 } 1757 mutex_init(rgep->genlock, NULL, MUTEX_DRIVER, 1758 DDI_INTR_PRI(rgep->intr_pri)); 1759 rw_init(rgep->errlock, NULL, RW_DRIVER, 1760 DDI_INTR_PRI(rgep->intr_pri)); 1761 mutex_init(rgep->tx_lock, NULL, MUTEX_DRIVER, 1762 DDI_INTR_PRI(rgep->intr_pri)); 1763 mutex_init(rgep->tc_lock, NULL, MUTEX_DRIVER, 1764 DDI_INTR_PRI(rgep->intr_pri)); 1765 mutex_init(rgep->rx_lock, NULL, MUTEX_DRIVER, 1766 DDI_INTR_PRI(rgep->intr_pri)); 1767 mutex_init(rgep->rc_lock, NULL, MUTEX_DRIVER, 1768 DDI_INTR_PRI(rgep->intr_pri)); 1769 1770 /* 1771 * Initialize rings 1772 */ 1773 err = rge_init_rings(rgep); 1774 if (err != DDI_SUCCESS) { 1775 rge_problem(rgep, "rge_init_rings() failed"); 1776 goto attach_fail; 1777 } 1778 rgep->progress |= PROGRESS_INIT; 1779 1780 /* 1781 * Now that mutex locks are initialized, enable interrupts. 1782 */ 1783 if (rgep->intr_cap & DDI_INTR_FLAG_BLOCK) { 1784 /* Call ddi_intr_block_enable() for MSI interrupts */ 1785 (void) ddi_intr_block_enable(rgep->htable, rgep->intr_cnt); 1786 } else { 1787 /* Call ddi_intr_enable for MSI or FIXED interrupts */ 1788 for (i = 0; i < rgep->intr_cnt; i++) { 1789 (void) ddi_intr_enable(rgep->htable[i]); 1790 } 1791 } 1792 1793 /* 1794 * Initialise link state variables 1795 * Stop, reset & reinitialise the chip. 1796 * Initialise the (internal) PHY. 1797 */ 1798 rgep->param_link_up = LINK_STATE_UNKNOWN; 1799 1800 /* 1801 * Reset chip & rings to initial state; also reset address 1802 * filtering, promiscuity, loopback mode. 1803 */ 1804 mutex_enter(rgep->genlock); 1805 (void) rge_chip_reset(rgep); 1806 rge_chip_sync(rgep, RGE_GET_MAC); 1807 bzero(rgep->mcast_hash, sizeof (rgep->mcast_hash)); 1808 bzero(rgep->mcast_refs, sizeof (rgep->mcast_refs)); 1809 rgep->promisc = B_FALSE; 1810 rgep->param_loop_mode = RGE_LOOP_NONE; 1811 mutex_exit(rgep->genlock); 1812 rge_phy_init(rgep); 1813 rgep->progress |= PROGRESS_PHY; 1814 1815 /* 1816 * Create & initialise named kstats 1817 */ 1818 rge_init_kstats(rgep, instance); 1819 rgep->progress |= PROGRESS_KSTATS; 1820 1821 if ((macp = mac_alloc(MAC_VERSION)) == NULL) 1822 goto attach_fail; 1823 macp->m_type_ident = MAC_PLUGIN_IDENT_ETHER; 1824 macp->m_driver = rgep; 1825 macp->m_dip = devinfo; 1826 macp->m_src_addr = rgep->netaddr; 1827 macp->m_callbacks = &rge_m_callbacks; 1828 macp->m_min_sdu = 0; 1829 macp->m_max_sdu = rgep->default_mtu; 1830 macp->m_margin = VLAN_TAGSZ; 1831 1832 /* 1833 * Finally, we're ready to register ourselves with the MAC layer 1834 * interface; if this succeeds, we're all ready to start() 1835 */ 1836 err = mac_register(macp, &rgep->mh); 1837 mac_free(macp); 1838 if (err != 0) 1839 goto attach_fail; 1840 1841 /* 1842 * Register a periodical handler. 1843 * reg_chip_cyclic() is invoked in kernel context. 1844 */ 1845 rgep->periodic_id = ddi_periodic_add(rge_chip_cyclic, rgep, 1846 RGE_CYCLIC_PERIOD, DDI_IPL_0); 1847 1848 rgep->progress |= PROGRESS_READY; 1849 return (DDI_SUCCESS); 1850 1851 attach_fail: 1852 rge_unattach(rgep); 1853 return (DDI_FAILURE); 1854 } 1855 1856 /* 1857 * rge_suspend() -- suspend transmit/receive for powerdown 1858 */ 1859 static int 1860 rge_suspend(rge_t *rgep) 1861 { 1862 /* 1863 * Stop processing and idle (powerdown) the PHY ... 1864 */ 1865 mutex_enter(rgep->genlock); 1866 rw_enter(rgep->errlock, RW_WRITER); 1867 1868 if (rgep->rge_mac_state != RGE_MAC_STARTED) { 1869 rw_exit(rgep->errlock); 1870 mutex_exit(rgep->genlock); 1871 return (DDI_SUCCESS); 1872 } 1873 1874 rgep->suspended = B_TRUE; 1875 rge_stop(rgep); 1876 rgep->rge_mac_state = RGE_MAC_STOPPED; 1877 1878 rw_exit(rgep->errlock); 1879 mutex_exit(rgep->genlock); 1880 1881 return (DDI_SUCCESS); 1882 } 1883 1884 /* 1885 * quiesce(9E) entry point. 1886 * 1887 * This function is called when the system is single-threaded at high 1888 * PIL with preemption disabled. Therefore, this function must not be 1889 * blocked. 1890 * 1891 * This function returns DDI_SUCCESS on success, or DDI_FAILURE on failure. 1892 * DDI_FAILURE indicates an error condition and should almost never happen. 1893 */ 1894 static int 1895 rge_quiesce(dev_info_t *devinfo) 1896 { 1897 rge_t *rgep = ddi_get_driver_private(devinfo); 1898 1899 if (rgep == NULL) 1900 return (DDI_FAILURE); 1901 1902 /* 1903 * Turn off debugging 1904 */ 1905 rge_debug = 0; 1906 rgep->debug = 0; 1907 1908 /* Stop the chip */ 1909 rge_chip_stop(rgep, B_FALSE); 1910 1911 return (DDI_SUCCESS); 1912 } 1913 1914 /* 1915 * detach(9E) -- Detach a device from the system 1916 */ 1917 static int 1918 rge_detach(dev_info_t *devinfo, ddi_detach_cmd_t cmd) 1919 { 1920 rge_t *rgep; 1921 1922 RGE_GTRACE(("rge_detach($%p, %d)", (void *)devinfo, cmd)); 1923 1924 rgep = ddi_get_driver_private(devinfo); 1925 1926 switch (cmd) { 1927 default: 1928 return (DDI_FAILURE); 1929 1930 case DDI_SUSPEND: 1931 return (rge_suspend(rgep)); 1932 1933 case DDI_DETACH: 1934 break; 1935 } 1936 1937 /* 1938 * If there is any posted buffer, the driver should reject to be 1939 * detached. Need notice upper layer to release them. 1940 */ 1941 if (!(rgep->chip_flags & CHIP_FLAG_FORCE_BCOPY) && 1942 rgep->rx_free != RGE_BUF_SLOTS) 1943 return (DDI_FAILURE); 1944 1945 /* 1946 * Unregister from the MAC layer subsystem. This can fail, in 1947 * particular if there are DLPI style-2 streams still open - 1948 * in which case we just return failure without shutting 1949 * down chip operations. 1950 */ 1951 if (mac_unregister(rgep->mh) != 0) 1952 return (DDI_FAILURE); 1953 1954 /* 1955 * All activity stopped, so we can clean up & exit 1956 */ 1957 rge_unattach(rgep); 1958 return (DDI_SUCCESS); 1959 } 1960 1961 1962 /* 1963 * ========== Module Loading Data & Entry Points ========== 1964 */ 1965 1966 #undef RGE_DBG 1967 #define RGE_DBG RGE_DBG_INIT /* debug flag for this code */ 1968 DDI_DEFINE_STREAM_OPS(rge_dev_ops, nulldev, nulldev, rge_attach, rge_detach, 1969 nodev, NULL, D_MP, NULL, rge_quiesce); 1970 1971 static struct modldrv rge_modldrv = { 1972 &mod_driverops, /* Type of module. This one is a driver */ 1973 rge_ident, /* short description */ 1974 &rge_dev_ops /* driver specific ops */ 1975 }; 1976 1977 static struct modlinkage modlinkage = { 1978 MODREV_1, (void *)&rge_modldrv, NULL 1979 }; 1980 1981 1982 int 1983 _info(struct modinfo *modinfop) 1984 { 1985 return (mod_info(&modlinkage, modinfop)); 1986 } 1987 1988 int 1989 _init(void) 1990 { 1991 int status; 1992 1993 mac_init_ops(&rge_dev_ops, "rge"); 1994 status = mod_install(&modlinkage); 1995 if (status == DDI_SUCCESS) 1996 mutex_init(rge_log_mutex, NULL, MUTEX_DRIVER, NULL); 1997 else 1998 mac_fini_ops(&rge_dev_ops); 1999 2000 return (status); 2001 } 2002 2003 int 2004 _fini(void) 2005 { 2006 int status; 2007 2008 status = mod_remove(&modlinkage); 2009 if (status == DDI_SUCCESS) { 2010 mac_fini_ops(&rge_dev_ops); 2011 mutex_destroy(rge_log_mutex); 2012 } 2013 return (status); 2014 } 2015