1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright 2010 Sun Microsystems, Inc. All rights reserved. 23 * Use is subject to license terms. 24 */ 25 26 27 #include <sys/types.h> 28 #include <sys/sunddi.h> 29 #include <sys/policy.h> 30 #include <sys/sdt.h> 31 #include "dmfe_impl.h" 32 33 /* 34 * This is the string displayed by modinfo, etc. 35 */ 36 static char dmfe_ident[] = "Davicom DM9102 Ethernet"; 37 38 39 /* 40 * NOTES: 41 * 42 * #defines: 43 * 44 * DMFE_PCI_RNUMBER is the register-set number to use for the operating 45 * registers. On an OBP-based machine, regset 0 refers to CONFIG space, 46 * regset 1 will be the operating registers in I/O space, and regset 2 47 * will be the operating registers in MEMORY space (preferred). If an 48 * expansion ROM is fitted, it may appear as a further register set. 49 * 50 * DMFE_SLOP defines the amount by which the chip may read beyond 51 * the end of a buffer or descriptor, apparently 6-8 dwords :( 52 * We have to make sure this doesn't cause it to access unallocated 53 * or unmapped memory. 54 * 55 * DMFE_BUF_SIZE must be at least (ETHERMAX + ETHERFCSL + DMFE_SLOP) 56 * rounded up to a multiple of 4. Here we choose a power of two for 57 * speed & simplicity at the cost of a bit more memory. 58 * 59 * However, the buffer length field in the TX/RX descriptors is only 60 * eleven bits, so even though we allocate DMFE_BUF_SIZE (2048) bytes 61 * per buffer, we tell the chip that they're only DMFE_BUF_SIZE_1 62 * (2000) bytes each. 63 * 64 * DMFE_DMA_MODE defines the mode (STREAMING/CONSISTENT) used for 65 * the data buffers. The descriptors are always set up in CONSISTENT 66 * mode. 67 * 68 * DMFE_HEADROOM defines how much space we'll leave in allocated 69 * mblks before the first valid data byte. This should be chosen 70 * to be 2 modulo 4, so that once the ethernet header (14 bytes) 71 * has been stripped off, the packet data will be 4-byte aligned. 72 * The remaining space can be used by upstream modules to prepend 73 * any headers required. 74 * 75 * Patchable globals: 76 * 77 * dmfe_bus_modes: the bus mode bits to be put into CSR0. 78 * Setting READ_MULTIPLE in this register seems to cause 79 * the chip to generate a READ LINE command with a parity 80 * error! Don't do it! 81 * 82 * dmfe_setup_desc1: the value to be put into descriptor word 1 83 * when sending a SETUP packet. 84 * 85 * Setting TX_LAST_DESC in desc1 in a setup packet seems 86 * to make the chip spontaneously reset internally - it 87 * attempts to give back the setup packet descriptor by 88 * writing to PCI address 00000000 - which may or may not 89 * get a MASTER ABORT - after which most of its registers 90 * seem to have either default values or garbage! 91 * 92 * TX_FIRST_DESC doesn't seem to have the same effect but 93 * it isn't needed on a setup packet so we'll leave it out 94 * too, just in case it has some other wierd side-effect. 95 * 96 * The default hardware packet filtering mode is now 97 * HASH_AND_PERFECT (imperfect filtering of multicast 98 * packets and perfect filtering of unicast packets). 99 * If this is found not to work reliably, setting the 100 * TX_FILTER_TYPE1 bit will cause a switchover to using 101 * HASH_ONLY mode (imperfect filtering of *all* packets). 102 * Software will then perform the additional filtering 103 * as required. 104 */ 105 106 #define DMFE_PCI_RNUMBER 2 107 #define DMFE_SLOP (8*sizeof (uint32_t)) 108 #define DMFE_BUF_SIZE 2048 109 #define DMFE_BUF_SIZE_1 2000 110 #define DMFE_DMA_MODE DDI_DMA_STREAMING 111 #define DMFE_HEADROOM 34 112 113 static uint32_t dmfe_bus_modes = TX_POLL_INTVL | CACHE_ALIGN; 114 static uint32_t dmfe_setup_desc1 = TX_SETUP_PACKET | SETUPBUF_SIZE | 115 TX_FILTER_TYPE0; 116 117 /* 118 * Some tunable parameters ... 119 * Number of RX/TX ring entries (128/128) 120 * Minimum number of TX ring slots to keep free (1) 121 * Low-water mark at which to try to reclaim TX ring slots (1) 122 * How often to take a TX-done interrupt (twice per ring cycle) 123 * Whether to reclaim TX ring entries on a TX-done interrupt (no) 124 */ 125 126 #define DMFE_TX_DESC 128 /* Should be a multiple of 4 <= 256 */ 127 #define DMFE_RX_DESC 128 /* Should be a multiple of 4 <= 256 */ 128 129 static uint32_t dmfe_rx_desc = DMFE_RX_DESC; 130 static uint32_t dmfe_tx_desc = DMFE_TX_DESC; 131 static uint32_t dmfe_tx_min_free = 1; 132 static uint32_t dmfe_tx_reclaim_level = 1; 133 static uint32_t dmfe_tx_int_factor = (DMFE_TX_DESC / 2) - 1; 134 static boolean_t dmfe_reclaim_on_done = B_FALSE; 135 136 /* 137 * Time-related parameters: 138 * 139 * We use a cyclic to provide a periodic callback; this is then used 140 * to check for TX-stall and poll the link status register. 141 * 142 * DMFE_TICK is the interval between cyclic callbacks, in microseconds. 143 * 144 * TX_STALL_TIME_100 is the timeout in microseconds between passing 145 * a packet to the chip for transmission and seeing that it's gone, 146 * when running at 100Mb/s. If we haven't reclaimed at least one 147 * descriptor in this time we assume the transmitter has stalled 148 * and reset the chip. 149 * 150 * TX_STALL_TIME_10 is the equivalent timeout when running at 10Mb/s. 151 * 152 * Patchable globals: 153 * 154 * dmfe_tick_us: DMFE_TICK 155 * dmfe_tx100_stall_us: TX_STALL_TIME_100 156 * dmfe_tx10_stall_us: TX_STALL_TIME_10 157 * 158 * These are then used in _init() to calculate: 159 * 160 * stall_100_tix[]: number of consecutive cyclic callbacks without a 161 * reclaim before the TX process is considered stalled, 162 * when running at 100Mb/s. The elements are indexed 163 * by transmit-engine-state. 164 * stall_10_tix[]: number of consecutive cyclic callbacks without a 165 * reclaim before the TX process is considered stalled, 166 * when running at 10Mb/s. The elements are indexed 167 * by transmit-engine-state. 168 */ 169 170 #define DMFE_TICK 25000 /* microseconds */ 171 #define TX_STALL_TIME_100 50000 /* microseconds */ 172 #define TX_STALL_TIME_10 200000 /* microseconds */ 173 174 static uint32_t dmfe_tick_us = DMFE_TICK; 175 static uint32_t dmfe_tx100_stall_us = TX_STALL_TIME_100; 176 static uint32_t dmfe_tx10_stall_us = TX_STALL_TIME_10; 177 178 /* 179 * Calculated from above in _init() 180 */ 181 182 static uint32_t stall_100_tix[TX_PROCESS_MAX_STATE+1]; 183 static uint32_t stall_10_tix[TX_PROCESS_MAX_STATE+1]; 184 185 /* 186 * Property names 187 */ 188 static char localmac_propname[] = "local-mac-address"; 189 static char opmode_propname[] = "opmode-reg-value"; 190 191 static int dmfe_m_start(void *); 192 static void dmfe_m_stop(void *); 193 static int dmfe_m_promisc(void *, boolean_t); 194 static int dmfe_m_multicst(void *, boolean_t, const uint8_t *); 195 static int dmfe_m_unicst(void *, const uint8_t *); 196 static void dmfe_m_ioctl(void *, queue_t *, mblk_t *); 197 static mblk_t *dmfe_m_tx(void *, mblk_t *); 198 static int dmfe_m_stat(void *, uint_t, uint64_t *); 199 static int dmfe_m_getprop(void *, const char *, mac_prop_id_t, 200 uint_t, void *); 201 static int dmfe_m_setprop(void *, const char *, mac_prop_id_t, 202 uint_t, const void *); 203 static void dmfe_m_propinfo(void *, const char *, mac_prop_id_t, 204 mac_prop_info_handle_t); 205 206 static mac_callbacks_t dmfe_m_callbacks = { 207 MC_IOCTL | MC_SETPROP | MC_GETPROP | MC_PROPINFO, 208 dmfe_m_stat, 209 dmfe_m_start, 210 dmfe_m_stop, 211 dmfe_m_promisc, 212 dmfe_m_multicst, 213 dmfe_m_unicst, 214 dmfe_m_tx, 215 NULL, 216 dmfe_m_ioctl, 217 NULL, /* getcapab */ 218 NULL, /* open */ 219 NULL, /* close */ 220 dmfe_m_setprop, 221 dmfe_m_getprop, 222 dmfe_m_propinfo 223 }; 224 225 226 /* 227 * Describes the chip's DMA engine 228 */ 229 static ddi_dma_attr_t dma_attr = { 230 DMA_ATTR_V0, /* dma_attr version */ 231 0, /* dma_attr_addr_lo */ 232 (uint32_t)0xFFFFFFFF, /* dma_attr_addr_hi */ 233 0x0FFFFFF, /* dma_attr_count_max */ 234 0x20, /* dma_attr_align */ 235 0x7F, /* dma_attr_burstsizes */ 236 1, /* dma_attr_minxfer */ 237 (uint32_t)0xFFFFFFFF, /* dma_attr_maxxfer */ 238 (uint32_t)0xFFFFFFFF, /* dma_attr_seg */ 239 1, /* dma_attr_sgllen */ 240 1, /* dma_attr_granular */ 241 0 /* dma_attr_flags */ 242 }; 243 244 /* 245 * DMA access attributes for registers and descriptors 246 */ 247 static ddi_device_acc_attr_t dmfe_reg_accattr = { 248 DDI_DEVICE_ATTR_V0, 249 DDI_STRUCTURE_LE_ACC, 250 DDI_STRICTORDER_ACC 251 }; 252 253 /* 254 * DMA access attributes for data: NOT to be byte swapped. 255 */ 256 static ddi_device_acc_attr_t dmfe_data_accattr = { 257 DDI_DEVICE_ATTR_V0, 258 DDI_NEVERSWAP_ACC, 259 DDI_STRICTORDER_ACC 260 }; 261 262 static uchar_t dmfe_broadcast_addr[ETHERADDRL] = { 263 0xff, 0xff, 0xff, 0xff, 0xff, 0xff 264 }; 265 266 267 /* 268 * ========== Lowest-level chip register & ring access routines ========== 269 */ 270 271 /* 272 * I/O register get/put routines 273 */ 274 uint32_t 275 dmfe_chip_get32(dmfe_t *dmfep, off_t offset) 276 { 277 uint32_t *addr; 278 279 addr = (void *)(dmfep->io_reg + offset); 280 return (ddi_get32(dmfep->io_handle, addr)); 281 } 282 283 void 284 dmfe_chip_put32(dmfe_t *dmfep, off_t offset, uint32_t value) 285 { 286 uint32_t *addr; 287 288 addr = (void *)(dmfep->io_reg + offset); 289 ddi_put32(dmfep->io_handle, addr, value); 290 } 291 292 /* 293 * TX/RX ring get/put routines 294 */ 295 static uint32_t 296 dmfe_ring_get32(dma_area_t *dma_p, uint_t index, uint_t offset) 297 { 298 uint32_t *addr; 299 300 addr = (void *)dma_p->mem_va; 301 return (ddi_get32(dma_p->acc_hdl, addr + index*DESC_SIZE + offset)); 302 } 303 304 static void 305 dmfe_ring_put32(dma_area_t *dma_p, uint_t index, uint_t offset, uint32_t value) 306 { 307 uint32_t *addr; 308 309 addr = (void *)dma_p->mem_va; 310 ddi_put32(dma_p->acc_hdl, addr + index*DESC_SIZE + offset, value); 311 } 312 313 /* 314 * Setup buffer get/put routines 315 */ 316 static uint32_t 317 dmfe_setup_get32(dma_area_t *dma_p, uint_t index) 318 { 319 uint32_t *addr; 320 321 addr = (void *)dma_p->setup_va; 322 return (ddi_get32(dma_p->acc_hdl, addr + index)); 323 } 324 325 static void 326 dmfe_setup_put32(dma_area_t *dma_p, uint_t index, uint32_t value) 327 { 328 uint32_t *addr; 329 330 addr = (void *)dma_p->setup_va; 331 ddi_put32(dma_p->acc_hdl, addr + index, value); 332 } 333 334 335 /* 336 * ========== Low-level chip & ring buffer manipulation ========== 337 */ 338 339 /* 340 * dmfe_set_opmode() -- function to set operating mode 341 */ 342 static void 343 dmfe_set_opmode(dmfe_t *dmfep) 344 { 345 ASSERT(mutex_owned(dmfep->oplock)); 346 347 dmfe_chip_put32(dmfep, OPN_MODE_REG, dmfep->opmode); 348 drv_usecwait(10); 349 } 350 351 /* 352 * dmfe_stop_chip() -- stop all chip processing & optionally reset the h/w 353 */ 354 static void 355 dmfe_stop_chip(dmfe_t *dmfep, enum chip_state newstate) 356 { 357 ASSERT(mutex_owned(dmfep->oplock)); 358 359 /* 360 * Stop the chip: 361 * disable all interrupts 362 * stop TX/RX processes 363 * clear the status bits for TX/RX stopped 364 * If required, reset the chip 365 * Record the new state 366 */ 367 dmfe_chip_put32(dmfep, INT_MASK_REG, 0); 368 dmfep->opmode &= ~(START_TRANSMIT | START_RECEIVE); 369 dmfe_set_opmode(dmfep); 370 dmfe_chip_put32(dmfep, STATUS_REG, TX_STOPPED_INT | RX_STOPPED_INT); 371 372 switch (newstate) { 373 default: 374 ASSERT(!"can't get here"); 375 return; 376 377 case CHIP_STOPPED: 378 case CHIP_ERROR: 379 break; 380 381 case CHIP_RESET: 382 dmfe_chip_put32(dmfep, BUS_MODE_REG, SW_RESET); 383 drv_usecwait(10); 384 dmfe_chip_put32(dmfep, BUS_MODE_REG, 0); 385 drv_usecwait(10); 386 dmfe_chip_put32(dmfep, BUS_MODE_REG, dmfe_bus_modes); 387 break; 388 } 389 390 dmfep->chip_state = newstate; 391 } 392 393 /* 394 * Initialize transmit and receive descriptor rings, and 395 * set the chip to point to the first entry in each ring 396 */ 397 static void 398 dmfe_init_rings(dmfe_t *dmfep) 399 { 400 dma_area_t *descp; 401 uint32_t pstart; 402 uint32_t pnext; 403 uint32_t pbuff; 404 uint32_t desc1; 405 int i; 406 407 /* 408 * You need all the locks in order to rewrite the descriptor rings 409 */ 410 ASSERT(mutex_owned(dmfep->oplock)); 411 ASSERT(mutex_owned(dmfep->rxlock)); 412 ASSERT(mutex_owned(dmfep->txlock)); 413 414 /* 415 * Program the RX ring entries 416 */ 417 descp = &dmfep->rx_desc; 418 pstart = descp->mem_dvma; 419 pnext = pstart + sizeof (struct rx_desc_type); 420 pbuff = dmfep->rx_buff.mem_dvma; 421 desc1 = RX_CHAINING | DMFE_BUF_SIZE_1; 422 423 for (i = 0; i < dmfep->rx.n_desc; ++i) { 424 dmfe_ring_put32(descp, i, RD_NEXT, pnext); 425 dmfe_ring_put32(descp, i, BUFFER1, pbuff); 426 dmfe_ring_put32(descp, i, DESC1, desc1); 427 dmfe_ring_put32(descp, i, DESC0, RX_OWN); 428 429 pnext += sizeof (struct rx_desc_type); 430 pbuff += DMFE_BUF_SIZE; 431 } 432 433 /* 434 * Fix up last entry & sync 435 */ 436 dmfe_ring_put32(descp, --i, RD_NEXT, pstart); 437 DMA_SYNC(descp, DDI_DMA_SYNC_FORDEV); 438 dmfep->rx.next_free = 0; 439 440 /* 441 * Set the base address of the RX descriptor list in CSR3 442 */ 443 dmfe_chip_put32(dmfep, RX_BASE_ADDR_REG, descp->mem_dvma); 444 445 /* 446 * Program the TX ring entries 447 */ 448 descp = &dmfep->tx_desc; 449 pstart = descp->mem_dvma; 450 pnext = pstart + sizeof (struct tx_desc_type); 451 pbuff = dmfep->tx_buff.mem_dvma; 452 desc1 = TX_CHAINING; 453 454 for (i = 0; i < dmfep->tx.n_desc; ++i) { 455 dmfe_ring_put32(descp, i, TD_NEXT, pnext); 456 dmfe_ring_put32(descp, i, BUFFER1, pbuff); 457 dmfe_ring_put32(descp, i, DESC1, desc1); 458 dmfe_ring_put32(descp, i, DESC0, 0); 459 460 pnext += sizeof (struct tx_desc_type); 461 pbuff += DMFE_BUF_SIZE; 462 } 463 464 /* 465 * Fix up last entry & sync 466 */ 467 dmfe_ring_put32(descp, --i, TD_NEXT, pstart); 468 DMA_SYNC(descp, DDI_DMA_SYNC_FORDEV); 469 dmfep->tx.n_free = dmfep->tx.n_desc; 470 dmfep->tx.next_free = dmfep->tx.next_busy = 0; 471 472 /* 473 * Set the base address of the TX descrptor list in CSR4 474 */ 475 dmfe_chip_put32(dmfep, TX_BASE_ADDR_REG, descp->mem_dvma); 476 } 477 478 /* 479 * dmfe_start_chip() -- start the chip transmitting and/or receiving 480 */ 481 static void 482 dmfe_start_chip(dmfe_t *dmfep, int mode) 483 { 484 ASSERT(mutex_owned(dmfep->oplock)); 485 486 dmfep->opmode |= mode; 487 dmfe_set_opmode(dmfep); 488 489 dmfe_chip_put32(dmfep, W_J_TIMER_REG, 0); 490 /* 491 * Enable VLAN length mode (allows packets to be 4 bytes Longer). 492 */ 493 dmfe_chip_put32(dmfep, W_J_TIMER_REG, VLAN_ENABLE); 494 495 /* 496 * Clear any pending process-stopped interrupts 497 */ 498 dmfe_chip_put32(dmfep, STATUS_REG, TX_STOPPED_INT | RX_STOPPED_INT); 499 dmfep->chip_state = mode & START_RECEIVE ? CHIP_TX_RX : 500 mode & START_TRANSMIT ? CHIP_TX_ONLY : CHIP_STOPPED; 501 } 502 503 /* 504 * dmfe_enable_interrupts() -- enable our favourite set of interrupts. 505 * 506 * Normal interrupts: 507 * We always enable: 508 * RX_PKTDONE_INT (packet received) 509 * TX_PKTDONE_INT (TX complete) 510 * We never enable: 511 * TX_ALLDONE_INT (next TX buffer not ready) 512 * 513 * Abnormal interrupts: 514 * We always enable: 515 * RX_STOPPED_INT 516 * TX_STOPPED_INT 517 * SYSTEM_ERR_INT 518 * RX_UNAVAIL_INT 519 * We never enable: 520 * RX_EARLY_INT 521 * RX_WATCHDOG_INT 522 * TX_JABBER_INT 523 * TX_EARLY_INT 524 * TX_UNDERFLOW_INT 525 * GP_TIMER_INT (not valid in -9 chips) 526 * LINK_STATUS_INT (not valid in -9 chips) 527 */ 528 static void 529 dmfe_enable_interrupts(dmfe_t *dmfep) 530 { 531 ASSERT(mutex_owned(dmfep->oplock)); 532 533 /* 534 * Put 'the standard set of interrupts' in the interrupt mask register 535 */ 536 dmfep->imask = RX_PKTDONE_INT | TX_PKTDONE_INT | 537 RX_STOPPED_INT | TX_STOPPED_INT | RX_UNAVAIL_INT | SYSTEM_ERR_INT; 538 539 dmfe_chip_put32(dmfep, INT_MASK_REG, 540 NORMAL_SUMMARY_INT | ABNORMAL_SUMMARY_INT | dmfep->imask); 541 dmfep->chip_state = CHIP_RUNNING; 542 } 543 544 /* 545 * ========== RX side routines ========== 546 */ 547 548 /* 549 * Function to update receive statistics on various errors 550 */ 551 static void 552 dmfe_update_rx_stats(dmfe_t *dmfep, uint32_t desc0) 553 { 554 ASSERT(mutex_owned(dmfep->rxlock)); 555 556 /* 557 * The error summary bit and the error bits that it summarises 558 * are only valid if this is the last fragment. Therefore, a 559 * fragment only contributes to the error statistics if both 560 * the last-fragment and error summary bits are set. 561 */ 562 if (((RX_LAST_DESC | RX_ERR_SUMMARY) & ~desc0) == 0) { 563 dmfep->rx_stats_ierrors += 1; 564 565 /* 566 * There are some other error bits in the descriptor for 567 * which there don't seem to be appropriate MAC statistics, 568 * notably RX_COLLISION and perhaps RX_DESC_ERR. The 569 * latter may not be possible if it is supposed to indicate 570 * that one buffer has been filled with a partial packet 571 * and the next buffer required for the rest of the packet 572 * was not available, as all our buffers are more than large 573 * enough for a whole packet without fragmenting. 574 */ 575 576 if (desc0 & RX_OVERFLOW) { 577 dmfep->rx_stats_overflow += 1; 578 579 } else if (desc0 & RX_RUNT_FRAME) 580 dmfep->rx_stats_short += 1; 581 582 if (desc0 & RX_CRC) 583 dmfep->rx_stats_fcs += 1; 584 585 if (desc0 & RX_FRAME2LONG) 586 dmfep->rx_stats_toolong += 1; 587 } 588 589 /* 590 * A receive watchdog timeout is counted as a MAC-level receive 591 * error. Strangely, it doesn't set the packet error summary bit, 592 * according to the chip data sheet :-? 593 */ 594 if (desc0 & RX_RCV_WD_TO) 595 dmfep->rx_stats_macrcv_errors += 1; 596 597 if (desc0 & RX_DRIBBLING) 598 dmfep->rx_stats_align += 1; 599 600 if (desc0 & RX_MII_ERR) 601 dmfep->rx_stats_macrcv_errors += 1; 602 } 603 604 /* 605 * Receive incoming packet(s) and pass them up ... 606 */ 607 static mblk_t * 608 dmfe_getp(dmfe_t *dmfep) 609 { 610 dma_area_t *descp; 611 mblk_t **tail; 612 mblk_t *head; 613 mblk_t *mp; 614 char *rxb; 615 uchar_t *dp; 616 uint32_t desc0; 617 uint32_t misses; 618 int packet_length; 619 int index; 620 621 mutex_enter(dmfep->rxlock); 622 623 /* 624 * Update the missed frame statistic from the on-chip counter. 625 */ 626 misses = dmfe_chip_get32(dmfep, MISSED_FRAME_REG); 627 dmfep->rx_stats_norcvbuf += (misses & MISSED_FRAME_MASK); 628 629 /* 630 * sync (all) receive descriptors before inspecting them 631 */ 632 descp = &dmfep->rx_desc; 633 DMA_SYNC(descp, DDI_DMA_SYNC_FORKERNEL); 634 635 /* 636 * We should own at least one RX entry, since we've had a 637 * receive interrupt, but let's not be dogmatic about it. 638 */ 639 index = dmfep->rx.next_free; 640 desc0 = dmfe_ring_get32(descp, index, DESC0); 641 642 DTRACE_PROBE1(rx__start, uint32_t, desc0); 643 for (head = NULL, tail = &head; (desc0 & RX_OWN) == 0; ) { 644 /* 645 * Maintain statistics for every descriptor returned 646 * to us by the chip ... 647 */ 648 dmfe_update_rx_stats(dmfep, desc0); 649 650 /* 651 * Check that the entry has both "packet start" and 652 * "packet end" flags. We really shouldn't get packet 653 * fragments, 'cos all the RX buffers are bigger than 654 * the largest valid packet. So we'll just drop any 655 * fragments we find & skip on to the next entry. 656 */ 657 if (((RX_FIRST_DESC | RX_LAST_DESC) & ~desc0) != 0) { 658 DTRACE_PROBE1(rx__frag, uint32_t, desc0); 659 goto skip; 660 } 661 662 /* 663 * A whole packet in one buffer. We have to check error 664 * status and packet length before forwarding it upstream. 665 */ 666 if (desc0 & RX_ERR_SUMMARY) { 667 DTRACE_PROBE1(rx__err, uint32_t, desc0); 668 goto skip; 669 } 670 671 packet_length = (desc0 >> 16) & 0x3fff; 672 if (packet_length > DMFE_MAX_PKT_SIZE) { 673 DTRACE_PROBE1(rx__toobig, int, packet_length); 674 goto skip; 675 } else if (packet_length < ETHERMIN) { 676 /* 677 * Note that VLAN packet would be even larger, 678 * but we don't worry about dropping runt VLAN 679 * frames. 680 * 681 * This check is probably redundant, as well, 682 * since the hardware should drop RUNT frames. 683 */ 684 DTRACE_PROBE1(rx__runt, int, packet_length); 685 goto skip; 686 } 687 688 /* 689 * Sync the data, so we can examine it; then check that 690 * the packet is really intended for us (remember that 691 * if we're using Imperfect Filtering, then the chip will 692 * receive unicast packets sent to stations whose addresses 693 * just happen to hash to the same value as our own; we 694 * discard these here so they don't get sent upstream ...) 695 */ 696 (void) ddi_dma_sync(dmfep->rx_buff.dma_hdl, 697 index * DMFE_BUF_SIZE, DMFE_BUF_SIZE, 698 DDI_DMA_SYNC_FORKERNEL); 699 rxb = &dmfep->rx_buff.mem_va[index*DMFE_BUF_SIZE]; 700 701 702 /* 703 * We do not bother to check that the packet is really for 704 * us, we let the MAC framework make that check instead. 705 * This is especially important if we ever want to support 706 * multiple MAC addresses. 707 */ 708 709 /* 710 * Packet looks good; get a buffer to copy it into. We 711 * allow some space at the front of the allocated buffer 712 * (HEADROOM) in case any upstream modules want to prepend 713 * some sort of header. The value has been carefully chosen 714 * So that it also has the side-effect of making the packet 715 * *contents* 4-byte aligned, as required by NCA! 716 */ 717 mp = allocb(DMFE_HEADROOM + packet_length, 0); 718 if (mp == NULL) { 719 DTRACE_PROBE(rx__no__buf); 720 dmfep->rx_stats_norcvbuf += 1; 721 goto skip; 722 } 723 724 /* 725 * Account for statistics of good packets. 726 */ 727 dmfep->rx_stats_ipackets += 1; 728 dmfep->rx_stats_rbytes += packet_length; 729 if (desc0 & RX_MULTI_FRAME) { 730 if (bcmp(rxb, dmfe_broadcast_addr, ETHERADDRL)) { 731 dmfep->rx_stats_multi += 1; 732 } else { 733 dmfep->rx_stats_bcast += 1; 734 } 735 } 736 737 /* 738 * Copy the packet into the STREAMS buffer 739 */ 740 dp = mp->b_rptr += DMFE_HEADROOM; 741 mp->b_cont = mp->b_next = NULL; 742 743 /* 744 * Don't worry about stripping the vlan tag, the MAC 745 * layer will take care of that for us. 746 */ 747 bcopy(rxb, dp, packet_length); 748 749 /* 750 * Fix up the packet length, and link it to the chain 751 */ 752 mp->b_wptr = mp->b_rptr + packet_length - ETHERFCSL; 753 *tail = mp; 754 tail = &mp->b_next; 755 756 skip: 757 /* 758 * Return ownership of ring entry & advance to next 759 */ 760 dmfe_ring_put32(descp, index, DESC0, RX_OWN); 761 index = NEXT(index, dmfep->rx.n_desc); 762 desc0 = dmfe_ring_get32(descp, index, DESC0); 763 } 764 765 /* 766 * Remember where to start looking next time ... 767 */ 768 dmfep->rx.next_free = index; 769 770 /* 771 * sync the receive descriptors that we've given back 772 * (actually, we sync all of them for simplicity), and 773 * wake the chip in case it had suspended receive 774 */ 775 DMA_SYNC(descp, DDI_DMA_SYNC_FORDEV); 776 dmfe_chip_put32(dmfep, RX_POLL_REG, 0); 777 778 mutex_exit(dmfep->rxlock); 779 return (head); 780 } 781 782 /* 783 * ========== Primary TX side routines ========== 784 */ 785 786 /* 787 * TX ring management: 788 * 789 * There are <tx.n_desc> entries in the ring, of which those from 790 * <tx.next_free> round to but not including <tx.next_busy> must 791 * be owned by the CPU. The number of such entries should equal 792 * <tx.n_free>; but there may also be some more entries which the 793 * chip has given back but which we haven't yet accounted for. 794 * The routine dmfe_reclaim_tx_desc() adjusts the indexes & counts 795 * as it discovers such entries. 796 * 797 * Initially, or when the ring is entirely free: 798 * C = Owned by CPU 799 * D = Owned by Davicom (DMFE) chip 800 * 801 * tx.next_free tx.n_desc = 16 802 * | 803 * v 804 * +---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+ 805 * | C | C | C | C | C | C | C | C | C | C | C | C | C | C | C | C | 806 * +---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+ 807 * ^ 808 * | 809 * tx.next_busy tx.n_free = 16 810 * 811 * On entry to reclaim() during normal use: 812 * 813 * tx.next_free tx.n_desc = 16 814 * | 815 * v 816 * +---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+ 817 * | C | C | C | C | C | C | D | D | D | C | C | C | C | C | C | C | 818 * +---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+ 819 * ^ 820 * | 821 * tx.next_busy tx.n_free = 9 822 * 823 * On exit from reclaim(): 824 * 825 * tx.next_free tx.n_desc = 16 826 * | 827 * v 828 * +---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+ 829 * | C | C | C | C | C | C | D | D | D | C | C | C | C | C | C | C | 830 * +---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+ 831 * ^ 832 * | 833 * tx.next_busy tx.n_free = 13 834 * 835 * The ring is considered "full" when only one entry is owned by 836 * the CPU; thus <tx.n_free> should always be >= 1. 837 * 838 * tx.next_free tx.n_desc = 16 839 * | 840 * v 841 * +---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+ 842 * | D | D | D | D | D | C | D | D | D | D | D | D | D | D | D | D | 843 * +---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+ 844 * ^ 845 * | 846 * tx.next_busy tx.n_free = 1 847 */ 848 849 /* 850 * Function to update transmit statistics on various errors 851 */ 852 static void 853 dmfe_update_tx_stats(dmfe_t *dmfep, int index, uint32_t desc0, uint32_t desc1) 854 { 855 uint32_t collisions; 856 uint32_t errbits; 857 uint32_t errsum; 858 859 ASSERT(mutex_owned(dmfep->txlock)); 860 861 collisions = ((desc0 >> 3) & 0x0f); 862 errsum = desc0 & TX_ERR_SUMMARY; 863 errbits = desc0 & (TX_UNDERFLOW | TX_LATE_COLL | TX_CARRIER_LOSS | 864 TX_NO_CARRIER | TX_EXCESS_COLL | TX_JABBER_TO); 865 if ((errsum == 0) != (errbits == 0)) { 866 dmfe_log(dmfep, "dubious TX error status 0x%x", desc0); 867 desc0 |= TX_ERR_SUMMARY; 868 } 869 870 if (desc0 & TX_ERR_SUMMARY) { 871 dmfep->tx_stats_oerrors += 1; 872 873 /* 874 * If we ever see a transmit jabber timeout, we count it 875 * as a MAC-level transmit error; but we probably won't 876 * see it as it causes an Abnormal interrupt and we reset 877 * the chip in order to recover 878 */ 879 if (desc0 & TX_JABBER_TO) { 880 dmfep->tx_stats_macxmt_errors += 1; 881 dmfep->tx_stats_jabber += 1; 882 } 883 884 if (desc0 & TX_UNDERFLOW) 885 dmfep->tx_stats_underflow += 1; 886 else if (desc0 & TX_LATE_COLL) 887 dmfep->tx_stats_xmtlatecoll += 1; 888 889 if (desc0 & (TX_CARRIER_LOSS | TX_NO_CARRIER)) 890 dmfep->tx_stats_nocarrier += 1; 891 892 if (desc0 & TX_EXCESS_COLL) { 893 dmfep->tx_stats_excoll += 1; 894 collisions = 16; 895 } 896 } else { 897 int bit = index % NBBY; 898 int byt = index / NBBY; 899 900 if (dmfep->tx_mcast[byt] & bit) { 901 dmfep->tx_mcast[byt] &= ~bit; 902 dmfep->tx_stats_multi += 1; 903 904 } else if (dmfep->tx_bcast[byt] & bit) { 905 dmfep->tx_bcast[byt] &= ~bit; 906 dmfep->tx_stats_bcast += 1; 907 } 908 909 dmfep->tx_stats_opackets += 1; 910 dmfep->tx_stats_obytes += desc1 & TX_BUFFER_SIZE1; 911 } 912 913 if (collisions == 1) 914 dmfep->tx_stats_first_coll += 1; 915 else if (collisions != 0) 916 dmfep->tx_stats_multi_coll += 1; 917 dmfep->tx_stats_collisions += collisions; 918 919 if (desc0 & TX_DEFERRED) 920 dmfep->tx_stats_defer += 1; 921 } 922 923 /* 924 * Reclaim all the ring entries that the chip has returned to us ... 925 * 926 * Returns B_FALSE if no entries could be reclaimed. Otherwise, reclaims 927 * as many as possible, restarts the TX stall timeout, and returns B_TRUE. 928 */ 929 static boolean_t 930 dmfe_reclaim_tx_desc(dmfe_t *dmfep) 931 { 932 dma_area_t *descp; 933 uint32_t desc0; 934 uint32_t desc1; 935 int i; 936 937 ASSERT(mutex_owned(dmfep->txlock)); 938 939 /* 940 * sync transmit descriptor ring before looking at it 941 */ 942 descp = &dmfep->tx_desc; 943 DMA_SYNC(descp, DDI_DMA_SYNC_FORKERNEL); 944 945 /* 946 * Early exit if there are no descriptors to reclaim, either 947 * because they're all reclaimed already, or because the next 948 * one is still owned by the chip ... 949 */ 950 i = dmfep->tx.next_busy; 951 if (i == dmfep->tx.next_free) 952 return (B_FALSE); 953 desc0 = dmfe_ring_get32(descp, i, DESC0); 954 if (desc0 & TX_OWN) 955 return (B_FALSE); 956 957 /* 958 * Reclaim as many descriptors as possible ... 959 */ 960 for (;;) { 961 desc1 = dmfe_ring_get32(descp, i, DESC1); 962 ASSERT((desc1 & (TX_SETUP_PACKET | TX_LAST_DESC)) != 0); 963 964 if ((desc1 & TX_SETUP_PACKET) == 0) { 965 /* 966 * Regular packet - just update stats 967 */ 968 dmfe_update_tx_stats(dmfep, i, desc0, desc1); 969 } 970 971 /* 972 * Update count & index; we're all done if the ring is 973 * now fully reclaimed, or the next entry if still owned 974 * by the chip ... 975 */ 976 dmfep->tx.n_free += 1; 977 i = NEXT(i, dmfep->tx.n_desc); 978 if (i == dmfep->tx.next_free) 979 break; 980 desc0 = dmfe_ring_get32(descp, i, DESC0); 981 if (desc0 & TX_OWN) 982 break; 983 } 984 985 dmfep->tx.next_busy = i; 986 dmfep->tx_pending_tix = 0; 987 return (B_TRUE); 988 } 989 990 /* 991 * Send the message in the message block chain <mp>. 992 * 993 * The message is freed if and only if its contents are successfully copied 994 * and queued for transmission (so that the return value is B_TRUE). 995 * If we can't queue the message, the return value is B_FALSE and 996 * the message is *not* freed. 997 * 998 * This routine handles the special case of <mp> == NULL, which indicates 999 * that we want to "send" the special "setup packet" allocated during 1000 * startup. We have to use some different flags in the packet descriptor 1001 * to say its a setup packet (from the global <dmfe_setup_desc1>), and the 1002 * setup packet *isn't* freed after use. 1003 */ 1004 static boolean_t 1005 dmfe_send_msg(dmfe_t *dmfep, mblk_t *mp) 1006 { 1007 dma_area_t *descp; 1008 mblk_t *bp; 1009 char *txb; 1010 uint32_t desc1; 1011 uint32_t index; 1012 size_t totlen; 1013 size_t mblen; 1014 uint32_t paddr; 1015 1016 /* 1017 * If the number of free slots is below the reclaim threshold 1018 * (soft limit), we'll try to reclaim some. If we fail, and 1019 * the number of free slots is also below the minimum required 1020 * (the hard limit, usually 1), then we can't send the packet. 1021 */ 1022 mutex_enter(dmfep->txlock); 1023 if (dmfep->suspended) 1024 return (B_FALSE); 1025 1026 if (dmfep->tx.n_free <= dmfe_tx_reclaim_level && 1027 dmfe_reclaim_tx_desc(dmfep) == B_FALSE && 1028 dmfep->tx.n_free <= dmfe_tx_min_free) { 1029 /* 1030 * Resource shortage - return B_FALSE so the packet 1031 * will be queued for retry after the next TX-done 1032 * interrupt. 1033 */ 1034 mutex_exit(dmfep->txlock); 1035 DTRACE_PROBE(tx__no__desc); 1036 return (B_FALSE); 1037 } 1038 1039 /* 1040 * There's a slot available, so claim it by incrementing 1041 * the next-free index and decrementing the free count. 1042 * If the ring is currently empty, we also restart the 1043 * stall-detect timer. The ASSERTions check that our 1044 * invariants still hold: 1045 * the next-free index must not match the next-busy index 1046 * there must still be at least one free entry 1047 * After this, we now have exclusive ownership of the ring 1048 * entry (and matching buffer) indicated by <index>, so we 1049 * don't need to hold the TX lock any longer 1050 */ 1051 index = dmfep->tx.next_free; 1052 dmfep->tx.next_free = NEXT(index, dmfep->tx.n_desc); 1053 ASSERT(dmfep->tx.next_free != dmfep->tx.next_busy); 1054 if (dmfep->tx.n_free-- == dmfep->tx.n_desc) 1055 dmfep->tx_pending_tix = 0; 1056 ASSERT(dmfep->tx.n_free >= 1); 1057 mutex_exit(dmfep->txlock); 1058 1059 /* 1060 * Check the ownership of the ring entry ... 1061 */ 1062 descp = &dmfep->tx_desc; 1063 ASSERT((dmfe_ring_get32(descp, index, DESC0) & TX_OWN) == 0); 1064 1065 if (mp == NULL) { 1066 /* 1067 * Indicates we should send a SETUP packet, which we do by 1068 * temporarily switching the BUFFER1 pointer in the ring 1069 * entry. The reclaim routine will restore BUFFER1 to its 1070 * usual value. 1071 * 1072 * Note that as the setup packet is tagged on the end of 1073 * the TX ring, when we sync the descriptor we're also 1074 * implicitly syncing the setup packet - hence, we don't 1075 * need a separate ddi_dma_sync() call here. 1076 */ 1077 desc1 = dmfe_setup_desc1; 1078 paddr = descp->setup_dvma; 1079 } else { 1080 /* 1081 * A regular packet; we copy the data into a pre-mapped 1082 * buffer, which avoids the overhead (and complication) 1083 * of mapping/unmapping STREAMS buffers and keeping hold 1084 * of them until the DMA has completed. 1085 * 1086 * Because all buffers are the same size, and larger 1087 * than the longest single valid message, we don't have 1088 * to bother about splitting the message across multiple 1089 * buffers. 1090 */ 1091 txb = &dmfep->tx_buff.mem_va[index*DMFE_BUF_SIZE]; 1092 totlen = 0; 1093 bp = mp; 1094 1095 /* 1096 * Copy all (remaining) mblks in the message ... 1097 */ 1098 for (; bp != NULL; bp = bp->b_cont) { 1099 mblen = MBLKL(bp); 1100 if ((totlen += mblen) <= DMFE_MAX_PKT_SIZE) { 1101 bcopy(bp->b_rptr, txb, mblen); 1102 txb += mblen; 1103 } 1104 } 1105 1106 /* 1107 * Is this a multicast or broadcast packet? We do 1108 * this so that we can track statistics accurately 1109 * when we reclaim it. 1110 */ 1111 txb = &dmfep->tx_buff.mem_va[index*DMFE_BUF_SIZE]; 1112 if (txb[0] & 0x1) { 1113 if (bcmp(txb, dmfe_broadcast_addr, ETHERADDRL) == 0) { 1114 dmfep->tx_bcast[index / NBBY] |= 1115 (1 << (index % NBBY)); 1116 } else { 1117 dmfep->tx_mcast[index / NBBY] |= 1118 (1 << (index % NBBY)); 1119 } 1120 } 1121 1122 /* 1123 * We'e reached the end of the chain; and we should have 1124 * collected no more than DMFE_MAX_PKT_SIZE bytes into our 1125 * buffer. Note that the <size> field in the descriptor is 1126 * only 11 bits, so bigger packets would be a problem! 1127 */ 1128 ASSERT(bp == NULL); 1129 ASSERT(totlen <= DMFE_MAX_PKT_SIZE); 1130 totlen &= TX_BUFFER_SIZE1; 1131 desc1 = TX_FIRST_DESC | TX_LAST_DESC | totlen; 1132 paddr = dmfep->tx_buff.mem_dvma + index*DMFE_BUF_SIZE; 1133 1134 (void) ddi_dma_sync(dmfep->tx_buff.dma_hdl, 1135 index * DMFE_BUF_SIZE, DMFE_BUF_SIZE, DDI_DMA_SYNC_FORDEV); 1136 } 1137 1138 /* 1139 * Update ring descriptor entries, sync them, and wake up the 1140 * transmit process 1141 */ 1142 if ((index & dmfe_tx_int_factor) == 0) 1143 desc1 |= TX_INT_ON_COMP; 1144 desc1 |= TX_CHAINING; 1145 dmfe_ring_put32(descp, index, BUFFER1, paddr); 1146 dmfe_ring_put32(descp, index, DESC1, desc1); 1147 dmfe_ring_put32(descp, index, DESC0, TX_OWN); 1148 DMA_SYNC(descp, DDI_DMA_SYNC_FORDEV); 1149 dmfe_chip_put32(dmfep, TX_POLL_REG, 0); 1150 1151 /* 1152 * Finally, free the message & return success 1153 */ 1154 if (mp) 1155 freemsg(mp); 1156 return (B_TRUE); 1157 } 1158 1159 /* 1160 * dmfe_m_tx() -- send a chain of packets 1161 * 1162 * Called when packet(s) are ready to be transmitted. A pointer to an 1163 * M_DATA message that contains the packet is passed to this routine. 1164 * The complete LLC header is contained in the message's first message 1165 * block, and the remainder of the packet is contained within 1166 * additional M_DATA message blocks linked to the first message block. 1167 * 1168 * Additional messages may be passed by linking with b_next. 1169 */ 1170 static mblk_t * 1171 dmfe_m_tx(void *arg, mblk_t *mp) 1172 { 1173 dmfe_t *dmfep = arg; /* private device info */ 1174 mblk_t *next; 1175 1176 ASSERT(mp != NULL); 1177 ASSERT(dmfep->mac_state == DMFE_MAC_STARTED); 1178 1179 if (dmfep->chip_state != CHIP_RUNNING) 1180 return (mp); 1181 1182 while (mp != NULL) { 1183 next = mp->b_next; 1184 mp->b_next = NULL; 1185 if (!dmfe_send_msg(dmfep, mp)) { 1186 mp->b_next = next; 1187 break; 1188 } 1189 mp = next; 1190 } 1191 1192 return (mp); 1193 } 1194 1195 /* 1196 * ========== Address-setting routines (TX-side) ========== 1197 */ 1198 1199 /* 1200 * Find the index of the relevant bit in the setup packet. 1201 * This must mirror the way the hardware will actually calculate it! 1202 */ 1203 static uint32_t 1204 dmfe_hash_index(const uint8_t *address) 1205 { 1206 uint32_t const POLY = HASH_POLY; 1207 uint32_t crc = HASH_CRC; 1208 uint32_t index; 1209 uint32_t msb; 1210 uchar_t currentbyte; 1211 int byteslength; 1212 int shift; 1213 int bit; 1214 1215 for (byteslength = 0; byteslength < ETHERADDRL; ++byteslength) { 1216 currentbyte = address[byteslength]; 1217 for (bit = 0; bit < 8; ++bit) { 1218 msb = crc >> 31; 1219 crc <<= 1; 1220 if (msb ^ (currentbyte & 1)) { 1221 crc ^= POLY; 1222 crc |= 0x00000001; 1223 } 1224 currentbyte >>= 1; 1225 } 1226 } 1227 1228 for (index = 0, bit = 23, shift = 8; shift >= 0; ++bit, --shift) 1229 index |= (((crc >> bit) & 1) << shift); 1230 1231 return (index); 1232 } 1233 1234 /* 1235 * Find and set/clear the relevant bit in the setup packet hash table 1236 * This must mirror the way the hardware will actually interpret it! 1237 */ 1238 static void 1239 dmfe_update_hash(dmfe_t *dmfep, uint32_t index, boolean_t val) 1240 { 1241 dma_area_t *descp; 1242 uint32_t tmp; 1243 1244 ASSERT(mutex_owned(dmfep->oplock)); 1245 1246 descp = &dmfep->tx_desc; 1247 tmp = dmfe_setup_get32(descp, index/16); 1248 if (val) 1249 tmp |= 1 << (index%16); 1250 else 1251 tmp &= ~(1 << (index%16)); 1252 dmfe_setup_put32(descp, index/16, tmp); 1253 } 1254 1255 /* 1256 * Update the refcount for the bit in the setup packet corresponding 1257 * to the specified address; if it changes between zero & nonzero, 1258 * also update the bitmap itself & return B_TRUE, so that the caller 1259 * knows to re-send the setup packet. Otherwise (only the refcount 1260 * changed), return B_FALSE 1261 */ 1262 static boolean_t 1263 dmfe_update_mcast(dmfe_t *dmfep, const uint8_t *mca, boolean_t val) 1264 { 1265 uint32_t index; 1266 uint8_t *refp; 1267 boolean_t change; 1268 1269 index = dmfe_hash_index(mca); 1270 refp = &dmfep->mcast_refs[index]; 1271 change = (val ? (*refp)++ : --(*refp)) == 0; 1272 1273 if (change) 1274 dmfe_update_hash(dmfep, index, val); 1275 1276 return (change); 1277 } 1278 1279 /* 1280 * "Transmit" the (possibly updated) magic setup packet 1281 */ 1282 static int 1283 dmfe_send_setup(dmfe_t *dmfep) 1284 { 1285 int status; 1286 1287 ASSERT(mutex_owned(dmfep->oplock)); 1288 1289 if (dmfep->suspended) 1290 return (0); 1291 1292 /* 1293 * If the chip isn't running, we can't really send the setup frame 1294 * now but it doesn't matter, 'cos it will be sent when the transmit 1295 * process is restarted (see dmfe_start()). 1296 */ 1297 if ((dmfep->opmode & START_TRANSMIT) == 0) 1298 return (0); 1299 1300 /* 1301 * "Send" the setup frame. If it fails (e.g. no resources), 1302 * set a flag; then the factotum will retry the "send". Once 1303 * it works, we can clear the flag no matter how many attempts 1304 * had previously failed. We tell the caller that it worked 1305 * whether it did or not; after all, it *will* work eventually. 1306 */ 1307 status = dmfe_send_msg(dmfep, NULL); 1308 dmfep->need_setup = status ? B_FALSE : B_TRUE; 1309 return (0); 1310 } 1311 1312 /* 1313 * dmfe_m_unicst() -- set the physical network address 1314 */ 1315 static int 1316 dmfe_m_unicst(void *arg, const uint8_t *macaddr) 1317 { 1318 dmfe_t *dmfep = arg; 1319 int status; 1320 int index; 1321 1322 /* 1323 * Update our current address and send out a new setup packet 1324 * 1325 * Here we accommodate the use of HASH_ONLY or HASH_AND_PERFECT 1326 * filtering modes (we don't support PERFECT_ONLY or INVERSE modes). 1327 * 1328 * It is said that there is a bug in the 21140 where it fails to 1329 * receive packes addresses to the specified perfect filter address. 1330 * If the same bug is present in the DM9102A, the TX_FILTER_TYPE1 1331 * bit should be set in the module variable dmfe_setup_desc1. 1332 * 1333 * If TX_FILTER_TYPE1 is set, we will use HASH_ONLY filtering. 1334 * In this mode, *all* incoming addresses are hashed and looked 1335 * up in the bitmap described by the setup packet. Therefore, 1336 * the bit representing the station address has to be added to 1337 * the table before sending it out. If the address is changed, 1338 * the old entry should be removed before the new entry is made. 1339 * 1340 * NOTE: in this mode, unicast packets that are not intended for 1341 * this station may be received; it is up to software to filter 1342 * them out afterwards! 1343 * 1344 * If TX_FILTER_TYPE1 is *not* set, we will use HASH_AND_PERFECT 1345 * filtering. In this mode, multicast addresses are hashed and 1346 * checked against the bitmap, while unicast addresses are simply 1347 * matched against the one physical address specified in the setup 1348 * packet. This means that we shouldn't receive unicast packets 1349 * that aren't intended for us (but software still has to filter 1350 * multicast packets just the same). 1351 * 1352 * Whichever mode we're using, we have to enter the broadcast 1353 * address into the multicast filter map too, so we do this on 1354 * the first time through after attach or reset. 1355 */ 1356 mutex_enter(dmfep->oplock); 1357 1358 if (dmfep->addr_set && dmfe_setup_desc1 & TX_FILTER_TYPE1) 1359 (void) dmfe_update_mcast(dmfep, dmfep->curr_addr, B_FALSE); 1360 if (dmfe_setup_desc1 & TX_FILTER_TYPE1) 1361 (void) dmfe_update_mcast(dmfep, macaddr, B_TRUE); 1362 if (!dmfep->addr_set) 1363 (void) dmfe_update_mcast(dmfep, dmfe_broadcast_addr, B_TRUE); 1364 1365 /* 1366 * Remember the new current address 1367 */ 1368 ethaddr_copy(macaddr, dmfep->curr_addr); 1369 dmfep->addr_set = B_TRUE; 1370 1371 /* 1372 * Install the new physical address into the proper position in 1373 * the setup frame; this is only used if we select hash+perfect 1374 * filtering, but we'll put it in anyway. The ugliness here is 1375 * down to the usual war of the egg :( 1376 */ 1377 for (index = 0; index < ETHERADDRL; index += 2) 1378 dmfe_setup_put32(&dmfep->tx_desc, SETUPBUF_PHYS+index/2, 1379 (macaddr[index+1] << 8) | macaddr[index]); 1380 1381 /* 1382 * Finally, we're ready to "transmit" the setup frame 1383 */ 1384 status = dmfe_send_setup(dmfep); 1385 mutex_exit(dmfep->oplock); 1386 1387 return (status); 1388 } 1389 1390 /* 1391 * dmfe_m_multicst() -- enable or disable a multicast address 1392 * 1393 * Program the hardware to enable/disable the multicast address 1394 * in "mca" (enable if add is true, otherwise disable it.) 1395 * We keep a refcount for each bit in the map, so that it still 1396 * works out properly if multiple addresses hash to the same bit. 1397 * dmfe_update_mcast() tells us whether the map actually changed; 1398 * if so, we have to re-"transmit" the magic setup packet. 1399 */ 1400 static int 1401 dmfe_m_multicst(void *arg, boolean_t add, const uint8_t *mca) 1402 { 1403 dmfe_t *dmfep = arg; /* private device info */ 1404 int status = 0; 1405 1406 mutex_enter(dmfep->oplock); 1407 if (dmfe_update_mcast(dmfep, mca, add)) 1408 status = dmfe_send_setup(dmfep); 1409 mutex_exit(dmfep->oplock); 1410 1411 return (status); 1412 } 1413 1414 1415 /* 1416 * ========== Internal state management entry points ========== 1417 */ 1418 1419 /* 1420 * These routines provide all the functionality required by the 1421 * corresponding MAC layer entry points, but don't update the MAC layer state 1422 * so they can be called internally without disturbing our record 1423 * of what MAC layer thinks we should be doing ... 1424 */ 1425 1426 /* 1427 * dmfe_stop() -- stop processing, don't reset h/w or rings 1428 */ 1429 static void 1430 dmfe_stop(dmfe_t *dmfep) 1431 { 1432 ASSERT(mutex_owned(dmfep->oplock)); 1433 1434 dmfe_stop_chip(dmfep, CHIP_STOPPED); 1435 } 1436 1437 /* 1438 * dmfe_reset() -- stop processing, reset h/w & rings to initial state 1439 */ 1440 static void 1441 dmfe_reset(dmfe_t *dmfep) 1442 { 1443 ASSERT(mutex_owned(dmfep->oplock)); 1444 ASSERT(mutex_owned(dmfep->rxlock)); 1445 ASSERT(mutex_owned(dmfep->txlock)); 1446 1447 dmfe_stop_chip(dmfep, CHIP_RESET); 1448 dmfe_init_rings(dmfep); 1449 } 1450 1451 /* 1452 * dmfe_start() -- start transmitting/receiving 1453 */ 1454 static void 1455 dmfe_start(dmfe_t *dmfep) 1456 { 1457 uint32_t gpsr; 1458 1459 ASSERT(mutex_owned(dmfep->oplock)); 1460 1461 ASSERT(dmfep->chip_state == CHIP_RESET || 1462 dmfep->chip_state == CHIP_STOPPED); 1463 1464 /* 1465 * Make opmode consistent with PHY duplex setting 1466 */ 1467 gpsr = dmfe_chip_get32(dmfep, PHY_STATUS_REG); 1468 if (gpsr & GPS_FULL_DUPLEX) 1469 dmfep->opmode |= FULL_DUPLEX; 1470 else 1471 dmfep->opmode &= ~FULL_DUPLEX; 1472 1473 /* 1474 * Start transmit processing 1475 * Set up the address filters 1476 * Start receive processing 1477 * Enable interrupts 1478 */ 1479 dmfe_start_chip(dmfep, START_TRANSMIT); 1480 (void) dmfe_send_setup(dmfep); 1481 drv_usecwait(10); 1482 dmfe_start_chip(dmfep, START_RECEIVE); 1483 dmfe_enable_interrupts(dmfep); 1484 } 1485 1486 /* 1487 * dmfe_restart - restart transmitting/receiving after error or suspend 1488 */ 1489 static void 1490 dmfe_restart(dmfe_t *dmfep) 1491 { 1492 ASSERT(mutex_owned(dmfep->oplock)); 1493 1494 /* 1495 * You need not only <oplock>, but also <rxlock> AND <txlock> 1496 * in order to reset the rings, but then <txlock> *mustn't* 1497 * be held across the call to dmfe_start() 1498 */ 1499 mutex_enter(dmfep->rxlock); 1500 mutex_enter(dmfep->txlock); 1501 dmfe_reset(dmfep); 1502 mutex_exit(dmfep->txlock); 1503 mutex_exit(dmfep->rxlock); 1504 if (dmfep->mac_state == DMFE_MAC_STARTED) { 1505 dmfe_start(dmfep); 1506 } 1507 } 1508 1509 1510 /* 1511 * ========== MAC-required management entry points ========== 1512 */ 1513 1514 /* 1515 * dmfe_m_stop() -- stop transmitting/receiving 1516 */ 1517 static void 1518 dmfe_m_stop(void *arg) 1519 { 1520 dmfe_t *dmfep = arg; /* private device info */ 1521 1522 /* 1523 * Just stop processing, then record new MAC state 1524 */ 1525 mii_stop(dmfep->mii); 1526 1527 mutex_enter(dmfep->oplock); 1528 if (!dmfep->suspended) 1529 dmfe_stop(dmfep); 1530 dmfep->mac_state = DMFE_MAC_STOPPED; 1531 mutex_exit(dmfep->oplock); 1532 } 1533 1534 /* 1535 * dmfe_m_start() -- start transmitting/receiving 1536 */ 1537 static int 1538 dmfe_m_start(void *arg) 1539 { 1540 dmfe_t *dmfep = arg; /* private device info */ 1541 1542 /* 1543 * Start processing and record new MAC state 1544 */ 1545 mutex_enter(dmfep->oplock); 1546 if (!dmfep->suspended) 1547 dmfe_start(dmfep); 1548 dmfep->mac_state = DMFE_MAC_STARTED; 1549 mutex_exit(dmfep->oplock); 1550 1551 mii_start(dmfep->mii); 1552 1553 return (0); 1554 } 1555 1556 /* 1557 * dmfe_m_promisc() -- set or reset promiscuous mode on the board 1558 * 1559 * Program the hardware to enable/disable promiscuous and/or 1560 * receive-all-multicast modes. Davicom don't document this 1561 * clearly, but it looks like we can do this on-the-fly (i.e. 1562 * without stopping & restarting the TX/RX processes). 1563 */ 1564 static int 1565 dmfe_m_promisc(void *arg, boolean_t on) 1566 { 1567 dmfe_t *dmfep = arg; 1568 1569 mutex_enter(dmfep->oplock); 1570 dmfep->opmode &= ~(PROMISC_MODE | PASS_MULTICAST); 1571 if (on) 1572 dmfep->opmode |= PROMISC_MODE; 1573 if (!dmfep->suspended) 1574 dmfe_set_opmode(dmfep); 1575 mutex_exit(dmfep->oplock); 1576 1577 return (0); 1578 } 1579 1580 /* 1581 * ========== Factotum, implemented as a softint handler ========== 1582 */ 1583 1584 /* 1585 * The factotum is woken up when there's something to do that we'd rather 1586 * not do from inside a (high-level?) hardware interrupt handler. Its 1587 * two main tasks are: 1588 * reset & restart the chip after an error 1589 * update & restart the chip after a link status change 1590 */ 1591 static uint_t 1592 dmfe_factotum(caddr_t arg) 1593 { 1594 dmfe_t *dmfep; 1595 1596 dmfep = (void *)arg; 1597 ASSERT(dmfep->dmfe_guard == DMFE_GUARD); 1598 1599 mutex_enter(dmfep->oplock); 1600 if (dmfep->suspended) { 1601 mutex_exit(dmfep->oplock); 1602 return (DDI_INTR_CLAIMED); 1603 } 1604 1605 dmfep->factotum_flag = 0; 1606 DRV_KS_INC(dmfep, KS_FACTOTUM_RUN); 1607 1608 /* 1609 * Check for chip error ... 1610 */ 1611 if (dmfep->chip_state == CHIP_ERROR) { 1612 /* 1613 * Error recovery required: reset the chip and the rings, 1614 * then, if it's supposed to be running, kick it off again. 1615 */ 1616 DRV_KS_INC(dmfep, KS_RECOVERY); 1617 dmfe_restart(dmfep); 1618 mutex_exit(dmfep->oplock); 1619 1620 mii_reset(dmfep->mii); 1621 1622 } else if (dmfep->need_setup) { 1623 (void) dmfe_send_setup(dmfep); 1624 mutex_exit(dmfep->oplock); 1625 } 1626 1627 return (DDI_INTR_CLAIMED); 1628 } 1629 1630 static void 1631 dmfe_wake_factotum(dmfe_t *dmfep, int ks_id, const char *why) 1632 { 1633 _NOTE(ARGUNUSED(why)); 1634 ASSERT(mutex_owned(dmfep->oplock)); 1635 DRV_KS_INC(dmfep, ks_id); 1636 1637 if (dmfep->factotum_flag++ == 0) 1638 ddi_trigger_softintr(dmfep->factotum_id); 1639 } 1640 1641 1642 /* 1643 * ========== Periodic Tasks (Cyclic handler & friends) ========== 1644 */ 1645 1646 /* 1647 * Periodic tick tasks, run from the cyclic handler 1648 * 1649 * Check for TX stall; flag an error and wake the factotum if so. 1650 */ 1651 static void 1652 dmfe_tick_stall_check(dmfe_t *dmfep, uint32_t gpsr, uint32_t istat) 1653 { 1654 boolean_t tx_stall; 1655 uint32_t tx_state; 1656 uint32_t limit; 1657 1658 ASSERT(mutex_owned(dmfep->oplock)); 1659 1660 /* 1661 * Check for transmit stall ... 1662 * 1663 * IF there's at least one packet in the ring, AND the timeout 1664 * has elapsed, AND we can't reclaim any descriptors, THEN we've 1665 * stalled; we return B_TRUE to trigger a reset-and-recover cycle. 1666 * 1667 * Note that the timeout limit is based on the transmit engine 1668 * state; we allow the transmitter longer to make progress in 1669 * some states than in others, based on observations of this 1670 * chip's actual behaviour in the lab. 1671 * 1672 * By observation, we find that on about 1 in 10000 passes through 1673 * here, the TX lock is already held. In that case, we'll skip 1674 * the check on this pass rather than wait. Most likely, the send 1675 * routine was holding the lock when the interrupt happened, and 1676 * we'll succeed next time through. In the event of a real stall, 1677 * the TX ring will fill up, after which the send routine won't be 1678 * called any more and then we're sure to get in. 1679 */ 1680 tx_stall = B_FALSE; 1681 if (mutex_tryenter(dmfep->txlock)) { 1682 if (dmfep->tx.n_free < dmfep->tx.n_desc) { 1683 tx_state = TX_PROCESS_STATE(istat); 1684 if (gpsr & GPS_LINK_100) 1685 limit = stall_100_tix[tx_state]; 1686 else 1687 limit = stall_10_tix[tx_state]; 1688 if (++dmfep->tx_pending_tix >= limit && 1689 dmfe_reclaim_tx_desc(dmfep) == B_FALSE) { 1690 dmfe_log(dmfep, "TX stall detected " 1691 "after %d ticks in state %d; " 1692 "automatic recovery initiated", 1693 dmfep->tx_pending_tix, tx_state); 1694 tx_stall = B_TRUE; 1695 } 1696 } 1697 mutex_exit(dmfep->txlock); 1698 } 1699 1700 if (tx_stall) { 1701 dmfe_stop_chip(dmfep, CHIP_ERROR); 1702 dmfe_wake_factotum(dmfep, KS_TX_STALL, "tick (TX stall)"); 1703 } 1704 } 1705 1706 /* 1707 * Cyclic callback handler 1708 */ 1709 static void 1710 dmfe_cyclic(void *arg) 1711 { 1712 dmfe_t *dmfep = arg; /* private device info */ 1713 uint32_t istat; 1714 uint32_t gpsr; 1715 1716 /* 1717 * If the chip's not RUNNING, there's nothing to do. 1718 * If we can't get the mutex straight away, we'll just 1719 * skip this pass; we'll back back soon enough anyway. 1720 */ 1721 if (mutex_tryenter(dmfep->oplock) == 0) 1722 return; 1723 if ((dmfep->suspended) || (dmfep->chip_state != CHIP_RUNNING)) { 1724 mutex_exit(dmfep->oplock); 1725 return; 1726 } 1727 1728 /* 1729 * Recheck chip state (it might have been stopped since we 1730 * checked above). If still running, call each of the *tick* 1731 * tasks. They will check for link change, TX stall, etc ... 1732 */ 1733 if (dmfep->chip_state == CHIP_RUNNING) { 1734 istat = dmfe_chip_get32(dmfep, STATUS_REG); 1735 gpsr = dmfe_chip_get32(dmfep, PHY_STATUS_REG); 1736 dmfe_tick_stall_check(dmfep, gpsr, istat); 1737 } 1738 1739 DRV_KS_INC(dmfep, KS_CYCLIC_RUN); 1740 mutex_exit(dmfep->oplock); 1741 } 1742 1743 /* 1744 * ========== Hardware interrupt handler ========== 1745 */ 1746 1747 /* 1748 * dmfe_interrupt() -- handle chip interrupts 1749 */ 1750 static uint_t 1751 dmfe_interrupt(caddr_t arg) 1752 { 1753 dmfe_t *dmfep; /* private device info */ 1754 uint32_t interrupts; 1755 uint32_t istat; 1756 const char *msg; 1757 mblk_t *mp; 1758 boolean_t warning_msg = B_TRUE; 1759 1760 dmfep = (void *)arg; 1761 1762 mutex_enter(dmfep->oplock); 1763 if (dmfep->suspended) { 1764 mutex_exit(dmfep->oplock); 1765 return (DDI_INTR_UNCLAIMED); 1766 } 1767 1768 /* 1769 * A quick check as to whether the interrupt was from this 1770 * device, before we even finish setting up all our local 1771 * variables. Note that reading the interrupt status register 1772 * doesn't have any unpleasant side effects such as clearing 1773 * the bits read, so it's quite OK to re-read it once we have 1774 * determined that we are going to service this interrupt and 1775 * grabbed the mutexen. 1776 */ 1777 istat = dmfe_chip_get32(dmfep, STATUS_REG); 1778 if ((istat & (NORMAL_SUMMARY_INT | ABNORMAL_SUMMARY_INT)) == 0) { 1779 1780 mutex_exit(dmfep->oplock); 1781 return (DDI_INTR_UNCLAIMED); 1782 } 1783 1784 DRV_KS_INC(dmfep, KS_INTERRUPT); 1785 1786 /* 1787 * Identify bits that represent enabled interrupts ... 1788 */ 1789 istat |= dmfe_chip_get32(dmfep, STATUS_REG); 1790 interrupts = istat & dmfep->imask; 1791 ASSERT(interrupts != 0); 1792 1793 DTRACE_PROBE1(intr, uint32_t, istat); 1794 1795 /* 1796 * Check for any interrupts other than TX/RX done. 1797 * If there are any, they are considered Abnormal 1798 * and will cause the chip to be reset. 1799 */ 1800 if (interrupts & ~(RX_PKTDONE_INT | TX_PKTDONE_INT)) { 1801 if (istat & ABNORMAL_SUMMARY_INT) { 1802 /* 1803 * Any Abnormal interrupts will lead to us 1804 * resetting the chip, so we don't bother 1805 * to clear each interrupt individually. 1806 * 1807 * Our main task here is to identify the problem, 1808 * by pointing out the most significant unexpected 1809 * bit. Additional bits may well be consequences 1810 * of the first problem, so we consider the possible 1811 * causes in order of severity. 1812 */ 1813 if (interrupts & SYSTEM_ERR_INT) { 1814 switch (istat & SYSTEM_ERR_BITS) { 1815 case SYSTEM_ERR_M_ABORT: 1816 msg = "Bus Master Abort"; 1817 break; 1818 1819 case SYSTEM_ERR_T_ABORT: 1820 msg = "Bus Target Abort"; 1821 break; 1822 1823 case SYSTEM_ERR_PARITY: 1824 msg = "Parity Error"; 1825 break; 1826 1827 default: 1828 msg = "Unknown System Bus Error"; 1829 break; 1830 } 1831 } else if (interrupts & RX_STOPPED_INT) { 1832 msg = "RX process stopped"; 1833 } else if (interrupts & RX_UNAVAIL_INT) { 1834 msg = "RX buffer unavailable"; 1835 warning_msg = B_FALSE; 1836 } else if (interrupts & RX_WATCHDOG_INT) { 1837 msg = "RX watchdog timeout?"; 1838 } else if (interrupts & RX_EARLY_INT) { 1839 msg = "RX early interrupt?"; 1840 } else if (interrupts & TX_STOPPED_INT) { 1841 msg = "TX process stopped"; 1842 } else if (interrupts & TX_JABBER_INT) { 1843 msg = "TX jabber timeout"; 1844 } else if (interrupts & TX_UNDERFLOW_INT) { 1845 msg = "TX underflow?"; 1846 } else if (interrupts & TX_EARLY_INT) { 1847 msg = "TX early interrupt?"; 1848 1849 } else if (interrupts & LINK_STATUS_INT) { 1850 msg = "Link status change?"; 1851 } else if (interrupts & GP_TIMER_INT) { 1852 msg = "Timer expired?"; 1853 } 1854 1855 if (warning_msg) 1856 dmfe_warning(dmfep, "abnormal interrupt, " 1857 "status 0x%x: %s", istat, msg); 1858 1859 /* 1860 * We don't want to run the entire reinitialisation 1861 * code out of this (high-level?) interrupt, so we 1862 * simply STOP the chip, and wake up the factotum 1863 * to reinitalise it ... 1864 */ 1865 dmfe_stop_chip(dmfep, CHIP_ERROR); 1866 dmfe_wake_factotum(dmfep, KS_CHIP_ERROR, 1867 "interrupt (error)"); 1868 } else { 1869 /* 1870 * We shouldn't really get here (it would mean 1871 * there were some unprocessed enabled bits but 1872 * they weren't Abnormal?), but we'll check just 1873 * in case ... 1874 */ 1875 DTRACE_PROBE1(intr__unexpected, uint32_t, istat); 1876 } 1877 } 1878 1879 /* 1880 * Acknowledge all the original bits - except in the case of an 1881 * error, when we leave them unacknowledged so that the recovery 1882 * code can see what was going on when the problem occurred ... 1883 */ 1884 if (dmfep->chip_state != CHIP_ERROR) { 1885 (void) dmfe_chip_put32(dmfep, STATUS_REG, istat); 1886 /* 1887 * Read-after-write forces completion on PCI bus. 1888 * 1889 */ 1890 (void) dmfe_chip_get32(dmfep, STATUS_REG); 1891 } 1892 1893 1894 /* 1895 * We've finished talking to the chip, so we can drop <oplock> 1896 * before handling the normal interrupts, which only involve 1897 * manipulation of descriptors ... 1898 */ 1899 mutex_exit(dmfep->oplock); 1900 1901 if (interrupts & RX_PKTDONE_INT) 1902 if ((mp = dmfe_getp(dmfep)) != NULL) 1903 mac_rx(dmfep->mh, NULL, mp); 1904 1905 if (interrupts & TX_PKTDONE_INT) { 1906 /* 1907 * The only reason for taking this interrupt is to give 1908 * MAC a chance to schedule queued packets after a 1909 * ring-full condition. To minimise the number of 1910 * redundant TX-Done interrupts, we only mark two of the 1911 * ring descriptors as 'interrupt-on-complete' - all the 1912 * others are simply handed back without an interrupt. 1913 */ 1914 if (dmfe_reclaim_on_done && mutex_tryenter(dmfep->txlock)) { 1915 (void) dmfe_reclaim_tx_desc(dmfep); 1916 mutex_exit(dmfep->txlock); 1917 } 1918 mac_tx_update(dmfep->mh); 1919 } 1920 1921 return (DDI_INTR_CLAIMED); 1922 } 1923 1924 /* 1925 * ========== Statistics update handler ========== 1926 */ 1927 1928 static int 1929 dmfe_m_stat(void *arg, uint_t stat, uint64_t *val) 1930 { 1931 dmfe_t *dmfep = arg; 1932 int rv = 0; 1933 1934 /* Let MII handle its own stats. */ 1935 if (mii_m_getstat(dmfep->mii, stat, val) == 0) { 1936 return (0); 1937 } 1938 1939 mutex_enter(dmfep->oplock); 1940 mutex_enter(dmfep->rxlock); 1941 mutex_enter(dmfep->txlock); 1942 1943 /* make sure we have all the stats collected */ 1944 (void) dmfe_reclaim_tx_desc(dmfep); 1945 1946 switch (stat) { 1947 1948 case MAC_STAT_IPACKETS: 1949 *val = dmfep->rx_stats_ipackets; 1950 break; 1951 1952 case MAC_STAT_MULTIRCV: 1953 *val = dmfep->rx_stats_multi; 1954 break; 1955 1956 case MAC_STAT_BRDCSTRCV: 1957 *val = dmfep->rx_stats_bcast; 1958 break; 1959 1960 case MAC_STAT_RBYTES: 1961 *val = dmfep->rx_stats_rbytes; 1962 break; 1963 1964 case MAC_STAT_IERRORS: 1965 *val = dmfep->rx_stats_ierrors; 1966 break; 1967 1968 case MAC_STAT_NORCVBUF: 1969 *val = dmfep->rx_stats_norcvbuf; 1970 break; 1971 1972 case MAC_STAT_COLLISIONS: 1973 *val = dmfep->tx_stats_collisions; 1974 break; 1975 1976 case MAC_STAT_OERRORS: 1977 *val = dmfep->tx_stats_oerrors; 1978 break; 1979 1980 case MAC_STAT_OPACKETS: 1981 *val = dmfep->tx_stats_opackets; 1982 break; 1983 1984 case MAC_STAT_MULTIXMT: 1985 *val = dmfep->tx_stats_multi; 1986 break; 1987 1988 case MAC_STAT_BRDCSTXMT: 1989 *val = dmfep->tx_stats_bcast; 1990 break; 1991 1992 case MAC_STAT_OBYTES: 1993 *val = dmfep->tx_stats_obytes; 1994 break; 1995 1996 case MAC_STAT_OVERFLOWS: 1997 *val = dmfep->rx_stats_overflow; 1998 break; 1999 2000 case MAC_STAT_UNDERFLOWS: 2001 *val = dmfep->tx_stats_underflow; 2002 break; 2003 2004 case ETHER_STAT_ALIGN_ERRORS: 2005 *val = dmfep->rx_stats_align; 2006 break; 2007 2008 case ETHER_STAT_FCS_ERRORS: 2009 *val = dmfep->rx_stats_fcs; 2010 break; 2011 2012 case ETHER_STAT_TOOLONG_ERRORS: 2013 *val = dmfep->rx_stats_toolong; 2014 break; 2015 2016 case ETHER_STAT_TOOSHORT_ERRORS: 2017 *val = dmfep->rx_stats_short; 2018 break; 2019 2020 case ETHER_STAT_MACRCV_ERRORS: 2021 *val = dmfep->rx_stats_macrcv_errors; 2022 break; 2023 2024 case ETHER_STAT_MACXMT_ERRORS: 2025 *val = dmfep->tx_stats_macxmt_errors; 2026 break; 2027 2028 case ETHER_STAT_JABBER_ERRORS: 2029 *val = dmfep->tx_stats_jabber; 2030 break; 2031 2032 case ETHER_STAT_CARRIER_ERRORS: 2033 *val = dmfep->tx_stats_nocarrier; 2034 break; 2035 2036 case ETHER_STAT_TX_LATE_COLLISIONS: 2037 *val = dmfep->tx_stats_xmtlatecoll; 2038 break; 2039 2040 case ETHER_STAT_EX_COLLISIONS: 2041 *val = dmfep->tx_stats_excoll; 2042 break; 2043 2044 case ETHER_STAT_DEFER_XMTS: 2045 *val = dmfep->tx_stats_defer; 2046 break; 2047 2048 case ETHER_STAT_FIRST_COLLISIONS: 2049 *val = dmfep->tx_stats_first_coll; 2050 break; 2051 2052 case ETHER_STAT_MULTI_COLLISIONS: 2053 *val = dmfep->tx_stats_multi_coll; 2054 break; 2055 2056 default: 2057 rv = ENOTSUP; 2058 } 2059 2060 mutex_exit(dmfep->txlock); 2061 mutex_exit(dmfep->rxlock); 2062 mutex_exit(dmfep->oplock); 2063 2064 return (rv); 2065 } 2066 2067 /* 2068 * ========== Ioctl handler & subfunctions ========== 2069 */ 2070 2071 static lb_property_t dmfe_loopmodes[] = { 2072 { normal, "normal", 0 }, 2073 { internal, "Internal", 1 }, 2074 { external, "External", 2 }, 2075 }; 2076 2077 /* 2078 * Specific dmfe IOCTLs, the mac module handles the generic ones. 2079 * Unfortunately, the DM9102 doesn't seem to work well with MII based 2080 * loopback, so we have to do something special for it. 2081 */ 2082 2083 static void 2084 dmfe_m_ioctl(void *arg, queue_t *wq, mblk_t *mp) 2085 { 2086 dmfe_t *dmfep = arg; 2087 struct iocblk *iocp; 2088 int rv = 0; 2089 lb_info_sz_t sz; 2090 int cmd; 2091 uint32_t mode; 2092 2093 iocp = (void *)mp->b_rptr; 2094 cmd = iocp->ioc_cmd; 2095 2096 if (mp->b_cont == NULL) { 2097 /* 2098 * All of these ioctls need data! 2099 */ 2100 miocnak(wq, mp, 0, EINVAL); 2101 return; 2102 } 2103 2104 switch (cmd) { 2105 case LB_GET_INFO_SIZE: 2106 if (iocp->ioc_count != sizeof (sz)) { 2107 rv = EINVAL; 2108 } else { 2109 sz = sizeof (dmfe_loopmodes); 2110 bcopy(&sz, mp->b_cont->b_rptr, sizeof (sz)); 2111 } 2112 break; 2113 2114 case LB_GET_INFO: 2115 if (iocp->ioc_count != sizeof (dmfe_loopmodes)) { 2116 rv = EINVAL; 2117 } else { 2118 bcopy(dmfe_loopmodes, mp->b_cont->b_rptr, 2119 iocp->ioc_count); 2120 } 2121 break; 2122 2123 case LB_GET_MODE: 2124 if (iocp->ioc_count != sizeof (mode)) { 2125 rv = EINVAL; 2126 } else { 2127 mutex_enter(dmfep->oplock); 2128 switch (dmfep->opmode & LOOPBACK_MODE_MASK) { 2129 case LOOPBACK_OFF: 2130 mode = 0; 2131 break; 2132 case LOOPBACK_INTERNAL: 2133 mode = 1; 2134 break; 2135 default: 2136 mode = 2; 2137 break; 2138 } 2139 mutex_exit(dmfep->oplock); 2140 bcopy(&mode, mp->b_cont->b_rptr, sizeof (mode)); 2141 } 2142 break; 2143 2144 case LB_SET_MODE: 2145 rv = secpolicy_net_config(iocp->ioc_cr, B_FALSE); 2146 if (rv != 0) 2147 break; 2148 if (iocp->ioc_count != sizeof (mode)) { 2149 rv = EINVAL; 2150 break; 2151 } 2152 bcopy(mp->b_cont->b_rptr, &mode, sizeof (mode)); 2153 2154 mutex_enter(dmfep->oplock); 2155 dmfep->opmode &= ~LOOPBACK_MODE_MASK; 2156 switch (mode) { 2157 case 2: 2158 dmfep->opmode |= LOOPBACK_PHY_D; 2159 break; 2160 case 1: 2161 dmfep->opmode |= LOOPBACK_INTERNAL; 2162 break; 2163 default: 2164 break; 2165 } 2166 if (!dmfep->suspended) { 2167 dmfe_restart(dmfep); 2168 } 2169 mutex_exit(dmfep->oplock); 2170 break; 2171 2172 default: 2173 rv = EINVAL; 2174 break; 2175 } 2176 2177 if (rv == 0) { 2178 miocack(wq, mp, iocp->ioc_count, 0); 2179 } else { 2180 miocnak(wq, mp, 0, rv); 2181 } 2182 } 2183 2184 int 2185 dmfe_m_getprop(void *arg, const char *name, mac_prop_id_t num, uint_t sz, 2186 void *val) 2187 { 2188 dmfe_t *dmfep = arg; 2189 2190 return (mii_m_getprop(dmfep->mii, name, num, sz, val)); 2191 } 2192 2193 int 2194 dmfe_m_setprop(void *arg, const char *name, mac_prop_id_t num, uint_t sz, 2195 const void *val) 2196 { 2197 dmfe_t *dmfep = arg; 2198 2199 return (mii_m_setprop(dmfep->mii, name, num, sz, val)); 2200 } 2201 2202 static void 2203 dmfe_m_propinfo(void *arg, const char *name, mac_prop_id_t num, 2204 mac_prop_info_handle_t mph) 2205 { 2206 dmfe_t *dmfep = arg; 2207 2208 mii_m_propinfo(dmfep->mii, name, num, mph); 2209 } 2210 2211 /* 2212 * ========== Per-instance setup/teardown code ========== 2213 */ 2214 2215 /* 2216 * Determine local MAC address & broadcast address for this interface 2217 */ 2218 static void 2219 dmfe_find_mac_address(dmfe_t *dmfep) 2220 { 2221 uchar_t *prop; 2222 uint_t propsize; 2223 int err; 2224 2225 /* 2226 * We have to find the "vendor's factory-set address". This is 2227 * the value of the property "local-mac-address", as set by OBP 2228 * (or a .conf file!) 2229 * 2230 * If the property is not there, then we try to find the factory 2231 * mac address from the devices serial EEPROM. 2232 */ 2233 bzero(dmfep->curr_addr, sizeof (dmfep->curr_addr)); 2234 err = ddi_prop_lookup_byte_array(DDI_DEV_T_ANY, dmfep->devinfo, 2235 DDI_PROP_DONTPASS, localmac_propname, &prop, &propsize); 2236 if (err == DDI_PROP_SUCCESS) { 2237 if (propsize == ETHERADDRL) 2238 ethaddr_copy(prop, dmfep->curr_addr); 2239 ddi_prop_free(prop); 2240 } else { 2241 /* no property set... check eeprom */ 2242 dmfe_read_eeprom(dmfep, EEPROM_EN_ADDR, dmfep->curr_addr, 2243 ETHERADDRL); 2244 } 2245 } 2246 2247 static int 2248 dmfe_alloc_dma_mem(dmfe_t *dmfep, size_t memsize, 2249 size_t setup, size_t slop, ddi_device_acc_attr_t *attr_p, 2250 uint_t dma_flags, dma_area_t *dma_p) 2251 { 2252 ddi_dma_cookie_t dma_cookie; 2253 uint_t ncookies; 2254 int err; 2255 2256 /* 2257 * Allocate handle 2258 */ 2259 err = ddi_dma_alloc_handle(dmfep->devinfo, &dma_attr, 2260 DDI_DMA_SLEEP, NULL, &dma_p->dma_hdl); 2261 if (err != DDI_SUCCESS) { 2262 dmfe_error(dmfep, "DMA handle allocation failed"); 2263 return (DDI_FAILURE); 2264 } 2265 2266 /* 2267 * Allocate memory 2268 */ 2269 err = ddi_dma_mem_alloc(dma_p->dma_hdl, memsize + setup + slop, 2270 attr_p, dma_flags & (DDI_DMA_CONSISTENT | DDI_DMA_STREAMING), 2271 DDI_DMA_SLEEP, NULL, 2272 &dma_p->mem_va, &dma_p->alength, &dma_p->acc_hdl); 2273 if (err != DDI_SUCCESS) { 2274 dmfe_error(dmfep, "DMA memory allocation failed: %d", err); 2275 return (DDI_FAILURE); 2276 } 2277 2278 /* 2279 * Bind the two together 2280 */ 2281 err = ddi_dma_addr_bind_handle(dma_p->dma_hdl, NULL, 2282 dma_p->mem_va, dma_p->alength, dma_flags, 2283 DDI_DMA_SLEEP, NULL, &dma_cookie, &ncookies); 2284 if (err != DDI_DMA_MAPPED) { 2285 dmfe_error(dmfep, "DMA mapping failed: %d", err); 2286 return (DDI_FAILURE); 2287 } 2288 if ((dma_p->ncookies = ncookies) != 1) { 2289 dmfe_error(dmfep, "Too many DMA cookeis: %d", ncookies); 2290 return (DDI_FAILURE); 2291 } 2292 2293 dma_p->mem_dvma = dma_cookie.dmac_address; 2294 if (setup > 0) { 2295 dma_p->setup_dvma = dma_p->mem_dvma + memsize; 2296 dma_p->setup_va = dma_p->mem_va + memsize; 2297 } else { 2298 dma_p->setup_dvma = 0; 2299 dma_p->setup_va = NULL; 2300 } 2301 2302 return (DDI_SUCCESS); 2303 } 2304 2305 /* 2306 * This function allocates the transmit and receive buffers and descriptors. 2307 */ 2308 static int 2309 dmfe_alloc_bufs(dmfe_t *dmfep) 2310 { 2311 size_t memsize; 2312 int err; 2313 2314 /* 2315 * Allocate memory & handles for TX descriptor ring 2316 */ 2317 memsize = dmfep->tx.n_desc * sizeof (struct tx_desc_type); 2318 err = dmfe_alloc_dma_mem(dmfep, memsize, SETUPBUF_SIZE, DMFE_SLOP, 2319 &dmfe_reg_accattr, DDI_DMA_RDWR | DDI_DMA_CONSISTENT, 2320 &dmfep->tx_desc); 2321 if (err != DDI_SUCCESS) { 2322 dmfe_error(dmfep, "TX descriptor allocation failed"); 2323 return (DDI_FAILURE); 2324 } 2325 2326 /* 2327 * Allocate memory & handles for TX buffers 2328 */ 2329 memsize = dmfep->tx.n_desc * DMFE_BUF_SIZE; 2330 err = dmfe_alloc_dma_mem(dmfep, memsize, 0, 0, 2331 &dmfe_data_accattr, DDI_DMA_WRITE | DMFE_DMA_MODE, 2332 &dmfep->tx_buff); 2333 if (err != DDI_SUCCESS) { 2334 dmfe_error(dmfep, "TX buffer allocation failed"); 2335 return (DDI_FAILURE); 2336 } 2337 2338 /* 2339 * Allocate memory & handles for RX descriptor ring 2340 */ 2341 memsize = dmfep->rx.n_desc * sizeof (struct rx_desc_type); 2342 err = dmfe_alloc_dma_mem(dmfep, memsize, 0, DMFE_SLOP, 2343 &dmfe_reg_accattr, DDI_DMA_RDWR | DDI_DMA_CONSISTENT, 2344 &dmfep->rx_desc); 2345 if (err != DDI_SUCCESS) { 2346 dmfe_error(dmfep, "RX descriptor allocation failed"); 2347 return (DDI_FAILURE); 2348 } 2349 2350 /* 2351 * Allocate memory & handles for RX buffers 2352 */ 2353 memsize = dmfep->rx.n_desc * DMFE_BUF_SIZE; 2354 err = dmfe_alloc_dma_mem(dmfep, memsize, 0, 0, 2355 &dmfe_data_accattr, DDI_DMA_READ | DMFE_DMA_MODE, &dmfep->rx_buff); 2356 if (err != DDI_SUCCESS) { 2357 dmfe_error(dmfep, "RX buffer allocation failed"); 2358 return (DDI_FAILURE); 2359 } 2360 2361 /* 2362 * Allocate bitmasks for tx packet type tracking 2363 */ 2364 dmfep->tx_mcast = kmem_zalloc(dmfep->tx.n_desc / NBBY, KM_SLEEP); 2365 dmfep->tx_bcast = kmem_zalloc(dmfep->tx.n_desc / NBBY, KM_SLEEP); 2366 2367 return (DDI_SUCCESS); 2368 } 2369 2370 static void 2371 dmfe_free_dma_mem(dma_area_t *dma_p) 2372 { 2373 if (dma_p->dma_hdl != NULL) { 2374 if (dma_p->ncookies) { 2375 (void) ddi_dma_unbind_handle(dma_p->dma_hdl); 2376 dma_p->ncookies = 0; 2377 } 2378 ddi_dma_free_handle(&dma_p->dma_hdl); 2379 dma_p->dma_hdl = NULL; 2380 dma_p->mem_dvma = 0; 2381 dma_p->setup_dvma = 0; 2382 } 2383 2384 if (dma_p->acc_hdl != NULL) { 2385 ddi_dma_mem_free(&dma_p->acc_hdl); 2386 dma_p->acc_hdl = NULL; 2387 dma_p->mem_va = NULL; 2388 dma_p->setup_va = NULL; 2389 } 2390 } 2391 2392 /* 2393 * This routine frees the transmit and receive buffers and descriptors. 2394 * Make sure the chip is stopped before calling it! 2395 */ 2396 static void 2397 dmfe_free_bufs(dmfe_t *dmfep) 2398 { 2399 dmfe_free_dma_mem(&dmfep->rx_buff); 2400 dmfe_free_dma_mem(&dmfep->rx_desc); 2401 dmfe_free_dma_mem(&dmfep->tx_buff); 2402 dmfe_free_dma_mem(&dmfep->tx_desc); 2403 if (dmfep->tx_mcast) 2404 kmem_free(dmfep->tx_mcast, dmfep->tx.n_desc / NBBY); 2405 if (dmfep->tx_bcast) 2406 kmem_free(dmfep->tx_bcast, dmfep->tx.n_desc / NBBY); 2407 } 2408 2409 static void 2410 dmfe_unattach(dmfe_t *dmfep) 2411 { 2412 /* 2413 * Clean up and free all DMFE data structures 2414 */ 2415 if (dmfep->cycid != NULL) { 2416 ddi_periodic_delete(dmfep->cycid); 2417 dmfep->cycid = NULL; 2418 } 2419 2420 if (dmfep->ksp_drv != NULL) 2421 kstat_delete(dmfep->ksp_drv); 2422 if (dmfep->progress & PROGRESS_HWINT) { 2423 ddi_remove_intr(dmfep->devinfo, 0, dmfep->iblk); 2424 } 2425 if (dmfep->progress & PROGRESS_SOFTINT) 2426 ddi_remove_softintr(dmfep->factotum_id); 2427 if (dmfep->mii != NULL) 2428 mii_free(dmfep->mii); 2429 if (dmfep->progress & PROGRESS_MUTEX) { 2430 mutex_destroy(dmfep->txlock); 2431 mutex_destroy(dmfep->rxlock); 2432 mutex_destroy(dmfep->oplock); 2433 } 2434 dmfe_free_bufs(dmfep); 2435 if (dmfep->io_handle != NULL) 2436 ddi_regs_map_free(&dmfep->io_handle); 2437 2438 kmem_free(dmfep, sizeof (*dmfep)); 2439 } 2440 2441 static int 2442 dmfe_config_init(dmfe_t *dmfep, chip_id_t *idp) 2443 { 2444 ddi_acc_handle_t handle; 2445 uint32_t regval; 2446 2447 if (pci_config_setup(dmfep->devinfo, &handle) != DDI_SUCCESS) 2448 return (DDI_FAILURE); 2449 2450 /* 2451 * Get vendor/device/revision. We expect (but don't check) that 2452 * (vendorid == DAVICOM_VENDOR_ID) && (deviceid == DEVICE_ID_9102) 2453 */ 2454 idp->vendor = pci_config_get16(handle, PCI_CONF_VENID); 2455 idp->device = pci_config_get16(handle, PCI_CONF_DEVID); 2456 idp->revision = pci_config_get8(handle, PCI_CONF_REVID); 2457 2458 /* 2459 * Turn on Bus Master Enable bit and ensure the device is not asleep 2460 */ 2461 regval = pci_config_get32(handle, PCI_CONF_COMM); 2462 pci_config_put32(handle, PCI_CONF_COMM, (regval | PCI_COMM_ME)); 2463 2464 regval = pci_config_get32(handle, PCI_DMFE_CONF_CFDD); 2465 pci_config_put32(handle, PCI_DMFE_CONF_CFDD, 2466 regval & ~(CFDD_SLEEP | CFDD_SNOOZE)); 2467 2468 pci_config_teardown(&handle); 2469 return (DDI_SUCCESS); 2470 } 2471 2472 struct ks_index { 2473 int index; 2474 char *name; 2475 }; 2476 2477 static const struct ks_index ks_drv_names[] = { 2478 { KS_INTERRUPT, "intr" }, 2479 { KS_CYCLIC_RUN, "cyclic_run" }, 2480 2481 { KS_TX_STALL, "tx_stall_detect" }, 2482 { KS_CHIP_ERROR, "chip_error_interrupt" }, 2483 2484 { KS_FACTOTUM_RUN, "factotum_run" }, 2485 { KS_RECOVERY, "factotum_recover" }, 2486 2487 { -1, NULL } 2488 }; 2489 2490 static void 2491 dmfe_init_kstats(dmfe_t *dmfep, int instance) 2492 { 2493 kstat_t *ksp; 2494 kstat_named_t *knp; 2495 const struct ks_index *ksip; 2496 2497 /* no need to create MII stats, the mac module already does it */ 2498 2499 /* Create and initialise driver-defined kstats */ 2500 ksp = kstat_create(DRIVER_NAME, instance, "dmfe_events", "net", 2501 KSTAT_TYPE_NAMED, KS_DRV_COUNT, KSTAT_FLAG_PERSISTENT); 2502 if (ksp != NULL) { 2503 for (knp = ksp->ks_data, ksip = ks_drv_names; 2504 ksip->name != NULL; ++ksip) { 2505 kstat_named_init(&knp[ksip->index], ksip->name, 2506 KSTAT_DATA_UINT64); 2507 } 2508 dmfep->ksp_drv = ksp; 2509 dmfep->knp_drv = knp; 2510 kstat_install(ksp); 2511 } else { 2512 dmfe_error(dmfep, "kstat_create() for dmfe_events failed"); 2513 } 2514 } 2515 2516 static int 2517 dmfe_resume(dev_info_t *devinfo) 2518 { 2519 dmfe_t *dmfep; /* Our private data */ 2520 chip_id_t chipid; 2521 boolean_t restart = B_FALSE; 2522 2523 dmfep = ddi_get_driver_private(devinfo); 2524 if (dmfep == NULL) 2525 return (DDI_FAILURE); 2526 2527 /* 2528 * Refuse to resume if the data structures aren't consistent 2529 */ 2530 if (dmfep->devinfo != devinfo) 2531 return (DDI_FAILURE); 2532 2533 /* 2534 * Refuse to resume if the chip's changed its identity (*boggle*) 2535 */ 2536 if (dmfe_config_init(dmfep, &chipid) != DDI_SUCCESS) 2537 return (DDI_FAILURE); 2538 if (chipid.vendor != dmfep->chipid.vendor) 2539 return (DDI_FAILURE); 2540 if (chipid.device != dmfep->chipid.device) 2541 return (DDI_FAILURE); 2542 if (chipid.revision != dmfep->chipid.revision) 2543 return (DDI_FAILURE); 2544 2545 mutex_enter(dmfep->oplock); 2546 mutex_enter(dmfep->txlock); 2547 dmfep->suspended = B_FALSE; 2548 mutex_exit(dmfep->txlock); 2549 2550 /* 2551 * All OK, reinitialise h/w & kick off MAC scheduling 2552 */ 2553 if (dmfep->mac_state == DMFE_MAC_STARTED) { 2554 dmfe_restart(dmfep); 2555 restart = B_TRUE; 2556 } 2557 mutex_exit(dmfep->oplock); 2558 2559 if (restart) { 2560 mii_resume(dmfep->mii); 2561 mac_tx_update(dmfep->mh); 2562 } 2563 return (DDI_SUCCESS); 2564 } 2565 2566 /* 2567 * attach(9E) -- Attach a device to the system 2568 * 2569 * Called once for each board successfully probed. 2570 */ 2571 static int 2572 dmfe_attach(dev_info_t *devinfo, ddi_attach_cmd_t cmd) 2573 { 2574 mac_register_t *macp; 2575 dmfe_t *dmfep; /* Our private data */ 2576 uint32_t csr6; 2577 int instance; 2578 int err; 2579 2580 instance = ddi_get_instance(devinfo); 2581 2582 switch (cmd) { 2583 default: 2584 return (DDI_FAILURE); 2585 2586 case DDI_RESUME: 2587 return (dmfe_resume(devinfo)); 2588 2589 case DDI_ATTACH: 2590 break; 2591 } 2592 2593 dmfep = kmem_zalloc(sizeof (*dmfep), KM_SLEEP); 2594 ddi_set_driver_private(devinfo, dmfep); 2595 dmfep->devinfo = devinfo; 2596 dmfep->dmfe_guard = DMFE_GUARD; 2597 2598 /* 2599 * Initialize more fields in DMFE private data 2600 * Determine the local MAC address 2601 */ 2602 #if DMFEDEBUG 2603 dmfep->debug = ddi_prop_get_int(DDI_DEV_T_ANY, devinfo, 0, 2604 debug_propname, dmfe_debug); 2605 #endif /* DMFEDEBUG */ 2606 dmfep->cycid = NULL; 2607 (void) snprintf(dmfep->ifname, sizeof (dmfep->ifname), "dmfe%d", 2608 instance); 2609 2610 /* 2611 * Check for custom "opmode-reg-value" property; 2612 * if none, use the defaults below for CSR6 ... 2613 */ 2614 csr6 = TX_THRESHOLD_HI | STORE_AND_FORWARD | EXT_MII_IF | OPN_25_MB1; 2615 dmfep->opmode = ddi_prop_get_int(DDI_DEV_T_ANY, devinfo, 2616 DDI_PROP_DONTPASS, opmode_propname, csr6); 2617 2618 /* 2619 * Read chip ID & set up config space command register(s) 2620 */ 2621 if (dmfe_config_init(dmfep, &dmfep->chipid) != DDI_SUCCESS) { 2622 dmfe_error(dmfep, "dmfe_config_init() failed"); 2623 goto attach_fail; 2624 } 2625 2626 /* 2627 * Map operating registers 2628 */ 2629 err = ddi_regs_map_setup(devinfo, DMFE_PCI_RNUMBER, 2630 &dmfep->io_reg, 0, 0, &dmfe_reg_accattr, &dmfep->io_handle); 2631 if (err != DDI_SUCCESS) { 2632 dmfe_error(dmfep, "ddi_regs_map_setup() failed"); 2633 goto attach_fail; 2634 } 2635 2636 /* 2637 * Get our MAC address. 2638 */ 2639 dmfe_find_mac_address(dmfep); 2640 2641 /* 2642 * Allocate the TX and RX descriptors/buffers. 2643 */ 2644 dmfep->tx.n_desc = dmfe_tx_desc; 2645 dmfep->rx.n_desc = dmfe_rx_desc; 2646 err = dmfe_alloc_bufs(dmfep); 2647 if (err != DDI_SUCCESS) { 2648 goto attach_fail; 2649 } 2650 2651 /* 2652 * Add the softint handler 2653 */ 2654 if (ddi_add_softintr(devinfo, DDI_SOFTINT_LOW, &dmfep->factotum_id, 2655 NULL, NULL, dmfe_factotum, (caddr_t)dmfep) != DDI_SUCCESS) { 2656 dmfe_error(dmfep, "ddi_add_softintr() failed"); 2657 goto attach_fail; 2658 } 2659 dmfep->progress |= PROGRESS_SOFTINT; 2660 2661 /* 2662 * Add the h/w interrupt handler & initialise mutexen 2663 */ 2664 if (ddi_get_iblock_cookie(devinfo, 0, &dmfep->iblk) != DDI_SUCCESS) { 2665 dmfe_error(dmfep, "ddi_get_iblock_cookie() failed"); 2666 goto attach_fail; 2667 } 2668 2669 mutex_init(dmfep->milock, NULL, MUTEX_DRIVER, NULL); 2670 mutex_init(dmfep->oplock, NULL, MUTEX_DRIVER, dmfep->iblk); 2671 mutex_init(dmfep->rxlock, NULL, MUTEX_DRIVER, dmfep->iblk); 2672 mutex_init(dmfep->txlock, NULL, MUTEX_DRIVER, dmfep->iblk); 2673 dmfep->progress |= PROGRESS_MUTEX; 2674 2675 if (ddi_add_intr(devinfo, 0, NULL, NULL, 2676 dmfe_interrupt, (caddr_t)dmfep) != DDI_SUCCESS) { 2677 dmfe_error(dmfep, "ddi_add_intr() failed"); 2678 goto attach_fail; 2679 } 2680 dmfep->progress |= PROGRESS_HWINT; 2681 2682 /* 2683 * Create & initialise named kstats 2684 */ 2685 dmfe_init_kstats(dmfep, instance); 2686 2687 /* 2688 * Reset & initialise the chip and the ring buffers 2689 * Initialise the (internal) PHY 2690 */ 2691 mutex_enter(dmfep->oplock); 2692 mutex_enter(dmfep->rxlock); 2693 mutex_enter(dmfep->txlock); 2694 2695 dmfe_reset(dmfep); 2696 2697 /* 2698 * Prepare the setup packet 2699 */ 2700 bzero(dmfep->tx_desc.setup_va, SETUPBUF_SIZE); 2701 bzero(dmfep->mcast_refs, MCASTBUF_SIZE); 2702 dmfep->addr_set = B_FALSE; 2703 dmfep->opmode &= ~(PROMISC_MODE | PASS_MULTICAST); 2704 dmfep->mac_state = DMFE_MAC_RESET; 2705 2706 mutex_exit(dmfep->txlock); 2707 mutex_exit(dmfep->rxlock); 2708 mutex_exit(dmfep->oplock); 2709 2710 if (dmfe_init_phy(dmfep) != B_TRUE) 2711 goto attach_fail; 2712 2713 /* 2714 * Send a reasonable setup frame. This configures our starting 2715 * address and the broadcast address. 2716 */ 2717 (void) dmfe_m_unicst(dmfep, dmfep->curr_addr); 2718 2719 /* 2720 * Initialize pointers to device specific functions which 2721 * will be used by the generic layer. 2722 */ 2723 if ((macp = mac_alloc(MAC_VERSION)) == NULL) 2724 goto attach_fail; 2725 macp->m_type_ident = MAC_PLUGIN_IDENT_ETHER; 2726 macp->m_driver = dmfep; 2727 macp->m_dip = devinfo; 2728 macp->m_src_addr = dmfep->curr_addr; 2729 macp->m_callbacks = &dmfe_m_callbacks; 2730 macp->m_min_sdu = 0; 2731 macp->m_max_sdu = ETHERMTU; 2732 macp->m_margin = VLAN_TAGSZ; 2733 2734 /* 2735 * Finally, we're ready to register ourselves with the MAC layer 2736 * interface; if this succeeds, we're all ready to start() 2737 */ 2738 err = mac_register(macp, &dmfep->mh); 2739 mac_free(macp); 2740 if (err != 0) 2741 goto attach_fail; 2742 ASSERT(dmfep->dmfe_guard == DMFE_GUARD); 2743 2744 /* 2745 * Install the cyclic callback that we use to check for link 2746 * status, transmit stall, etc. The cyclic callback (dmfe_cyclic()) 2747 * is invoked in kernel context then. 2748 */ 2749 ASSERT(dmfep->cycid == NULL); 2750 dmfep->cycid = ddi_periodic_add(dmfe_cyclic, dmfep, 2751 dmfe_tick_us * 1000, DDI_IPL_0); 2752 return (DDI_SUCCESS); 2753 2754 attach_fail: 2755 dmfe_unattach(dmfep); 2756 return (DDI_FAILURE); 2757 } 2758 2759 /* 2760 * dmfe_suspend() -- suspend transmit/receive for powerdown 2761 */ 2762 static int 2763 dmfe_suspend(dmfe_t *dmfep) 2764 { 2765 /* 2766 * Just stop processing ... 2767 */ 2768 mii_suspend(dmfep->mii); 2769 mutex_enter(dmfep->oplock); 2770 dmfe_stop(dmfep); 2771 2772 mutex_enter(dmfep->txlock); 2773 dmfep->suspended = B_TRUE; 2774 mutex_exit(dmfep->txlock); 2775 mutex_exit(dmfep->oplock); 2776 2777 return (DDI_SUCCESS); 2778 } 2779 2780 /* 2781 * detach(9E) -- Detach a device from the system 2782 */ 2783 static int 2784 dmfe_detach(dev_info_t *devinfo, ddi_detach_cmd_t cmd) 2785 { 2786 dmfe_t *dmfep; 2787 2788 dmfep = ddi_get_driver_private(devinfo); 2789 2790 switch (cmd) { 2791 default: 2792 return (DDI_FAILURE); 2793 2794 case DDI_SUSPEND: 2795 return (dmfe_suspend(dmfep)); 2796 2797 case DDI_DETACH: 2798 break; 2799 } 2800 2801 /* 2802 * Unregister from the MAC subsystem. This can fail, in 2803 * particular if there are DLPI style-2 streams still open - 2804 * in which case we just return failure without shutting 2805 * down chip operations. 2806 */ 2807 if (mac_unregister(dmfep->mh) != DDI_SUCCESS) 2808 return (DDI_FAILURE); 2809 2810 /* 2811 * All activity stopped, so we can clean up & exit 2812 */ 2813 dmfe_unattach(dmfep); 2814 return (DDI_SUCCESS); 2815 } 2816 2817 2818 /* 2819 * ========== Module Loading Data & Entry Points ========== 2820 */ 2821 2822 DDI_DEFINE_STREAM_OPS(dmfe_dev_ops, nulldev, nulldev, dmfe_attach, dmfe_detach, 2823 nodev, NULL, D_MP, NULL, ddi_quiesce_not_supported); 2824 2825 static struct modldrv dmfe_modldrv = { 2826 &mod_driverops, /* Type of module. This one is a driver */ 2827 dmfe_ident, /* short description */ 2828 &dmfe_dev_ops /* driver specific ops */ 2829 }; 2830 2831 static struct modlinkage modlinkage = { 2832 MODREV_1, (void *)&dmfe_modldrv, NULL 2833 }; 2834 2835 int 2836 _info(struct modinfo *modinfop) 2837 { 2838 return (mod_info(&modlinkage, modinfop)); 2839 } 2840 2841 int 2842 _init(void) 2843 { 2844 uint32_t tmp100; 2845 uint32_t tmp10; 2846 int i; 2847 int status; 2848 2849 /* Calculate global timing parameters */ 2850 tmp100 = (dmfe_tx100_stall_us+dmfe_tick_us-1)/dmfe_tick_us; 2851 tmp10 = (dmfe_tx10_stall_us+dmfe_tick_us-1)/dmfe_tick_us; 2852 2853 for (i = 0; i <= TX_PROCESS_MAX_STATE; ++i) { 2854 switch (i) { 2855 case TX_PROCESS_STATE(TX_PROCESS_FETCH_DATA): 2856 case TX_PROCESS_STATE(TX_PROCESS_WAIT_END): 2857 /* 2858 * The chip doesn't spontaneously recover from 2859 * a stall in these states, so we reset early 2860 */ 2861 stall_100_tix[i] = tmp100; 2862 stall_10_tix[i] = tmp10; 2863 break; 2864 2865 case TX_PROCESS_STATE(TX_PROCESS_SUSPEND): 2866 default: 2867 /* 2868 * The chip has been seen to spontaneously recover 2869 * after an apparent stall in the SUSPEND state, 2870 * so we'll allow it rather longer to do so. As 2871 * stalls in other states have not been observed, 2872 * we'll use long timeouts for them too ... 2873 */ 2874 stall_100_tix[i] = tmp100 * 20; 2875 stall_10_tix[i] = tmp10 * 20; 2876 break; 2877 } 2878 } 2879 2880 mac_init_ops(&dmfe_dev_ops, "dmfe"); 2881 status = mod_install(&modlinkage); 2882 if (status == DDI_SUCCESS) 2883 dmfe_log_init(); 2884 2885 return (status); 2886 } 2887 2888 int 2889 _fini(void) 2890 { 2891 int status; 2892 2893 status = mod_remove(&modlinkage); 2894 if (status == DDI_SUCCESS) { 2895 mac_fini_ops(&dmfe_dev_ops); 2896 dmfe_log_fini(); 2897 } 2898 2899 return (status); 2900 } 2901