1 /*
2 * CDDL HEADER START
3 *
4 * The contents of this file are subject to the terms of the
5 * Common Development and Distribution License (the "License").
6 * You may not use this file except in compliance with the License.
7 *
8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9 * or http://www.opensolaris.org/os/licensing.
10 * See the License for the specific language governing permissions
11 * and limitations under the License.
12 *
13 * When distributing Covered Code, include this CDDL HEADER in each
14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15 * If applicable, add the following below this CDDL HEADER, with the
16 * fields enclosed by brackets "[]" replaced with your own identifying
17 * information: Portions Copyright [yyyy] [name of copyright owner]
18 *
19 * CDDL HEADER END
20 */
21
22 /*
23 * Copyright (c) 2009, 2010, Oracle and/or its affiliates. All rights reserved.
24 */
25 #include <sys/stream.h>
26 #include <sys/strsun.h>
27 #include <sys/stat.h>
28 #include <sys/pci.h>
29 #include <sys/modctl.h>
30 #include <sys/kstat.h>
31 #include <sys/ethernet.h>
32 #include <sys/devops.h>
33 #include <sys/debug.h>
34 #include <sys/conf.h>
35 #include <sys/sysmacros.h>
36 #include <sys/dditypes.h>
37 #include <sys/ddi.h>
38 #include <sys/sunddi.h>
39 #include <sys/miiregs.h>
40 #include <sys/byteorder.h>
41 #include <sys/cyclic.h>
42 #include <sys/note.h>
43 #include <sys/crc32.h>
44 #include <sys/mac_provider.h>
45 #include <sys/mac_ether.h>
46 #include <sys/vlan.h>
47 #include <sys/errno.h>
48 #include <sys/sdt.h>
49 #include <sys/strsubr.h>
50
51 #include "bfe.h"
52 #include "bfe_hw.h"
53
54
55 /*
56 * Broadcom BCM4401 chipsets use two rings :
57 *
58 * - One TX : For sending packets down the wire.
59 * - One RX : For receving packets.
60 *
61 * Each ring can have any number of descriptors (configured during attach).
62 * As of now we configure only 128 descriptor per ring (TX/RX). Each descriptor
63 * has address (desc_addr) and control (desc_ctl) which holds a DMA buffer for
64 * the packet and control information (like start/end of frame or end of table).
65 * The descriptor table is allocated first and then a DMA buffer (for a packet)
66 * is allocated and linked to each descriptor.
67 *
68 * Each descriptor entry is bfe_desc_t structure in bfe. During TX/RX
69 * interrupt, the stat register will point to current descriptor being
70 * processed.
71 *
72 * Here's an example of TX and RX ring :
73 *
74 * TX:
75 *
76 * Base of the descriptor table is programmed using BFE_DMATX_CTRL control
77 * register. Each 'addr' points to DMA buffer (or packet data buffer) to
78 * be transmitted and 'ctl' has the length of the packet (usually MTU).
79 *
80 * ----------------------|
81 * | addr |Descriptor 0 |
82 * | ctl | |
83 * ----------------------|
84 * | addr |Descriptor 1 | SOF (start of the frame)
85 * | ctl | |
86 * ----------------------|
87 * | ... |Descriptor... | EOF (end of the frame)
88 * | ... | |
89 * ----------------------|
90 * | addr |Descritor 127 |
91 * | ctl | EOT | EOT (End of Table)
92 * ----------------------|
93 *
94 * 'r_curr_desc' : pointer to current descriptor which can be used to transmit
95 * a packet.
96 * 'r_avail_desc' : decremented whenever a packet is being sent.
97 * 'r_cons_desc' : incremented whenever a packet is sent down the wire and
98 * notified by an interrupt to bfe driver.
99 *
100 * RX:
101 *
102 * Base of the descriptor table is programmed using BFE_DMARX_CTRL control
103 * register. Each 'addr' points to DMA buffer (or packet data buffer). 'ctl'
104 * contains the size of the DMA buffer and all the DMA buffers are
105 * pre-allocated during attach and hence the maxmium size of the packet is
106 * also known (r_buf_len from the bfe_rint_t structure). During RX interrupt
107 * the packet length is embedded in bfe_header_t which is added by the
108 * chip in the beginning of the packet.
109 *
110 * ----------------------|
111 * | addr |Descriptor 0 |
112 * | ctl | |
113 * ----------------------|
114 * | addr |Descriptor 1 |
115 * | ctl | |
116 * ----------------------|
117 * | ... |Descriptor... |
118 * | ... | |
119 * ----------------------|
120 * | addr |Descriptor 127|
121 * | ctl | EOT | EOT (End of Table)
122 * ----------------------|
123 *
124 * 'r_curr_desc' : pointer to current descriptor while receving a packet.
125 *
126 */
127
128 #define MODULE_NAME "bfe"
129
130 /*
131 * Used for checking PHY (link state, speed)
132 */
133 #define BFE_TIMEOUT_INTERVAL (1000 * 1000 * 1000)
134
135
136 /*
137 * Chip restart action and reason for restart
138 */
139 #define BFE_ACTION_RESTART 0x1 /* For restarting the chip */
140 #define BFE_ACTION_RESTART_SETPROP 0x2 /* restart due to setprop */
141 #define BFE_ACTION_RESTART_FAULT 0x4 /* restart due to fault */
142 #define BFE_ACTION_RESTART_PKT 0x8 /* restart due to pkt timeout */
143
144 static char bfe_ident[] = "bfe driver for Broadcom BCM4401 chipsets";
145
146 /*
147 * Function Prototypes for bfe driver.
148 */
149 static int bfe_check_link(bfe_t *);
150 static void bfe_report_link(bfe_t *);
151 static void bfe_chip_halt(bfe_t *);
152 static void bfe_chip_reset(bfe_t *);
153 static void bfe_tx_desc_init(bfe_ring_t *);
154 static void bfe_rx_desc_init(bfe_ring_t *);
155 static void bfe_set_rx_mode(bfe_t *);
156 static void bfe_enable_chip_intrs(bfe_t *);
157 static void bfe_chip_restart(bfe_t *);
158 static void bfe_init_vars(bfe_t *);
159 static void bfe_clear_stats(bfe_t *);
160 static void bfe_gather_stats(bfe_t *);
161 static void bfe_error(dev_info_t *, char *, ...);
162 static int bfe_mac_getprop(void *, const char *, mac_prop_id_t, uint_t,
163 void *);
164 static int bfe_mac_setprop(void *, const char *, mac_prop_id_t, uint_t,
165 const void *);
166 static int bfe_tx_reclaim(bfe_ring_t *);
167 int bfe_mac_set_ether_addr(void *, const uint8_t *);
168
169
170 /*
171 * Macros for ddi_dma_sync().
172 */
173 #define SYNC_DESC(r, s, l, d) \
174 (void) ddi_dma_sync(r->r_desc_dma_handle, \
175 (off_t)(s * sizeof (bfe_desc_t)), \
176 (size_t)(l * sizeof (bfe_desc_t)), \
177 d)
178
179 #define SYNC_BUF(r, s, b, l, d) \
180 (void) ddi_dma_sync(r->r_buf_dma[s].handle, \
181 (off_t)(b), (size_t)(l), d)
182
183 /*
184 * Supported Broadcom BCM4401 Cards.
185 */
186 static bfe_cards_t bfe_cards[] = {
187 { 0x14e4, 0x170c, "BCM4401 100Base-TX"},
188 };
189
190
191 /*
192 * DMA attributes for device registers, packet data (buffer) and
193 * descriptor table.
194 */
195 static struct ddi_device_acc_attr bfe_dev_attr = {
196 DDI_DEVICE_ATTR_V0,
197 DDI_STRUCTURE_LE_ACC,
198 DDI_STRICTORDER_ACC
199 };
200
201 static struct ddi_device_acc_attr bfe_buf_attr = {
202 DDI_DEVICE_ATTR_V0,
203 DDI_NEVERSWAP_ACC, /* native endianness */
204 DDI_STRICTORDER_ACC
205 };
206
207 static ddi_dma_attr_t bfe_dma_attr_buf = {
208 DMA_ATTR_V0, /* dma_attr_version */
209 0, /* dma_attr_addr_lo */
210 BFE_PCI_DMA - 1, /* dma_attr_addr_hi */
211 0x1fff, /* dma_attr_count_max */
212 8, /* dma_attr_align */
213 0, /* dma_attr_burstsizes */
214 1, /* dma_attr_minxfer */
215 0x1fff, /* dma_attr_maxxfer */
216 BFE_PCI_DMA - 1, /* dma_attr_seg */
217 1, /* dma_attr_sgllen */
218 1, /* dma_attr_granular */
219 0 /* dma_attr_flags */
220 };
221
222 static ddi_dma_attr_t bfe_dma_attr_desc = {
223 DMA_ATTR_V0, /* dma_attr_version */
224 0, /* dma_attr_addr_lo */
225 BFE_PCI_DMA - 1, /* dma_attr_addr_hi */
226 BFE_PCI_DMA - 1, /* dma_attr_count_max */
227 BFE_DESC_ALIGN, /* dma_attr_align */
228 0, /* dma_attr_burstsizes */
229 1, /* dma_attr_minxfer */
230 BFE_PCI_DMA - 1, /* dma_attr_maxxfer */
231 BFE_PCI_DMA - 1, /* dma_attr_seg */
232 1, /* dma_attr_sgllen */
233 1, /* dma_attr_granular */
234 0 /* dma_attr_flags */
235 };
236
237 /*
238 * Ethernet broadcast addresses.
239 */
240 static uchar_t bfe_broadcast[ETHERADDRL] = {
241 0xff, 0xff, 0xff, 0xff, 0xff, 0xff
242 };
243
244 #define ASSERT_ALL_LOCKS(bfe) { \
245 ASSERT(mutex_owned(&bfe->bfe_tx_ring.r_lock)); \
246 ASSERT(rw_write_held(&bfe->bfe_rwlock)); \
247 }
248
249 /*
250 * Debugging and error reproting code.
251 */
252 static void
bfe_error(dev_info_t * dip,char * fmt,...)253 bfe_error(dev_info_t *dip, char *fmt, ...)
254 {
255 va_list ap;
256 char buf[256];
257
258 va_start(ap, fmt);
259 (void) vsnprintf(buf, sizeof (buf), fmt, ap);
260 va_end(ap);
261
262 if (dip) {
263 cmn_err(CE_WARN, "%s%d: %s",
264 ddi_driver_name(dip), ddi_get_instance(dip), buf);
265 } else {
266 cmn_err(CE_WARN, "bfe: %s", buf);
267 }
268 }
269
270 /*
271 * Grabs all necessary locks to block any other operation on the chip.
272 */
273 static void
bfe_grab_locks(bfe_t * bfe)274 bfe_grab_locks(bfe_t *bfe)
275 {
276 bfe_ring_t *tx = &bfe->bfe_tx_ring;
277
278 /*
279 * Grab all the locks.
280 * - bfe_rwlock : locks down whole chip including RX.
281 * - tx's r_lock : locks down only TX side.
282 */
283 rw_enter(&bfe->bfe_rwlock, RW_WRITER);
284 mutex_enter(&tx->r_lock);
285
286 /*
287 * Note that we don't use RX's r_lock.
288 */
289 }
290
291 /*
292 * Release lock on chip/drver.
293 */
294 static void
bfe_release_locks(bfe_t * bfe)295 bfe_release_locks(bfe_t *bfe)
296 {
297 bfe_ring_t *tx = &bfe->bfe_tx_ring;
298
299 /*
300 * Release all the locks in the order in which they were grabbed.
301 */
302 mutex_exit(&tx->r_lock);
303 rw_exit(&bfe->bfe_rwlock);
304 }
305
306
307 /*
308 * It's used to make sure that the write to device register was successful.
309 */
310 static int
bfe_wait_bit(bfe_t * bfe,uint32_t reg,uint32_t bit,ulong_t t,const int clear)311 bfe_wait_bit(bfe_t *bfe, uint32_t reg, uint32_t bit,
312 ulong_t t, const int clear)
313 {
314 ulong_t i;
315 uint32_t v;
316
317 for (i = 0; i < t; i++) {
318 v = INL(bfe, reg);
319
320 if (clear && !(v & bit))
321 break;
322
323 if (!clear && (v & bit))
324 break;
325
326 drv_usecwait(10);
327 }
328
329 /* if device still didn't see the value */
330 if (i == t)
331 return (-1);
332
333 return (0);
334 }
335
336 /*
337 * PHY functions (read, write, stop, reset and startup)
338 */
339 static int
bfe_read_phy(bfe_t * bfe,uint32_t reg)340 bfe_read_phy(bfe_t *bfe, uint32_t reg)
341 {
342 OUTL(bfe, BFE_EMAC_ISTAT, BFE_EMAC_INT_MII);
343 OUTL(bfe, BFE_MDIO_DATA, (BFE_MDIO_SB_START |
344 (BFE_MDIO_OP_READ << BFE_MDIO_OP_SHIFT) |
345 (bfe->bfe_phy_addr << BFE_MDIO_PMD_SHIFT) |
346 (reg << BFE_MDIO_RA_SHIFT) |
347 (BFE_MDIO_TA_VALID << BFE_MDIO_TA_SHIFT)));
348
349 (void) bfe_wait_bit(bfe, BFE_EMAC_ISTAT, BFE_EMAC_INT_MII, 10, 0);
350
351 return ((INL(bfe, BFE_MDIO_DATA) & BFE_MDIO_DATA_DATA));
352 }
353
354 static void
bfe_write_phy(bfe_t * bfe,uint32_t reg,uint32_t val)355 bfe_write_phy(bfe_t *bfe, uint32_t reg, uint32_t val)
356 {
357 OUTL(bfe, BFE_EMAC_ISTAT, BFE_EMAC_INT_MII);
358 OUTL(bfe, BFE_MDIO_DATA, (BFE_MDIO_SB_START |
359 (BFE_MDIO_OP_WRITE << BFE_MDIO_OP_SHIFT) |
360 (bfe->bfe_phy_addr << BFE_MDIO_PMD_SHIFT) |
361 (reg << BFE_MDIO_RA_SHIFT) |
362 (BFE_MDIO_TA_VALID << BFE_MDIO_TA_SHIFT) |
363 (val & BFE_MDIO_DATA_DATA)));
364
365 (void) bfe_wait_bit(bfe, BFE_EMAC_ISTAT, BFE_EMAC_INT_MII, 10, 0);
366 }
367
368 /*
369 * It resets the PHY layer.
370 */
371 static int
bfe_reset_phy(bfe_t * bfe)372 bfe_reset_phy(bfe_t *bfe)
373 {
374 uint32_t i;
375
376 bfe_write_phy(bfe, MII_CONTROL, MII_CONTROL_RESET);
377 drv_usecwait(100);
378 for (i = 0; i < 10; i++) {
379 if (bfe_read_phy(bfe, MII_CONTROL) &
380 MII_CONTROL_RESET) {
381 drv_usecwait(500);
382 continue;
383 }
384
385 break;
386 }
387
388 if (i == 10) {
389 bfe_error(bfe->bfe_dip, "Timeout waiting for PHY to reset");
390 bfe->bfe_phy_state = BFE_PHY_RESET_TIMEOUT;
391 return (BFE_FAILURE);
392 }
393
394 bfe->bfe_phy_state = BFE_PHY_RESET_DONE;
395
396 return (BFE_SUCCESS);
397 }
398
399 /*
400 * Make sure timer function is out of our way and especially during
401 * detach.
402 */
403 static void
bfe_stop_timer(bfe_t * bfe)404 bfe_stop_timer(bfe_t *bfe)
405 {
406 if (bfe->bfe_periodic_id) {
407 ddi_periodic_delete(bfe->bfe_periodic_id);
408 bfe->bfe_periodic_id = NULL;
409 }
410 }
411
412 /*
413 * Stops the PHY
414 */
415 static void
bfe_stop_phy(bfe_t * bfe)416 bfe_stop_phy(bfe_t *bfe)
417 {
418 bfe_write_phy(bfe, MII_CONTROL, MII_CONTROL_PWRDN |
419 MII_CONTROL_ISOLATE);
420
421 bfe->bfe_chip.link = LINK_STATE_UNKNOWN;
422 bfe->bfe_chip.speed = 0;
423 bfe->bfe_chip.duplex = LINK_DUPLEX_UNKNOWN;
424
425 bfe->bfe_phy_state = BFE_PHY_STOPPED;
426
427 /*
428 * Report the link status to MAC layer.
429 */
430 if (bfe->bfe_machdl != NULL)
431 (void) bfe_report_link(bfe);
432 }
433
434 static int
bfe_probe_phy(bfe_t * bfe)435 bfe_probe_phy(bfe_t *bfe)
436 {
437 int phy;
438 uint32_t status;
439
440 if (bfe->bfe_phy_addr) {
441 status = bfe_read_phy(bfe, MII_STATUS);
442 if (status != 0xffff && status != 0) {
443 bfe_write_phy(bfe, MII_CONTROL, 0);
444 return (BFE_SUCCESS);
445 }
446 }
447
448 for (phy = 0; phy < 32; phy++) {
449 bfe->bfe_phy_addr = phy;
450 status = bfe_read_phy(bfe, MII_STATUS);
451 if (status != 0xffff && status != 0) {
452 bfe_write_phy(bfe, MII_CONTROL, 0);
453 return (BFE_SUCCESS);
454 }
455 }
456
457 return (BFE_FAILURE);
458 }
459
460 /*
461 * This timeout function fires at BFE_TIMEOUT_INTERVAL to check the link
462 * status.
463 */
464 static void
bfe_timeout(void * arg)465 bfe_timeout(void *arg)
466 {
467 bfe_t *bfe = (bfe_t *)arg;
468 int resched = 0;
469
470 /*
471 * We don't grab any lock because bfe can't go away.
472 * untimeout() will wait for this timeout instance to complete.
473 */
474 if (bfe->bfe_chip_action & BFE_ACTION_RESTART) {
475 /*
476 * Restart the chip.
477 */
478 bfe_grab_locks(bfe);
479 bfe_chip_restart(bfe);
480 bfe->bfe_chip_action &= ~BFE_ACTION_RESTART;
481 bfe->bfe_chip_action &= ~BFE_ACTION_RESTART_FAULT;
482 bfe->bfe_chip_action &= ~BFE_ACTION_RESTART_PKT;
483 bfe_release_locks(bfe);
484 mac_tx_update(bfe->bfe_machdl);
485 /* Restart will register a new timeout */
486 return;
487 }
488
489 rw_enter(&bfe->bfe_rwlock, RW_READER);
490
491 if (bfe->bfe_chip_state == BFE_CHIP_ACTIVE) {
492 hrtime_t hr;
493
494 hr = gethrtime();
495 if (bfe->bfe_tx_stall_time != 0 &&
496 hr > bfe->bfe_tx_stall_time) {
497 DTRACE_PROBE2(chip__restart, int, bfe->bfe_unit,
498 char *, "pkt timeout");
499 bfe->bfe_chip_action |=
500 (BFE_ACTION_RESTART | BFE_ACTION_RESTART_PKT);
501 bfe->bfe_tx_stall_time = 0;
502 }
503 }
504
505 if (bfe->bfe_phy_state == BFE_PHY_STARTED) {
506 /*
507 * Report the link status to MAC layer if link status changed.
508 */
509 if (bfe_check_link(bfe)) {
510 bfe_report_link(bfe);
511 if (bfe->bfe_chip.link == LINK_STATE_UP) {
512 uint32_t val, flow;
513
514 val = INL(bfe, BFE_TX_CTRL);
515 val &= ~BFE_TX_DUPLEX;
516 if (bfe->bfe_chip.duplex == LINK_DUPLEX_FULL) {
517 val |= BFE_TX_DUPLEX;
518 flow = INL(bfe, BFE_RXCONF);
519 flow &= ~BFE_RXCONF_FLOW;
520 OUTL(bfe, BFE_RXCONF, flow);
521
522 flow = INL(bfe, BFE_MAC_FLOW);
523 flow &= ~(BFE_FLOW_RX_HIWAT);
524 OUTL(bfe, BFE_MAC_FLOW, flow);
525 }
526
527 resched = 1;
528
529 OUTL(bfe, BFE_TX_CTRL, val);
530 DTRACE_PROBE1(link__up,
531 int, bfe->bfe_unit);
532 }
533 }
534 }
535
536 rw_exit(&bfe->bfe_rwlock);
537
538 if (resched)
539 mac_tx_update(bfe->bfe_machdl);
540 }
541
542 /*
543 * Starts PHY layer.
544 */
545 static int
bfe_startup_phy(bfe_t * bfe)546 bfe_startup_phy(bfe_t *bfe)
547 {
548 uint16_t bmsr, bmcr, anar;
549 int prog, s;
550 int phyid1, phyid2;
551
552 if (bfe_probe_phy(bfe) == BFE_FAILURE) {
553 bfe->bfe_phy_state = BFE_PHY_NOTFOUND;
554 return (BFE_FAILURE);
555 }
556
557 (void) bfe_reset_phy(bfe);
558
559 phyid1 = bfe_read_phy(bfe, MII_PHYIDH);
560 phyid2 = bfe_read_phy(bfe, MII_PHYIDL);
561 bfe->bfe_phy_id = (phyid1 << 16) | phyid2;
562
563 bmsr = bfe_read_phy(bfe, MII_STATUS);
564 anar = bfe_read_phy(bfe, MII_AN_ADVERT);
565
566 again:
567 anar &= ~(MII_ABILITY_100BASE_T4 |
568 MII_ABILITY_100BASE_TX_FD | MII_ABILITY_100BASE_TX |
569 MII_ABILITY_10BASE_T_FD | MII_ABILITY_10BASE_T);
570
571 /*
572 * Supported hardware modes are in bmsr.
573 */
574 bfe->bfe_chip.bmsr = bmsr;
575
576 /*
577 * Assume no capabilities are supported in the hardware.
578 */
579 bfe->bfe_cap_aneg = bfe->bfe_cap_100T4 =
580 bfe->bfe_cap_100fdx = bfe->bfe_cap_100hdx =
581 bfe->bfe_cap_10fdx = bfe->bfe_cap_10hdx = 0;
582
583 /*
584 * Assume property is set.
585 */
586 s = 1;
587 if (!(bfe->bfe_chip_action & BFE_ACTION_RESTART_SETPROP)) {
588 /*
589 * Property is not set which means bfe_mac_setprop()
590 * is not called on us.
591 */
592 s = 0;
593 }
594
595 bmcr = prog = 0;
596
597 if (bmsr & MII_STATUS_100_BASEX_FD) {
598 bfe->bfe_cap_100fdx = 1;
599 if (s == 0) {
600 anar |= MII_ABILITY_100BASE_TX_FD;
601 bfe->bfe_adv_100fdx = 1;
602 prog++;
603 } else if (bfe->bfe_adv_100fdx) {
604 anar |= MII_ABILITY_100BASE_TX_FD;
605 prog++;
606 }
607 }
608
609 if (bmsr & MII_STATUS_100_BASE_T4) {
610 bfe->bfe_cap_100T4 = 1;
611 if (s == 0) {
612 anar |= MII_ABILITY_100BASE_T4;
613 bfe->bfe_adv_100T4 = 1;
614 prog++;
615 } else if (bfe->bfe_adv_100T4) {
616 anar |= MII_ABILITY_100BASE_T4;
617 prog++;
618 }
619 }
620
621 if (bmsr & MII_STATUS_100_BASEX) {
622 bfe->bfe_cap_100hdx = 1;
623 if (s == 0) {
624 anar |= MII_ABILITY_100BASE_TX;
625 bfe->bfe_adv_100hdx = 1;
626 prog++;
627 } else if (bfe->bfe_adv_100hdx) {
628 anar |= MII_ABILITY_100BASE_TX;
629 prog++;
630 }
631 }
632
633 if (bmsr & MII_STATUS_10_FD) {
634 bfe->bfe_cap_10fdx = 1;
635 if (s == 0) {
636 anar |= MII_ABILITY_10BASE_T_FD;
637 bfe->bfe_adv_10fdx = 1;
638 prog++;
639 } else if (bfe->bfe_adv_10fdx) {
640 anar |= MII_ABILITY_10BASE_T_FD;
641 prog++;
642 }
643 }
644
645 if (bmsr & MII_STATUS_10) {
646 bfe->bfe_cap_10hdx = 1;
647 if (s == 0) {
648 anar |= MII_ABILITY_10BASE_T;
649 bfe->bfe_adv_10hdx = 1;
650 prog++;
651 } else if (bfe->bfe_adv_10hdx) {
652 anar |= MII_ABILITY_10BASE_T;
653 prog++;
654 }
655 }
656
657 if (bmsr & MII_STATUS_CANAUTONEG) {
658 bfe->bfe_cap_aneg = 1;
659 if (s == 0) {
660 bfe->bfe_adv_aneg = 1;
661 }
662 }
663
664 if (prog == 0) {
665 if (s == 0) {
666 bfe_error(bfe->bfe_dip,
667 "No valid link mode selected. Powering down PHY");
668 bfe_stop_phy(bfe);
669 bfe_report_link(bfe);
670 return (BFE_FAILURE);
671 }
672
673 /*
674 * If property is set then user would have goofed up. So we
675 * go back to default properties.
676 */
677 bfe->bfe_chip_action &= ~BFE_ACTION_RESTART_SETPROP;
678 goto again;
679 }
680
681 if (bfe->bfe_adv_aneg && (bmsr & MII_STATUS_CANAUTONEG)) {
682 bmcr = (MII_CONTROL_ANE | MII_CONTROL_RSAN);
683 } else {
684 if (bfe->bfe_adv_100fdx)
685 bmcr = (MII_CONTROL_100MB | MII_CONTROL_FDUPLEX);
686 else if (bfe->bfe_adv_100hdx)
687 bmcr = MII_CONTROL_100MB;
688 else if (bfe->bfe_adv_10fdx)
689 bmcr = MII_CONTROL_FDUPLEX;
690 else
691 bmcr = 0; /* 10HDX */
692 }
693
694 if (prog)
695 bfe_write_phy(bfe, MII_AN_ADVERT, anar);
696
697 if (bmcr)
698 bfe_write_phy(bfe, MII_CONTROL, bmcr);
699
700 bfe->bfe_mii_anar = anar;
701 bfe->bfe_mii_bmcr = bmcr;
702 bfe->bfe_phy_state = BFE_PHY_STARTED;
703
704 if (bfe->bfe_periodic_id == NULL) {
705 bfe->bfe_periodic_id = ddi_periodic_add(bfe_timeout,
706 (void *)bfe, BFE_TIMEOUT_INTERVAL, DDI_IPL_0);
707
708 DTRACE_PROBE1(first__timeout, int, bfe->bfe_unit);
709 }
710
711 DTRACE_PROBE4(phy_started, int, bfe->bfe_unit,
712 int, bmsr, int, bmcr, int, anar);
713
714 return (BFE_SUCCESS);
715 }
716
717 /*
718 * Reports link status back to MAC Layer.
719 */
720 static void
bfe_report_link(bfe_t * bfe)721 bfe_report_link(bfe_t *bfe)
722 {
723 mac_link_update(bfe->bfe_machdl, bfe->bfe_chip.link);
724 }
725
726 /*
727 * Reads PHY/MII registers and get the link status for us.
728 */
729 static int
bfe_check_link(bfe_t * bfe)730 bfe_check_link(bfe_t *bfe)
731 {
732 uint16_t bmsr, bmcr, anar, anlpar;
733 int speed, duplex, link;
734
735 speed = bfe->bfe_chip.speed;
736 duplex = bfe->bfe_chip.duplex;
737 link = bfe->bfe_chip.link;
738
739 bmsr = bfe_read_phy(bfe, MII_STATUS);
740 bfe->bfe_mii_bmsr = bmsr;
741
742 bmcr = bfe_read_phy(bfe, MII_CONTROL);
743
744 anar = bfe_read_phy(bfe, MII_AN_ADVERT);
745 bfe->bfe_mii_anar = anar;
746
747 anlpar = bfe_read_phy(bfe, MII_AN_LPABLE);
748 bfe->bfe_mii_anlpar = anlpar;
749
750 bfe->bfe_mii_exp = bfe_read_phy(bfe, MII_AN_EXPANSION);
751
752 /*
753 * If exp register is not present in PHY.
754 */
755 if (bfe->bfe_mii_exp == 0xffff) {
756 bfe->bfe_mii_exp = 0;
757 }
758
759 if ((bmsr & MII_STATUS_LINKUP) == 0) {
760 bfe->bfe_chip.link = LINK_STATE_DOWN;
761 bfe->bfe_chip.speed = 0;
762 bfe->bfe_chip.duplex = LINK_DUPLEX_UNKNOWN;
763 goto done;
764 }
765
766 bfe->bfe_chip.link = LINK_STATE_UP;
767
768 if (!(bmcr & MII_CONTROL_ANE)) {
769 /* Forced mode */
770 if (bmcr & MII_CONTROL_100MB)
771 bfe->bfe_chip.speed = 100000000;
772 else
773 bfe->bfe_chip.speed = 10000000;
774
775 if (bmcr & MII_CONTROL_FDUPLEX)
776 bfe->bfe_chip.duplex = LINK_DUPLEX_FULL;
777 else
778 bfe->bfe_chip.duplex = LINK_DUPLEX_HALF;
779
780 } else if ((!(bmsr & MII_STATUS_CANAUTONEG)) ||
781 (!(bmsr & MII_STATUS_ANDONE))) {
782 bfe->bfe_chip.speed = 0;
783 bfe->bfe_chip.duplex = LINK_DUPLEX_UNKNOWN;
784 } else if (anar & anlpar & MII_ABILITY_100BASE_TX_FD) {
785 bfe->bfe_chip.speed = 100000000;
786 bfe->bfe_chip.duplex = LINK_DUPLEX_FULL;
787 } else if (anar & anlpar & MII_ABILITY_100BASE_T4) {
788 bfe->bfe_chip.speed = 100000000;
789 bfe->bfe_chip.duplex = LINK_DUPLEX_HALF;
790 } else if (anar & anlpar & MII_ABILITY_100BASE_TX) {
791 bfe->bfe_chip.speed = 100000000;
792 bfe->bfe_chip.duplex = LINK_DUPLEX_HALF;
793 } else if (anar & anlpar & MII_ABILITY_10BASE_T_FD) {
794 bfe->bfe_chip.speed = 10000000;
795 bfe->bfe_chip.duplex = LINK_DUPLEX_FULL;
796 } else if (anar & anlpar & MII_ABILITY_10BASE_T) {
797 bfe->bfe_chip.speed = 10000000;
798 bfe->bfe_chip.duplex = LINK_DUPLEX_HALF;
799 } else {
800 bfe->bfe_chip.speed = 0;
801 bfe->bfe_chip.duplex = LINK_DUPLEX_UNKNOWN;
802 }
803
804 done:
805 /*
806 * If speed or link status or duplex mode changed then report to
807 * MAC layer which is done by the caller.
808 */
809 if (speed != bfe->bfe_chip.speed ||
810 duplex != bfe->bfe_chip.duplex ||
811 link != bfe->bfe_chip.link) {
812 return (1);
813 }
814
815 return (0);
816 }
817
818 static void
bfe_cam_write(bfe_t * bfe,uchar_t * d,int index)819 bfe_cam_write(bfe_t *bfe, uchar_t *d, int index)
820 {
821 uint32_t v;
822
823 v = ((uint32_t)d[2] << 24);
824 v |= ((uint32_t)d[3] << 16);
825 v |= ((uint32_t)d[4] << 8);
826 v |= (uint32_t)d[5];
827
828 OUTL(bfe, BFE_CAM_DATA_LO, v);
829 v = (BFE_CAM_HI_VALID |
830 (((uint32_t)d[0]) << 8) |
831 (((uint32_t)d[1])));
832
833 OUTL(bfe, BFE_CAM_DATA_HI, v);
834 OUTL(bfe, BFE_CAM_CTRL, (BFE_CAM_WRITE |
835 ((uint32_t)index << BFE_CAM_INDEX_SHIFT)));
836 (void) bfe_wait_bit(bfe, BFE_CAM_CTRL, BFE_CAM_BUSY, 10, 1);
837 }
838
839 /*
840 * Chip related functions (halt, reset, start).
841 */
842 static void
bfe_chip_halt(bfe_t * bfe)843 bfe_chip_halt(bfe_t *bfe)
844 {
845 /*
846 * Disables interrupts.
847 */
848 OUTL(bfe, BFE_INTR_MASK, 0);
849 FLUSH(bfe, BFE_INTR_MASK);
850
851 OUTL(bfe, BFE_ENET_CTRL, BFE_ENET_DISABLE);
852
853 /*
854 * Wait until TX and RX finish their job.
855 */
856 (void) bfe_wait_bit(bfe, BFE_ENET_CTRL, BFE_ENET_DISABLE, 20, 1);
857
858 /*
859 * Disables DMA engine.
860 */
861 OUTL(bfe, BFE_DMARX_CTRL, 0);
862 OUTL(bfe, BFE_DMATX_CTRL, 0);
863
864 drv_usecwait(10);
865
866 bfe->bfe_chip_state = BFE_CHIP_HALT;
867 }
868
869 static void
bfe_chip_restart(bfe_t * bfe)870 bfe_chip_restart(bfe_t *bfe)
871 {
872 DTRACE_PROBE2(chip__restart, int, bfe->bfe_unit,
873 int, bfe->bfe_chip_action);
874
875 /*
876 * Halt chip and PHY.
877 */
878 bfe_chip_halt(bfe);
879 bfe_stop_phy(bfe);
880 bfe->bfe_chip_state = BFE_CHIP_STOPPED;
881
882 /*
883 * Init variables.
884 */
885 bfe_init_vars(bfe);
886
887 /*
888 * Reset chip and start PHY.
889 */
890 bfe_chip_reset(bfe);
891
892 /*
893 * DMA descriptor rings.
894 */
895 bfe_tx_desc_init(&bfe->bfe_tx_ring);
896 bfe_rx_desc_init(&bfe->bfe_rx_ring);
897
898 bfe->bfe_chip_state = BFE_CHIP_ACTIVE;
899 bfe_set_rx_mode(bfe);
900 bfe_enable_chip_intrs(bfe);
901 }
902
903 /*
904 * Disables core by stopping the clock.
905 */
906 static void
bfe_core_disable(bfe_t * bfe)907 bfe_core_disable(bfe_t *bfe)
908 {
909 if ((INL(bfe, BFE_SBTMSLOW) & BFE_RESET))
910 return;
911
912 OUTL(bfe, BFE_SBTMSLOW, (BFE_REJECT | BFE_CLOCK));
913 (void) bfe_wait_bit(bfe, BFE_SBTMSLOW, BFE_REJECT, 100, 0);
914 (void) bfe_wait_bit(bfe, BFE_SBTMSHIGH, BFE_BUSY, 100, 1);
915 OUTL(bfe, BFE_SBTMSLOW, (BFE_FGC | BFE_CLOCK | BFE_REJECT | BFE_RESET));
916 FLUSH(bfe, BFE_SBTMSLOW);
917 drv_usecwait(10);
918 OUTL(bfe, BFE_SBTMSLOW, (BFE_REJECT | BFE_RESET));
919 drv_usecwait(10);
920 }
921
922 /*
923 * Resets core.
924 */
925 static void
bfe_core_reset(bfe_t * bfe)926 bfe_core_reset(bfe_t *bfe)
927 {
928 uint32_t val;
929
930 /*
931 * First disable the core.
932 */
933 bfe_core_disable(bfe);
934
935 OUTL(bfe, BFE_SBTMSLOW, (BFE_RESET | BFE_CLOCK | BFE_FGC));
936 FLUSH(bfe, BFE_SBTMSLOW);
937 drv_usecwait(1);
938
939 if (INL(bfe, BFE_SBTMSHIGH) & BFE_SERR)
940 OUTL(bfe, BFE_SBTMSHIGH, 0);
941
942 val = INL(bfe, BFE_SBIMSTATE);
943 if (val & (BFE_IBE | BFE_TO))
944 OUTL(bfe, BFE_SBIMSTATE, val & ~(BFE_IBE | BFE_TO));
945
946 OUTL(bfe, BFE_SBTMSLOW, (BFE_CLOCK | BFE_FGC));
947 FLUSH(bfe, BFE_SBTMSLOW);
948 drv_usecwait(1);
949
950 OUTL(bfe, BFE_SBTMSLOW, BFE_CLOCK);
951 FLUSH(bfe, BFE_SBTMSLOW);
952 drv_usecwait(1);
953 }
954
955 static void
bfe_setup_config(bfe_t * bfe,uint32_t cores)956 bfe_setup_config(bfe_t *bfe, uint32_t cores)
957 {
958 uint32_t bar_orig, val;
959
960 /*
961 * Change bar0 window to map sbtopci registers.
962 */
963 bar_orig = pci_config_get32(bfe->bfe_conf_handle, BFE_BAR0_WIN);
964 pci_config_put32(bfe->bfe_conf_handle, BFE_BAR0_WIN, BFE_REG_PCI);
965
966 /* Just read it and don't do anything */
967 val = INL(bfe, BFE_SBIDHIGH) & BFE_IDH_CORE;
968
969 val = INL(bfe, BFE_SBINTVEC);
970 val |= cores;
971 OUTL(bfe, BFE_SBINTVEC, val);
972
973 val = INL(bfe, BFE_SSB_PCI_TRANS_2);
974 val |= BFE_SSB_PCI_PREF | BFE_SSB_PCI_BURST;
975 OUTL(bfe, BFE_SSB_PCI_TRANS_2, val);
976
977 /*
978 * Restore bar0 window mapping.
979 */
980 pci_config_put32(bfe->bfe_conf_handle, BFE_BAR0_WIN, bar_orig);
981 }
982
983 /*
984 * Resets chip and starts PHY.
985 */
986 static void
bfe_chip_reset(bfe_t * bfe)987 bfe_chip_reset(bfe_t *bfe)
988 {
989 uint32_t val;
990
991 /* Set the interrupt vector for the enet core */
992 bfe_setup_config(bfe, BFE_INTVEC_ENET0);
993
994 /* check if core is up */
995 val = INL(bfe, BFE_SBTMSLOW) &
996 (BFE_RESET | BFE_REJECT | BFE_CLOCK);
997
998 if (val == BFE_CLOCK) {
999 OUTL(bfe, BFE_RCV_LAZY, 0);
1000 OUTL(bfe, BFE_ENET_CTRL, BFE_ENET_DISABLE);
1001 (void) bfe_wait_bit(bfe, BFE_ENET_CTRL,
1002 BFE_ENET_DISABLE, 10, 1);
1003 OUTL(bfe, BFE_DMATX_CTRL, 0);
1004 FLUSH(bfe, BFE_DMARX_STAT);
1005 drv_usecwait(20000); /* 20 milli seconds */
1006 if (INL(bfe, BFE_DMARX_STAT) & BFE_STAT_EMASK) {
1007 (void) bfe_wait_bit(bfe, BFE_DMARX_STAT, BFE_STAT_SIDLE,
1008 10, 0);
1009 }
1010 OUTL(bfe, BFE_DMARX_CTRL, 0);
1011 }
1012
1013 bfe_core_reset(bfe);
1014 bfe_clear_stats(bfe);
1015
1016 OUTL(bfe, BFE_MDIO_CTRL, 0x8d);
1017 val = INL(bfe, BFE_DEVCTRL);
1018 if (!(val & BFE_IPP))
1019 OUTL(bfe, BFE_ENET_CTRL, BFE_ENET_EPSEL);
1020 else if (INL(bfe, BFE_DEVCTRL & BFE_EPR)) {
1021 OUTL_AND(bfe, BFE_DEVCTRL, ~BFE_EPR);
1022 drv_usecwait(20000); /* 20 milli seconds */
1023 }
1024
1025 OUTL_OR(bfe, BFE_MAC_CTRL, BFE_CTRL_CRC32_ENAB | BFE_CTRL_LED);
1026
1027 OUTL_AND(bfe, BFE_MAC_CTRL, ~BFE_CTRL_PDOWN);
1028
1029 OUTL(bfe, BFE_RCV_LAZY, ((1 << BFE_LAZY_FC_SHIFT) &
1030 BFE_LAZY_FC_MASK));
1031
1032 OUTL_OR(bfe, BFE_RCV_LAZY, 0);
1033
1034 OUTL(bfe, BFE_RXMAXLEN, bfe->bfe_rx_ring.r_buf_len);
1035 OUTL(bfe, BFE_TXMAXLEN, bfe->bfe_tx_ring.r_buf_len);
1036
1037 OUTL(bfe, BFE_TX_WMARK, 56);
1038
1039 /* Program DMA channels */
1040 OUTL(bfe, BFE_DMATX_CTRL, BFE_TX_CTRL_ENABLE);
1041
1042 /*
1043 * DMA addresses need to be added to BFE_PCI_DMA
1044 */
1045 OUTL(bfe, BFE_DMATX_ADDR,
1046 bfe->bfe_tx_ring.r_desc_cookie.dmac_laddress + BFE_PCI_DMA);
1047
1048 OUTL(bfe, BFE_DMARX_CTRL, (BFE_RX_OFFSET << BFE_RX_CTRL_ROSHIFT)
1049 | BFE_RX_CTRL_ENABLE);
1050
1051 OUTL(bfe, BFE_DMARX_ADDR,
1052 bfe->bfe_rx_ring.r_desc_cookie.dmac_laddress + BFE_PCI_DMA);
1053
1054 (void) bfe_startup_phy(bfe);
1055
1056 bfe->bfe_chip_state = BFE_CHIP_INITIALIZED;
1057 }
1058
1059 /*
1060 * It enables interrupts. Should be the last step while starting chip.
1061 */
1062 static void
bfe_enable_chip_intrs(bfe_t * bfe)1063 bfe_enable_chip_intrs(bfe_t *bfe)
1064 {
1065 /* Enable the chip and core */
1066 OUTL(bfe, BFE_ENET_CTRL, BFE_ENET_ENABLE);
1067
1068 /* Enable interrupts */
1069 OUTL(bfe, BFE_INTR_MASK, BFE_IMASK_DEF);
1070 }
1071
1072 /*
1073 * Common code to take care of setting RX side mode (filter).
1074 */
1075 static void
bfe_set_rx_mode(bfe_t * bfe)1076 bfe_set_rx_mode(bfe_t *bfe)
1077 {
1078 uint32_t val;
1079 int i;
1080 ether_addr_t mac[ETHERADDRL] = {0, 0, 0, 0, 0, 0};
1081
1082 /*
1083 * We don't touch RX filter if we were asked to suspend. It's fine
1084 * if chip is not active (no interface is plumbed on us).
1085 */
1086 if (bfe->bfe_chip_state == BFE_CHIP_SUSPENDED)
1087 return;
1088
1089 val = INL(bfe, BFE_RXCONF);
1090
1091 val &= ~BFE_RXCONF_PROMISC;
1092 val &= ~BFE_RXCONF_DBCAST;
1093
1094 if ((bfe->bfe_chip_mode & BFE_RX_MODE_ENABLE) == 0) {
1095 OUTL(bfe, BFE_CAM_CTRL, 0);
1096 FLUSH(bfe, BFE_CAM_CTRL);
1097 } else if (bfe->bfe_chip_mode & BFE_RX_MODE_PROMISC) {
1098 val |= BFE_RXCONF_PROMISC;
1099 val &= ~BFE_RXCONF_DBCAST;
1100 } else {
1101 if (bfe->bfe_chip_state == BFE_CHIP_ACTIVE) {
1102 /* Flush everything */
1103 OUTL(bfe, BFE_RXCONF, val |
1104 BFE_RXCONF_PROMISC | BFE_RXCONF_ALLMULTI);
1105 FLUSH(bfe, BFE_RXCONF);
1106 }
1107
1108 /* Disable CAM */
1109 OUTL(bfe, BFE_CAM_CTRL, 0);
1110 FLUSH(bfe, BFE_CAM_CTRL);
1111
1112 /*
1113 * We receive all multicast packets.
1114 */
1115 val |= BFE_RXCONF_ALLMULTI;
1116
1117 for (i = 0; i < BFE_MAX_MULTICAST_TABLE - 1; i++) {
1118 bfe_cam_write(bfe, (uchar_t *)mac, i);
1119 }
1120
1121 bfe_cam_write(bfe, bfe->bfe_ether_addr, i);
1122
1123 /* Enable CAM */
1124 OUTL_OR(bfe, BFE_CAM_CTRL, BFE_CAM_ENABLE);
1125 FLUSH(bfe, BFE_CAM_CTRL);
1126 }
1127
1128 DTRACE_PROBE2(rx__mode__filter, int, bfe->bfe_unit,
1129 int, val);
1130
1131 OUTL(bfe, BFE_RXCONF, val);
1132 FLUSH(bfe, BFE_RXCONF);
1133 }
1134
1135 /*
1136 * Reset various variable values to initial state.
1137 */
1138 static void
bfe_init_vars(bfe_t * bfe)1139 bfe_init_vars(bfe_t *bfe)
1140 {
1141 bfe->bfe_chip_mode = BFE_RX_MODE_ENABLE;
1142
1143 /* Initial assumption */
1144 bfe->bfe_chip.link = LINK_STATE_UNKNOWN;
1145 bfe->bfe_chip.speed = 0;
1146 bfe->bfe_chip.duplex = LINK_DUPLEX_UNKNOWN;
1147
1148 bfe->bfe_periodic_id = NULL;
1149 bfe->bfe_chip_state = BFE_CHIP_UNINITIALIZED;
1150
1151 bfe->bfe_tx_stall_time = 0;
1152 }
1153
1154 /*
1155 * Initializes TX side descriptor entries (bfe_desc_t). Each descriptor entry
1156 * has control (desc_ctl) and address (desc_addr) member.
1157 */
1158 static void
bfe_tx_desc_init(bfe_ring_t * r)1159 bfe_tx_desc_init(bfe_ring_t *r)
1160 {
1161 int i;
1162 uint32_t v;
1163
1164 for (i = 0; i < r->r_ndesc; i++) {
1165 PUT_DESC(r, (uint32_t *)&(r->r_desc[i].desc_ctl),
1166 (r->r_buf_dma[i].len & BFE_DESC_LEN));
1167
1168 /*
1169 * DMA addresses need to be added to BFE_PCI_DMA
1170 */
1171 PUT_DESC(r, (uint32_t *)&(r->r_desc[i].desc_addr),
1172 (r->r_buf_dma[i].cookie.dmac_laddress + BFE_PCI_DMA));
1173 }
1174
1175 v = GET_DESC(r, (uint32_t *)&(r->r_desc[i - 1].desc_ctl));
1176 PUT_DESC(r, (uint32_t *)&(r->r_desc[i - 1].desc_ctl),
1177 v | BFE_DESC_EOT);
1178
1179 (void) SYNC_DESC(r, 0, r->r_ndesc, DDI_DMA_SYNC_FORDEV);
1180
1181 r->r_curr_desc = 0;
1182 r->r_avail_desc = TX_NUM_DESC;
1183 r->r_cons_desc = 0;
1184 }
1185
1186 /*
1187 * Initializes RX side descriptor entries (bfe_desc_t). Each descriptor entry
1188 * has control (desc_ctl) and address (desc_addr) member.
1189 */
1190 static void
bfe_rx_desc_init(bfe_ring_t * r)1191 bfe_rx_desc_init(bfe_ring_t *r)
1192 {
1193 int i;
1194 uint32_t v;
1195
1196 for (i = 0; i < r->r_ndesc; i++) {
1197 PUT_DESC(r, (uint32_t *)&(r->r_desc[i].desc_ctl),
1198 (r->r_buf_dma[i].len& BFE_DESC_LEN));
1199
1200 PUT_DESC(r, (uint32_t *)&(r->r_desc[i].desc_addr),
1201 (r->r_buf_dma[i].cookie.dmac_laddress + BFE_PCI_DMA));
1202
1203 /* Initialize rx header (len, flags) */
1204 bzero(r->r_buf_dma[i].addr, sizeof (bfe_rx_header_t));
1205
1206 (void) SYNC_BUF(r, i, 0, sizeof (bfe_rx_header_t),
1207 DDI_DMA_SYNC_FORDEV);
1208 }
1209
1210 v = GET_DESC(r, (uint32_t *)&(r->r_desc[i - 1].desc_ctl));
1211 PUT_DESC(r, (uint32_t *)&(r->r_desc[i - 1].desc_ctl),
1212 v | BFE_DESC_EOT);
1213
1214 (void) SYNC_DESC(r, 0, r->r_ndesc, DDI_DMA_SYNC_FORDEV);
1215
1216 /* TAIL of RX Descriptor */
1217 OUTL(r->r_bfe, BFE_DMARX_PTR, ((i) * sizeof (bfe_desc_t)));
1218
1219 r->r_curr_desc = 0;
1220 r->r_avail_desc = RX_NUM_DESC;
1221 }
1222
1223 static int
bfe_chip_start(bfe_t * bfe)1224 bfe_chip_start(bfe_t *bfe)
1225 {
1226 ASSERT_ALL_LOCKS(bfe);
1227
1228 /*
1229 * Stop the chip first & then Reset the chip. At last enable interrupts.
1230 */
1231 bfe_chip_halt(bfe);
1232 bfe_stop_phy(bfe);
1233
1234 /*
1235 * Reset chip and start PHY.
1236 */
1237 bfe_chip_reset(bfe);
1238
1239 /*
1240 * Initailize Descriptor Rings.
1241 */
1242 bfe_tx_desc_init(&bfe->bfe_tx_ring);
1243 bfe_rx_desc_init(&bfe->bfe_rx_ring);
1244
1245 bfe->bfe_chip_state = BFE_CHIP_ACTIVE;
1246 bfe->bfe_chip_mode |= BFE_RX_MODE_ENABLE;
1247 bfe_set_rx_mode(bfe);
1248 bfe_enable_chip_intrs(bfe);
1249
1250 /* Check link, speed and duplex mode */
1251 (void) bfe_check_link(bfe);
1252
1253 return (DDI_SUCCESS);
1254 }
1255
1256
1257 /*
1258 * Clear chip statistics.
1259 */
1260 static void
bfe_clear_stats(bfe_t * bfe)1261 bfe_clear_stats(bfe_t *bfe)
1262 {
1263 ulong_t r;
1264
1265 OUTL(bfe, BFE_MIB_CTRL, BFE_MIB_CLR_ON_READ);
1266
1267 /*
1268 * Stat registers are cleared by reading.
1269 */
1270 for (r = BFE_TX_GOOD_O; r <= BFE_TX_PAUSE; r += 4)
1271 (void) INL(bfe, r);
1272
1273 for (r = BFE_RX_GOOD_O; r <= BFE_RX_NPAUSE; r += 4)
1274 (void) INL(bfe, r);
1275 }
1276
1277 /*
1278 * Collect chip statistics.
1279 */
1280 static void
bfe_gather_stats(bfe_t * bfe)1281 bfe_gather_stats(bfe_t *bfe)
1282 {
1283 ulong_t r;
1284 uint32_t *v;
1285 uint32_t txerr = 0, rxerr = 0, coll = 0;
1286
1287 v = &bfe->bfe_hw_stats.tx_good_octets;
1288 for (r = BFE_TX_GOOD_O; r <= BFE_TX_PAUSE; r += 4) {
1289 *v += INL(bfe, r);
1290 v++;
1291 }
1292
1293 v = &bfe->bfe_hw_stats.rx_good_octets;
1294 for (r = BFE_RX_GOOD_O; r <= BFE_RX_NPAUSE; r += 4) {
1295 *v += INL(bfe, r);
1296 v++;
1297 }
1298
1299 /*
1300 * TX :
1301 * -------
1302 * tx_good_octets, tx_good_pkts, tx_octets
1303 * tx_pkts, tx_broadcast_pkts, tx_multicast_pkts
1304 * tx_len_64, tx_len_65_to_127, tx_len_128_to_255
1305 * tx_len_256_to_511, tx_len_512_to_1023, tx_len_1024_to_max
1306 * tx_jabber_pkts, tx_oversize_pkts, tx_fragment_pkts
1307 * tx_underruns, tx_total_cols, tx_single_cols
1308 * tx_multiple_cols, tx_excessive_cols, tx_late_cols
1309 * tx_defered, tx_carrier_lost, tx_pause_pkts
1310 *
1311 * RX :
1312 * -------
1313 * rx_good_octets, rx_good_pkts, rx_octets
1314 * rx_pkts, rx_broadcast_pkts, rx_multicast_pkts
1315 * rx_len_64, rx_len_65_to_127, rx_len_128_to_255
1316 * rx_len_256_to_511, rx_len_512_to_1023, rx_len_1024_to_max
1317 * rx_jabber_pkts, rx_oversize_pkts, rx_fragment_pkts
1318 * rx_missed_pkts, rx_crc_align_errs, rx_undersize
1319 * rx_crc_errs, rx_align_errs, rx_symbol_errs
1320 * rx_pause_pkts, rx_nonpause_pkts
1321 */
1322
1323 bfe->bfe_stats.ether_stat_carrier_errors =
1324 bfe->bfe_hw_stats.tx_carrier_lost;
1325
1326 /* txerr += bfe->bfe_hw_stats.tx_carrier_lost; */
1327
1328 bfe->bfe_stats.ether_stat_ex_collisions =
1329 bfe->bfe_hw_stats.tx_excessive_cols;
1330 txerr += bfe->bfe_hw_stats.tx_excessive_cols;
1331 coll += bfe->bfe_hw_stats.tx_excessive_cols;
1332
1333 bfe->bfe_stats.ether_stat_fcs_errors =
1334 bfe->bfe_hw_stats.rx_crc_errs;
1335 rxerr += bfe->bfe_hw_stats.rx_crc_errs;
1336
1337 bfe->bfe_stats.ether_stat_first_collisions =
1338 bfe->bfe_hw_stats.tx_single_cols;
1339 coll += bfe->bfe_hw_stats.tx_single_cols;
1340 bfe->bfe_stats.ether_stat_multi_collisions =
1341 bfe->bfe_hw_stats.tx_multiple_cols;
1342 coll += bfe->bfe_hw_stats.tx_multiple_cols;
1343
1344 bfe->bfe_stats.ether_stat_toolong_errors =
1345 bfe->bfe_hw_stats.rx_oversize_pkts;
1346 rxerr += bfe->bfe_hw_stats.rx_oversize_pkts;
1347
1348 bfe->bfe_stats.ether_stat_tooshort_errors =
1349 bfe->bfe_hw_stats.rx_undersize;
1350 rxerr += bfe->bfe_hw_stats.rx_undersize;
1351
1352 bfe->bfe_stats.ether_stat_tx_late_collisions +=
1353 bfe->bfe_hw_stats.tx_late_cols;
1354
1355 bfe->bfe_stats.ether_stat_defer_xmts +=
1356 bfe->bfe_hw_stats.tx_defered;
1357
1358 bfe->bfe_stats.ether_stat_macrcv_errors += rxerr;
1359 bfe->bfe_stats.ether_stat_macxmt_errors += txerr;
1360
1361 bfe->bfe_stats.collisions += coll;
1362 }
1363
1364 /*
1365 * Gets the state for dladm command and all.
1366 */
1367 int
bfe_mac_getstat(void * arg,uint_t stat,uint64_t * val)1368 bfe_mac_getstat(void *arg, uint_t stat, uint64_t *val)
1369 {
1370 bfe_t *bfe = (bfe_t *)arg;
1371 uint64_t v;
1372 int err = 0;
1373
1374 rw_enter(&bfe->bfe_rwlock, RW_READER);
1375
1376
1377 switch (stat) {
1378 default:
1379 err = ENOTSUP;
1380 break;
1381
1382 case MAC_STAT_IFSPEED:
1383 /*
1384 * MAC layer will ask for IFSPEED first and hence we
1385 * collect it only once.
1386 */
1387 if (bfe->bfe_chip_state == BFE_CHIP_ACTIVE) {
1388 /*
1389 * Update stats from the hardware.
1390 */
1391 bfe_gather_stats(bfe);
1392 }
1393 v = bfe->bfe_chip.speed;
1394 break;
1395
1396 case ETHER_STAT_ADV_CAP_100T4:
1397 v = bfe->bfe_adv_100T4;
1398 break;
1399
1400 case ETHER_STAT_ADV_CAP_100FDX:
1401 v = (bfe->bfe_mii_anar & MII_ABILITY_100BASE_TX_FD) != 0;
1402 break;
1403
1404 case ETHER_STAT_ADV_CAP_100HDX:
1405 v = (bfe->bfe_mii_anar & MII_ABILITY_100BASE_TX) != 0;
1406 break;
1407
1408 case ETHER_STAT_ADV_CAP_10FDX:
1409 v = (bfe->bfe_mii_anar & MII_ABILITY_10BASE_T_FD) != 0;
1410 break;
1411
1412 case ETHER_STAT_ADV_CAP_10HDX:
1413 v = (bfe->bfe_mii_anar & MII_ABILITY_10BASE_T) != 0;
1414 break;
1415
1416 case ETHER_STAT_ADV_CAP_ASMPAUSE:
1417 v = 0;
1418 break;
1419
1420 case ETHER_STAT_ADV_CAP_AUTONEG:
1421 v = bfe->bfe_adv_aneg;
1422 break;
1423
1424 case ETHER_STAT_ADV_CAP_PAUSE:
1425 v = (bfe->bfe_mii_anar & MII_ABILITY_PAUSE) != 0;
1426 break;
1427
1428 case ETHER_STAT_ADV_REMFAULT:
1429 v = (bfe->bfe_mii_anar & MII_AN_ADVERT_REMFAULT) != 0;
1430 break;
1431
1432 case ETHER_STAT_ALIGN_ERRORS:
1433 /* MIB */
1434 v = bfe->bfe_stats.ether_stat_align_errors;
1435 break;
1436
1437 case ETHER_STAT_CAP_100T4:
1438 v = (bfe->bfe_mii_bmsr & MII_STATUS_100_BASE_T4) != 0;
1439 break;
1440
1441 case ETHER_STAT_CAP_100FDX:
1442 v = (bfe->bfe_mii_bmsr & MII_STATUS_100_BASEX_FD) != 0;
1443 break;
1444
1445 case ETHER_STAT_CAP_100HDX:
1446 v = (bfe->bfe_mii_bmsr & MII_STATUS_100_BASEX) != 0;
1447 break;
1448
1449 case ETHER_STAT_CAP_10FDX:
1450 v = (bfe->bfe_mii_bmsr & MII_STATUS_10_FD) != 0;
1451 break;
1452
1453 case ETHER_STAT_CAP_10HDX:
1454 v = (bfe->bfe_mii_bmsr & MII_STATUS_10) != 0;
1455 break;
1456
1457 case ETHER_STAT_CAP_ASMPAUSE:
1458 v = 0;
1459 break;
1460
1461 case ETHER_STAT_CAP_AUTONEG:
1462 v = ((bfe->bfe_mii_bmsr & MII_STATUS_CANAUTONEG) != 0);
1463 break;
1464
1465 case ETHER_STAT_CAP_PAUSE:
1466 v = 1;
1467 break;
1468
1469 case ETHER_STAT_CAP_REMFAULT:
1470 v = (bfe->bfe_mii_bmsr & MII_STATUS_REMFAULT) != 0;
1471 break;
1472
1473 case ETHER_STAT_CARRIER_ERRORS:
1474 v = bfe->bfe_stats.ether_stat_carrier_errors;
1475 break;
1476
1477 case ETHER_STAT_JABBER_ERRORS:
1478 err = ENOTSUP;
1479 break;
1480
1481 case ETHER_STAT_DEFER_XMTS:
1482 v = bfe->bfe_stats.ether_stat_defer_xmts;
1483 break;
1484
1485 case ETHER_STAT_EX_COLLISIONS:
1486 /* MIB */
1487 v = bfe->bfe_stats.ether_stat_ex_collisions;
1488 break;
1489
1490 case ETHER_STAT_FCS_ERRORS:
1491 /* MIB */
1492 v = bfe->bfe_stats.ether_stat_fcs_errors;
1493 break;
1494
1495 case ETHER_STAT_FIRST_COLLISIONS:
1496 /* MIB */
1497 v = bfe->bfe_stats.ether_stat_first_collisions;
1498 break;
1499
1500 case ETHER_STAT_LINK_ASMPAUSE:
1501 v = 0;
1502 break;
1503
1504 case ETHER_STAT_LINK_AUTONEG:
1505 v = (bfe->bfe_mii_bmcr & MII_CONTROL_ANE) != 0 &&
1506 (bfe->bfe_mii_bmsr & MII_STATUS_ANDONE) != 0;
1507 break;
1508
1509 case ETHER_STAT_LINK_DUPLEX:
1510 v = bfe->bfe_chip.duplex;
1511 break;
1512
1513 case ETHER_STAT_LP_CAP_100T4:
1514 v = (bfe->bfe_mii_anlpar & MII_ABILITY_100BASE_T4) != 0;
1515 break;
1516
1517 case ETHER_STAT_LP_CAP_100FDX:
1518 v = (bfe->bfe_mii_anlpar & MII_ABILITY_100BASE_TX_FD) != 0;
1519 break;
1520
1521 case ETHER_STAT_LP_CAP_100HDX:
1522 v = (bfe->bfe_mii_anlpar & MII_ABILITY_100BASE_TX) != 0;
1523 break;
1524
1525 case ETHER_STAT_LP_CAP_10FDX:
1526 v = (bfe->bfe_mii_anlpar & MII_ABILITY_10BASE_T_FD) != 0;
1527 break;
1528
1529 case ETHER_STAT_LP_CAP_10HDX:
1530 v = (bfe->bfe_mii_anlpar & MII_ABILITY_10BASE_T) != 0;
1531 break;
1532
1533 case ETHER_STAT_LP_CAP_ASMPAUSE:
1534 v = 0;
1535 break;
1536
1537 case ETHER_STAT_LP_CAP_AUTONEG:
1538 v = (bfe->bfe_mii_exp & MII_AN_EXP_LPCANAN) != 0;
1539 break;
1540
1541 case ETHER_STAT_LP_CAP_PAUSE:
1542 v = (bfe->bfe_mii_anlpar & MII_ABILITY_PAUSE) != 0;
1543 break;
1544
1545 case ETHER_STAT_LP_REMFAULT:
1546 v = (bfe->bfe_mii_anlpar & MII_STATUS_REMFAULT) != 0;
1547 break;
1548
1549 case ETHER_STAT_MACRCV_ERRORS:
1550 v = bfe->bfe_stats.ether_stat_macrcv_errors;
1551 break;
1552
1553 case ETHER_STAT_MACXMT_ERRORS:
1554 v = bfe->bfe_stats.ether_stat_macxmt_errors;
1555 break;
1556
1557 case ETHER_STAT_MULTI_COLLISIONS:
1558 v = bfe->bfe_stats.ether_stat_multi_collisions;
1559 break;
1560
1561 case ETHER_STAT_SQE_ERRORS:
1562 err = ENOTSUP;
1563 break;
1564
1565 case ETHER_STAT_TOOLONG_ERRORS:
1566 v = bfe->bfe_stats.ether_stat_toolong_errors;
1567 break;
1568
1569 case ETHER_STAT_TOOSHORT_ERRORS:
1570 v = bfe->bfe_stats.ether_stat_tooshort_errors;
1571 break;
1572
1573 case ETHER_STAT_TX_LATE_COLLISIONS:
1574 v = bfe->bfe_stats.ether_stat_tx_late_collisions;
1575 break;
1576
1577 case ETHER_STAT_XCVR_ADDR:
1578 v = bfe->bfe_phy_addr;
1579 break;
1580
1581 case ETHER_STAT_XCVR_ID:
1582 v = bfe->bfe_phy_id;
1583 break;
1584
1585 case MAC_STAT_BRDCSTRCV:
1586 v = bfe->bfe_stats.brdcstrcv;
1587 break;
1588
1589 case MAC_STAT_BRDCSTXMT:
1590 v = bfe->bfe_stats.brdcstxmt;
1591 break;
1592
1593 case MAC_STAT_MULTIXMT:
1594 v = bfe->bfe_stats.multixmt;
1595 break;
1596
1597 case MAC_STAT_COLLISIONS:
1598 v = bfe->bfe_stats.collisions;
1599 break;
1600
1601 case MAC_STAT_IERRORS:
1602 v = bfe->bfe_stats.ierrors;
1603 break;
1604
1605 case MAC_STAT_IPACKETS:
1606 v = bfe->bfe_stats.ipackets;
1607 break;
1608
1609 case MAC_STAT_MULTIRCV:
1610 v = bfe->bfe_stats.multircv;
1611 break;
1612
1613 case MAC_STAT_NORCVBUF:
1614 v = bfe->bfe_stats.norcvbuf;
1615 break;
1616
1617 case MAC_STAT_NOXMTBUF:
1618 v = bfe->bfe_stats.noxmtbuf;
1619 break;
1620
1621 case MAC_STAT_OBYTES:
1622 v = bfe->bfe_stats.obytes;
1623 break;
1624
1625 case MAC_STAT_OERRORS:
1626 /* MIB */
1627 v = bfe->bfe_stats.ether_stat_macxmt_errors;
1628 break;
1629
1630 case MAC_STAT_OPACKETS:
1631 v = bfe->bfe_stats.opackets;
1632 break;
1633
1634 case MAC_STAT_RBYTES:
1635 v = bfe->bfe_stats.rbytes;
1636 break;
1637
1638 case MAC_STAT_UNDERFLOWS:
1639 v = bfe->bfe_stats.underflows;
1640 break;
1641
1642 case MAC_STAT_OVERFLOWS:
1643 v = bfe->bfe_stats.overflows;
1644 break;
1645 }
1646
1647 rw_exit(&bfe->bfe_rwlock);
1648
1649 *val = v;
1650 return (err);
1651 }
1652
1653 int
bfe_mac_getprop(void * arg,const char * name,mac_prop_id_t num,uint_t sz,void * val)1654 bfe_mac_getprop(void *arg, const char *name, mac_prop_id_t num, uint_t sz,
1655 void *val)
1656 {
1657 bfe_t *bfe = (bfe_t *)arg;
1658 int err = 0;
1659
1660 switch (num) {
1661 case MAC_PROP_DUPLEX:
1662 ASSERT(sz >= sizeof (link_duplex_t));
1663 bcopy(&bfe->bfe_chip.duplex, val, sizeof (link_duplex_t));
1664 break;
1665
1666 case MAC_PROP_SPEED:
1667 ASSERT(sz >= sizeof (uint64_t));
1668 bcopy(&bfe->bfe_chip.speed, val, sizeof (uint64_t));
1669 break;
1670
1671 case MAC_PROP_AUTONEG:
1672 *(uint8_t *)val = bfe->bfe_adv_aneg;
1673 break;
1674
1675 case MAC_PROP_ADV_100FDX_CAP:
1676 *(uint8_t *)val = bfe->bfe_adv_100fdx;
1677 break;
1678
1679 case MAC_PROP_EN_100FDX_CAP:
1680 *(uint8_t *)val = bfe->bfe_adv_100fdx;
1681 break;
1682
1683 case MAC_PROP_ADV_100HDX_CAP:
1684 *(uint8_t *)val = bfe->bfe_adv_100hdx;
1685 break;
1686
1687 case MAC_PROP_EN_100HDX_CAP:
1688 *(uint8_t *)val = bfe->bfe_adv_100hdx;
1689 break;
1690
1691 case MAC_PROP_ADV_10FDX_CAP:
1692 *(uint8_t *)val = bfe->bfe_adv_10fdx;
1693 break;
1694
1695 case MAC_PROP_EN_10FDX_CAP:
1696 *(uint8_t *)val = bfe->bfe_adv_10fdx;
1697 break;
1698
1699 case MAC_PROP_ADV_10HDX_CAP:
1700 *(uint8_t *)val = bfe->bfe_adv_10hdx;
1701 break;
1702
1703 case MAC_PROP_EN_10HDX_CAP:
1704 *(uint8_t *)val = bfe->bfe_adv_10hdx;
1705 break;
1706
1707 case MAC_PROP_ADV_100T4_CAP:
1708 *(uint8_t *)val = bfe->bfe_adv_100T4;
1709 break;
1710
1711 case MAC_PROP_EN_100T4_CAP:
1712 *(uint8_t *)val = bfe->bfe_adv_100T4;
1713 break;
1714
1715 default:
1716 err = ENOTSUP;
1717 }
1718
1719 return (err);
1720 }
1721
1722
1723 static void
bfe_mac_propinfo(void * arg,const char * name,mac_prop_id_t num,mac_prop_info_handle_t prh)1724 bfe_mac_propinfo(void *arg, const char *name, mac_prop_id_t num,
1725 mac_prop_info_handle_t prh)
1726 {
1727 bfe_t *bfe = (bfe_t *)arg;
1728
1729 switch (num) {
1730 case MAC_PROP_DUPLEX:
1731 case MAC_PROP_SPEED:
1732 case MAC_PROP_ADV_100FDX_CAP:
1733 case MAC_PROP_ADV_100HDX_CAP:
1734 case MAC_PROP_ADV_10FDX_CAP:
1735 case MAC_PROP_ADV_10HDX_CAP:
1736 case MAC_PROP_ADV_100T4_CAP:
1737 case MAC_PROP_EN_100T4_CAP:
1738 mac_prop_info_set_perm(prh, MAC_PROP_PERM_READ);
1739 break;
1740
1741 case MAC_PROP_AUTONEG:
1742 mac_prop_info_set_default_uint8(prh, bfe->bfe_cap_aneg);
1743 break;
1744
1745 case MAC_PROP_EN_100FDX_CAP:
1746 mac_prop_info_set_default_uint8(prh, bfe->bfe_cap_100fdx);
1747 break;
1748
1749 case MAC_PROP_EN_100HDX_CAP:
1750 mac_prop_info_set_default_uint8(prh, bfe->bfe_cap_100hdx);
1751 break;
1752
1753 case MAC_PROP_EN_10FDX_CAP:
1754 mac_prop_info_set_default_uint8(prh, bfe->bfe_cap_10fdx);
1755 break;
1756
1757 case MAC_PROP_EN_10HDX_CAP:
1758 mac_prop_info_set_default_uint8(prh, bfe->bfe_cap_10hdx);
1759 break;
1760 }
1761 }
1762
1763
1764 /*ARGSUSED*/
1765 int
bfe_mac_setprop(void * arg,const char * name,mac_prop_id_t num,uint_t sz,const void * val)1766 bfe_mac_setprop(void *arg, const char *name, mac_prop_id_t num, uint_t sz,
1767 const void *val)
1768 {
1769 bfe_t *bfe = (bfe_t *)arg;
1770 uint8_t *advp;
1771 uint8_t *capp;
1772 int r = 0;
1773
1774 switch (num) {
1775 case MAC_PROP_EN_100FDX_CAP:
1776 advp = &bfe->bfe_adv_100fdx;
1777 capp = &bfe->bfe_cap_100fdx;
1778 break;
1779
1780 case MAC_PROP_EN_100HDX_CAP:
1781 advp = &bfe->bfe_adv_100hdx;
1782 capp = &bfe->bfe_cap_100hdx;
1783 break;
1784
1785 case MAC_PROP_EN_10FDX_CAP:
1786 advp = &bfe->bfe_adv_10fdx;
1787 capp = &bfe->bfe_cap_10fdx;
1788 break;
1789
1790 case MAC_PROP_EN_10HDX_CAP:
1791 advp = &bfe->bfe_adv_10hdx;
1792 capp = &bfe->bfe_cap_10hdx;
1793 break;
1794
1795 case MAC_PROP_AUTONEG:
1796 advp = &bfe->bfe_adv_aneg;
1797 capp = &bfe->bfe_cap_aneg;
1798 break;
1799
1800 default:
1801 return (ENOTSUP);
1802 }
1803
1804 if (*capp == 0)
1805 return (ENOTSUP);
1806
1807 bfe_grab_locks(bfe);
1808
1809 if (*advp != *(const uint8_t *)val) {
1810 *advp = *(const uint8_t *)val;
1811
1812 bfe->bfe_chip_action = BFE_ACTION_RESTART_SETPROP;
1813 if (bfe->bfe_chip_state == BFE_CHIP_ACTIVE) {
1814 /*
1815 * We need to stop the timer before grabbing locks
1816 * otherwise we can land-up in deadlock with untimeout.
1817 */
1818 bfe_stop_timer(bfe);
1819
1820 bfe->bfe_chip_action |= BFE_ACTION_RESTART;
1821
1822 bfe_chip_restart(bfe);
1823
1824 /*
1825 * We leave SETPROP because properties can be
1826 * temporary.
1827 */
1828 bfe->bfe_chip_action &= ~(BFE_ACTION_RESTART);
1829 r = 1;
1830 }
1831 }
1832
1833 bfe_release_locks(bfe);
1834
1835 /* kick-off a potential stopped downstream */
1836 if (r)
1837 mac_tx_update(bfe->bfe_machdl);
1838
1839 return (0);
1840 }
1841
1842
1843 int
bfe_mac_set_ether_addr(void * arg,const uint8_t * ea)1844 bfe_mac_set_ether_addr(void *arg, const uint8_t *ea)
1845 {
1846 bfe_t *bfe = (bfe_t *)arg;
1847
1848 bfe_grab_locks(bfe);
1849 bcopy(ea, bfe->bfe_ether_addr, ETHERADDRL);
1850 bfe_set_rx_mode(bfe);
1851 bfe_release_locks(bfe);
1852 return (0);
1853 }
1854
1855 int
bfe_mac_start(void * arg)1856 bfe_mac_start(void *arg)
1857 {
1858 bfe_t *bfe = (bfe_t *)arg;
1859
1860 bfe_grab_locks(bfe);
1861 if (bfe_chip_start(bfe) == DDI_FAILURE) {
1862 bfe_release_locks(bfe);
1863 return (EINVAL);
1864 }
1865
1866 bfe_release_locks(bfe);
1867
1868 mac_tx_update(bfe->bfe_machdl);
1869
1870 return (0);
1871 }
1872
1873 void
bfe_mac_stop(void * arg)1874 bfe_mac_stop(void *arg)
1875 {
1876 bfe_t *bfe = (bfe_t *)arg;
1877
1878 /*
1879 * We need to stop the timer before grabbing locks otherwise
1880 * we can land-up in deadlock with untimeout.
1881 */
1882 bfe_stop_timer(bfe);
1883
1884 bfe_grab_locks(bfe);
1885
1886 /*
1887 * First halt the chip by disabling interrupts.
1888 */
1889 bfe_chip_halt(bfe);
1890 bfe_stop_phy(bfe);
1891
1892 bfe->bfe_chip_state = BFE_CHIP_STOPPED;
1893
1894 /*
1895 * This will leave the PHY running.
1896 */
1897 bfe_chip_reset(bfe);
1898
1899 /*
1900 * Disable RX register.
1901 */
1902 bfe->bfe_chip_mode &= ~BFE_RX_MODE_ENABLE;
1903 bfe_set_rx_mode(bfe);
1904
1905 bfe_release_locks(bfe);
1906 }
1907
1908 /*
1909 * Send a packet down the wire.
1910 */
1911 static int
bfe_send_a_packet(bfe_t * bfe,mblk_t * mp)1912 bfe_send_a_packet(bfe_t *bfe, mblk_t *mp)
1913 {
1914 bfe_ring_t *r = &bfe->bfe_tx_ring;
1915 uint32_t cur = r->r_curr_desc;
1916 uint32_t next;
1917 size_t pktlen = msgsize(mp);
1918 uchar_t *buf;
1919 uint32_t v;
1920
1921 ASSERT(MUTEX_HELD(&r->r_lock));
1922 ASSERT(mp != NULL);
1923
1924 if (pktlen > r->r_buf_len) {
1925 freemsg(mp);
1926 return (BFE_SUCCESS);
1927 }
1928
1929 /*
1930 * There is a big reason why we don't check for '0'. It becomes easy
1931 * for us to not roll over the ring since we are based on producer (tx)
1932 * and consumer (reclaim by an interrupt) model. Especially when we
1933 * run out of TX descriptor, chip will send a single interrupt and
1934 * both producer and consumer counter will be same. So we keep a
1935 * difference of 1 always.
1936 */
1937 if (r->r_avail_desc <= 1) {
1938 bfe->bfe_stats.noxmtbuf++;
1939 bfe->bfe_tx_resched = 1;
1940 return (BFE_FAILURE);
1941 }
1942
1943 /*
1944 * Get the DMA buffer to hold packet.
1945 */
1946 buf = (uchar_t *)r->r_buf_dma[cur].addr;
1947
1948 mcopymsg(mp, buf); /* it also frees mp */
1949
1950 /*
1951 * Gather statistics.
1952 */
1953 if (buf[0] & 0x1) {
1954 if (bcmp(buf, bfe_broadcast, ETHERADDRL) != 0)
1955 bfe->bfe_stats.multixmt++;
1956 else
1957 bfe->bfe_stats.brdcstxmt++;
1958 }
1959 bfe->bfe_stats.opackets++;
1960 bfe->bfe_stats.obytes += pktlen;
1961
1962
1963 /*
1964 * Program the DMA descriptor (start and end of frame are same).
1965 */
1966 next = cur;
1967 v = (pktlen & BFE_DESC_LEN) | BFE_DESC_IOC | BFE_DESC_SOF |
1968 BFE_DESC_EOF;
1969
1970 if (cur == (TX_NUM_DESC - 1))
1971 v |= BFE_DESC_EOT;
1972
1973 PUT_DESC(r, (uint32_t *)&(r->r_desc[cur].desc_ctl), v);
1974
1975 /*
1976 * DMA addresses need to be added to BFE_PCI_DMA
1977 */
1978 PUT_DESC(r, (uint32_t *)&(r->r_desc[cur].desc_addr),
1979 (r->r_buf_dma[cur].cookie.dmac_laddress + BFE_PCI_DMA));
1980
1981 /*
1982 * Sync the packet data for the device.
1983 */
1984 (void) SYNC_BUF(r, cur, 0, pktlen, DDI_DMA_SYNC_FORDEV);
1985
1986 /* Move to next descriptor slot */
1987 BFE_INC_SLOT(next, TX_NUM_DESC);
1988
1989 (void) SYNC_DESC(r, 0, r->r_ndesc, DDI_DMA_SYNC_FORDEV);
1990
1991 r->r_curr_desc = next;
1992
1993 /*
1994 * The order should be 1,2,3,... for BFE_DMATX_PTR if 0,1,2,3,...
1995 * descriptor slot are being programmed.
1996 */
1997 OUTL(bfe, BFE_DMATX_PTR, next * sizeof (bfe_desc_t));
1998 FLUSH(bfe, BFE_DMATX_PTR);
1999
2000 r->r_avail_desc--;
2001
2002 /*
2003 * Let timeout know that it must reset the chip if a
2004 * packet is not sent down the wire for more than 5 seconds.
2005 */
2006 bfe->bfe_tx_stall_time = gethrtime() + (5 * 1000000000ULL);
2007
2008 return (BFE_SUCCESS);
2009 }
2010
2011 mblk_t *
bfe_mac_transmit_packet(void * arg,mblk_t * mp)2012 bfe_mac_transmit_packet(void *arg, mblk_t *mp)
2013 {
2014 bfe_t *bfe = (bfe_t *)arg;
2015 bfe_ring_t *r = &bfe->bfe_tx_ring;
2016 mblk_t *nmp;
2017
2018 mutex_enter(&r->r_lock);
2019
2020 if (bfe->bfe_chip_state != BFE_CHIP_ACTIVE) {
2021 DTRACE_PROBE1(tx__chip__not__active, int, bfe->bfe_unit);
2022
2023 freemsgchain(mp);
2024 mutex_exit(&r->r_lock);
2025 return (NULL);
2026 }
2027
2028
2029 while (mp != NULL) {
2030 nmp = mp->b_next;
2031 mp->b_next = NULL;
2032
2033 if (bfe_send_a_packet(bfe, mp) == BFE_FAILURE) {
2034 mp->b_next = nmp;
2035 break;
2036 }
2037 mp = nmp;
2038 }
2039
2040 mutex_exit(&r->r_lock);
2041
2042 return (mp);
2043 }
2044
2045 int
bfe_mac_set_promisc(void * arg,boolean_t promiscflag)2046 bfe_mac_set_promisc(void *arg, boolean_t promiscflag)
2047 {
2048 bfe_t *bfe = (bfe_t *)arg;
2049
2050 bfe_grab_locks(bfe);
2051 if (bfe->bfe_chip_state != BFE_CHIP_ACTIVE) {
2052 bfe_release_locks(bfe);
2053 return (EIO);
2054 }
2055
2056 if (promiscflag) {
2057 /* Set Promiscous on */
2058 bfe->bfe_chip_mode |= BFE_RX_MODE_PROMISC;
2059 } else {
2060 bfe->bfe_chip_mode &= ~BFE_RX_MODE_PROMISC;
2061 }
2062
2063 bfe_set_rx_mode(bfe);
2064 bfe_release_locks(bfe);
2065
2066 return (0);
2067 }
2068
2069 int
bfe_mac_set_multicast(void * arg,boolean_t add,const uint8_t * macaddr)2070 bfe_mac_set_multicast(void *arg, boolean_t add, const uint8_t *macaddr)
2071 {
2072 /*
2073 * It was too much of pain to implement multicast in CAM. Instead
2074 * we never disable multicast filter.
2075 */
2076 return (0);
2077 }
2078
2079 static mac_callbacks_t bfe_mac_callbacks = {
2080 MC_SETPROP | MC_GETPROP | MC_PROPINFO,
2081 bfe_mac_getstat, /* gets stats */
2082 bfe_mac_start, /* starts mac */
2083 bfe_mac_stop, /* stops mac */
2084 bfe_mac_set_promisc, /* sets promisc mode for snoop */
2085 bfe_mac_set_multicast, /* multicast implementation */
2086 bfe_mac_set_ether_addr, /* sets ethernet address (unicast) */
2087 bfe_mac_transmit_packet, /* transmits packet */
2088 NULL,
2089 NULL, /* ioctl */
2090 NULL, /* getcap */
2091 NULL, /* open */
2092 NULL, /* close */
2093 bfe_mac_setprop,
2094 bfe_mac_getprop,
2095 bfe_mac_propinfo
2096 };
2097
2098 static void
bfe_error_handler(bfe_t * bfe,int intr_mask)2099 bfe_error_handler(bfe_t *bfe, int intr_mask)
2100 {
2101 uint32_t v;
2102
2103 if (intr_mask & BFE_ISTAT_RFO) {
2104 bfe->bfe_stats.overflows++;
2105 bfe->bfe_chip_action |=
2106 (BFE_ACTION_RESTART | BFE_ACTION_RESTART_FAULT);
2107 goto action;
2108 }
2109
2110 if (intr_mask & BFE_ISTAT_TFU) {
2111 bfe->bfe_stats.underflows++;
2112 return;
2113 }
2114
2115 /* Descriptor Protocol Error */
2116 if (intr_mask & BFE_ISTAT_DPE) {
2117 bfe_error(bfe->bfe_dip,
2118 "Descriptor Protocol Error. Halting Chip");
2119 bfe->bfe_chip_action |=
2120 (BFE_ACTION_RESTART | BFE_ACTION_RESTART_FAULT);
2121 goto action;
2122 }
2123
2124 /* Descriptor Error */
2125 if (intr_mask & BFE_ISTAT_DSCE) {
2126 bfe_error(bfe->bfe_dip, "Descriptor Error. Restarting Chip");
2127 goto action;
2128 }
2129
2130 /* Receive Descr. Underflow */
2131 if (intr_mask & BFE_ISTAT_RDU) {
2132 bfe_error(bfe->bfe_dip,
2133 "Receive Descriptor Underflow. Restarting Chip");
2134 bfe->bfe_stats.ether_stat_macrcv_errors++;
2135 bfe->bfe_chip_action |=
2136 (BFE_ACTION_RESTART | BFE_ACTION_RESTART_FAULT);
2137 goto action;
2138 }
2139
2140 v = INL(bfe, BFE_DMATX_STAT);
2141
2142 /* Error while sending a packet */
2143 if (v & BFE_STAT_EMASK) {
2144 bfe->bfe_stats.ether_stat_macxmt_errors++;
2145 bfe_error(bfe->bfe_dip,
2146 "Error while sending a packet. Restarting Chip");
2147 }
2148
2149 /* Error while receiving a packet */
2150 v = INL(bfe, BFE_DMARX_STAT);
2151 if (v & BFE_RX_FLAG_ERRORS) {
2152 bfe->bfe_stats.ierrors++;
2153 bfe_error(bfe->bfe_dip,
2154 "Error while receiving a packet. Restarting Chip");
2155 }
2156
2157
2158 bfe->bfe_chip_action |=
2159 (BFE_ACTION_RESTART | BFE_ACTION_RESTART_FAULT);
2160
2161 action:
2162 bfe_chip_halt(bfe);
2163 }
2164
2165 /*
2166 * It will recycle a RX descriptor slot.
2167 */
2168 static void
bfe_rx_desc_buf_reinit(bfe_t * bfe,uint_t slot)2169 bfe_rx_desc_buf_reinit(bfe_t *bfe, uint_t slot)
2170 {
2171 bfe_ring_t *r = &bfe->bfe_rx_ring;
2172 uint32_t v;
2173
2174 slot %= RX_NUM_DESC;
2175
2176 bzero(r->r_buf_dma[slot].addr, sizeof (bfe_rx_header_t));
2177
2178 (void) SYNC_BUF(r, slot, 0, BFE_RX_OFFSET, DDI_DMA_SYNC_FORDEV);
2179
2180 v = r->r_buf_dma[slot].len & BFE_DESC_LEN;
2181 if (slot == (RX_NUM_DESC - 1))
2182 v |= BFE_DESC_EOT;
2183
2184 PUT_DESC(r, (uint32_t *)&(r->r_desc[slot].desc_ctl), v);
2185
2186 /*
2187 * DMA addresses need to be added to BFE_PCI_DMA
2188 */
2189 PUT_DESC(r, (uint32_t *)&(r->r_desc[slot].desc_addr),
2190 (r->r_buf_dma[slot].cookie.dmac_laddress + BFE_PCI_DMA));
2191 }
2192
2193 /*
2194 * Gets called from interrupt context to handle RX interrupt.
2195 */
2196 static mblk_t *
bfe_receive(bfe_t * bfe,int intr_mask)2197 bfe_receive(bfe_t *bfe, int intr_mask)
2198 {
2199 int rxstat, current;
2200 mblk_t *mp = NULL, *rx_head, *rx_tail;
2201 uchar_t *rx_header;
2202 uint16_t len;
2203 uchar_t *bp;
2204 bfe_ring_t *r = &bfe->bfe_rx_ring;
2205 int i;
2206
2207 rxstat = INL(bfe, BFE_DMARX_STAT);
2208 current = (rxstat & BFE_STAT_CDMASK) / sizeof (bfe_desc_t);
2209 i = r->r_curr_desc;
2210
2211 rx_head = rx_tail = NULL;
2212
2213 DTRACE_PROBE3(receive, int, bfe->bfe_unit,
2214 int, r->r_curr_desc,
2215 int, current);
2216
2217 for (i = r->r_curr_desc; i != current;
2218 BFE_INC_SLOT(i, RX_NUM_DESC)) {
2219
2220 /*
2221 * Sync the buffer associated with the descriptor table entry.
2222 */
2223 (void) SYNC_BUF(r, i, 0, r->r_buf_dma[i].len,
2224 DDI_DMA_SYNC_FORKERNEL);
2225
2226 rx_header = (void *)r->r_buf_dma[i].addr;
2227
2228 /*
2229 * We do this to make sure we are endian neutral. Chip is
2230 * big endian.
2231 *
2232 * The header looks like :-
2233 *
2234 * Offset 0 -> uint16_t len
2235 * Offset 2 -> uint16_t flags
2236 * Offset 4 -> uint16_t pad[12]
2237 */
2238 len = (rx_header[1] << 8) | rx_header[0];
2239 len -= 4; /* CRC bytes need to be removed */
2240
2241 /*
2242 * Don't receive this packet if pkt length is greater than
2243 * MTU + VLAN_TAGSZ.
2244 */
2245 if (len > r->r_buf_len) {
2246 /* Recycle slot for later use */
2247 bfe_rx_desc_buf_reinit(bfe, i);
2248 continue;
2249 }
2250
2251 if ((mp = allocb(len + VLAN_TAGSZ, BPRI_MED)) != NULL) {
2252 mp->b_rptr += VLAN_TAGSZ;
2253 bp = mp->b_rptr;
2254 mp->b_wptr = bp + len;
2255
2256 /* sizeof (bfe_rx_header_t) + 2 */
2257 bcopy(r->r_buf_dma[i].addr +
2258 BFE_RX_OFFSET, bp, len);
2259
2260 mp->b_next = NULL;
2261 if (rx_tail == NULL)
2262 rx_head = rx_tail = mp;
2263 else {
2264 rx_tail->b_next = mp;
2265 rx_tail = mp;
2266 }
2267
2268 /* Number of packets received so far */
2269 bfe->bfe_stats.ipackets++;
2270
2271 /* Total bytes of packets received so far */
2272 bfe->bfe_stats.rbytes += len;
2273
2274 if (bcmp(mp->b_rptr, bfe_broadcast, ETHERADDRL) == 0)
2275 bfe->bfe_stats.brdcstrcv++;
2276 else
2277 bfe->bfe_stats.multircv++;
2278 } else {
2279 bfe->bfe_stats.norcvbuf++;
2280 /* Recycle the slot for later use */
2281 bfe_rx_desc_buf_reinit(bfe, i);
2282 break;
2283 }
2284
2285 /*
2286 * Reinitialize the current descriptor slot's buffer so that
2287 * it can be reused.
2288 */
2289 bfe_rx_desc_buf_reinit(bfe, i);
2290 }
2291
2292 r->r_curr_desc = i;
2293
2294 (void) SYNC_DESC(r, 0, r->r_ndesc, DDI_DMA_SYNC_FORDEV);
2295
2296 return (rx_head);
2297 }
2298
2299 static int
bfe_tx_reclaim(bfe_ring_t * r)2300 bfe_tx_reclaim(bfe_ring_t *r)
2301 {
2302 uint32_t cur, start;
2303 uint32_t v;
2304
2305 cur = INL(r->r_bfe, BFE_DMATX_STAT) & BFE_STAT_CDMASK;
2306 cur = cur / sizeof (bfe_desc_t);
2307
2308 /*
2309 * Start with the last descriptor consumed by the chip.
2310 */
2311 start = r->r_cons_desc;
2312
2313 DTRACE_PROBE3(tx__reclaim, int, r->r_bfe->bfe_unit,
2314 int, start,
2315 int, cur);
2316
2317 /*
2318 * There will be at least one descriptor to process.
2319 */
2320 while (start != cur) {
2321 r->r_avail_desc++;
2322 v = r->r_buf_dma[start].len & BFE_DESC_LEN;
2323 if (start == (TX_NUM_DESC - 1))
2324 v |= BFE_DESC_EOT;
2325
2326 PUT_DESC(r, (uint32_t *)&(r->r_desc[start].desc_ctl), v);
2327 PUT_DESC(r, (uint32_t *)&(r->r_desc[start].desc_addr),
2328 (r->r_buf_dma[start].cookie.dmac_laddress + BFE_PCI_DMA));
2329
2330 /* Move to next descriptor in TX ring */
2331 BFE_INC_SLOT(start, TX_NUM_DESC);
2332 }
2333
2334 (void) ddi_dma_sync(r->r_desc_dma_handle,
2335 0, (r->r_ndesc * sizeof (bfe_desc_t)),
2336 DDI_DMA_SYNC_FORDEV);
2337
2338 r->r_cons_desc = start; /* consumed pointer */
2339 r->r_bfe->bfe_tx_stall_time = 0;
2340
2341 return (cur);
2342 }
2343
2344 static int
bfe_tx_done(bfe_t * bfe,int intr_mask)2345 bfe_tx_done(bfe_t *bfe, int intr_mask)
2346 {
2347 bfe_ring_t *r = &bfe->bfe_tx_ring;
2348 int resched = 0;
2349
2350 mutex_enter(&r->r_lock);
2351 (void) bfe_tx_reclaim(r);
2352
2353 if (bfe->bfe_tx_resched) {
2354 resched = 1;
2355 bfe->bfe_tx_resched = 0;
2356 }
2357 mutex_exit(&r->r_lock);
2358
2359 return (resched);
2360 }
2361
2362 /*
2363 * ISR for interrupt handling
2364 */
2365 static uint_t
bfe_interrupt(caddr_t arg1,caddr_t arg2)2366 bfe_interrupt(caddr_t arg1, caddr_t arg2)
2367 {
2368 bfe_t *bfe = (void *)arg1;
2369 uint32_t intr_stat;
2370 mblk_t *rx_head = NULL;
2371 int resched = 0;
2372
2373 /*
2374 * Grab the lock to avoid stopping the chip while this interrupt
2375 * is handled.
2376 */
2377 rw_enter(&bfe->bfe_rwlock, RW_READER);
2378
2379 /*
2380 * It's necessary to read intr stat again because masking interrupt
2381 * register does not really mask interrupts coming from the chip.
2382 */
2383 intr_stat = INL(bfe, BFE_INTR_STAT);
2384 intr_stat &= BFE_IMASK_DEF;
2385 OUTL(bfe, BFE_INTR_STAT, intr_stat);
2386 (void) INL(bfe, BFE_INTR_STAT);
2387
2388 if (intr_stat == 0) {
2389 rw_exit(&bfe->bfe_rwlock);
2390 return (DDI_INTR_UNCLAIMED);
2391 }
2392
2393 DTRACE_PROBE2(bfe__interrupt, int, bfe->bfe_unit,
2394 int, intr_stat);
2395
2396 if (bfe->bfe_chip_state != BFE_CHIP_ACTIVE) {
2397 /*
2398 * If chip is suspended then we just return.
2399 */
2400 if (bfe->bfe_chip_state == BFE_CHIP_SUSPENDED) {
2401 rw_exit(&bfe->bfe_rwlock);
2402 DTRACE_PROBE1(interrupt__chip__is__suspend, int,
2403 bfe->bfe_unit);
2404 return (DDI_INTR_CLAIMED);
2405 }
2406
2407 /*
2408 * Halt the chip again i.e basically disable interrupts.
2409 */
2410 bfe_chip_halt(bfe);
2411 rw_exit(&bfe->bfe_rwlock);
2412 DTRACE_PROBE1(interrupt__chip__not__active, int,
2413 bfe->bfe_unit);
2414 return (DDI_INTR_CLAIMED);
2415 }
2416
2417 /* A packet was received */
2418 if (intr_stat & BFE_ISTAT_RX) {
2419 rx_head = bfe_receive(bfe, intr_stat);
2420 }
2421
2422 /* A packet was sent down the wire */
2423 if (intr_stat & BFE_ISTAT_TX) {
2424 resched = bfe_tx_done(bfe, intr_stat);
2425 }
2426
2427 /* There was an error */
2428 if (intr_stat & BFE_ISTAT_ERRORS) {
2429 bfe_error_handler(bfe, intr_stat);
2430 }
2431
2432 rw_exit(&bfe->bfe_rwlock);
2433
2434 /*
2435 * Pass the list of packets received from chip to MAC layer.
2436 */
2437 if (rx_head) {
2438 mac_rx(bfe->bfe_machdl, 0, rx_head);
2439 }
2440
2441 /*
2442 * Let the MAC start sending pkts to a potential stopped stream.
2443 */
2444 if (resched)
2445 mac_tx_update(bfe->bfe_machdl);
2446
2447 return (DDI_INTR_CLAIMED);
2448 }
2449
2450 /*
2451 * Removes registered interrupt handler.
2452 */
2453 static void
bfe_remove_intr(bfe_t * bfe)2454 bfe_remove_intr(bfe_t *bfe)
2455 {
2456 (void) ddi_intr_remove_handler(bfe->bfe_intrhdl);
2457 (void) ddi_intr_free(bfe->bfe_intrhdl);
2458 }
2459
2460 /*
2461 * Add an interrupt for the driver.
2462 */
2463 static int
bfe_add_intr(bfe_t * bfe)2464 bfe_add_intr(bfe_t *bfe)
2465 {
2466 int nintrs = 1;
2467 int ret;
2468
2469 ret = ddi_intr_alloc(bfe->bfe_dip, &bfe->bfe_intrhdl,
2470 DDI_INTR_TYPE_FIXED, /* type */
2471 0, /* inumber */
2472 1, /* count */
2473 &nintrs, /* actual nintrs */
2474 DDI_INTR_ALLOC_STRICT);
2475
2476 if (ret != DDI_SUCCESS) {
2477 bfe_error(bfe->bfe_dip, "ddi_intr_alloc() failed"
2478 " : ret : %d", ret);
2479 return (DDI_FAILURE);
2480 }
2481
2482 ret = ddi_intr_add_handler(bfe->bfe_intrhdl, bfe_interrupt, bfe, NULL);
2483 if (ret != DDI_SUCCESS) {
2484 bfe_error(bfe->bfe_dip, "ddi_intr_add_handler() failed");
2485 (void) ddi_intr_free(bfe->bfe_intrhdl);
2486 return (DDI_FAILURE);
2487 }
2488
2489 ret = ddi_intr_get_pri(bfe->bfe_intrhdl, &bfe->bfe_intrpri);
2490 if (ret != DDI_SUCCESS) {
2491 bfe_error(bfe->bfe_dip, "ddi_intr_get_pri() failed");
2492 bfe_remove_intr(bfe);
2493 return (DDI_FAILURE);
2494 }
2495
2496 return (DDI_SUCCESS);
2497 }
2498
2499
2500 /*
2501 * Identify chipset family.
2502 */
2503 static int
bfe_identify_hardware(bfe_t * bfe)2504 bfe_identify_hardware(bfe_t *bfe)
2505 {
2506 uint16_t vid, did;
2507 int i;
2508
2509 vid = pci_config_get16(bfe->bfe_conf_handle, PCI_CONF_VENID);
2510 did = pci_config_get16(bfe->bfe_conf_handle, PCI_CONF_DEVID);
2511
2512 for (i = 0; i < (sizeof (bfe_cards) / sizeof (bfe_cards_t)); i++) {
2513 if (bfe_cards[i].vendor_id == vid &&
2514 bfe_cards[i].device_id == did) {
2515 return (BFE_SUCCESS);
2516 }
2517 }
2518
2519 bfe_error(bfe->bfe_dip, "bfe driver is attaching to unknown pci%d,%d"
2520 " vendor/device-id card", vid, did);
2521
2522 return (BFE_SUCCESS);
2523 }
2524
2525 /*
2526 * Maps device registers.
2527 */
2528 static int
bfe_regs_map(bfe_t * bfe)2529 bfe_regs_map(bfe_t *bfe)
2530 {
2531 dev_info_t *dip = bfe->bfe_dip;
2532 int ret;
2533
2534 ret = ddi_regs_map_setup(dip, 1, &bfe->bfe_mem_regset.addr, 0, 0,
2535 &bfe_dev_attr, &bfe->bfe_mem_regset.hdl);
2536
2537 if (ret != DDI_SUCCESS) {
2538 bfe_error(bfe->bfe_dip, "ddi_regs_map_setup failed");
2539 return (DDI_FAILURE);
2540 }
2541
2542 return (DDI_SUCCESS);
2543 }
2544
2545 static void
bfe_unmap_regs(bfe_t * bfe)2546 bfe_unmap_regs(bfe_t *bfe)
2547 {
2548 ddi_regs_map_free(&bfe->bfe_mem_regset.hdl);
2549 }
2550
2551 static int
bfe_get_chip_config(bfe_t * bfe)2552 bfe_get_chip_config(bfe_t *bfe)
2553 {
2554 uint32_t prom[BFE_EEPROM_SIZE];
2555 int i;
2556
2557 /*
2558 * Read EEPROM in prom[]
2559 */
2560 for (i = 0; i < BFE_EEPROM_SIZE; i++) {
2561 prom[i] = INL(bfe, BFE_EEPROM_BASE + i * sizeof (uint32_t));
2562 }
2563
2564 bfe->bfe_dev_addr[0] = bfe->bfe_ether_addr[0] =
2565 INB(bfe, BFE_EEPROM_BASE + 79);
2566
2567 bfe->bfe_dev_addr[1] = bfe->bfe_ether_addr[1] =
2568 INB(bfe, BFE_EEPROM_BASE + 78);
2569
2570 bfe->bfe_dev_addr[2] = bfe->bfe_ether_addr[2] =
2571 INB(bfe, BFE_EEPROM_BASE + 81);
2572
2573 bfe->bfe_dev_addr[3] = bfe->bfe_ether_addr[3] =
2574 INB(bfe, BFE_EEPROM_BASE + 80);
2575
2576 bfe->bfe_dev_addr[4] = bfe->bfe_ether_addr[4] =
2577 INB(bfe, BFE_EEPROM_BASE + 83);
2578
2579 bfe->bfe_dev_addr[5] = bfe->bfe_ether_addr[5] =
2580 INB(bfe, BFE_EEPROM_BASE + 82);
2581
2582 bfe->bfe_phy_addr = -1;
2583
2584 return (DDI_SUCCESS);
2585 }
2586
2587 /*
2588 * Ring Management routines
2589 */
2590 static int
bfe_ring_buf_alloc(bfe_t * bfe,bfe_ring_t * r,int slot,int d)2591 bfe_ring_buf_alloc(bfe_t *bfe, bfe_ring_t *r, int slot, int d)
2592 {
2593 int err;
2594 uint_t count = 0;
2595
2596 err = ddi_dma_alloc_handle(bfe->bfe_dip,
2597 &bfe_dma_attr_buf, DDI_DMA_SLEEP, NULL,
2598 &r->r_buf_dma[slot].handle);
2599
2600 if (err != DDI_SUCCESS) {
2601 bfe_error(bfe->bfe_dip, " bfe_ring_buf_alloc() :"
2602 " alloc_handle failed");
2603 goto fail0;
2604 }
2605
2606 err = ddi_dma_mem_alloc(r->r_buf_dma[slot].handle,
2607 r->r_buf_len, &bfe_buf_attr, DDI_DMA_STREAMING,
2608 DDI_DMA_SLEEP, NULL, &r->r_buf_dma[slot].addr,
2609 &r->r_buf_dma[slot].len,
2610 &r->r_buf_dma[slot].acchdl);
2611
2612 if (err != DDI_SUCCESS) {
2613 bfe_error(bfe->bfe_dip, " bfe_ring_buf_alloc() :"
2614 " mem_alloc failed :%d", err);
2615 goto fail1;
2616 }
2617
2618 err = ddi_dma_addr_bind_handle(r->r_buf_dma[slot].handle,
2619 NULL, r->r_buf_dma[slot].addr,
2620 r->r_buf_dma[slot].len,
2621 (DDI_DMA_RDWR | DDI_DMA_STREAMING),
2622 DDI_DMA_SLEEP, NULL,
2623 &r->r_buf_dma[slot].cookie,
2624 &count);
2625
2626 if (err != DDI_DMA_MAPPED) {
2627 bfe_error(bfe->bfe_dip, " bfe_ring_buf_alloc() :"
2628 " bind_handle failed");
2629 goto fail2;
2630 }
2631
2632 if (count > 1) {
2633 bfe_error(bfe->bfe_dip, " bfe_ring_buf_alloc() :"
2634 " more than one DMA cookie");
2635 (void) ddi_dma_unbind_handle(r->r_buf_dma[slot].handle);
2636 goto fail2;
2637 }
2638
2639 return (DDI_SUCCESS);
2640 fail2:
2641 ddi_dma_mem_free(&r->r_buf_dma[slot].acchdl);
2642 fail1:
2643 ddi_dma_free_handle(&r->r_buf_dma[slot].handle);
2644 fail0:
2645 return (DDI_FAILURE);
2646 }
2647
2648 static void
bfe_ring_buf_free(bfe_ring_t * r,int slot)2649 bfe_ring_buf_free(bfe_ring_t *r, int slot)
2650 {
2651 if (r->r_buf_dma == NULL)
2652 return;
2653
2654 (void) ddi_dma_unbind_handle(r->r_buf_dma[slot].handle);
2655 ddi_dma_mem_free(&r->r_buf_dma[slot].acchdl);
2656 ddi_dma_free_handle(&r->r_buf_dma[slot].handle);
2657 }
2658
2659 static void
bfe_buffer_free(bfe_ring_t * r)2660 bfe_buffer_free(bfe_ring_t *r)
2661 {
2662 int i;
2663
2664 for (i = 0; i < r->r_ndesc; i++) {
2665 bfe_ring_buf_free(r, i);
2666 }
2667 }
2668
2669 static void
bfe_ring_desc_free(bfe_ring_t * r)2670 bfe_ring_desc_free(bfe_ring_t *r)
2671 {
2672 (void) ddi_dma_unbind_handle(r->r_desc_dma_handle);
2673 ddi_dma_mem_free(&r->r_desc_acc_handle);
2674 ddi_dma_free_handle(&r->r_desc_dma_handle);
2675 kmem_free(r->r_buf_dma, r->r_ndesc * sizeof (bfe_dma_t));
2676
2677 r->r_buf_dma = NULL;
2678 r->r_desc = NULL;
2679 }
2680
2681
2682 static int
bfe_ring_desc_alloc(bfe_t * bfe,bfe_ring_t * r,int d)2683 bfe_ring_desc_alloc(bfe_t *bfe, bfe_ring_t *r, int d)
2684 {
2685 int err, i, fail = 0;
2686 caddr_t ring;
2687 size_t size_krnl = 0, size_dma = 0, ring_len = 0;
2688 ddi_dma_cookie_t cookie;
2689 uint_t count = 0;
2690
2691 ASSERT(bfe != NULL);
2692
2693 size_krnl = r->r_ndesc * sizeof (bfe_dma_t);
2694 size_dma = r->r_ndesc * sizeof (bfe_desc_t);
2695 r->r_buf_dma = kmem_zalloc(size_krnl, KM_SLEEP);
2696
2697
2698 err = ddi_dma_alloc_handle(bfe->bfe_dip, &bfe_dma_attr_desc,
2699 DDI_DMA_SLEEP, NULL, &r->r_desc_dma_handle);
2700
2701 if (err != DDI_SUCCESS) {
2702 bfe_error(bfe->bfe_dip, "bfe_ring_desc_alloc() failed on"
2703 " ddi_dma_alloc_handle()");
2704 kmem_free(r->r_buf_dma, size_krnl);
2705 return (DDI_FAILURE);
2706 }
2707
2708
2709 err = ddi_dma_mem_alloc(r->r_desc_dma_handle,
2710 size_dma, &bfe_buf_attr,
2711 DDI_DMA_CONSISTENT, DDI_DMA_SLEEP, NULL,
2712 &ring, &ring_len, &r->r_desc_acc_handle);
2713
2714 if (err != DDI_SUCCESS) {
2715 bfe_error(bfe->bfe_dip, "bfe_ring_desc_alloc() failed on"
2716 " ddi_dma_mem_alloc()");
2717 ddi_dma_free_handle(&r->r_desc_dma_handle);
2718 kmem_free(r->r_buf_dma, size_krnl);
2719 return (DDI_FAILURE);
2720 }
2721
2722 err = ddi_dma_addr_bind_handle(r->r_desc_dma_handle,
2723 NULL, ring, ring_len,
2724 DDI_DMA_RDWR | DDI_DMA_CONSISTENT,
2725 DDI_DMA_SLEEP, NULL,
2726 &cookie, &count);
2727
2728 if (err != DDI_SUCCESS) {
2729 bfe_error(bfe->bfe_dip, "bfe_ring_desc_alloc() failed on"
2730 " ddi_dma_addr_bind_handle()");
2731 ddi_dma_mem_free(&r->r_desc_acc_handle);
2732 ddi_dma_free_handle(&r->r_desc_dma_handle);
2733 kmem_free(r->r_buf_dma, size_krnl);
2734 return (DDI_FAILURE);
2735 }
2736
2737 /*
2738 * We don't want to have multiple cookies. Descriptor should be
2739 * aligned to PAGESIZE boundary.
2740 */
2741 ASSERT(count == 1);
2742
2743 /* The actual descriptor for the ring */
2744 r->r_desc_len = ring_len;
2745 r->r_desc_cookie = cookie;
2746
2747 r->r_desc = (void *)ring;
2748
2749 bzero(r->r_desc, size_dma);
2750 bzero(r->r_desc, ring_len);
2751
2752 /* For each descriptor, allocate a DMA buffer */
2753 fail = 0;
2754 for (i = 0; i < r->r_ndesc; i++) {
2755 if (bfe_ring_buf_alloc(bfe, r, i, d) != DDI_SUCCESS) {
2756 i--;
2757 fail = 1;
2758 break;
2759 }
2760 }
2761
2762 if (fail) {
2763 while (i-- >= 0) {
2764 bfe_ring_buf_free(r, i);
2765 }
2766
2767 /* We don't need the descriptor anymore */
2768 bfe_ring_desc_free(r);
2769 return (DDI_FAILURE);
2770 }
2771
2772 return (DDI_SUCCESS);
2773 }
2774
2775 static int
bfe_rings_alloc(bfe_t * bfe)2776 bfe_rings_alloc(bfe_t *bfe)
2777 {
2778 /* TX */
2779 mutex_init(&bfe->bfe_tx_ring.r_lock, NULL, MUTEX_DRIVER, NULL);
2780 bfe->bfe_tx_ring.r_lockp = &bfe->bfe_tx_ring.r_lock;
2781 bfe->bfe_tx_ring.r_buf_len = BFE_MTU + sizeof (struct ether_header) +
2782 VLAN_TAGSZ + ETHERFCSL;
2783 bfe->bfe_tx_ring.r_ndesc = TX_NUM_DESC;
2784 bfe->bfe_tx_ring.r_bfe = bfe;
2785 bfe->bfe_tx_ring.r_avail_desc = TX_NUM_DESC;
2786
2787 /* RX */
2788 mutex_init(&bfe->bfe_rx_ring.r_lock, NULL, MUTEX_DRIVER, NULL);
2789 bfe->bfe_rx_ring.r_lockp = &bfe->bfe_rx_ring.r_lock;
2790 bfe->bfe_rx_ring.r_buf_len = BFE_MTU + sizeof (struct ether_header) +
2791 VLAN_TAGSZ + ETHERFCSL + RX_HEAD_ROOM;
2792 bfe->bfe_rx_ring.r_ndesc = RX_NUM_DESC;
2793 bfe->bfe_rx_ring.r_bfe = bfe;
2794 bfe->bfe_rx_ring.r_avail_desc = RX_NUM_DESC;
2795
2796 /* Allocate TX Ring */
2797 if (bfe_ring_desc_alloc(bfe, &bfe->bfe_tx_ring,
2798 DDI_DMA_WRITE) != DDI_SUCCESS)
2799 return (DDI_FAILURE);
2800
2801 /* Allocate RX Ring */
2802 if (bfe_ring_desc_alloc(bfe, &bfe->bfe_rx_ring,
2803 DDI_DMA_READ) != DDI_SUCCESS) {
2804 cmn_err(CE_NOTE, "RX ring allocation failed");
2805 bfe_ring_desc_free(&bfe->bfe_tx_ring);
2806 return (DDI_FAILURE);
2807 }
2808
2809 bfe->bfe_tx_ring.r_flags = BFE_RING_ALLOCATED;
2810 bfe->bfe_rx_ring.r_flags = BFE_RING_ALLOCATED;
2811
2812 return (DDI_SUCCESS);
2813 }
2814
2815 static int
bfe_resume(dev_info_t * dip)2816 bfe_resume(dev_info_t *dip)
2817 {
2818 bfe_t *bfe;
2819 int err = DDI_SUCCESS;
2820
2821 if ((bfe = ddi_get_driver_private(dip)) == NULL) {
2822 bfe_error(dip, "Unexpected error (no driver private data)"
2823 " while resume");
2824 return (DDI_FAILURE);
2825 }
2826
2827 /*
2828 * Grab all the locks first.
2829 */
2830 bfe_grab_locks(bfe);
2831 bfe->bfe_chip_state = BFE_CHIP_RESUME;
2832
2833 bfe_init_vars(bfe);
2834 /* PHY will also start running */
2835 bfe_chip_reset(bfe);
2836 if (bfe_chip_start(bfe) == DDI_FAILURE) {
2837 bfe_error(dip, "Could not resume chip");
2838 err = DDI_FAILURE;
2839 }
2840
2841 bfe_release_locks(bfe);
2842
2843 if (err == DDI_SUCCESS)
2844 mac_tx_update(bfe->bfe_machdl);
2845
2846 return (err);
2847 }
2848
2849 static int
bfe_attach(dev_info_t * dip,ddi_attach_cmd_t cmd)2850 bfe_attach(dev_info_t *dip, ddi_attach_cmd_t cmd)
2851 {
2852 int unit;
2853 bfe_t *bfe;
2854 mac_register_t *macreg;
2855 int ret;
2856
2857 switch (cmd) {
2858 case DDI_RESUME:
2859 return (bfe_resume(dip));
2860
2861 case DDI_ATTACH:
2862 break;
2863
2864 default:
2865 return (DDI_FAILURE);
2866 }
2867
2868
2869 unit = ddi_get_instance(dip);
2870
2871 bfe = kmem_zalloc(sizeof (bfe_t), KM_SLEEP);
2872 bfe->bfe_dip = dip;
2873 bfe->bfe_unit = unit;
2874
2875 if (pci_config_setup(dip, &bfe->bfe_conf_handle) != DDI_SUCCESS) {
2876 bfe_error(dip, "pci_config_setup failed");
2877 goto fail0;
2878 }
2879
2880 /*
2881 * Enable IO space, Bus Master and Memory Space accessess.
2882 */
2883 ret = pci_config_get16(bfe->bfe_conf_handle, PCI_CONF_COMM);
2884 pci_config_put16(bfe->bfe_conf_handle, PCI_CONF_COMM,
2885 PCI_COMM_IO | PCI_COMM_MAE | PCI_COMM_ME | ret);
2886
2887 ddi_set_driver_private(dip, bfe);
2888
2889 /* Identify hardware */
2890 if (bfe_identify_hardware(bfe) == BFE_FAILURE) {
2891 bfe_error(dip, "Could not identify device");
2892 goto fail1;
2893 }
2894
2895 if (bfe_regs_map(bfe) != DDI_SUCCESS) {
2896 bfe_error(dip, "Could not map device registers");
2897 goto fail1;
2898 }
2899
2900 (void) bfe_get_chip_config(bfe);
2901
2902 /*
2903 * Register with MAC layer
2904 */
2905 if ((macreg = mac_alloc(MAC_VERSION)) == NULL) {
2906 bfe_error(dip, "mac_alloc() failed");
2907 goto fail2;
2908 }
2909
2910 macreg->m_type_ident = MAC_PLUGIN_IDENT_ETHER;
2911 macreg->m_driver = bfe;
2912 macreg->m_dip = dip;
2913 macreg->m_instance = unit;
2914 macreg->m_src_addr = bfe->bfe_ether_addr;
2915 macreg->m_callbacks = &bfe_mac_callbacks;
2916 macreg->m_min_sdu = 0;
2917 macreg->m_max_sdu = ETHERMTU;
2918 macreg->m_margin = VLAN_TAGSZ;
2919
2920 if ((ret = mac_register(macreg, &bfe->bfe_machdl)) != 0) {
2921 bfe_error(dip, "mac_register() failed with %d error", ret);
2922 mac_free(macreg);
2923 goto fail2;
2924 }
2925
2926 mac_free(macreg);
2927
2928 rw_init(&bfe->bfe_rwlock, NULL, RW_DRIVER,
2929 DDI_INTR_PRI(bfe->bfe_intrpri));
2930
2931 if (bfe_add_intr(bfe) != DDI_SUCCESS) {
2932 bfe_error(dip, "Could not add interrupt");
2933 goto fail3;
2934 }
2935
2936 if (bfe_rings_alloc(bfe) != DDI_SUCCESS) {
2937 bfe_error(dip, "Could not allocate TX/RX Ring");
2938 goto fail4;
2939 }
2940
2941 /* Init and then reset the chip */
2942 bfe->bfe_chip_action = 0;
2943 bfe_init_vars(bfe);
2944
2945 /* PHY will also start running */
2946 bfe_chip_reset(bfe);
2947
2948 /*
2949 * Even though we enable the interrupts here but chip's interrupt
2950 * is not enabled yet. It will be enabled once we plumb the interface.
2951 */
2952 if (ddi_intr_enable(bfe->bfe_intrhdl) != DDI_SUCCESS) {
2953 bfe_error(dip, "Could not enable interrupt");
2954 goto fail4;
2955 }
2956
2957 return (DDI_SUCCESS);
2958
2959 fail4:
2960 bfe_remove_intr(bfe);
2961 fail3:
2962 (void) mac_unregister(bfe->bfe_machdl);
2963 fail2:
2964 bfe_unmap_regs(bfe);
2965 fail1:
2966 pci_config_teardown(&bfe->bfe_conf_handle);
2967 fail0:
2968 kmem_free(bfe, sizeof (bfe_t));
2969 return (DDI_FAILURE);
2970 }
2971
2972 static int
bfe_detach(dev_info_t * devinfo,ddi_detach_cmd_t cmd)2973 bfe_detach(dev_info_t *devinfo, ddi_detach_cmd_t cmd)
2974 {
2975 bfe_t *bfe;
2976
2977 bfe = ddi_get_driver_private(devinfo);
2978
2979 switch (cmd) {
2980 case DDI_DETACH:
2981 /*
2982 * We need to stop the timer before grabbing locks otherwise
2983 * we can land-up in deadlock with untimeout.
2984 */
2985 bfe_stop_timer(bfe);
2986
2987 /*
2988 * First unregister with MAC layer before stopping DMA
2989 * engine.
2990 */
2991 if (mac_unregister(bfe->bfe_machdl) != DDI_SUCCESS)
2992 return (DDI_FAILURE);
2993
2994 bfe->bfe_machdl = NULL;
2995
2996 /*
2997 * Quiesce the chip first.
2998 */
2999 bfe_grab_locks(bfe);
3000 bfe_chip_halt(bfe);
3001 bfe_stop_phy(bfe);
3002 bfe_release_locks(bfe);
3003
3004 (void) ddi_intr_disable(bfe->bfe_intrhdl);
3005
3006 /* Make sure timer is gone. */
3007 bfe_stop_timer(bfe);
3008
3009 /*
3010 * Free the DMA resources for buffer and then descriptors
3011 */
3012 if (bfe->bfe_tx_ring.r_flags == BFE_RING_ALLOCATED) {
3013 /* TX */
3014 bfe_buffer_free(&bfe->bfe_tx_ring);
3015 bfe_ring_desc_free(&bfe->bfe_tx_ring);
3016 }
3017
3018 if (bfe->bfe_rx_ring.r_flags == BFE_RING_ALLOCATED) {
3019 /* RX */
3020 bfe_buffer_free(&bfe->bfe_rx_ring);
3021 bfe_ring_desc_free(&bfe->bfe_rx_ring);
3022 }
3023
3024 bfe_remove_intr(bfe);
3025 bfe_unmap_regs(bfe);
3026 pci_config_teardown(&bfe->bfe_conf_handle);
3027
3028 mutex_destroy(&bfe->bfe_tx_ring.r_lock);
3029 mutex_destroy(&bfe->bfe_rx_ring.r_lock);
3030 rw_destroy(&bfe->bfe_rwlock);
3031
3032 kmem_free(bfe, sizeof (bfe_t));
3033
3034 ddi_set_driver_private(devinfo, NULL);
3035 return (DDI_SUCCESS);
3036
3037 case DDI_SUSPEND:
3038 /*
3039 * We need to stop the timer before grabbing locks otherwise
3040 * we can land-up in deadlock with untimeout.
3041 */
3042 bfe_stop_timer(bfe);
3043
3044 /*
3045 * Grab all the locks first.
3046 */
3047 bfe_grab_locks(bfe);
3048 bfe_chip_halt(bfe);
3049 bfe_stop_phy(bfe);
3050 bfe->bfe_chip_state = BFE_CHIP_SUSPENDED;
3051 bfe_release_locks(bfe);
3052
3053 return (DDI_SUCCESS);
3054
3055 default:
3056 return (DDI_FAILURE);
3057 }
3058 }
3059
3060 /*
3061 * Quiesce the card for fast reboot
3062 */
3063 int
bfe_quiesce(dev_info_t * dev_info)3064 bfe_quiesce(dev_info_t *dev_info)
3065 {
3066 bfe_t *bfe;
3067
3068 bfe = ddi_get_driver_private(dev_info);
3069
3070 bfe_chip_halt(bfe);
3071 bfe_stop_phy(bfe);
3072 bfe->bfe_chip_state = BFE_CHIP_QUIESCED;
3073
3074 return (DDI_SUCCESS);
3075 }
3076
3077 static struct cb_ops bfe_cb_ops = {
3078 nulldev, /* cb_open */
3079 nulldev, /* cb_close */
3080 nodev, /* cb_strategy */
3081 nodev, /* cb_print */
3082 nodev, /* cb_dump */
3083 nodev, /* cb_read */
3084 nodev, /* cb_write */
3085 nodev, /* cb_ioctl */
3086 nodev, /* cb_devmap */
3087 nodev, /* cb_mmap */
3088 nodev, /* cb_segmap */
3089 nochpoll, /* cb_chpoll */
3090 ddi_prop_op, /* cb_prop_op */
3091 NULL, /* cb_stream */
3092 D_MP | D_HOTPLUG, /* cb_flag */
3093 CB_REV, /* cb_rev */
3094 nodev, /* cb_aread */
3095 nodev /* cb_awrite */
3096 };
3097
3098 static struct dev_ops bfe_dev_ops = {
3099 DEVO_REV, /* devo_rev */
3100 0, /* devo_refcnt */
3101 NULL, /* devo_getinfo */
3102 nulldev, /* devo_identify */
3103 nulldev, /* devo_probe */
3104 bfe_attach, /* devo_attach */
3105 bfe_detach, /* devo_detach */
3106 nodev, /* devo_reset */
3107 &bfe_cb_ops, /* devo_cb_ops */
3108 NULL, /* devo_bus_ops */
3109 ddi_power, /* devo_power */
3110 bfe_quiesce /* devo_quiesce */
3111 };
3112
3113 static struct modldrv bfe_modldrv = {
3114 &mod_driverops,
3115 bfe_ident,
3116 &bfe_dev_ops
3117 };
3118
3119 static struct modlinkage modlinkage = {
3120 MODREV_1, (void *)&bfe_modldrv, NULL
3121 };
3122
3123 int
_info(struct modinfo * modinfop)3124 _info(struct modinfo *modinfop)
3125 {
3126 return (mod_info(&modlinkage, modinfop));
3127 }
3128
3129 int
_init(void)3130 _init(void)
3131 {
3132 int status;
3133
3134 mac_init_ops(&bfe_dev_ops, MODULE_NAME);
3135 status = mod_install(&modlinkage);
3136 if (status == DDI_FAILURE)
3137 mac_fini_ops(&bfe_dev_ops);
3138 return (status);
3139 }
3140
3141 int
_fini(void)3142 _fini(void)
3143 {
3144 int status;
3145
3146 status = mod_remove(&modlinkage);
3147 if (status == 0) {
3148 mac_fini_ops(&bfe_dev_ops);
3149 }
3150 return (status);
3151 }
3152