xref: /illumos-gate/usr/src/uts/common/io/igb/igb_gld.c (revision 2efb3bf9c7f4cf34038896f1431531c93d3f57c2)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 
22 /*
23  * Copyright(c) 2007-2010 Intel Corporation. All rights reserved.
24  */
25 
26 /*
27  * Copyright (c) 2008, 2010, Oracle and/or its affiliates. All rights reserved.
28  * Copyright 2013, Nexenta Systems, Inc. All rights reserved.
29  */
30 
31 #include "igb_sw.h"
32 
33 int
34 igb_m_stat(void *arg, uint_t stat, uint64_t *val)
35 {
36 	igb_t *igb = (igb_t *)arg;
37 	struct e1000_hw *hw = &igb->hw;
38 	igb_stat_t *igb_ks;
39 	uint32_t low_val, high_val;
40 
41 	igb_ks = (igb_stat_t *)igb->igb_ks->ks_data;
42 
43 	mutex_enter(&igb->gen_lock);
44 
45 	if (igb->igb_state & IGB_SUSPENDED) {
46 		mutex_exit(&igb->gen_lock);
47 		return (ECANCELED);
48 	}
49 
50 	switch (stat) {
51 	case MAC_STAT_IFSPEED:
52 		*val = igb->link_speed * 1000000ull;
53 		break;
54 
55 	case MAC_STAT_MULTIRCV:
56 		igb_ks->mprc.value.ui64 +=
57 		    E1000_READ_REG(hw, E1000_MPRC);
58 		*val = igb_ks->mprc.value.ui64;
59 		break;
60 
61 	case MAC_STAT_BRDCSTRCV:
62 		igb_ks->bprc.value.ui64 +=
63 		    E1000_READ_REG(hw, E1000_BPRC);
64 		*val = igb_ks->bprc.value.ui64;
65 		break;
66 
67 	case MAC_STAT_MULTIXMT:
68 		igb_ks->mptc.value.ui64 +=
69 		    E1000_READ_REG(hw, E1000_MPTC);
70 		*val = igb_ks->mptc.value.ui64;
71 		break;
72 
73 	case MAC_STAT_BRDCSTXMT:
74 		igb_ks->bptc.value.ui64 +=
75 		    E1000_READ_REG(hw, E1000_BPTC);
76 		*val = igb_ks->bptc.value.ui64;
77 		break;
78 
79 	case MAC_STAT_NORCVBUF:
80 		igb_ks->rnbc.value.ui64 +=
81 		    E1000_READ_REG(hw, E1000_RNBC);
82 		*val = igb_ks->rnbc.value.ui64;
83 		break;
84 
85 	case MAC_STAT_IERRORS:
86 		igb_ks->rxerrc.value.ui64 +=
87 		    E1000_READ_REG(hw, E1000_RXERRC);
88 		igb_ks->algnerrc.value.ui64 +=
89 		    E1000_READ_REG(hw, E1000_ALGNERRC);
90 		igb_ks->rlec.value.ui64 +=
91 		    E1000_READ_REG(hw, E1000_RLEC);
92 		igb_ks->crcerrs.value.ui64 +=
93 		    E1000_READ_REG(hw, E1000_CRCERRS);
94 		igb_ks->cexterr.value.ui64 +=
95 		    E1000_READ_REG(hw, E1000_CEXTERR);
96 		*val = igb_ks->rxerrc.value.ui64 +
97 		    igb_ks->algnerrc.value.ui64 +
98 		    igb_ks->rlec.value.ui64 +
99 		    igb_ks->crcerrs.value.ui64 +
100 		    igb_ks->cexterr.value.ui64;
101 		break;
102 
103 	case MAC_STAT_NOXMTBUF:
104 		*val = 0;
105 		break;
106 
107 	case MAC_STAT_OERRORS:
108 		igb_ks->ecol.value.ui64 +=
109 		    E1000_READ_REG(hw, E1000_ECOL);
110 		*val = igb_ks->ecol.value.ui64;
111 		break;
112 
113 	case MAC_STAT_COLLISIONS:
114 		igb_ks->colc.value.ui64 +=
115 		    E1000_READ_REG(hw, E1000_COLC);
116 		*val = igb_ks->colc.value.ui64;
117 		break;
118 
119 	case MAC_STAT_RBYTES:
120 		/*
121 		 * The 64-bit register will reset whenever the upper
122 		 * 32 bits are read. So we need to read the lower
123 		 * 32 bits first, then read the upper 32 bits.
124 		 */
125 		low_val = E1000_READ_REG(hw, E1000_TORL);
126 		high_val = E1000_READ_REG(hw, E1000_TORH);
127 		igb_ks->tor.value.ui64 +=
128 		    (uint64_t)high_val << 32 | (uint64_t)low_val;
129 		*val = igb_ks->tor.value.ui64;
130 		break;
131 
132 	case MAC_STAT_IPACKETS:
133 		igb_ks->tpr.value.ui64 +=
134 		    E1000_READ_REG(hw, E1000_TPR);
135 		*val = igb_ks->tpr.value.ui64;
136 		break;
137 
138 	case MAC_STAT_OBYTES:
139 		/*
140 		 * The 64-bit register will reset whenever the upper
141 		 * 32 bits are read. So we need to read the lower
142 		 * 32 bits first, then read the upper 32 bits.
143 		 */
144 		low_val = E1000_READ_REG(hw, E1000_TOTL);
145 		high_val = E1000_READ_REG(hw, E1000_TOTH);
146 		igb_ks->tot.value.ui64 +=
147 		    (uint64_t)high_val << 32 | (uint64_t)low_val;
148 		*val = igb_ks->tot.value.ui64;
149 		break;
150 
151 	case MAC_STAT_OPACKETS:
152 		igb_ks->tpt.value.ui64 +=
153 		    E1000_READ_REG(hw, E1000_TPT);
154 		*val = igb_ks->tpt.value.ui64;
155 		break;
156 
157 	/* RFC 1643 stats */
158 	case ETHER_STAT_ALIGN_ERRORS:
159 		igb_ks->algnerrc.value.ui64 +=
160 		    E1000_READ_REG(hw, E1000_ALGNERRC);
161 		*val = igb_ks->algnerrc.value.ui64;
162 		break;
163 
164 	case ETHER_STAT_FCS_ERRORS:
165 		igb_ks->crcerrs.value.ui64 +=
166 		    E1000_READ_REG(hw, E1000_CRCERRS);
167 		*val = igb_ks->crcerrs.value.ui64;
168 		break;
169 
170 	case ETHER_STAT_FIRST_COLLISIONS:
171 		igb_ks->scc.value.ui64 +=
172 		    E1000_READ_REG(hw, E1000_SCC);
173 		*val = igb_ks->scc.value.ui64;
174 		break;
175 
176 	case ETHER_STAT_MULTI_COLLISIONS:
177 		igb_ks->mcc.value.ui64 +=
178 		    E1000_READ_REG(hw, E1000_MCC);
179 		*val = igb_ks->mcc.value.ui64;
180 		break;
181 
182 	case ETHER_STAT_SQE_ERRORS:
183 		igb_ks->sec.value.ui64 +=
184 		    E1000_READ_REG(hw, E1000_SEC);
185 		*val = igb_ks->sec.value.ui64;
186 		break;
187 
188 	case ETHER_STAT_DEFER_XMTS:
189 		igb_ks->dc.value.ui64 +=
190 		    E1000_READ_REG(hw, E1000_DC);
191 		*val = igb_ks->dc.value.ui64;
192 		break;
193 
194 	case ETHER_STAT_TX_LATE_COLLISIONS:
195 		igb_ks->latecol.value.ui64 +=
196 		    E1000_READ_REG(hw, E1000_LATECOL);
197 		*val = igb_ks->latecol.value.ui64;
198 		break;
199 
200 	case ETHER_STAT_EX_COLLISIONS:
201 		igb_ks->ecol.value.ui64 +=
202 		    E1000_READ_REG(hw, E1000_ECOL);
203 		*val = igb_ks->ecol.value.ui64;
204 		break;
205 
206 	case ETHER_STAT_MACXMT_ERRORS:
207 		igb_ks->ecol.value.ui64 +=
208 		    E1000_READ_REG(hw, E1000_ECOL);
209 		*val = igb_ks->ecol.value.ui64;
210 		break;
211 
212 	case ETHER_STAT_CARRIER_ERRORS:
213 		igb_ks->cexterr.value.ui64 +=
214 		    E1000_READ_REG(hw, E1000_CEXTERR);
215 		*val = igb_ks->cexterr.value.ui64;
216 		break;
217 
218 	case ETHER_STAT_TOOLONG_ERRORS:
219 		igb_ks->roc.value.ui64 +=
220 		    E1000_READ_REG(hw, E1000_ROC);
221 		*val = igb_ks->roc.value.ui64;
222 		break;
223 
224 	case ETHER_STAT_MACRCV_ERRORS:
225 		igb_ks->rxerrc.value.ui64 +=
226 		    E1000_READ_REG(hw, E1000_RXERRC);
227 		*val = igb_ks->rxerrc.value.ui64;
228 		break;
229 
230 	/* MII/GMII stats */
231 	case ETHER_STAT_XCVR_ADDR:
232 		/* The Internal PHY's MDI address for each MAC is 1 */
233 		*val = 1;
234 		break;
235 
236 	case ETHER_STAT_XCVR_ID:
237 		*val = hw->phy.id | hw->phy.revision;
238 		break;
239 
240 	case ETHER_STAT_XCVR_INUSE:
241 		switch (igb->link_speed) {
242 		case SPEED_1000:
243 			*val =
244 			    (hw->phy.media_type == e1000_media_type_copper) ?
245 			    XCVR_1000T : XCVR_1000X;
246 			break;
247 		case SPEED_100:
248 			*val =
249 			    (hw->phy.media_type == e1000_media_type_copper) ?
250 			    (igb->param_100t4_cap == 1) ?
251 			    XCVR_100T4 : XCVR_100T2 : XCVR_100X;
252 			break;
253 		case SPEED_10:
254 			*val = XCVR_10;
255 			break;
256 		default:
257 			*val = XCVR_NONE;
258 			break;
259 		}
260 		break;
261 
262 	case ETHER_STAT_CAP_1000FDX:
263 		*val = igb->param_1000fdx_cap;
264 		break;
265 
266 	case ETHER_STAT_CAP_1000HDX:
267 		*val = igb->param_1000hdx_cap;
268 		break;
269 
270 	case ETHER_STAT_CAP_100FDX:
271 		*val = igb->param_100fdx_cap;
272 		break;
273 
274 	case ETHER_STAT_CAP_100HDX:
275 		*val = igb->param_100hdx_cap;
276 		break;
277 
278 	case ETHER_STAT_CAP_10FDX:
279 		*val = igb->param_10fdx_cap;
280 		break;
281 
282 	case ETHER_STAT_CAP_10HDX:
283 		*val = igb->param_10hdx_cap;
284 		break;
285 
286 	case ETHER_STAT_CAP_ASMPAUSE:
287 		*val = igb->param_asym_pause_cap;
288 		break;
289 
290 	case ETHER_STAT_CAP_PAUSE:
291 		*val = igb->param_pause_cap;
292 		break;
293 
294 	case ETHER_STAT_CAP_AUTONEG:
295 		*val = igb->param_autoneg_cap;
296 		break;
297 
298 	case ETHER_STAT_ADV_CAP_1000FDX:
299 		*val = igb->param_adv_1000fdx_cap;
300 		break;
301 
302 	case ETHER_STAT_ADV_CAP_1000HDX:
303 		*val = igb->param_adv_1000hdx_cap;
304 		break;
305 
306 	case ETHER_STAT_ADV_CAP_100FDX:
307 		*val = igb->param_adv_100fdx_cap;
308 		break;
309 
310 	case ETHER_STAT_ADV_CAP_100HDX:
311 		*val = igb->param_adv_100hdx_cap;
312 		break;
313 
314 	case ETHER_STAT_ADV_CAP_10FDX:
315 		*val = igb->param_adv_10fdx_cap;
316 		break;
317 
318 	case ETHER_STAT_ADV_CAP_10HDX:
319 		*val = igb->param_adv_10hdx_cap;
320 		break;
321 
322 	case ETHER_STAT_ADV_CAP_ASMPAUSE:
323 		*val = igb->param_adv_asym_pause_cap;
324 		break;
325 
326 	case ETHER_STAT_ADV_CAP_PAUSE:
327 		*val = igb->param_adv_pause_cap;
328 		break;
329 
330 	case ETHER_STAT_ADV_CAP_AUTONEG:
331 		*val = hw->mac.autoneg;
332 		break;
333 
334 	case ETHER_STAT_LP_CAP_1000FDX:
335 		*val = igb->param_lp_1000fdx_cap;
336 		break;
337 
338 	case ETHER_STAT_LP_CAP_1000HDX:
339 		*val = igb->param_lp_1000hdx_cap;
340 		break;
341 
342 	case ETHER_STAT_LP_CAP_100FDX:
343 		*val = igb->param_lp_100fdx_cap;
344 		break;
345 
346 	case ETHER_STAT_LP_CAP_100HDX:
347 		*val = igb->param_lp_100hdx_cap;
348 		break;
349 
350 	case ETHER_STAT_LP_CAP_10FDX:
351 		*val = igb->param_lp_10fdx_cap;
352 		break;
353 
354 	case ETHER_STAT_LP_CAP_10HDX:
355 		*val = igb->param_lp_10hdx_cap;
356 		break;
357 
358 	case ETHER_STAT_LP_CAP_ASMPAUSE:
359 		*val = igb->param_lp_asym_pause_cap;
360 		break;
361 
362 	case ETHER_STAT_LP_CAP_PAUSE:
363 		*val = igb->param_lp_pause_cap;
364 		break;
365 
366 	case ETHER_STAT_LP_CAP_AUTONEG:
367 		*val = igb->param_lp_autoneg_cap;
368 		break;
369 
370 	case ETHER_STAT_LINK_ASMPAUSE:
371 		*val = igb->param_asym_pause_cap;
372 		break;
373 
374 	case ETHER_STAT_LINK_PAUSE:
375 		*val = igb->param_pause_cap;
376 		break;
377 
378 	case ETHER_STAT_LINK_AUTONEG:
379 		*val = hw->mac.autoneg;
380 		break;
381 
382 	case ETHER_STAT_LINK_DUPLEX:
383 		*val = (igb->link_duplex == FULL_DUPLEX) ?
384 		    LINK_DUPLEX_FULL : LINK_DUPLEX_HALF;
385 		break;
386 
387 	case ETHER_STAT_TOOSHORT_ERRORS:
388 		igb_ks->ruc.value.ui64 +=
389 		    E1000_READ_REG(hw, E1000_RUC);
390 		*val = igb_ks->ruc.value.ui64;
391 		break;
392 
393 	case ETHER_STAT_CAP_REMFAULT:
394 		*val = igb->param_rem_fault;
395 		break;
396 
397 	case ETHER_STAT_ADV_REMFAULT:
398 		*val = igb->param_adv_rem_fault;
399 		break;
400 
401 	case ETHER_STAT_LP_REMFAULT:
402 		*val = igb->param_lp_rem_fault;
403 		break;
404 
405 	case ETHER_STAT_JABBER_ERRORS:
406 		igb_ks->rjc.value.ui64 +=
407 		    E1000_READ_REG(hw, E1000_RJC);
408 		*val = igb_ks->rjc.value.ui64;
409 		break;
410 
411 	case ETHER_STAT_CAP_100T4:
412 		*val = igb->param_100t4_cap;
413 		break;
414 
415 	case ETHER_STAT_ADV_CAP_100T4:
416 		*val = igb->param_adv_100t4_cap;
417 		break;
418 
419 	case ETHER_STAT_LP_CAP_100T4:
420 		*val = igb->param_lp_100t4_cap;
421 		break;
422 
423 	default:
424 		mutex_exit(&igb->gen_lock);
425 		return (ENOTSUP);
426 	}
427 
428 	mutex_exit(&igb->gen_lock);
429 
430 	if (igb_check_acc_handle(igb->osdep.reg_handle) != DDI_FM_OK) {
431 		ddi_fm_service_impact(igb->dip, DDI_SERVICE_DEGRADED);
432 		return (EIO);
433 	}
434 
435 	return (0);
436 }
437 
438 /*
439  * Bring the device out of the reset/quiesced state that it
440  * was in when the interface was registered.
441  */
442 int
443 igb_m_start(void *arg)
444 {
445 	igb_t *igb = (igb_t *)arg;
446 
447 	mutex_enter(&igb->gen_lock);
448 
449 	if (igb->igb_state & IGB_SUSPENDED) {
450 		mutex_exit(&igb->gen_lock);
451 		return (ECANCELED);
452 	}
453 
454 	if (igb_start(igb, B_TRUE) != IGB_SUCCESS) {
455 		mutex_exit(&igb->gen_lock);
456 		return (EIO);
457 	}
458 
459 	atomic_or_32(&igb->igb_state, IGB_STARTED);
460 
461 	mutex_exit(&igb->gen_lock);
462 
463 	/*
464 	 * Enable and start the watchdog timer
465 	 */
466 	igb_enable_watchdog_timer(igb);
467 
468 	return (0);
469 }
470 
471 /*
472  * Stop the device and put it in a reset/quiesced state such
473  * that the interface can be unregistered.
474  */
475 void
476 igb_m_stop(void *arg)
477 {
478 	igb_t *igb = (igb_t *)arg;
479 
480 	mutex_enter(&igb->gen_lock);
481 
482 	if (igb->igb_state & IGB_SUSPENDED) {
483 		mutex_exit(&igb->gen_lock);
484 		return;
485 	}
486 
487 	atomic_and_32(&igb->igb_state, ~IGB_STARTED);
488 
489 	igb_stop(igb, B_TRUE);
490 
491 	mutex_exit(&igb->gen_lock);
492 
493 	/*
494 	 * Disable and stop the watchdog timer
495 	 */
496 	igb_disable_watchdog_timer(igb);
497 }
498 
499 /*
500  * Set the promiscuity of the device.
501  */
502 int
503 igb_m_promisc(void *arg, boolean_t on)
504 {
505 	igb_t *igb = (igb_t *)arg;
506 	uint32_t reg_val;
507 
508 	mutex_enter(&igb->gen_lock);
509 
510 	if (igb->igb_state & IGB_SUSPENDED) {
511 		mutex_exit(&igb->gen_lock);
512 		return (ECANCELED);
513 	}
514 
515 	reg_val = E1000_READ_REG(&igb->hw, E1000_RCTL);
516 
517 	if (on)
518 		reg_val |= (E1000_RCTL_UPE | E1000_RCTL_MPE);
519 	else
520 		reg_val &= (~(E1000_RCTL_UPE | E1000_RCTL_MPE));
521 
522 	E1000_WRITE_REG(&igb->hw, E1000_RCTL, reg_val);
523 
524 	mutex_exit(&igb->gen_lock);
525 
526 	if (igb_check_acc_handle(igb->osdep.reg_handle) != DDI_FM_OK) {
527 		ddi_fm_service_impact(igb->dip, DDI_SERVICE_DEGRADED);
528 		return (EIO);
529 	}
530 
531 	return (0);
532 }
533 
534 /*
535  * Add/remove the addresses to/from the set of multicast
536  * addresses for which the device will receive packets.
537  */
538 int
539 igb_m_multicst(void *arg, boolean_t add, const uint8_t *mcst_addr)
540 {
541 	igb_t *igb = (igb_t *)arg;
542 	int result;
543 
544 	mutex_enter(&igb->gen_lock);
545 
546 	if (igb->igb_state & IGB_SUSPENDED) {
547 		mutex_exit(&igb->gen_lock);
548 		return (ECANCELED);
549 	}
550 
551 	result = (add) ? igb_multicst_add(igb, mcst_addr)
552 	    : igb_multicst_remove(igb, mcst_addr);
553 
554 	mutex_exit(&igb->gen_lock);
555 
556 	return (result);
557 }
558 
559 /*
560  * Pass on M_IOCTL messages passed to the DLD, and support
561  * private IOCTLs for debugging and ndd.
562  */
563 void
564 igb_m_ioctl(void *arg, queue_t *q, mblk_t *mp)
565 {
566 	igb_t *igb = (igb_t *)arg;
567 	struct iocblk *iocp;
568 	enum ioc_reply status;
569 
570 	iocp = (struct iocblk *)(uintptr_t)mp->b_rptr;
571 	iocp->ioc_error = 0;
572 
573 	mutex_enter(&igb->gen_lock);
574 	if (igb->igb_state & IGB_SUSPENDED) {
575 		mutex_exit(&igb->gen_lock);
576 		miocnak(q, mp, 0, EINVAL);
577 		return;
578 	}
579 	mutex_exit(&igb->gen_lock);
580 
581 	switch (iocp->ioc_cmd) {
582 	case LB_GET_INFO_SIZE:
583 	case LB_GET_INFO:
584 	case LB_GET_MODE:
585 	case LB_SET_MODE:
586 		status = igb_loopback_ioctl(igb, iocp, mp);
587 		break;
588 
589 	default:
590 		status = IOC_INVAL;
591 		break;
592 	}
593 
594 	/*
595 	 * Decide how to reply
596 	 */
597 	switch (status) {
598 	default:
599 	case IOC_INVAL:
600 		/*
601 		 * Error, reply with a NAK and EINVAL or the specified error
602 		 */
603 		miocnak(q, mp, 0, iocp->ioc_error == 0 ?
604 		    EINVAL : iocp->ioc_error);
605 		break;
606 
607 	case IOC_DONE:
608 		/*
609 		 * OK, reply already sent
610 		 */
611 		break;
612 
613 	case IOC_ACK:
614 		/*
615 		 * OK, reply with an ACK
616 		 */
617 		miocack(q, mp, 0, 0);
618 		break;
619 
620 	case IOC_REPLY:
621 		/*
622 		 * OK, send prepared reply as ACK or NAK
623 		 */
624 		mp->b_datap->db_type = iocp->ioc_error == 0 ?
625 		    M_IOCACK : M_IOCNAK;
626 		qreply(q, mp);
627 		break;
628 	}
629 }
630 
631 /*
632  * Add a MAC address to the target RX group.
633  */
634 static int
635 igb_addmac(void *arg, const uint8_t *mac_addr)
636 {
637 	igb_rx_group_t *rx_group = (igb_rx_group_t *)arg;
638 	igb_t *igb = rx_group->igb;
639 	struct e1000_hw *hw = &igb->hw;
640 	int i, slot;
641 
642 	mutex_enter(&igb->gen_lock);
643 
644 	if (igb->igb_state & IGB_SUSPENDED) {
645 		mutex_exit(&igb->gen_lock);
646 		return (ECANCELED);
647 	}
648 
649 	if (igb->unicst_avail == 0) {
650 		/* no slots available */
651 		mutex_exit(&igb->gen_lock);
652 		return (ENOSPC);
653 	}
654 
655 	/*
656 	 * The slots from 0 to igb->num_rx_groups are reserved slots which
657 	 * are 1 to 1 mapped with group index directly. The other slots are
658 	 * shared between the all of groups. While adding a MAC address,
659 	 * it will try to set the reserved slots first, then the shared slots.
660 	 */
661 	slot = -1;
662 	if (igb->unicst_addr[rx_group->index].mac.set == 1) {
663 		/*
664 		 * The reserved slot for current group is used, find the free
665 		 * slots in the shared slots.
666 		 */
667 		for (i = igb->num_rx_groups; i < igb->unicst_total; i++) {
668 			if (igb->unicst_addr[i].mac.set == 0) {
669 				slot = i;
670 				break;
671 			}
672 		}
673 	} else
674 		slot = rx_group->index;
675 
676 	if (slot == -1) {
677 		/* no slots available in the shared slots */
678 		mutex_exit(&igb->gen_lock);
679 		return (ENOSPC);
680 	}
681 
682 	/* Set VMDq according to the mode supported by hardware. */
683 	e1000_rar_set_vmdq(hw, mac_addr, slot, igb->vmdq_mode, rx_group->index);
684 
685 	bcopy(mac_addr, igb->unicst_addr[slot].mac.addr, ETHERADDRL);
686 	igb->unicst_addr[slot].mac.group_index = rx_group->index;
687 	igb->unicst_addr[slot].mac.set = 1;
688 	igb->unicst_avail--;
689 
690 	mutex_exit(&igb->gen_lock);
691 
692 	return (0);
693 }
694 
695 /*
696  * Remove a MAC address from the specified RX group.
697  */
698 static int
699 igb_remmac(void *arg, const uint8_t *mac_addr)
700 {
701 	igb_rx_group_t *rx_group = (igb_rx_group_t *)arg;
702 	igb_t *igb = rx_group->igb;
703 	struct e1000_hw *hw = &igb->hw;
704 	int slot;
705 
706 	mutex_enter(&igb->gen_lock);
707 
708 	if (igb->igb_state & IGB_SUSPENDED) {
709 		mutex_exit(&igb->gen_lock);
710 		return (ECANCELED);
711 	}
712 
713 	slot = igb_unicst_find(igb, mac_addr);
714 	if (slot == -1) {
715 		mutex_exit(&igb->gen_lock);
716 		return (EINVAL);
717 	}
718 
719 	if (igb->unicst_addr[slot].mac.set == 0) {
720 		mutex_exit(&igb->gen_lock);
721 		return (EINVAL);
722 	}
723 
724 	/* Clear the MAC ddress in the slot */
725 	e1000_rar_clear(hw, slot);
726 	igb->unicst_addr[slot].mac.set = 0;
727 	igb->unicst_avail++;
728 
729 	mutex_exit(&igb->gen_lock);
730 
731 	return (0);
732 }
733 
734 /*
735  * Enable interrupt on the specificed rx ring.
736  */
737 int
738 igb_rx_ring_intr_enable(mac_intr_handle_t intrh)
739 {
740 	igb_rx_ring_t *rx_ring = (igb_rx_ring_t *)intrh;
741 	igb_t *igb = rx_ring->igb;
742 	struct e1000_hw *hw = &igb->hw;
743 	uint32_t index = rx_ring->index;
744 
745 	if (igb->intr_type == DDI_INTR_TYPE_MSIX) {
746 		/* Interrupt enabling for MSI-X */
747 		igb->eims_mask |= (E1000_EICR_RX_QUEUE0 << index);
748 		E1000_WRITE_REG(hw, E1000_EIMS, igb->eims_mask);
749 		E1000_WRITE_REG(hw, E1000_EIAC, igb->eims_mask);
750 	} else {
751 		ASSERT(index == 0);
752 		/* Interrupt enabling for MSI and legacy */
753 		igb->ims_mask |= E1000_IMS_RXT0;
754 		E1000_WRITE_REG(hw, E1000_IMS, igb->ims_mask);
755 	}
756 
757 	E1000_WRITE_FLUSH(hw);
758 
759 	return (0);
760 }
761 
762 /*
763  * Disable interrupt on the specificed rx ring.
764  */
765 int
766 igb_rx_ring_intr_disable(mac_intr_handle_t intrh)
767 {
768 	igb_rx_ring_t *rx_ring = (igb_rx_ring_t *)intrh;
769 	igb_t *igb = rx_ring->igb;
770 	struct e1000_hw *hw = &igb->hw;
771 	uint32_t index = rx_ring->index;
772 
773 	if (igb->intr_type == DDI_INTR_TYPE_MSIX) {
774 		/* Interrupt disabling for MSI-X */
775 		igb->eims_mask &= ~(E1000_EICR_RX_QUEUE0 << index);
776 		E1000_WRITE_REG(hw, E1000_EIMC,
777 		    (E1000_EICR_RX_QUEUE0 << index));
778 		E1000_WRITE_REG(hw, E1000_EIAC, igb->eims_mask);
779 	} else {
780 		ASSERT(index == 0);
781 		/* Interrupt disabling for MSI and legacy */
782 		igb->ims_mask &= ~E1000_IMS_RXT0;
783 		E1000_WRITE_REG(hw, E1000_IMC, E1000_IMS_RXT0);
784 	}
785 
786 	E1000_WRITE_FLUSH(hw);
787 
788 	return (0);
789 }
790 
791 /*
792  * Get the global ring index by a ring index within a group.
793  */
794 int
795 igb_get_rx_ring_index(igb_t *igb, int gindex, int rindex)
796 {
797 	igb_rx_ring_t *rx_ring;
798 	int i;
799 
800 	for (i = 0; i < igb->num_rx_rings; i++) {
801 		rx_ring = &igb->rx_rings[i];
802 		if (rx_ring->group_index == gindex)
803 			rindex--;
804 		if (rindex < 0)
805 			return (i);
806 	}
807 
808 	return (-1);
809 }
810 
811 static int
812 igb_ring_start(mac_ring_driver_t rh, uint64_t mr_gen_num)
813 {
814 	igb_rx_ring_t *rx_ring = (igb_rx_ring_t *)rh;
815 
816 	mutex_enter(&rx_ring->rx_lock);
817 	rx_ring->ring_gen_num = mr_gen_num;
818 	mutex_exit(&rx_ring->rx_lock);
819 	return (0);
820 }
821 
822 /*
823  * Callback funtion for MAC layer to register all rings.
824  */
825 /* ARGSUSED */
826 void
827 igb_fill_ring(void *arg, mac_ring_type_t rtype, const int rg_index,
828     const int index, mac_ring_info_t *infop, mac_ring_handle_t rh)
829 {
830 	igb_t *igb = (igb_t *)arg;
831 	mac_intr_t *mintr = &infop->mri_intr;
832 
833 	switch (rtype) {
834 	case MAC_RING_TYPE_RX: {
835 		igb_rx_ring_t *rx_ring;
836 		int global_index;
837 
838 		/*
839 		 * 'index' is the ring index within the group.
840 		 * We need the global ring index by searching in group.
841 		 */
842 		global_index = igb_get_rx_ring_index(igb, rg_index, index);
843 
844 		ASSERT(global_index >= 0);
845 
846 		rx_ring = &igb->rx_rings[global_index];
847 		rx_ring->ring_handle = rh;
848 
849 		infop->mri_driver = (mac_ring_driver_t)rx_ring;
850 		infop->mri_start = igb_ring_start;
851 		infop->mri_stop = NULL;
852 		infop->mri_poll = (mac_ring_poll_t)igb_rx_ring_poll;
853 		infop->mri_stat = igb_rx_ring_stat;
854 
855 		mintr->mi_handle = (mac_intr_handle_t)rx_ring;
856 		mintr->mi_enable = igb_rx_ring_intr_enable;
857 		mintr->mi_disable = igb_rx_ring_intr_disable;
858 		if (igb->intr_type & (DDI_INTR_TYPE_MSIX | DDI_INTR_TYPE_MSI)) {
859 			mintr->mi_ddi_handle =
860 			    igb->htable[rx_ring->intr_vector];
861 		}
862 		break;
863 	}
864 	case MAC_RING_TYPE_TX: {
865 		ASSERT(index < igb->num_tx_rings);
866 
867 		igb_tx_ring_t *tx_ring = &igb->tx_rings[index];
868 		tx_ring->ring_handle = rh;
869 
870 		infop->mri_driver = (mac_ring_driver_t)tx_ring;
871 		infop->mri_start = NULL;
872 		infop->mri_stop = NULL;
873 		infop->mri_tx = igb_tx_ring_send;
874 		infop->mri_stat = igb_tx_ring_stat;
875 		if (igb->intr_type & (DDI_INTR_TYPE_MSIX | DDI_INTR_TYPE_MSI)) {
876 			mintr->mi_ddi_handle =
877 			    igb->htable[tx_ring->intr_vector];
878 		}
879 		break;
880 	}
881 	default:
882 		break;
883 	}
884 }
885 
886 void
887 igb_fill_group(void *arg, mac_ring_type_t rtype, const int index,
888     mac_group_info_t *infop, mac_group_handle_t gh)
889 {
890 	igb_t *igb = (igb_t *)arg;
891 
892 	switch (rtype) {
893 	case MAC_RING_TYPE_RX: {
894 		igb_rx_group_t *rx_group;
895 
896 		ASSERT((index >= 0) && (index < igb->num_rx_groups));
897 
898 		rx_group = &igb->rx_groups[index];
899 		rx_group->group_handle = gh;
900 
901 		infop->mgi_driver = (mac_group_driver_t)rx_group;
902 		infop->mgi_start = NULL;
903 		infop->mgi_stop = NULL;
904 		infop->mgi_addmac = igb_addmac;
905 		infop->mgi_remmac = igb_remmac;
906 		infop->mgi_count = (igb->num_rx_rings / igb->num_rx_groups);
907 
908 		break;
909 	}
910 	case MAC_RING_TYPE_TX:
911 		break;
912 	default:
913 		break;
914 	}
915 }
916 
917 /*
918  * Obtain the MAC's capabilities and associated data from
919  * the driver.
920  */
921 boolean_t
922 igb_m_getcapab(void *arg, mac_capab_t cap, void *cap_data)
923 {
924 	igb_t *igb = (igb_t *)arg;
925 
926 	switch (cap) {
927 	case MAC_CAPAB_HCKSUM: {
928 		uint32_t *tx_hcksum_flags = cap_data;
929 
930 		/*
931 		 * We advertise our capabilities only if tx hcksum offload is
932 		 * enabled.  On receive, the stack will accept checksummed
933 		 * packets anyway, even if we haven't said we can deliver
934 		 * them.
935 		 */
936 		if (!igb->tx_hcksum_enable)
937 			return (B_FALSE);
938 
939 		*tx_hcksum_flags = HCKSUM_INET_PARTIAL | HCKSUM_IPHDRCKSUM;
940 		break;
941 	}
942 	case MAC_CAPAB_LSO: {
943 		mac_capab_lso_t *cap_lso = cap_data;
944 
945 		if (igb->lso_enable) {
946 			cap_lso->lso_flags = LSO_TX_BASIC_TCP_IPV4;
947 			cap_lso->lso_basic_tcp_ipv4.lso_max = IGB_LSO_MAXLEN;
948 			break;
949 		} else {
950 			return (B_FALSE);
951 		}
952 	}
953 	case MAC_CAPAB_RINGS: {
954 		mac_capab_rings_t *cap_rings = cap_data;
955 
956 		switch (cap_rings->mr_type) {
957 		case MAC_RING_TYPE_RX:
958 			cap_rings->mr_group_type = MAC_GROUP_TYPE_STATIC;
959 			cap_rings->mr_rnum = igb->num_rx_rings;
960 			cap_rings->mr_gnum = igb->num_rx_groups;
961 			cap_rings->mr_rget = igb_fill_ring;
962 			cap_rings->mr_gget = igb_fill_group;
963 			cap_rings->mr_gaddring = NULL;
964 			cap_rings->mr_gremring = NULL;
965 
966 			break;
967 		case MAC_RING_TYPE_TX:
968 			cap_rings->mr_group_type = MAC_GROUP_TYPE_STATIC;
969 			cap_rings->mr_rnum = igb->num_tx_rings;
970 			cap_rings->mr_gnum = 0;
971 			cap_rings->mr_rget = igb_fill_ring;
972 			cap_rings->mr_gget = NULL;
973 
974 			break;
975 		default:
976 			break;
977 		}
978 		break;
979 	}
980 
981 	default:
982 		return (B_FALSE);
983 	}
984 	return (B_TRUE);
985 }
986 
987 int
988 igb_m_setprop(void *arg, const char *pr_name, mac_prop_id_t pr_num,
989     uint_t pr_valsize, const void *pr_val)
990 {
991 	igb_t *igb = (igb_t *)arg;
992 	struct e1000_hw *hw = &igb->hw;
993 	int err = 0;
994 	uint32_t flow_control;
995 	uint32_t cur_mtu, new_mtu;
996 	uint32_t rx_size;
997 	uint32_t tx_size;
998 
999 	mutex_enter(&igb->gen_lock);
1000 	if (igb->igb_state & IGB_SUSPENDED) {
1001 		mutex_exit(&igb->gen_lock);
1002 		return (ECANCELED);
1003 	}
1004 
1005 	if (igb->loopback_mode != IGB_LB_NONE && igb_param_locked(pr_num)) {
1006 		/*
1007 		 * All en_* parameters are locked (read-only)
1008 		 * while the device is in any sort of loopback mode.
1009 		 */
1010 		mutex_exit(&igb->gen_lock);
1011 		return (EBUSY);
1012 	}
1013 
1014 	switch (pr_num) {
1015 	case MAC_PROP_EN_1000FDX_CAP:
1016 		/* read/write on copper, read-only on serdes */
1017 		if (hw->phy.media_type != e1000_media_type_copper) {
1018 			err = ENOTSUP;
1019 			break;
1020 		}
1021 		igb->param_en_1000fdx_cap = *(uint8_t *)pr_val;
1022 		igb->param_adv_1000fdx_cap = *(uint8_t *)pr_val;
1023 		goto setup_link;
1024 	case MAC_PROP_EN_100FDX_CAP:
1025 		if (hw->phy.media_type != e1000_media_type_copper) {
1026 			err = ENOTSUP;
1027 			break;
1028 		}
1029 		igb->param_en_100fdx_cap = *(uint8_t *)pr_val;
1030 		igb->param_adv_100fdx_cap = *(uint8_t *)pr_val;
1031 		goto setup_link;
1032 	case MAC_PROP_EN_100HDX_CAP:
1033 		if (hw->phy.media_type != e1000_media_type_copper) {
1034 			err = ENOTSUP;
1035 			break;
1036 		}
1037 		igb->param_en_100hdx_cap = *(uint8_t *)pr_val;
1038 		igb->param_adv_100hdx_cap = *(uint8_t *)pr_val;
1039 		goto setup_link;
1040 	case MAC_PROP_EN_10FDX_CAP:
1041 		if (hw->phy.media_type != e1000_media_type_copper) {
1042 			err = ENOTSUP;
1043 			break;
1044 		}
1045 		igb->param_en_10fdx_cap = *(uint8_t *)pr_val;
1046 		igb->param_adv_10fdx_cap = *(uint8_t *)pr_val;
1047 		goto setup_link;
1048 	case MAC_PROP_EN_10HDX_CAP:
1049 		if (hw->phy.media_type != e1000_media_type_copper) {
1050 			err = ENOTSUP;
1051 			break;
1052 		}
1053 		igb->param_en_10hdx_cap = *(uint8_t *)pr_val;
1054 		igb->param_adv_10hdx_cap = *(uint8_t *)pr_val;
1055 		goto setup_link;
1056 	case MAC_PROP_AUTONEG:
1057 		if (hw->phy.media_type != e1000_media_type_copper) {
1058 			err = ENOTSUP;
1059 			break;
1060 		}
1061 		igb->param_adv_autoneg_cap = *(uint8_t *)pr_val;
1062 		goto setup_link;
1063 	case MAC_PROP_FLOWCTRL:
1064 		bcopy(pr_val, &flow_control, sizeof (flow_control));
1065 
1066 		switch (flow_control) {
1067 		default:
1068 			err = EINVAL;
1069 			break;
1070 		case LINK_FLOWCTRL_NONE:
1071 			hw->fc.requested_mode = e1000_fc_none;
1072 			break;
1073 		case LINK_FLOWCTRL_RX:
1074 			hw->fc.requested_mode = e1000_fc_rx_pause;
1075 			break;
1076 		case LINK_FLOWCTRL_TX:
1077 			hw->fc.requested_mode = e1000_fc_tx_pause;
1078 			break;
1079 		case LINK_FLOWCTRL_BI:
1080 			hw->fc.requested_mode = e1000_fc_full;
1081 			break;
1082 		}
1083 setup_link:
1084 		if (err == 0) {
1085 			if (igb_setup_link(igb, B_TRUE) != IGB_SUCCESS)
1086 				err = EINVAL;
1087 		}
1088 		break;
1089 	case MAC_PROP_ADV_1000FDX_CAP:
1090 	case MAC_PROP_ADV_1000HDX_CAP:
1091 	case MAC_PROP_ADV_100T4_CAP:
1092 	case MAC_PROP_ADV_100FDX_CAP:
1093 	case MAC_PROP_ADV_100HDX_CAP:
1094 	case MAC_PROP_ADV_10FDX_CAP:
1095 	case MAC_PROP_ADV_10HDX_CAP:
1096 	case MAC_PROP_EN_1000HDX_CAP:
1097 	case MAC_PROP_EN_100T4_CAP:
1098 	case MAC_PROP_STATUS:
1099 	case MAC_PROP_SPEED:
1100 	case MAC_PROP_DUPLEX:
1101 		err = ENOTSUP; /* read-only prop. Can't set this. */
1102 		break;
1103 	case MAC_PROP_MTU:
1104 		/* adapter must be stopped for an MTU change */
1105 		if (igb->igb_state & IGB_STARTED) {
1106 			err = EBUSY;
1107 			break;
1108 		}
1109 
1110 		cur_mtu = igb->default_mtu;
1111 		bcopy(pr_val, &new_mtu, sizeof (new_mtu));
1112 		if (new_mtu == cur_mtu) {
1113 			err = 0;
1114 			break;
1115 		}
1116 
1117 		if (new_mtu < MIN_MTU || new_mtu > MAX_MTU) {
1118 			err = EINVAL;
1119 			break;
1120 		}
1121 
1122 		err = mac_maxsdu_update(igb->mac_hdl, new_mtu);
1123 		if (err == 0) {
1124 			igb->default_mtu = new_mtu;
1125 			igb->max_frame_size = igb->default_mtu +
1126 			    sizeof (struct ether_vlan_header) + ETHERFCSL;
1127 
1128 			/*
1129 			 * Set rx buffer size
1130 			 */
1131 			rx_size = igb->max_frame_size + IPHDR_ALIGN_ROOM;
1132 			igb->rx_buf_size = ((rx_size >> 10) + ((rx_size &
1133 			    (((uint32_t)1 << 10) - 1)) > 0 ? 1 : 0)) << 10;
1134 
1135 			/*
1136 			 * Set tx buffer size
1137 			 */
1138 			tx_size = igb->max_frame_size;
1139 			igb->tx_buf_size = ((tx_size >> 10) + ((tx_size &
1140 			    (((uint32_t)1 << 10) - 1)) > 0 ? 1 : 0)) << 10;
1141 		}
1142 		break;
1143 	case MAC_PROP_PRIVATE:
1144 		err = igb_set_priv_prop(igb, pr_name, pr_valsize, pr_val);
1145 		break;
1146 	default:
1147 		err = EINVAL;
1148 		break;
1149 	}
1150 
1151 	mutex_exit(&igb->gen_lock);
1152 
1153 	if (igb_check_acc_handle(igb->osdep.reg_handle) != DDI_FM_OK) {
1154 		ddi_fm_service_impact(igb->dip, DDI_SERVICE_DEGRADED);
1155 		return (EIO);
1156 	}
1157 
1158 	return (err);
1159 }
1160 
1161 int
1162 igb_m_getprop(void *arg, const char *pr_name, mac_prop_id_t pr_num,
1163     uint_t pr_valsize, void *pr_val)
1164 {
1165 	igb_t *igb = (igb_t *)arg;
1166 	struct e1000_hw *hw = &igb->hw;
1167 	int err = 0;
1168 	uint32_t flow_control;
1169 	uint64_t tmp = 0;
1170 
1171 	switch (pr_num) {
1172 	case MAC_PROP_DUPLEX:
1173 		ASSERT(pr_valsize >= sizeof (link_duplex_t));
1174 		bcopy(&igb->link_duplex, pr_val, sizeof (link_duplex_t));
1175 		break;
1176 	case MAC_PROP_SPEED:
1177 		ASSERT(pr_valsize >= sizeof (uint64_t));
1178 		tmp = igb->link_speed * 1000000ull;
1179 		bcopy(&tmp, pr_val, sizeof (tmp));
1180 		break;
1181 	case MAC_PROP_AUTONEG:
1182 		ASSERT(pr_valsize >= sizeof (uint8_t));
1183 		*(uint8_t *)pr_val = igb->param_adv_autoneg_cap;
1184 		break;
1185 	case MAC_PROP_FLOWCTRL:
1186 		ASSERT(pr_valsize >= sizeof (uint32_t));
1187 		switch (hw->fc.requested_mode) {
1188 			case e1000_fc_none:
1189 				flow_control = LINK_FLOWCTRL_NONE;
1190 				break;
1191 			case e1000_fc_rx_pause:
1192 				flow_control = LINK_FLOWCTRL_RX;
1193 				break;
1194 			case e1000_fc_tx_pause:
1195 				flow_control = LINK_FLOWCTRL_TX;
1196 				break;
1197 			case e1000_fc_full:
1198 				flow_control = LINK_FLOWCTRL_BI;
1199 				break;
1200 		}
1201 		bcopy(&flow_control, pr_val, sizeof (flow_control));
1202 		break;
1203 	case MAC_PROP_ADV_1000FDX_CAP:
1204 		*(uint8_t *)pr_val = igb->param_adv_1000fdx_cap;
1205 		break;
1206 	case MAC_PROP_EN_1000FDX_CAP:
1207 		*(uint8_t *)pr_val = igb->param_en_1000fdx_cap;
1208 		break;
1209 	case MAC_PROP_ADV_1000HDX_CAP:
1210 		*(uint8_t *)pr_val = igb->param_adv_1000hdx_cap;
1211 		break;
1212 	case MAC_PROP_EN_1000HDX_CAP:
1213 		*(uint8_t *)pr_val = igb->param_en_1000hdx_cap;
1214 		break;
1215 	case MAC_PROP_ADV_100T4_CAP:
1216 		*(uint8_t *)pr_val = igb->param_adv_100t4_cap;
1217 		break;
1218 	case MAC_PROP_EN_100T4_CAP:
1219 		*(uint8_t *)pr_val = igb->param_en_100t4_cap;
1220 		break;
1221 	case MAC_PROP_ADV_100FDX_CAP:
1222 		*(uint8_t *)pr_val = igb->param_adv_100fdx_cap;
1223 		break;
1224 	case MAC_PROP_EN_100FDX_CAP:
1225 		*(uint8_t *)pr_val = igb->param_en_100fdx_cap;
1226 		break;
1227 	case MAC_PROP_ADV_100HDX_CAP:
1228 		*(uint8_t *)pr_val = igb->param_adv_100hdx_cap;
1229 		break;
1230 	case MAC_PROP_EN_100HDX_CAP:
1231 		*(uint8_t *)pr_val = igb->param_en_100hdx_cap;
1232 		break;
1233 	case MAC_PROP_ADV_10FDX_CAP:
1234 		*(uint8_t *)pr_val = igb->param_adv_10fdx_cap;
1235 		break;
1236 	case MAC_PROP_EN_10FDX_CAP:
1237 		*(uint8_t *)pr_val = igb->param_en_10fdx_cap;
1238 		break;
1239 	case MAC_PROP_ADV_10HDX_CAP:
1240 		*(uint8_t *)pr_val = igb->param_adv_10hdx_cap;
1241 		break;
1242 	case MAC_PROP_EN_10HDX_CAP:
1243 		*(uint8_t *)pr_val = igb->param_en_10hdx_cap;
1244 		break;
1245 	case MAC_PROP_PRIVATE:
1246 		err = igb_get_priv_prop(igb, pr_name, pr_valsize, pr_val);
1247 		break;
1248 	default:
1249 		err = EINVAL;
1250 		break;
1251 	}
1252 	return (err);
1253 }
1254 
1255 void
1256 igb_m_propinfo(void *arg, const char *pr_name, mac_prop_id_t pr_num,
1257     mac_prop_info_handle_t prh)
1258 {
1259 	igb_t *igb = (igb_t *)arg;
1260 	struct e1000_hw *hw = &igb->hw;
1261 	uint16_t phy_status, phy_ext_status;
1262 
1263 	switch (pr_num) {
1264 	case MAC_PROP_DUPLEX:
1265 	case MAC_PROP_SPEED:
1266 	case MAC_PROP_ADV_1000FDX_CAP:
1267 	case MAC_PROP_ADV_1000HDX_CAP:
1268 	case MAC_PROP_EN_1000HDX_CAP:
1269 	case MAC_PROP_ADV_100T4_CAP:
1270 	case MAC_PROP_EN_100T4_CAP:
1271 		mac_prop_info_set_perm(prh, MAC_PROP_PERM_READ);
1272 		break;
1273 
1274 	case MAC_PROP_EN_1000FDX_CAP:
1275 		if (hw->phy.media_type != e1000_media_type_copper) {
1276 			mac_prop_info_set_perm(prh, MAC_PROP_PERM_READ);
1277 		} else {
1278 			(void) e1000_read_phy_reg(hw, PHY_EXT_STATUS,
1279 			    &phy_ext_status);
1280 			mac_prop_info_set_default_uint8(prh,
1281 			    ((phy_ext_status & IEEE_ESR_1000T_FD_CAPS) ||
1282 			    (phy_ext_status & IEEE_ESR_1000X_FD_CAPS)) ? 1 : 0);
1283 		}
1284 		break;
1285 
1286 	case MAC_PROP_ADV_100FDX_CAP:
1287 	case MAC_PROP_EN_100FDX_CAP:
1288 		if (hw->phy.media_type != e1000_media_type_copper) {
1289 			mac_prop_info_set_perm(prh, MAC_PROP_PERM_READ);
1290 		} else {
1291 			(void) e1000_read_phy_reg(hw, PHY_STATUS, &phy_status);
1292 			mac_prop_info_set_default_uint8(prh,
1293 			    ((phy_status & MII_SR_100X_FD_CAPS) ||
1294 			    (phy_status & MII_SR_100T2_FD_CAPS)) ? 1 : 0);
1295 		}
1296 		break;
1297 
1298 	case MAC_PROP_ADV_100HDX_CAP:
1299 	case MAC_PROP_EN_100HDX_CAP:
1300 		if (hw->phy.media_type != e1000_media_type_copper) {
1301 			mac_prop_info_set_perm(prh, MAC_PROP_PERM_READ);
1302 		} else {
1303 			(void) e1000_read_phy_reg(hw, PHY_STATUS, &phy_status);
1304 			mac_prop_info_set_default_uint8(prh,
1305 			    ((phy_status & MII_SR_100X_HD_CAPS) ||
1306 			    (phy_status & MII_SR_100T2_HD_CAPS)) ? 1 : 0);
1307 		}
1308 		break;
1309 
1310 	case MAC_PROP_ADV_10FDX_CAP:
1311 	case MAC_PROP_EN_10FDX_CAP:
1312 		if (hw->phy.media_type != e1000_media_type_copper) {
1313 			mac_prop_info_set_perm(prh, MAC_PROP_PERM_READ);
1314 		} else {
1315 			(void) e1000_read_phy_reg(hw, PHY_STATUS, &phy_status);
1316 			mac_prop_info_set_default_uint8(prh,
1317 			    (phy_status & MII_SR_10T_FD_CAPS) ? 1 : 0);
1318 		}
1319 		break;
1320 
1321 	case MAC_PROP_ADV_10HDX_CAP:
1322 	case MAC_PROP_EN_10HDX_CAP:
1323 		if (hw->phy.media_type != e1000_media_type_copper) {
1324 			mac_prop_info_set_perm(prh, MAC_PROP_PERM_READ);
1325 		} else {
1326 			(void) e1000_read_phy_reg(hw, PHY_STATUS, &phy_status);
1327 			mac_prop_info_set_default_uint8(prh,
1328 			    (phy_status & MII_SR_10T_HD_CAPS) ? 1 : 0);
1329 		}
1330 		break;
1331 
1332 	case MAC_PROP_AUTONEG:
1333 		if (hw->phy.media_type != e1000_media_type_copper) {
1334 			mac_prop_info_set_perm(prh, MAC_PROP_PERM_READ);
1335 		} else {
1336 			(void) e1000_read_phy_reg(hw, PHY_STATUS, &phy_status);
1337 			mac_prop_info_set_default_uint8(prh,
1338 			    (phy_status & MII_SR_AUTONEG_CAPS) ? 1 : 0);
1339 		}
1340 		break;
1341 
1342 	case MAC_PROP_FLOWCTRL:
1343 		mac_prop_info_set_default_link_flowctrl(prh, LINK_FLOWCTRL_BI);
1344 		break;
1345 
1346 	case MAC_PROP_MTU:
1347 		mac_prop_info_set_range_uint32(prh, MIN_MTU, MAX_MTU);
1348 		break;
1349 
1350 	case MAC_PROP_PRIVATE:
1351 		igb_priv_prop_info(igb, pr_name, prh);
1352 		break;
1353 	}
1354 
1355 }
1356 
1357 boolean_t
1358 igb_param_locked(mac_prop_id_t pr_num)
1359 {
1360 	/*
1361 	 * All en_* parameters are locked (read-only) while
1362 	 * the device is in any sort of loopback mode ...
1363 	 */
1364 	switch (pr_num) {
1365 		case MAC_PROP_EN_1000FDX_CAP:
1366 		case MAC_PROP_EN_1000HDX_CAP:
1367 		case MAC_PROP_EN_100T4_CAP:
1368 		case MAC_PROP_EN_100FDX_CAP:
1369 		case MAC_PROP_EN_100HDX_CAP:
1370 		case MAC_PROP_EN_10FDX_CAP:
1371 		case MAC_PROP_EN_10HDX_CAP:
1372 		case MAC_PROP_AUTONEG:
1373 		case MAC_PROP_FLOWCTRL:
1374 			return (B_TRUE);
1375 	}
1376 	return (B_FALSE);
1377 }
1378 
1379 /* ARGSUSED */
1380 int
1381 igb_set_priv_prop(igb_t *igb, const char *pr_name,
1382     uint_t pr_valsize, const void *pr_val)
1383 {
1384 	int err = 0;
1385 	long result;
1386 	struct e1000_hw *hw = &igb->hw;
1387 	int i;
1388 
1389 	if (strcmp(pr_name, "_eee_support") == 0) {
1390 		if (pr_val == NULL)
1391 			return (EINVAL);
1392 		(void) ddi_strtol(pr_val, (char **)NULL, 0, &result);
1393 		switch (result) {
1394 		case 0:
1395 		case 1:
1396 			if (hw->mac.type != e1000_i350) {
1397 				/*
1398 				 * For now, only supported on I350.
1399 				 * Add new mac.type values (or use < instead)
1400 				 * as new cards offer up EEE.
1401 				 */
1402 				return (ENXIO);
1403 			}
1404 			/* Must set this prior to the set call. */
1405 			hw->dev_spec._82575.eee_disable = !result;
1406 			if (e1000_set_eee_i350(hw) != E1000_SUCCESS)
1407 				err = EIO;
1408 			break;
1409 		default:
1410 			err = EINVAL;
1411 			/* FALLTHRU */
1412 		}
1413 		return (err);
1414 	}
1415 	if (strcmp(pr_name, "_tx_copy_thresh") == 0) {
1416 		if (pr_val == NULL) {
1417 			err = EINVAL;
1418 			return (err);
1419 		}
1420 		(void) ddi_strtol(pr_val, (char **)NULL, 0, &result);
1421 		if (result < MIN_TX_COPY_THRESHOLD ||
1422 		    result > MAX_TX_COPY_THRESHOLD)
1423 			err = EINVAL;
1424 		else {
1425 			igb->tx_copy_thresh = (uint32_t)result;
1426 		}
1427 		return (err);
1428 	}
1429 	if (strcmp(pr_name, "_tx_recycle_thresh") == 0) {
1430 		if (pr_val == NULL) {
1431 			err = EINVAL;
1432 			return (err);
1433 		}
1434 		(void) ddi_strtol(pr_val, (char **)NULL, 0, &result);
1435 		if (result < MIN_TX_RECYCLE_THRESHOLD ||
1436 		    result > MAX_TX_RECYCLE_THRESHOLD)
1437 			err = EINVAL;
1438 		else {
1439 			igb->tx_recycle_thresh = (uint32_t)result;
1440 		}
1441 		return (err);
1442 	}
1443 	if (strcmp(pr_name, "_tx_overload_thresh") == 0) {
1444 		if (pr_val == NULL) {
1445 			err = EINVAL;
1446 			return (err);
1447 		}
1448 		(void) ddi_strtol(pr_val, (char **)NULL, 0, &result);
1449 		if (result < MIN_TX_OVERLOAD_THRESHOLD ||
1450 		    result > MAX_TX_OVERLOAD_THRESHOLD)
1451 			err = EINVAL;
1452 		else {
1453 			igb->tx_overload_thresh = (uint32_t)result;
1454 		}
1455 		return (err);
1456 	}
1457 	if (strcmp(pr_name, "_tx_resched_thresh") == 0) {
1458 		if (pr_val == NULL) {
1459 			err = EINVAL;
1460 			return (err);
1461 		}
1462 		(void) ddi_strtol(pr_val, (char **)NULL, 0, &result);
1463 		if (result < MIN_TX_RESCHED_THRESHOLD ||
1464 		    result > MAX_TX_RESCHED_THRESHOLD ||
1465 		    result > igb->tx_ring_size)
1466 			err = EINVAL;
1467 		else {
1468 			igb->tx_resched_thresh = (uint32_t)result;
1469 		}
1470 		return (err);
1471 	}
1472 	if (strcmp(pr_name, "_rx_copy_thresh") == 0) {
1473 		if (pr_val == NULL) {
1474 			err = EINVAL;
1475 			return (err);
1476 		}
1477 		(void) ddi_strtol(pr_val, (char **)NULL, 0, &result);
1478 		if (result < MIN_RX_COPY_THRESHOLD ||
1479 		    result > MAX_RX_COPY_THRESHOLD)
1480 			err = EINVAL;
1481 		else {
1482 			igb->rx_copy_thresh = (uint32_t)result;
1483 		}
1484 		return (err);
1485 	}
1486 	if (strcmp(pr_name, "_rx_limit_per_intr") == 0) {
1487 		if (pr_val == NULL) {
1488 			err = EINVAL;
1489 			return (err);
1490 		}
1491 		(void) ddi_strtol(pr_val, (char **)NULL, 0, &result);
1492 		if (result < MIN_RX_LIMIT_PER_INTR ||
1493 		    result > MAX_RX_LIMIT_PER_INTR)
1494 			err = EINVAL;
1495 		else {
1496 			igb->rx_limit_per_intr = (uint32_t)result;
1497 		}
1498 		return (err);
1499 	}
1500 	if (strcmp(pr_name, "_intr_throttling") == 0) {
1501 		if (pr_val == NULL) {
1502 			err = EINVAL;
1503 			return (err);
1504 		}
1505 		(void) ddi_strtol(pr_val, (char **)NULL, 0, &result);
1506 
1507 		if (result < igb->capab->min_intr_throttle ||
1508 		    result > igb->capab->max_intr_throttle)
1509 			err = EINVAL;
1510 		else {
1511 			igb->intr_throttling[0] = (uint32_t)result;
1512 
1513 			for (i = 0; i < MAX_NUM_EITR; i++)
1514 				igb->intr_throttling[i] =
1515 				    igb->intr_throttling[0];
1516 
1517 			/* Set interrupt throttling rate */
1518 			for (i = 0; i < igb->intr_cnt; i++)
1519 				E1000_WRITE_REG(hw, E1000_EITR(i),
1520 				    igb->intr_throttling[i]);
1521 		}
1522 		return (err);
1523 	}
1524 	return (ENOTSUP);
1525 }
1526 
1527 int
1528 igb_get_priv_prop(igb_t *igb, const char *pr_name, uint_t pr_valsize,
1529     void *pr_val)
1530 {
1531 	int value;
1532 
1533 	if (strcmp(pr_name, "_adv_pause_cap") == 0) {
1534 		value = igb->param_adv_pause_cap;
1535 	} else if (strcmp(pr_name, "_adv_asym_pause_cap") == 0) {
1536 		value = igb->param_adv_asym_pause_cap;
1537 	} else if (strcmp(pr_name, "_eee_support") == 0) {
1538 		/*
1539 		 * For now, only supported on I350.  Add new mac.type values
1540 		 * (or use < instead) as new cards offer up EEE.
1541 		 */
1542 		value = (igb->hw.mac.type != e1000_i350) ? 0 :
1543 		    !(igb->hw.dev_spec._82575.eee_disable);
1544 	} else if (strcmp(pr_name, "_tx_copy_thresh") == 0) {
1545 		value = igb->tx_copy_thresh;
1546 	} else if (strcmp(pr_name, "_tx_recycle_thresh") == 0) {
1547 		value = igb->tx_recycle_thresh;
1548 	} else if (strcmp(pr_name, "_tx_overload_thresh") == 0) {
1549 		value = igb->tx_overload_thresh;
1550 	} else if (strcmp(pr_name, "_tx_resched_thresh") == 0) {
1551 		value = igb->tx_resched_thresh;
1552 	} else if (strcmp(pr_name, "_rx_copy_thresh") == 0) {
1553 		value = igb->rx_copy_thresh;
1554 	} else if (strcmp(pr_name, "_rx_limit_per_intr") == 0) {
1555 		value = igb->rx_limit_per_intr;
1556 	} else if (strcmp(pr_name, "_intr_throttling") == 0) {
1557 		value = igb->intr_throttling[0];
1558 	} else {
1559 		return (ENOTSUP);
1560 	}
1561 
1562 	(void) snprintf(pr_val, pr_valsize, "%d", value);
1563 	return (0);
1564 }
1565 
1566 void
1567 igb_priv_prop_info(igb_t *igb, const char *pr_name, mac_prop_info_handle_t prh)
1568 {
1569 	char valstr[64];
1570 	int value;
1571 
1572 	if (strcmp(pr_name, "_adv_pause_cap") == 0 ||
1573 	    strcmp(pr_name, "_adv_asym_pause_cap") == 0) {
1574 		mac_prop_info_set_perm(prh, MAC_PROP_PERM_READ);
1575 		return;
1576 	} else if (strcmp(pr_name, "_tx_copy_thresh") == 0) {
1577 		value = DEFAULT_TX_COPY_THRESHOLD;
1578 	} else if (strcmp(pr_name, "_tx_recycle_thresh") == 0) {
1579 		value = DEFAULT_TX_RECYCLE_THRESHOLD;
1580 	} else if (strcmp(pr_name, "_tx_overload_thresh") == 0) {
1581 		value = DEFAULT_TX_OVERLOAD_THRESHOLD;
1582 	} else if (strcmp(pr_name, "_tx_resched_thresh") == 0) {
1583 		value = DEFAULT_TX_RESCHED_THRESHOLD;
1584 	} else if (strcmp(pr_name, "_rx_copy_thresh") == 0) {
1585 		value = DEFAULT_RX_COPY_THRESHOLD;
1586 	} else if (strcmp(pr_name, "_rx_limit_per_intr") == 0) {
1587 		value = DEFAULT_RX_LIMIT_PER_INTR;
1588 	} else 	if (strcmp(pr_name, "_intr_throttling") == 0) {
1589 		value = igb->capab->def_intr_throttle;
1590 	} else {
1591 		return;
1592 	}
1593 
1594 	(void) snprintf(valstr, sizeof (valstr), "%d", value);
1595 	mac_prop_info_set_default_str(prh, valstr);
1596 }
1597