xref: /freebsd/sys/dev/ixgbe/ixgbe_82598.c (revision c9c67103c371140fa440f0be08eaa8f308b23239)
1 /******************************************************************************
2 
3   Copyright (c) 2001-2013, Intel Corporation
4   All rights reserved.
5 
6   Redistribution and use in source and binary forms, with or without
7   modification, are permitted provided that the following conditions are met:
8 
9    1. Redistributions of source code must retain the above copyright notice,
10       this list of conditions and the following disclaimer.
11 
12    2. Redistributions in binary form must reproduce the above copyright
13       notice, this list of conditions and the following disclaimer in the
14       documentation and/or other materials provided with the distribution.
15 
16    3. Neither the name of the Intel Corporation nor the names of its
17       contributors may be used to endorse or promote products derived from
18       this software without specific prior written permission.
19 
20   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
21   AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22   IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23   ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
24   LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25   CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26   SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27   INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28   CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29   ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
30   POSSIBILITY OF SUCH DAMAGE.
31 
32 ******************************************************************************/
33 /*$FreeBSD$*/
34 
35 #include "ixgbe_type.h"
36 #include "ixgbe_82598.h"
37 #include "ixgbe_api.h"
38 #include "ixgbe_common.h"
39 #include "ixgbe_phy.h"
40 
41 static s32 ixgbe_get_link_capabilities_82598(struct ixgbe_hw *hw,
42 					     ixgbe_link_speed *speed,
43 					     bool *autoneg);
44 static enum ixgbe_media_type ixgbe_get_media_type_82598(struct ixgbe_hw *hw);
45 static s32 ixgbe_start_mac_link_82598(struct ixgbe_hw *hw,
46 				      bool autoneg_wait_to_complete);
47 static s32 ixgbe_check_mac_link_82598(struct ixgbe_hw *hw,
48 				      ixgbe_link_speed *speed, bool *link_up,
49 				      bool link_up_wait_to_complete);
50 static s32 ixgbe_setup_mac_link_82598(struct ixgbe_hw *hw,
51 				      ixgbe_link_speed speed,
52 				      bool autoneg_wait_to_complete);
53 static s32 ixgbe_setup_copper_link_82598(struct ixgbe_hw *hw,
54 					 ixgbe_link_speed speed,
55 					 bool autoneg_wait_to_complete);
56 static s32 ixgbe_reset_hw_82598(struct ixgbe_hw *hw);
57 static s32 ixgbe_clear_vmdq_82598(struct ixgbe_hw *hw, u32 rar, u32 vmdq);
58 static s32 ixgbe_clear_vfta_82598(struct ixgbe_hw *hw);
59 static void ixgbe_set_rxpba_82598(struct ixgbe_hw *hw, int num_pb,
60 				  u32 headroom, int strategy);
61 static s32 ixgbe_read_i2c_sff8472_82598(struct ixgbe_hw *hw, u8 byte_offset,
62 					u8 *sff8472_data);
63 /**
64  *  ixgbe_set_pcie_completion_timeout - set pci-e completion timeout
65  *  @hw: pointer to the HW structure
66  *
67  *  The defaults for 82598 should be in the range of 50us to 50ms,
68  *  however the hardware default for these parts is 500us to 1ms which is less
69  *  than the 10ms recommended by the pci-e spec.  To address this we need to
70  *  increase the value to either 10ms to 250ms for capability version 1 config,
71  *  or 16ms to 55ms for version 2.
72  **/
73 void ixgbe_set_pcie_completion_timeout(struct ixgbe_hw *hw)
74 {
75 	u32 gcr = IXGBE_READ_REG(hw, IXGBE_GCR);
76 	u16 pcie_devctl2;
77 
78 	/* only take action if timeout value is defaulted to 0 */
79 	if (gcr & IXGBE_GCR_CMPL_TMOUT_MASK)
80 		goto out;
81 
82 	/*
83 	 * if capababilities version is type 1 we can write the
84 	 * timeout of 10ms to 250ms through the GCR register
85 	 */
86 	if (!(gcr & IXGBE_GCR_CAP_VER2)) {
87 		gcr |= IXGBE_GCR_CMPL_TMOUT_10ms;
88 		goto out;
89 	}
90 
91 	/*
92 	 * for version 2 capabilities we need to write the config space
93 	 * directly in order to set the completion timeout value for
94 	 * 16ms to 55ms
95 	 */
96 	pcie_devctl2 = IXGBE_READ_PCIE_WORD(hw, IXGBE_PCI_DEVICE_CONTROL2);
97 	pcie_devctl2 |= IXGBE_PCI_DEVICE_CONTROL2_16ms;
98 	IXGBE_WRITE_PCIE_WORD(hw, IXGBE_PCI_DEVICE_CONTROL2, pcie_devctl2);
99 out:
100 	/* disable completion timeout resend */
101 	gcr &= ~IXGBE_GCR_CMPL_TMOUT_RESEND;
102 	IXGBE_WRITE_REG(hw, IXGBE_GCR, gcr);
103 }
104 
105 /**
106  *  ixgbe_init_ops_82598 - Inits func ptrs and MAC type
107  *  @hw: pointer to hardware structure
108  *
109  *  Initialize the function pointers and assign the MAC type for 82598.
110  *  Does not touch the hardware.
111  **/
112 s32 ixgbe_init_ops_82598(struct ixgbe_hw *hw)
113 {
114 	struct ixgbe_mac_info *mac = &hw->mac;
115 	struct ixgbe_phy_info *phy = &hw->phy;
116 	s32 ret_val;
117 
118 	DEBUGFUNC("ixgbe_init_ops_82598");
119 
120 	ret_val = ixgbe_init_phy_ops_generic(hw);
121 	ret_val = ixgbe_init_ops_generic(hw);
122 
123 	/* PHY */
124 	phy->ops.init = &ixgbe_init_phy_ops_82598;
125 
126 	/* MAC */
127 	mac->ops.start_hw = &ixgbe_start_hw_82598;
128 	mac->ops.enable_relaxed_ordering = &ixgbe_enable_relaxed_ordering_82598;
129 	mac->ops.reset_hw = &ixgbe_reset_hw_82598;
130 	mac->ops.get_media_type = &ixgbe_get_media_type_82598;
131 	mac->ops.get_supported_physical_layer =
132 				&ixgbe_get_supported_physical_layer_82598;
133 	mac->ops.read_analog_reg8 = &ixgbe_read_analog_reg8_82598;
134 	mac->ops.write_analog_reg8 = &ixgbe_write_analog_reg8_82598;
135 	mac->ops.set_lan_id = &ixgbe_set_lan_id_multi_port_pcie_82598;
136 
137 	/* RAR, Multicast, VLAN */
138 	mac->ops.set_vmdq = &ixgbe_set_vmdq_82598;
139 	mac->ops.clear_vmdq = &ixgbe_clear_vmdq_82598;
140 	mac->ops.set_vfta = &ixgbe_set_vfta_82598;
141 	mac->ops.set_vlvf = NULL;
142 	mac->ops.clear_vfta = &ixgbe_clear_vfta_82598;
143 
144 	/* Flow Control */
145 	mac->ops.fc_enable = &ixgbe_fc_enable_82598;
146 
147 	mac->mcft_size		= 128;
148 	mac->vft_size		= 128;
149 	mac->num_rar_entries	= 16;
150 	mac->rx_pb_size		= 512;
151 	mac->max_tx_queues	= 32;
152 	mac->max_rx_queues	= 64;
153 	mac->max_msix_vectors	= ixgbe_get_pcie_msix_count_generic(hw);
154 
155 	/* SFP+ Module */
156 	phy->ops.read_i2c_eeprom = &ixgbe_read_i2c_eeprom_82598;
157 	phy->ops.read_i2c_sff8472 = &ixgbe_read_i2c_sff8472_82598;
158 
159 	/* Link */
160 	mac->ops.check_link = &ixgbe_check_mac_link_82598;
161 	mac->ops.setup_link = &ixgbe_setup_mac_link_82598;
162 	mac->ops.flap_tx_laser = NULL;
163 	mac->ops.get_link_capabilities = &ixgbe_get_link_capabilities_82598;
164 	mac->ops.setup_rxpba = &ixgbe_set_rxpba_82598;
165 
166 	/* Manageability interface */
167 	mac->ops.set_fw_drv_ver = NULL;
168 
169 	mac->ops.get_rtrup2tc = NULL;
170 
171 	return ret_val;
172 }
173 
174 /**
175  *  ixgbe_init_phy_ops_82598 - PHY/SFP specific init
176  *  @hw: pointer to hardware structure
177  *
178  *  Initialize any function pointers that were not able to be
179  *  set during init_shared_code because the PHY/SFP type was
180  *  not known.  Perform the SFP init if necessary.
181  *
182  **/
183 s32 ixgbe_init_phy_ops_82598(struct ixgbe_hw *hw)
184 {
185 	struct ixgbe_mac_info *mac = &hw->mac;
186 	struct ixgbe_phy_info *phy = &hw->phy;
187 	s32 ret_val = IXGBE_SUCCESS;
188 	u16 list_offset, data_offset;
189 
190 	DEBUGFUNC("ixgbe_init_phy_ops_82598");
191 
192 	/* Identify the PHY */
193 	phy->ops.identify(hw);
194 
195 	/* Overwrite the link function pointers if copper PHY */
196 	if (mac->ops.get_media_type(hw) == ixgbe_media_type_copper) {
197 		mac->ops.setup_link = &ixgbe_setup_copper_link_82598;
198 		mac->ops.get_link_capabilities =
199 				&ixgbe_get_copper_link_capabilities_generic;
200 	}
201 
202 	switch (hw->phy.type) {
203 	case ixgbe_phy_tn:
204 		phy->ops.setup_link = &ixgbe_setup_phy_link_tnx;
205 		phy->ops.check_link = &ixgbe_check_phy_link_tnx;
206 		phy->ops.get_firmware_version =
207 					&ixgbe_get_phy_firmware_version_tnx;
208 		break;
209 	case ixgbe_phy_nl:
210 		phy->ops.reset = &ixgbe_reset_phy_nl;
211 
212 		/* Call SFP+ identify routine to get the SFP+ module type */
213 		ret_val = phy->ops.identify_sfp(hw);
214 		if (ret_val != IXGBE_SUCCESS)
215 			goto out;
216 		else if (hw->phy.sfp_type == ixgbe_sfp_type_unknown) {
217 			ret_val = IXGBE_ERR_SFP_NOT_SUPPORTED;
218 			goto out;
219 		}
220 
221 		/* Check to see if SFP+ module is supported */
222 		ret_val = ixgbe_get_sfp_init_sequence_offsets(hw,
223 							      &list_offset,
224 							      &data_offset);
225 		if (ret_val != IXGBE_SUCCESS) {
226 			ret_val = IXGBE_ERR_SFP_NOT_SUPPORTED;
227 			goto out;
228 		}
229 		break;
230 	default:
231 		break;
232 	}
233 
234 out:
235 	return ret_val;
236 }
237 
238 /**
239  *  ixgbe_start_hw_82598 - Prepare hardware for Tx/Rx
240  *  @hw: pointer to hardware structure
241  *
242  *  Starts the hardware using the generic start_hw function.
243  *  Disables relaxed ordering Then set pcie completion timeout
244  *
245  **/
246 s32 ixgbe_start_hw_82598(struct ixgbe_hw *hw)
247 {
248 	u32 regval;
249 	u32 i;
250 	s32 ret_val = IXGBE_SUCCESS;
251 
252 	DEBUGFUNC("ixgbe_start_hw_82598");
253 
254 	ret_val = ixgbe_start_hw_generic(hw);
255 
256 	/* Disable relaxed ordering */
257 	for (i = 0; ((i < hw->mac.max_tx_queues) &&
258 	     (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
259 		regval = IXGBE_READ_REG(hw, IXGBE_DCA_TXCTRL(i));
260 		regval &= ~IXGBE_DCA_TXCTRL_DESC_WRO_EN;
261 		IXGBE_WRITE_REG(hw, IXGBE_DCA_TXCTRL(i), regval);
262 	}
263 
264 	for (i = 0; ((i < hw->mac.max_rx_queues) &&
265 	     (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
266 		regval = IXGBE_READ_REG(hw, IXGBE_DCA_RXCTRL(i));
267 		regval &= ~(IXGBE_DCA_RXCTRL_DATA_WRO_EN |
268 			    IXGBE_DCA_RXCTRL_HEAD_WRO_EN);
269 		IXGBE_WRITE_REG(hw, IXGBE_DCA_RXCTRL(i), regval);
270 	}
271 
272 	/* set the completion timeout for interface */
273 	if (ret_val == IXGBE_SUCCESS)
274 		ixgbe_set_pcie_completion_timeout(hw);
275 
276 	return ret_val;
277 }
278 
279 /**
280  *  ixgbe_get_link_capabilities_82598 - Determines link capabilities
281  *  @hw: pointer to hardware structure
282  *  @speed: pointer to link speed
283  *  @autoneg: boolean auto-negotiation value
284  *
285  *  Determines the link capabilities by reading the AUTOC register.
286  **/
287 static s32 ixgbe_get_link_capabilities_82598(struct ixgbe_hw *hw,
288 					     ixgbe_link_speed *speed,
289 					     bool *autoneg)
290 {
291 	s32 status = IXGBE_SUCCESS;
292 	u32 autoc = 0;
293 
294 	DEBUGFUNC("ixgbe_get_link_capabilities_82598");
295 
296 	/*
297 	 * Determine link capabilities based on the stored value of AUTOC,
298 	 * which represents EEPROM defaults.  If AUTOC value has not been
299 	 * stored, use the current register value.
300 	 */
301 	if (hw->mac.orig_link_settings_stored)
302 		autoc = hw->mac.orig_autoc;
303 	else
304 		autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
305 
306 	switch (autoc & IXGBE_AUTOC_LMS_MASK) {
307 	case IXGBE_AUTOC_LMS_1G_LINK_NO_AN:
308 		*speed = IXGBE_LINK_SPEED_1GB_FULL;
309 		*autoneg = FALSE;
310 		break;
311 
312 	case IXGBE_AUTOC_LMS_10G_LINK_NO_AN:
313 		*speed = IXGBE_LINK_SPEED_10GB_FULL;
314 		*autoneg = FALSE;
315 		break;
316 
317 	case IXGBE_AUTOC_LMS_1G_AN:
318 		*speed = IXGBE_LINK_SPEED_1GB_FULL;
319 		*autoneg = TRUE;
320 		break;
321 
322 	case IXGBE_AUTOC_LMS_KX4_AN:
323 	case IXGBE_AUTOC_LMS_KX4_AN_1G_AN:
324 		*speed = IXGBE_LINK_SPEED_UNKNOWN;
325 		if (autoc & IXGBE_AUTOC_KX4_SUPP)
326 			*speed |= IXGBE_LINK_SPEED_10GB_FULL;
327 		if (autoc & IXGBE_AUTOC_KX_SUPP)
328 			*speed |= IXGBE_LINK_SPEED_1GB_FULL;
329 		*autoneg = TRUE;
330 		break;
331 
332 	default:
333 		status = IXGBE_ERR_LINK_SETUP;
334 		break;
335 	}
336 
337 	return status;
338 }
339 
340 /**
341  *  ixgbe_get_media_type_82598 - Determines media type
342  *  @hw: pointer to hardware structure
343  *
344  *  Returns the media type (fiber, copper, backplane)
345  **/
346 static enum ixgbe_media_type ixgbe_get_media_type_82598(struct ixgbe_hw *hw)
347 {
348 	enum ixgbe_media_type media_type;
349 
350 	DEBUGFUNC("ixgbe_get_media_type_82598");
351 
352 	/* Detect if there is a copper PHY attached. */
353 	switch (hw->phy.type) {
354 	case ixgbe_phy_cu_unknown:
355 	case ixgbe_phy_tn:
356 		media_type = ixgbe_media_type_copper;
357 		goto out;
358 	default:
359 		break;
360 	}
361 
362 	/* Media type for I82598 is based on device ID */
363 	switch (hw->device_id) {
364 	case IXGBE_DEV_ID_82598:
365 	case IXGBE_DEV_ID_82598_BX:
366 		/* Default device ID is mezzanine card KX/KX4 */
367 		media_type = ixgbe_media_type_backplane;
368 		break;
369 	case IXGBE_DEV_ID_82598AF_DUAL_PORT:
370 	case IXGBE_DEV_ID_82598AF_SINGLE_PORT:
371 	case IXGBE_DEV_ID_82598_DA_DUAL_PORT:
372 	case IXGBE_DEV_ID_82598_SR_DUAL_PORT_EM:
373 	case IXGBE_DEV_ID_82598EB_XF_LR:
374 	case IXGBE_DEV_ID_82598EB_SFP_LOM:
375 		media_type = ixgbe_media_type_fiber;
376 		break;
377 	case IXGBE_DEV_ID_82598EB_CX4:
378 	case IXGBE_DEV_ID_82598_CX4_DUAL_PORT:
379 		media_type = ixgbe_media_type_cx4;
380 		break;
381 	case IXGBE_DEV_ID_82598AT:
382 	case IXGBE_DEV_ID_82598AT2:
383 		media_type = ixgbe_media_type_copper;
384 		break;
385 	default:
386 		media_type = ixgbe_media_type_unknown;
387 		break;
388 	}
389 out:
390 	return media_type;
391 }
392 
393 /**
394  *  ixgbe_fc_enable_82598 - Enable flow control
395  *  @hw: pointer to hardware structure
396  *
397  *  Enable flow control according to the current settings.
398  **/
399 s32 ixgbe_fc_enable_82598(struct ixgbe_hw *hw)
400 {
401 	s32 ret_val = IXGBE_SUCCESS;
402 	u32 fctrl_reg;
403 	u32 rmcs_reg;
404 	u32 reg;
405 	u32 fcrtl, fcrth;
406 	u32 link_speed = 0;
407 	int i;
408 	bool link_up;
409 
410 	DEBUGFUNC("ixgbe_fc_enable_82598");
411 
412 	/* Validate the water mark configuration */
413 	if (!hw->fc.pause_time) {
414 		ret_val = IXGBE_ERR_INVALID_LINK_SETTINGS;
415 		goto out;
416 	}
417 
418 	/* Low water mark of zero causes XOFF floods */
419 	for (i = 0; i < IXGBE_DCB_MAX_TRAFFIC_CLASS; i++) {
420 		if ((hw->fc.current_mode & ixgbe_fc_tx_pause) &&
421 		    hw->fc.high_water[i]) {
422 			if (!hw->fc.low_water[i] ||
423 			    hw->fc.low_water[i] >= hw->fc.high_water[i]) {
424 				DEBUGOUT("Invalid water mark configuration\n");
425 				ret_val = IXGBE_ERR_INVALID_LINK_SETTINGS;
426 				goto out;
427 			}
428 		}
429 	}
430 
431 	/*
432 	 * On 82598 having Rx FC on causes resets while doing 1G
433 	 * so if it's on turn it off once we know link_speed. For
434 	 * more details see 82598 Specification update.
435 	 */
436 	hw->mac.ops.check_link(hw, &link_speed, &link_up, FALSE);
437 	if (link_up && link_speed == IXGBE_LINK_SPEED_1GB_FULL) {
438 		switch (hw->fc.requested_mode) {
439 		case ixgbe_fc_full:
440 			hw->fc.requested_mode = ixgbe_fc_tx_pause;
441 			break;
442 		case ixgbe_fc_rx_pause:
443 			hw->fc.requested_mode = ixgbe_fc_none;
444 			break;
445 		default:
446 			/* no change */
447 			break;
448 		}
449 	}
450 
451 	/* Negotiate the fc mode to use */
452 	ixgbe_fc_autoneg(hw);
453 
454 	/* Disable any previous flow control settings */
455 	fctrl_reg = IXGBE_READ_REG(hw, IXGBE_FCTRL);
456 	fctrl_reg &= ~(IXGBE_FCTRL_RFCE | IXGBE_FCTRL_RPFCE);
457 
458 	rmcs_reg = IXGBE_READ_REG(hw, IXGBE_RMCS);
459 	rmcs_reg &= ~(IXGBE_RMCS_TFCE_PRIORITY | IXGBE_RMCS_TFCE_802_3X);
460 
461 	/*
462 	 * The possible values of fc.current_mode are:
463 	 * 0: Flow control is completely disabled
464 	 * 1: Rx flow control is enabled (we can receive pause frames,
465 	 *    but not send pause frames).
466 	 * 2: Tx flow control is enabled (we can send pause frames but
467 	 *     we do not support receiving pause frames).
468 	 * 3: Both Rx and Tx flow control (symmetric) are enabled.
469 	 * other: Invalid.
470 	 */
471 	switch (hw->fc.current_mode) {
472 	case ixgbe_fc_none:
473 		/*
474 		 * Flow control is disabled by software override or autoneg.
475 		 * The code below will actually disable it in the HW.
476 		 */
477 		break;
478 	case ixgbe_fc_rx_pause:
479 		/*
480 		 * Rx Flow control is enabled and Tx Flow control is
481 		 * disabled by software override. Since there really
482 		 * isn't a way to advertise that we are capable of RX
483 		 * Pause ONLY, we will advertise that we support both
484 		 * symmetric and asymmetric Rx PAUSE.  Later, we will
485 		 * disable the adapter's ability to send PAUSE frames.
486 		 */
487 		fctrl_reg |= IXGBE_FCTRL_RFCE;
488 		break;
489 	case ixgbe_fc_tx_pause:
490 		/*
491 		 * Tx Flow control is enabled, and Rx Flow control is
492 		 * disabled by software override.
493 		 */
494 		rmcs_reg |= IXGBE_RMCS_TFCE_802_3X;
495 		break;
496 	case ixgbe_fc_full:
497 		/* Flow control (both Rx and Tx) is enabled by SW override. */
498 		fctrl_reg |= IXGBE_FCTRL_RFCE;
499 		rmcs_reg |= IXGBE_RMCS_TFCE_802_3X;
500 		break;
501 	default:
502 		DEBUGOUT("Flow control param set incorrectly\n");
503 		ret_val = IXGBE_ERR_CONFIG;
504 		goto out;
505 		break;
506 	}
507 
508 	/* Set 802.3x based flow control settings. */
509 	fctrl_reg |= IXGBE_FCTRL_DPF;
510 	IXGBE_WRITE_REG(hw, IXGBE_FCTRL, fctrl_reg);
511 	IXGBE_WRITE_REG(hw, IXGBE_RMCS, rmcs_reg);
512 
513 	/* Set up and enable Rx high/low water mark thresholds, enable XON. */
514 	for (i = 0; i < IXGBE_DCB_MAX_TRAFFIC_CLASS; i++) {
515 		if ((hw->fc.current_mode & ixgbe_fc_tx_pause) &&
516 		    hw->fc.high_water[i]) {
517 			fcrtl = (hw->fc.low_water[i] << 10) | IXGBE_FCRTL_XONE;
518 			fcrth = (hw->fc.high_water[i] << 10) | IXGBE_FCRTH_FCEN;
519 			IXGBE_WRITE_REG(hw, IXGBE_FCRTL(i), fcrtl);
520 			IXGBE_WRITE_REG(hw, IXGBE_FCRTH(i), fcrth);
521 		} else {
522 			IXGBE_WRITE_REG(hw, IXGBE_FCRTL(i), 0);
523 			IXGBE_WRITE_REG(hw, IXGBE_FCRTH(i), 0);
524 		}
525 
526 	}
527 
528 	/* Configure pause time (2 TCs per register) */
529 	reg = hw->fc.pause_time * 0x00010001;
530 	for (i = 0; i < (IXGBE_DCB_MAX_TRAFFIC_CLASS / 2); i++)
531 		IXGBE_WRITE_REG(hw, IXGBE_FCTTV(i), reg);
532 
533 	/* Configure flow control refresh threshold value */
534 	IXGBE_WRITE_REG(hw, IXGBE_FCRTV, hw->fc.pause_time / 2);
535 
536 out:
537 	return ret_val;
538 }
539 
540 /**
541  *  ixgbe_start_mac_link_82598 - Configures MAC link settings
542  *  @hw: pointer to hardware structure
543  *
544  *  Configures link settings based on values in the ixgbe_hw struct.
545  *  Restarts the link.  Performs autonegotiation if needed.
546  **/
547 static s32 ixgbe_start_mac_link_82598(struct ixgbe_hw *hw,
548 				      bool autoneg_wait_to_complete)
549 {
550 	u32 autoc_reg;
551 	u32 links_reg;
552 	u32 i;
553 	s32 status = IXGBE_SUCCESS;
554 
555 	DEBUGFUNC("ixgbe_start_mac_link_82598");
556 
557 	/* Restart link */
558 	autoc_reg = IXGBE_READ_REG(hw, IXGBE_AUTOC);
559 	autoc_reg |= IXGBE_AUTOC_AN_RESTART;
560 	IXGBE_WRITE_REG(hw, IXGBE_AUTOC, autoc_reg);
561 
562 	/* Only poll for autoneg to complete if specified to do so */
563 	if (autoneg_wait_to_complete) {
564 		if ((autoc_reg & IXGBE_AUTOC_LMS_MASK) ==
565 		     IXGBE_AUTOC_LMS_KX4_AN ||
566 		    (autoc_reg & IXGBE_AUTOC_LMS_MASK) ==
567 		     IXGBE_AUTOC_LMS_KX4_AN_1G_AN) {
568 			links_reg = 0; /* Just in case Autoneg time = 0 */
569 			for (i = 0; i < IXGBE_AUTO_NEG_TIME; i++) {
570 				links_reg = IXGBE_READ_REG(hw, IXGBE_LINKS);
571 				if (links_reg & IXGBE_LINKS_KX_AN_COMP)
572 					break;
573 				msec_delay(100);
574 			}
575 			if (!(links_reg & IXGBE_LINKS_KX_AN_COMP)) {
576 				status = IXGBE_ERR_AUTONEG_NOT_COMPLETE;
577 				DEBUGOUT("Autonegotiation did not complete.\n");
578 			}
579 		}
580 	}
581 
582 	/* Add delay to filter out noises during initial link setup */
583 	msec_delay(50);
584 
585 	return status;
586 }
587 
588 /**
589  *  ixgbe_validate_link_ready - Function looks for phy link
590  *  @hw: pointer to hardware structure
591  *
592  *  Function indicates success when phy link is available. If phy is not ready
593  *  within 5 seconds of MAC indicating link, the function returns error.
594  **/
595 static s32 ixgbe_validate_link_ready(struct ixgbe_hw *hw)
596 {
597 	u32 timeout;
598 	u16 an_reg;
599 
600 	if (hw->device_id != IXGBE_DEV_ID_82598AT2)
601 		return IXGBE_SUCCESS;
602 
603 	for (timeout = 0;
604 	     timeout < IXGBE_VALIDATE_LINK_READY_TIMEOUT; timeout++) {
605 		hw->phy.ops.read_reg(hw, IXGBE_MDIO_AUTO_NEG_STATUS,
606 				     IXGBE_MDIO_AUTO_NEG_DEV_TYPE, &an_reg);
607 
608 		if ((an_reg & IXGBE_MII_AUTONEG_COMPLETE) &&
609 		    (an_reg & IXGBE_MII_AUTONEG_LINK_UP))
610 			break;
611 
612 		msec_delay(100);
613 	}
614 
615 	if (timeout == IXGBE_VALIDATE_LINK_READY_TIMEOUT) {
616 		DEBUGOUT("Link was indicated but link is down\n");
617 		return IXGBE_ERR_LINK_SETUP;
618 	}
619 
620 	return IXGBE_SUCCESS;
621 }
622 
623 /**
624  *  ixgbe_check_mac_link_82598 - Get link/speed status
625  *  @hw: pointer to hardware structure
626  *  @speed: pointer to link speed
627  *  @link_up: TRUE is link is up, FALSE otherwise
628  *  @link_up_wait_to_complete: bool used to wait for link up or not
629  *
630  *  Reads the links register to determine if link is up and the current speed
631  **/
632 static s32 ixgbe_check_mac_link_82598(struct ixgbe_hw *hw,
633 				      ixgbe_link_speed *speed, bool *link_up,
634 				      bool link_up_wait_to_complete)
635 {
636 	u32 links_reg;
637 	u32 i;
638 	u16 link_reg, adapt_comp_reg;
639 
640 	DEBUGFUNC("ixgbe_check_mac_link_82598");
641 
642 	/*
643 	 * SERDES PHY requires us to read link status from undocumented
644 	 * register 0xC79F.  Bit 0 set indicates link is up/ready; clear
645 	 * indicates link down.  OxC00C is read to check that the XAUI lanes
646 	 * are active.  Bit 0 clear indicates active; set indicates inactive.
647 	 */
648 	if (hw->phy.type == ixgbe_phy_nl) {
649 		hw->phy.ops.read_reg(hw, 0xC79F, IXGBE_TWINAX_DEV, &link_reg);
650 		hw->phy.ops.read_reg(hw, 0xC79F, IXGBE_TWINAX_DEV, &link_reg);
651 		hw->phy.ops.read_reg(hw, 0xC00C, IXGBE_TWINAX_DEV,
652 				     &adapt_comp_reg);
653 		if (link_up_wait_to_complete) {
654 			for (i = 0; i < IXGBE_LINK_UP_TIME; i++) {
655 				if ((link_reg & 1) &&
656 				    ((adapt_comp_reg & 1) == 0)) {
657 					*link_up = TRUE;
658 					break;
659 				} else {
660 					*link_up = FALSE;
661 				}
662 				msec_delay(100);
663 				hw->phy.ops.read_reg(hw, 0xC79F,
664 						     IXGBE_TWINAX_DEV,
665 						     &link_reg);
666 				hw->phy.ops.read_reg(hw, 0xC00C,
667 						     IXGBE_TWINAX_DEV,
668 						     &adapt_comp_reg);
669 			}
670 		} else {
671 			if ((link_reg & 1) && ((adapt_comp_reg & 1) == 0))
672 				*link_up = TRUE;
673 			else
674 				*link_up = FALSE;
675 		}
676 
677 		if (*link_up == FALSE)
678 			goto out;
679 	}
680 
681 	links_reg = IXGBE_READ_REG(hw, IXGBE_LINKS);
682 	if (link_up_wait_to_complete) {
683 		for (i = 0; i < IXGBE_LINK_UP_TIME; i++) {
684 			if (links_reg & IXGBE_LINKS_UP) {
685 				*link_up = TRUE;
686 				break;
687 			} else {
688 				*link_up = FALSE;
689 			}
690 			msec_delay(100);
691 			links_reg = IXGBE_READ_REG(hw, IXGBE_LINKS);
692 		}
693 	} else {
694 		if (links_reg & IXGBE_LINKS_UP)
695 			*link_up = TRUE;
696 		else
697 			*link_up = FALSE;
698 	}
699 
700 	if (links_reg & IXGBE_LINKS_SPEED)
701 		*speed = IXGBE_LINK_SPEED_10GB_FULL;
702 	else
703 		*speed = IXGBE_LINK_SPEED_1GB_FULL;
704 
705 	if ((hw->device_id == IXGBE_DEV_ID_82598AT2) && (*link_up == TRUE) &&
706 	    (ixgbe_validate_link_ready(hw) != IXGBE_SUCCESS))
707 		*link_up = FALSE;
708 
709 out:
710 	return IXGBE_SUCCESS;
711 }
712 
713 /**
714  *  ixgbe_setup_mac_link_82598 - Set MAC link speed
715  *  @hw: pointer to hardware structure
716  *  @speed: new link speed
717  *  @autoneg_wait_to_complete: TRUE when waiting for completion is needed
718  *
719  *  Set the link speed in the AUTOC register and restarts link.
720  **/
721 static s32 ixgbe_setup_mac_link_82598(struct ixgbe_hw *hw,
722 				      ixgbe_link_speed speed,
723 				      bool autoneg_wait_to_complete)
724 {
725 	bool autoneg = FALSE;
726 	s32 status = IXGBE_SUCCESS;
727 	ixgbe_link_speed link_capabilities = IXGBE_LINK_SPEED_UNKNOWN;
728 	u32 curr_autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
729 	u32 autoc = curr_autoc;
730 	u32 link_mode = autoc & IXGBE_AUTOC_LMS_MASK;
731 
732 	DEBUGFUNC("ixgbe_setup_mac_link_82598");
733 
734 	/* Check to see if speed passed in is supported. */
735 	ixgbe_get_link_capabilities(hw, &link_capabilities, &autoneg);
736 	speed &= link_capabilities;
737 
738 	if (speed == IXGBE_LINK_SPEED_UNKNOWN)
739 		status = IXGBE_ERR_LINK_SETUP;
740 
741 	/* Set KX4/KX support according to speed requested */
742 	else if (link_mode == IXGBE_AUTOC_LMS_KX4_AN ||
743 		 link_mode == IXGBE_AUTOC_LMS_KX4_AN_1G_AN) {
744 		autoc &= ~IXGBE_AUTOC_KX4_KX_SUPP_MASK;
745 		if (speed & IXGBE_LINK_SPEED_10GB_FULL)
746 			autoc |= IXGBE_AUTOC_KX4_SUPP;
747 		if (speed & IXGBE_LINK_SPEED_1GB_FULL)
748 			autoc |= IXGBE_AUTOC_KX_SUPP;
749 		if (autoc != curr_autoc)
750 			IXGBE_WRITE_REG(hw, IXGBE_AUTOC, autoc);
751 	}
752 
753 	if (status == IXGBE_SUCCESS) {
754 		/*
755 		 * Setup and restart the link based on the new values in
756 		 * ixgbe_hw This will write the AUTOC register based on the new
757 		 * stored values
758 		 */
759 		status = ixgbe_start_mac_link_82598(hw,
760 						    autoneg_wait_to_complete);
761 	}
762 
763 	return status;
764 }
765 
766 
767 /**
768  *  ixgbe_setup_copper_link_82598 - Set the PHY autoneg advertised field
769  *  @hw: pointer to hardware structure
770  *  @speed: new link speed
771  *  @autoneg_wait_to_complete: TRUE if waiting is needed to complete
772  *
773  *  Sets the link speed in the AUTOC register in the MAC and restarts link.
774  **/
775 static s32 ixgbe_setup_copper_link_82598(struct ixgbe_hw *hw,
776 					 ixgbe_link_speed speed,
777 					 bool autoneg_wait_to_complete)
778 {
779 	s32 status;
780 
781 	DEBUGFUNC("ixgbe_setup_copper_link_82598");
782 
783 	/* Setup the PHY according to input speed */
784 	status = hw->phy.ops.setup_link_speed(hw, speed,
785 					      autoneg_wait_to_complete);
786 	/* Set up MAC */
787 	ixgbe_start_mac_link_82598(hw, autoneg_wait_to_complete);
788 
789 	return status;
790 }
791 
792 /**
793  *  ixgbe_reset_hw_82598 - Performs hardware reset
794  *  @hw: pointer to hardware structure
795  *
796  *  Resets the hardware by resetting the transmit and receive units, masks and
797  *  clears all interrupts, performing a PHY reset, and performing a link (MAC)
798  *  reset.
799  **/
800 static s32 ixgbe_reset_hw_82598(struct ixgbe_hw *hw)
801 {
802 	s32 status = IXGBE_SUCCESS;
803 	s32 phy_status = IXGBE_SUCCESS;
804 	u32 ctrl;
805 	u32 gheccr;
806 	u32 i;
807 	u32 autoc;
808 	u8  analog_val;
809 
810 	DEBUGFUNC("ixgbe_reset_hw_82598");
811 
812 	/* Call adapter stop to disable tx/rx and clear interrupts */
813 	status = hw->mac.ops.stop_adapter(hw);
814 	if (status != IXGBE_SUCCESS)
815 		goto reset_hw_out;
816 
817 	/*
818 	 * Power up the Atlas Tx lanes if they are currently powered down.
819 	 * Atlas Tx lanes are powered down for MAC loopback tests, but
820 	 * they are not automatically restored on reset.
821 	 */
822 	hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_LPBK, &analog_val);
823 	if (analog_val & IXGBE_ATLAS_PDN_TX_REG_EN) {
824 		/* Enable Tx Atlas so packets can be transmitted again */
825 		hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_LPBK,
826 					     &analog_val);
827 		analog_val &= ~IXGBE_ATLAS_PDN_TX_REG_EN;
828 		hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_LPBK,
829 					      analog_val);
830 
831 		hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_10G,
832 					     &analog_val);
833 		analog_val &= ~IXGBE_ATLAS_PDN_TX_10G_QL_ALL;
834 		hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_10G,
835 					      analog_val);
836 
837 		hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_1G,
838 					     &analog_val);
839 		analog_val &= ~IXGBE_ATLAS_PDN_TX_1G_QL_ALL;
840 		hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_1G,
841 					      analog_val);
842 
843 		hw->mac.ops.read_analog_reg8(hw, IXGBE_ATLAS_PDN_AN,
844 					     &analog_val);
845 		analog_val &= ~IXGBE_ATLAS_PDN_TX_AN_QL_ALL;
846 		hw->mac.ops.write_analog_reg8(hw, IXGBE_ATLAS_PDN_AN,
847 					      analog_val);
848 	}
849 
850 	/* Reset PHY */
851 	if (hw->phy.reset_disable == FALSE) {
852 		/* PHY ops must be identified and initialized prior to reset */
853 
854 		/* Init PHY and function pointers, perform SFP setup */
855 		phy_status = hw->phy.ops.init(hw);
856 		if (phy_status == IXGBE_ERR_SFP_NOT_SUPPORTED)
857 			goto reset_hw_out;
858 		if (phy_status == IXGBE_ERR_SFP_NOT_PRESENT)
859 			goto mac_reset_top;
860 
861 		hw->phy.ops.reset(hw);
862 	}
863 
864 mac_reset_top:
865 	/*
866 	 * Issue global reset to the MAC.  This needs to be a SW reset.
867 	 * If link reset is used, it might reset the MAC when mng is using it
868 	 */
869 	ctrl = IXGBE_READ_REG(hw, IXGBE_CTRL) | IXGBE_CTRL_RST;
870 	IXGBE_WRITE_REG(hw, IXGBE_CTRL, ctrl);
871 	IXGBE_WRITE_FLUSH(hw);
872 
873 	/* Poll for reset bit to self-clear indicating reset is complete */
874 	for (i = 0; i < 10; i++) {
875 		usec_delay(1);
876 		ctrl = IXGBE_READ_REG(hw, IXGBE_CTRL);
877 		if (!(ctrl & IXGBE_CTRL_RST))
878 			break;
879 	}
880 	if (ctrl & IXGBE_CTRL_RST) {
881 		status = IXGBE_ERR_RESET_FAILED;
882 		DEBUGOUT("Reset polling failed to complete.\n");
883 	}
884 
885 	msec_delay(50);
886 
887 	/*
888 	 * Double resets are required for recovery from certain error
889 	 * conditions.  Between resets, it is necessary to stall to allow time
890 	 * for any pending HW events to complete.
891 	 */
892 	if (hw->mac.flags & IXGBE_FLAGS_DOUBLE_RESET_REQUIRED) {
893 		hw->mac.flags &= ~IXGBE_FLAGS_DOUBLE_RESET_REQUIRED;
894 		goto mac_reset_top;
895 	}
896 
897 	gheccr = IXGBE_READ_REG(hw, IXGBE_GHECCR);
898 	gheccr &= ~((1 << 21) | (1 << 18) | (1 << 9) | (1 << 6));
899 	IXGBE_WRITE_REG(hw, IXGBE_GHECCR, gheccr);
900 
901 	/*
902 	 * Store the original AUTOC value if it has not been
903 	 * stored off yet.  Otherwise restore the stored original
904 	 * AUTOC value since the reset operation sets back to deaults.
905 	 */
906 	autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
907 	if (hw->mac.orig_link_settings_stored == FALSE) {
908 		hw->mac.orig_autoc = autoc;
909 		hw->mac.orig_link_settings_stored = TRUE;
910 	} else if (autoc != hw->mac.orig_autoc) {
911 		IXGBE_WRITE_REG(hw, IXGBE_AUTOC, hw->mac.orig_autoc);
912 	}
913 
914 	/* Store the permanent mac address */
915 	hw->mac.ops.get_mac_addr(hw, hw->mac.perm_addr);
916 
917 	/*
918 	 * Store MAC address from RAR0, clear receive address registers, and
919 	 * clear the multicast table
920 	 */
921 	hw->mac.ops.init_rx_addrs(hw);
922 
923 reset_hw_out:
924 	if (phy_status != IXGBE_SUCCESS)
925 		status = phy_status;
926 
927 	return status;
928 }
929 
930 /**
931  *  ixgbe_set_vmdq_82598 - Associate a VMDq set index with a rx address
932  *  @hw: pointer to hardware struct
933  *  @rar: receive address register index to associate with a VMDq index
934  *  @vmdq: VMDq set index
935  **/
936 s32 ixgbe_set_vmdq_82598(struct ixgbe_hw *hw, u32 rar, u32 vmdq)
937 {
938 	u32 rar_high;
939 	u32 rar_entries = hw->mac.num_rar_entries;
940 
941 	DEBUGFUNC("ixgbe_set_vmdq_82598");
942 
943 	/* Make sure we are using a valid rar index range */
944 	if (rar >= rar_entries) {
945 		DEBUGOUT1("RAR index %d is out of range.\n", rar);
946 		return IXGBE_ERR_INVALID_ARGUMENT;
947 	}
948 
949 	rar_high = IXGBE_READ_REG(hw, IXGBE_RAH(rar));
950 	rar_high &= ~IXGBE_RAH_VIND_MASK;
951 	rar_high |= ((vmdq << IXGBE_RAH_VIND_SHIFT) & IXGBE_RAH_VIND_MASK);
952 	IXGBE_WRITE_REG(hw, IXGBE_RAH(rar), rar_high);
953 	return IXGBE_SUCCESS;
954 }
955 
956 /**
957  *  ixgbe_clear_vmdq_82598 - Disassociate a VMDq set index from an rx address
958  *  @hw: pointer to hardware struct
959  *  @rar: receive address register index to associate with a VMDq index
960  *  @vmdq: VMDq clear index (not used in 82598, but elsewhere)
961  **/
962 static s32 ixgbe_clear_vmdq_82598(struct ixgbe_hw *hw, u32 rar, u32 vmdq)
963 {
964 	u32 rar_high;
965 	u32 rar_entries = hw->mac.num_rar_entries;
966 
967 	UNREFERENCED_1PARAMETER(vmdq);
968 
969 	/* Make sure we are using a valid rar index range */
970 	if (rar >= rar_entries) {
971 		DEBUGOUT1("RAR index %d is out of range.\n", rar);
972 		return IXGBE_ERR_INVALID_ARGUMENT;
973 	}
974 
975 	rar_high = IXGBE_READ_REG(hw, IXGBE_RAH(rar));
976 	if (rar_high & IXGBE_RAH_VIND_MASK) {
977 		rar_high &= ~IXGBE_RAH_VIND_MASK;
978 		IXGBE_WRITE_REG(hw, IXGBE_RAH(rar), rar_high);
979 	}
980 
981 	return IXGBE_SUCCESS;
982 }
983 
984 /**
985  *  ixgbe_set_vfta_82598 - Set VLAN filter table
986  *  @hw: pointer to hardware structure
987  *  @vlan: VLAN id to write to VLAN filter
988  *  @vind: VMDq output index that maps queue to VLAN id in VFTA
989  *  @vlan_on: boolean flag to turn on/off VLAN in VFTA
990  *
991  *  Turn on/off specified VLAN in the VLAN filter table.
992  **/
993 s32 ixgbe_set_vfta_82598(struct ixgbe_hw *hw, u32 vlan, u32 vind,
994 			 bool vlan_on)
995 {
996 	u32 regindex;
997 	u32 bitindex;
998 	u32 bits;
999 	u32 vftabyte;
1000 
1001 	DEBUGFUNC("ixgbe_set_vfta_82598");
1002 
1003 	if (vlan > 4095)
1004 		return IXGBE_ERR_PARAM;
1005 
1006 	/* Determine 32-bit word position in array */
1007 	regindex = (vlan >> 5) & 0x7F;   /* upper seven bits */
1008 
1009 	/* Determine the location of the (VMD) queue index */
1010 	vftabyte =  ((vlan >> 3) & 0x03); /* bits (4:3) indicating byte array */
1011 	bitindex = (vlan & 0x7) << 2;    /* lower 3 bits indicate nibble */
1012 
1013 	/* Set the nibble for VMD queue index */
1014 	bits = IXGBE_READ_REG(hw, IXGBE_VFTAVIND(vftabyte, regindex));
1015 	bits &= (~(0x0F << bitindex));
1016 	bits |= (vind << bitindex);
1017 	IXGBE_WRITE_REG(hw, IXGBE_VFTAVIND(vftabyte, regindex), bits);
1018 
1019 	/* Determine the location of the bit for this VLAN id */
1020 	bitindex = vlan & 0x1F;   /* lower five bits */
1021 
1022 	bits = IXGBE_READ_REG(hw, IXGBE_VFTA(regindex));
1023 	if (vlan_on)
1024 		/* Turn on this VLAN id */
1025 		bits |= (1 << bitindex);
1026 	else
1027 		/* Turn off this VLAN id */
1028 		bits &= ~(1 << bitindex);
1029 	IXGBE_WRITE_REG(hw, IXGBE_VFTA(regindex), bits);
1030 
1031 	return IXGBE_SUCCESS;
1032 }
1033 
1034 /**
1035  *  ixgbe_clear_vfta_82598 - Clear VLAN filter table
1036  *  @hw: pointer to hardware structure
1037  *
1038  *  Clears the VLAN filer table, and the VMDq index associated with the filter
1039  **/
1040 static s32 ixgbe_clear_vfta_82598(struct ixgbe_hw *hw)
1041 {
1042 	u32 offset;
1043 	u32 vlanbyte;
1044 
1045 	DEBUGFUNC("ixgbe_clear_vfta_82598");
1046 
1047 	for (offset = 0; offset < hw->mac.vft_size; offset++)
1048 		IXGBE_WRITE_REG(hw, IXGBE_VFTA(offset), 0);
1049 
1050 	for (vlanbyte = 0; vlanbyte < 4; vlanbyte++)
1051 		for (offset = 0; offset < hw->mac.vft_size; offset++)
1052 			IXGBE_WRITE_REG(hw, IXGBE_VFTAVIND(vlanbyte, offset),
1053 					0);
1054 
1055 	return IXGBE_SUCCESS;
1056 }
1057 
1058 /**
1059  *  ixgbe_read_analog_reg8_82598 - Reads 8 bit Atlas analog register
1060  *  @hw: pointer to hardware structure
1061  *  @reg: analog register to read
1062  *  @val: read value
1063  *
1064  *  Performs read operation to Atlas analog register specified.
1065  **/
1066 s32 ixgbe_read_analog_reg8_82598(struct ixgbe_hw *hw, u32 reg, u8 *val)
1067 {
1068 	u32  atlas_ctl;
1069 
1070 	DEBUGFUNC("ixgbe_read_analog_reg8_82598");
1071 
1072 	IXGBE_WRITE_REG(hw, IXGBE_ATLASCTL,
1073 			IXGBE_ATLASCTL_WRITE_CMD | (reg << 8));
1074 	IXGBE_WRITE_FLUSH(hw);
1075 	usec_delay(10);
1076 	atlas_ctl = IXGBE_READ_REG(hw, IXGBE_ATLASCTL);
1077 	*val = (u8)atlas_ctl;
1078 
1079 	return IXGBE_SUCCESS;
1080 }
1081 
1082 /**
1083  *  ixgbe_write_analog_reg8_82598 - Writes 8 bit Atlas analog register
1084  *  @hw: pointer to hardware structure
1085  *  @reg: atlas register to write
1086  *  @val: value to write
1087  *
1088  *  Performs write operation to Atlas analog register specified.
1089  **/
1090 s32 ixgbe_write_analog_reg8_82598(struct ixgbe_hw *hw, u32 reg, u8 val)
1091 {
1092 	u32  atlas_ctl;
1093 
1094 	DEBUGFUNC("ixgbe_write_analog_reg8_82598");
1095 
1096 	atlas_ctl = (reg << 8) | val;
1097 	IXGBE_WRITE_REG(hw, IXGBE_ATLASCTL, atlas_ctl);
1098 	IXGBE_WRITE_FLUSH(hw);
1099 	usec_delay(10);
1100 
1101 	return IXGBE_SUCCESS;
1102 }
1103 
1104 /**
1105  *  ixgbe_read_i2c_phy_82598 - Reads 8 bit word over I2C interface.
1106  *  @hw: pointer to hardware structure
1107  *  @dev_addr: address to read from
1108  *  @byte_offset: byte offset to read from dev_addr
1109  *  @eeprom_data: value read
1110  *
1111  *  Performs 8 byte read operation to SFP module's EEPROM over I2C interface.
1112  **/
1113 static s32 ixgbe_read_i2c_phy_82598(struct ixgbe_hw *hw, u8 dev_addr,
1114 				    u8 byte_offset, u8 *eeprom_data)
1115 {
1116 	s32 status = IXGBE_SUCCESS;
1117 	u16 sfp_addr = 0;
1118 	u16 sfp_data = 0;
1119 	u16 sfp_stat = 0;
1120 	u16 gssr;
1121 	u32 i;
1122 
1123 	DEBUGFUNC("ixgbe_read_i2c_phy_82598");
1124 
1125 	if (IXGBE_READ_REG(hw, IXGBE_STATUS) & IXGBE_STATUS_LAN_ID_1)
1126 		gssr = IXGBE_GSSR_PHY1_SM;
1127 	else
1128 		gssr = IXGBE_GSSR_PHY0_SM;
1129 
1130 	if (hw->mac.ops.acquire_swfw_sync(hw, gssr) != IXGBE_SUCCESS)
1131 		return IXGBE_ERR_SWFW_SYNC;
1132 
1133 	if (hw->phy.type == ixgbe_phy_nl) {
1134 		/*
1135 		 * NetLogic phy SDA/SCL registers are at addresses 0xC30A to
1136 		 * 0xC30D. These registers are used to talk to the SFP+
1137 		 * module's EEPROM through the SDA/SCL (I2C) interface.
1138 		 */
1139 		sfp_addr = (dev_addr << 8) + byte_offset;
1140 		sfp_addr = (sfp_addr | IXGBE_I2C_EEPROM_READ_MASK);
1141 		hw->phy.ops.write_reg_mdi(hw,
1142 					  IXGBE_MDIO_PMA_PMD_SDA_SCL_ADDR,
1143 					  IXGBE_MDIO_PMA_PMD_DEV_TYPE,
1144 					  sfp_addr);
1145 
1146 		/* Poll status */
1147 		for (i = 0; i < 100; i++) {
1148 			hw->phy.ops.read_reg_mdi(hw,
1149 						IXGBE_MDIO_PMA_PMD_SDA_SCL_STAT,
1150 						IXGBE_MDIO_PMA_PMD_DEV_TYPE,
1151 						&sfp_stat);
1152 			sfp_stat = sfp_stat & IXGBE_I2C_EEPROM_STATUS_MASK;
1153 			if (sfp_stat != IXGBE_I2C_EEPROM_STATUS_IN_PROGRESS)
1154 				break;
1155 			msec_delay(10);
1156 		}
1157 
1158 		if (sfp_stat != IXGBE_I2C_EEPROM_STATUS_PASS) {
1159 			DEBUGOUT("EEPROM read did not pass.\n");
1160 			status = IXGBE_ERR_SFP_NOT_PRESENT;
1161 			goto out;
1162 		}
1163 
1164 		/* Read data */
1165 		hw->phy.ops.read_reg_mdi(hw, IXGBE_MDIO_PMA_PMD_SDA_SCL_DATA,
1166 					IXGBE_MDIO_PMA_PMD_DEV_TYPE, &sfp_data);
1167 
1168 		*eeprom_data = (u8)(sfp_data >> 8);
1169 	} else {
1170 		status = IXGBE_ERR_PHY;
1171 	}
1172 
1173 out:
1174 	hw->mac.ops.release_swfw_sync(hw, gssr);
1175 	return status;
1176 }
1177 
1178 /**
1179  *  ixgbe_read_i2c_eeprom_82598 - Reads 8 bit word over I2C interface.
1180  *  @hw: pointer to hardware structure
1181  *  @byte_offset: EEPROM byte offset to read
1182  *  @eeprom_data: value read
1183  *
1184  *  Performs 8 byte read operation to SFP module's EEPROM over I2C interface.
1185  **/
1186 s32 ixgbe_read_i2c_eeprom_82598(struct ixgbe_hw *hw, u8 byte_offset,
1187 				u8 *eeprom_data)
1188 {
1189 	return ixgbe_read_i2c_phy_82598(hw, IXGBE_I2C_EEPROM_DEV_ADDR,
1190 					byte_offset, eeprom_data);
1191 }
1192 
1193 /**
1194  *  ixgbe_read_i2c_sff8472_82598 - Reads 8 bit word over I2C interface.
1195  *  @hw: pointer to hardware structure
1196  *  @byte_offset: byte offset at address 0xA2
1197  *  @eeprom_data: value read
1198  *
1199  *  Performs 8 byte read operation to SFP module's SFF-8472 data over I2C
1200  **/
1201 static s32 ixgbe_read_i2c_sff8472_82598(struct ixgbe_hw *hw, u8 byte_offset,
1202 					u8 *sff8472_data)
1203 {
1204 	return ixgbe_read_i2c_phy_82598(hw, IXGBE_I2C_EEPROM_DEV_ADDR2,
1205 					byte_offset, sff8472_data);
1206 }
1207 
1208 /**
1209  *  ixgbe_get_supported_physical_layer_82598 - Returns physical layer type
1210  *  @hw: pointer to hardware structure
1211  *
1212  *  Determines physical layer capabilities of the current configuration.
1213  **/
1214 u32 ixgbe_get_supported_physical_layer_82598(struct ixgbe_hw *hw)
1215 {
1216 	u32 physical_layer = IXGBE_PHYSICAL_LAYER_UNKNOWN;
1217 	u32 autoc = IXGBE_READ_REG(hw, IXGBE_AUTOC);
1218 	u32 pma_pmd_10g = autoc & IXGBE_AUTOC_10G_PMA_PMD_MASK;
1219 	u32 pma_pmd_1g = autoc & IXGBE_AUTOC_1G_PMA_PMD_MASK;
1220 	u16 ext_ability = 0;
1221 
1222 	DEBUGFUNC("ixgbe_get_supported_physical_layer_82598");
1223 
1224 	hw->phy.ops.identify(hw);
1225 
1226 	/* Copper PHY must be checked before AUTOC LMS to determine correct
1227 	 * physical layer because 10GBase-T PHYs use LMS = KX4/KX */
1228 	switch (hw->phy.type) {
1229 	case ixgbe_phy_tn:
1230 	case ixgbe_phy_cu_unknown:
1231 		hw->phy.ops.read_reg(hw, IXGBE_MDIO_PHY_EXT_ABILITY,
1232 		IXGBE_MDIO_PMA_PMD_DEV_TYPE, &ext_ability);
1233 		if (ext_ability & IXGBE_MDIO_PHY_10GBASET_ABILITY)
1234 			physical_layer |= IXGBE_PHYSICAL_LAYER_10GBASE_T;
1235 		if (ext_ability & IXGBE_MDIO_PHY_1000BASET_ABILITY)
1236 			physical_layer |= IXGBE_PHYSICAL_LAYER_1000BASE_T;
1237 		if (ext_ability & IXGBE_MDIO_PHY_100BASETX_ABILITY)
1238 			physical_layer |= IXGBE_PHYSICAL_LAYER_100BASE_TX;
1239 		goto out;
1240 	default:
1241 		break;
1242 	}
1243 
1244 	switch (autoc & IXGBE_AUTOC_LMS_MASK) {
1245 	case IXGBE_AUTOC_LMS_1G_AN:
1246 	case IXGBE_AUTOC_LMS_1G_LINK_NO_AN:
1247 		if (pma_pmd_1g == IXGBE_AUTOC_1G_KX)
1248 			physical_layer = IXGBE_PHYSICAL_LAYER_1000BASE_KX;
1249 		else
1250 			physical_layer = IXGBE_PHYSICAL_LAYER_1000BASE_BX;
1251 		break;
1252 	case IXGBE_AUTOC_LMS_10G_LINK_NO_AN:
1253 		if (pma_pmd_10g == IXGBE_AUTOC_10G_CX4)
1254 			physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_CX4;
1255 		else if (pma_pmd_10g == IXGBE_AUTOC_10G_KX4)
1256 			physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_KX4;
1257 		else /* XAUI */
1258 			physical_layer = IXGBE_PHYSICAL_LAYER_UNKNOWN;
1259 		break;
1260 	case IXGBE_AUTOC_LMS_KX4_AN:
1261 	case IXGBE_AUTOC_LMS_KX4_AN_1G_AN:
1262 		if (autoc & IXGBE_AUTOC_KX_SUPP)
1263 			physical_layer |= IXGBE_PHYSICAL_LAYER_1000BASE_KX;
1264 		if (autoc & IXGBE_AUTOC_KX4_SUPP)
1265 			physical_layer |= IXGBE_PHYSICAL_LAYER_10GBASE_KX4;
1266 		break;
1267 	default:
1268 		break;
1269 	}
1270 
1271 	if (hw->phy.type == ixgbe_phy_nl) {
1272 		hw->phy.ops.identify_sfp(hw);
1273 
1274 		switch (hw->phy.sfp_type) {
1275 		case ixgbe_sfp_type_da_cu:
1276 			physical_layer = IXGBE_PHYSICAL_LAYER_SFP_PLUS_CU;
1277 			break;
1278 		case ixgbe_sfp_type_sr:
1279 			physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_SR;
1280 			break;
1281 		case ixgbe_sfp_type_lr:
1282 			physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_LR;
1283 			break;
1284 		default:
1285 			physical_layer = IXGBE_PHYSICAL_LAYER_UNKNOWN;
1286 			break;
1287 		}
1288 	}
1289 
1290 	switch (hw->device_id) {
1291 	case IXGBE_DEV_ID_82598_DA_DUAL_PORT:
1292 		physical_layer = IXGBE_PHYSICAL_LAYER_SFP_PLUS_CU;
1293 		break;
1294 	case IXGBE_DEV_ID_82598AF_DUAL_PORT:
1295 	case IXGBE_DEV_ID_82598AF_SINGLE_PORT:
1296 	case IXGBE_DEV_ID_82598_SR_DUAL_PORT_EM:
1297 		physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_SR;
1298 		break;
1299 	case IXGBE_DEV_ID_82598EB_XF_LR:
1300 		physical_layer = IXGBE_PHYSICAL_LAYER_10GBASE_LR;
1301 		break;
1302 	default:
1303 		break;
1304 	}
1305 
1306 out:
1307 	return physical_layer;
1308 }
1309 
1310 /**
1311  *  ixgbe_set_lan_id_multi_port_pcie_82598 - Set LAN id for PCIe multiple
1312  *  port devices.
1313  *  @hw: pointer to the HW structure
1314  *
1315  *  Calls common function and corrects issue with some single port devices
1316  *  that enable LAN1 but not LAN0.
1317  **/
1318 void ixgbe_set_lan_id_multi_port_pcie_82598(struct ixgbe_hw *hw)
1319 {
1320 	struct ixgbe_bus_info *bus = &hw->bus;
1321 	u16 pci_gen = 0;
1322 	u16 pci_ctrl2 = 0;
1323 
1324 	DEBUGFUNC("ixgbe_set_lan_id_multi_port_pcie_82598");
1325 
1326 	ixgbe_set_lan_id_multi_port_pcie(hw);
1327 
1328 	/* check if LAN0 is disabled */
1329 	hw->eeprom.ops.read(hw, IXGBE_PCIE_GENERAL_PTR, &pci_gen);
1330 	if ((pci_gen != 0) && (pci_gen != 0xFFFF)) {
1331 
1332 		hw->eeprom.ops.read(hw, pci_gen + IXGBE_PCIE_CTRL2, &pci_ctrl2);
1333 
1334 		/* if LAN0 is completely disabled force function to 0 */
1335 		if ((pci_ctrl2 & IXGBE_PCIE_CTRL2_LAN_DISABLE) &&
1336 		    !(pci_ctrl2 & IXGBE_PCIE_CTRL2_DISABLE_SELECT) &&
1337 		    !(pci_ctrl2 & IXGBE_PCIE_CTRL2_DUMMY_ENABLE)) {
1338 
1339 			bus->func = 0;
1340 		}
1341 	}
1342 }
1343 
1344 /**
1345  *  ixgbe_enable_relaxed_ordering_82598 - enable relaxed ordering
1346  *  @hw: pointer to hardware structure
1347  *
1348  **/
1349 void ixgbe_enable_relaxed_ordering_82598(struct ixgbe_hw *hw)
1350 {
1351 	u32 regval;
1352 	u32 i;
1353 
1354 	DEBUGFUNC("ixgbe_enable_relaxed_ordering_82598");
1355 
1356 	/* Enable relaxed ordering */
1357 	for (i = 0; ((i < hw->mac.max_tx_queues) &&
1358 	     (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
1359 		regval = IXGBE_READ_REG(hw, IXGBE_DCA_TXCTRL(i));
1360 		regval |= IXGBE_DCA_TXCTRL_DESC_WRO_EN;
1361 		IXGBE_WRITE_REG(hw, IXGBE_DCA_TXCTRL(i), regval);
1362 	}
1363 
1364 	for (i = 0; ((i < hw->mac.max_rx_queues) &&
1365 	     (i < IXGBE_DCA_MAX_QUEUES_82598)); i++) {
1366 		regval = IXGBE_READ_REG(hw, IXGBE_DCA_RXCTRL(i));
1367 		regval |= IXGBE_DCA_RXCTRL_DATA_WRO_EN |
1368 			  IXGBE_DCA_RXCTRL_HEAD_WRO_EN;
1369 		IXGBE_WRITE_REG(hw, IXGBE_DCA_RXCTRL(i), regval);
1370 	}
1371 
1372 }
1373 
1374 /**
1375  * ixgbe_set_rxpba_82598 - Initialize RX packet buffer
1376  * @hw: pointer to hardware structure
1377  * @num_pb: number of packet buffers to allocate
1378  * @headroom: reserve n KB of headroom
1379  * @strategy: packet buffer allocation strategy
1380  **/
1381 static void ixgbe_set_rxpba_82598(struct ixgbe_hw *hw, int num_pb,
1382 				  u32 headroom, int strategy)
1383 {
1384 	u32 rxpktsize = IXGBE_RXPBSIZE_64KB;
1385 	u8 i = 0;
1386 	UNREFERENCED_1PARAMETER(headroom);
1387 
1388 	if (!num_pb)
1389 		return;
1390 
1391 	/* Setup Rx packet buffer sizes */
1392 	switch (strategy) {
1393 	case PBA_STRATEGY_WEIGHTED:
1394 		/* Setup the first four at 80KB */
1395 		rxpktsize = IXGBE_RXPBSIZE_80KB;
1396 		for (; i < 4; i++)
1397 			IXGBE_WRITE_REG(hw, IXGBE_RXPBSIZE(i), rxpktsize);
1398 		/* Setup the last four at 48KB...don't re-init i */
1399 		rxpktsize = IXGBE_RXPBSIZE_48KB;
1400 		/* Fall Through */
1401 	case PBA_STRATEGY_EQUAL:
1402 	default:
1403 		/* Divide the remaining Rx packet buffer evenly among the TCs */
1404 		for (; i < IXGBE_MAX_PACKET_BUFFERS; i++)
1405 			IXGBE_WRITE_REG(hw, IXGBE_RXPBSIZE(i), rxpktsize);
1406 		break;
1407 	}
1408 
1409 	/* Setup Tx packet buffer sizes */
1410 	for (i = 0; i < IXGBE_MAX_PACKET_BUFFERS; i++)
1411 		IXGBE_WRITE_REG(hw, IXGBE_TXPBSIZE(i), IXGBE_TXPBSIZE_40KB);
1412 
1413 	return;
1414 }
1415