xref: /linux/drivers/infiniband/hw/hfi1/platform.c (revision ec63e2a4897075e427c121d863bd89c44578094f)
1 /*
2  * Copyright(c) 2015, 2016 Intel Corporation.
3  *
4  * This file is provided under a dual BSD/GPLv2 license.  When using or
5  * redistributing this file, you may do so under either license.
6  *
7  * GPL LICENSE SUMMARY
8  *
9  * This program is free software; you can redistribute it and/or modify
10  * it under the terms of version 2 of the GNU General Public License as
11  * published by the Free Software Foundation.
12  *
13  * This program is distributed in the hope that it will be useful, but
14  * WITHOUT ANY WARRANTY; without even the implied warranty of
15  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
16  * General Public License for more details.
17  *
18  * BSD LICENSE
19  *
20  * Redistribution and use in source and binary forms, with or without
21  * modification, are permitted provided that the following conditions
22  * are met:
23  *
24  *  - Redistributions of source code must retain the above copyright
25  *    notice, this list of conditions and the following disclaimer.
26  *  - Redistributions in binary form must reproduce the above copyright
27  *    notice, this list of conditions and the following disclaimer in
28  *    the documentation and/or other materials provided with the
29  *    distribution.
30  *  - Neither the name of Intel Corporation nor the names of its
31  *    contributors may be used to endorse or promote products derived
32  *    from this software without specific prior written permission.
33  *
34  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
35  * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
36  * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
37  * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
38  * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
39  * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
40  * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
41  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
42  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
43  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
44  * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
45  *
46  */
47 
48 #include <linux/firmware.h>
49 
50 #include "hfi.h"
51 #include "efivar.h"
52 #include "eprom.h"
53 
54 #define DEFAULT_PLATFORM_CONFIG_NAME "hfi1_platform.dat"
55 
56 static int validate_scratch_checksum(struct hfi1_devdata *dd)
57 {
58 	u64 checksum = 0, temp_scratch = 0;
59 	int i, j, version;
60 
61 	temp_scratch = read_csr(dd, ASIC_CFG_SCRATCH);
62 	version = (temp_scratch & BITMAP_VERSION_SMASK) >> BITMAP_VERSION_SHIFT;
63 
64 	/* Prevent power on default of all zeroes from passing checksum */
65 	if (!version) {
66 		dd_dev_err(dd, "%s: Config bitmap uninitialized\n", __func__);
67 		dd_dev_err(dd,
68 			   "%s: Please update your BIOS to support active channels\n",
69 			   __func__);
70 		return 0;
71 	}
72 
73 	/*
74 	 * ASIC scratch 0 only contains the checksum and bitmap version as
75 	 * fields of interest, both of which are handled separately from the
76 	 * loop below, so skip it
77 	 */
78 	checksum += version;
79 	for (i = 1; i < ASIC_NUM_SCRATCH; i++) {
80 		temp_scratch = read_csr(dd, ASIC_CFG_SCRATCH + (8 * i));
81 		for (j = sizeof(u64); j != 0; j -= 2) {
82 			checksum += (temp_scratch & 0xFFFF);
83 			temp_scratch >>= 16;
84 		}
85 	}
86 
87 	while (checksum >> 16)
88 		checksum = (checksum & CHECKSUM_MASK) + (checksum >> 16);
89 
90 	temp_scratch = read_csr(dd, ASIC_CFG_SCRATCH);
91 	temp_scratch &= CHECKSUM_SMASK;
92 	temp_scratch >>= CHECKSUM_SHIFT;
93 
94 	if (checksum + temp_scratch == 0xFFFF)
95 		return 1;
96 
97 	dd_dev_err(dd, "%s: Configuration bitmap corrupted\n", __func__);
98 	return 0;
99 }
100 
101 static void save_platform_config_fields(struct hfi1_devdata *dd)
102 {
103 	struct hfi1_pportdata *ppd = dd->pport;
104 	u64 temp_scratch = 0, temp_dest = 0;
105 
106 	temp_scratch = read_csr(dd, ASIC_CFG_SCRATCH_1);
107 
108 	temp_dest = temp_scratch &
109 		    (dd->hfi1_id ? PORT1_PORT_TYPE_SMASK :
110 		     PORT0_PORT_TYPE_SMASK);
111 	ppd->port_type = temp_dest >>
112 			 (dd->hfi1_id ? PORT1_PORT_TYPE_SHIFT :
113 			  PORT0_PORT_TYPE_SHIFT);
114 
115 	temp_dest = temp_scratch &
116 		    (dd->hfi1_id ? PORT1_LOCAL_ATTEN_SMASK :
117 		     PORT0_LOCAL_ATTEN_SMASK);
118 	ppd->local_atten = temp_dest >>
119 			   (dd->hfi1_id ? PORT1_LOCAL_ATTEN_SHIFT :
120 			    PORT0_LOCAL_ATTEN_SHIFT);
121 
122 	temp_dest = temp_scratch &
123 		    (dd->hfi1_id ? PORT1_REMOTE_ATTEN_SMASK :
124 		     PORT0_REMOTE_ATTEN_SMASK);
125 	ppd->remote_atten = temp_dest >>
126 			    (dd->hfi1_id ? PORT1_REMOTE_ATTEN_SHIFT :
127 			     PORT0_REMOTE_ATTEN_SHIFT);
128 
129 	temp_dest = temp_scratch &
130 		    (dd->hfi1_id ? PORT1_DEFAULT_ATTEN_SMASK :
131 		     PORT0_DEFAULT_ATTEN_SMASK);
132 	ppd->default_atten = temp_dest >>
133 			     (dd->hfi1_id ? PORT1_DEFAULT_ATTEN_SHIFT :
134 			      PORT0_DEFAULT_ATTEN_SHIFT);
135 
136 	temp_scratch = read_csr(dd, dd->hfi1_id ? ASIC_CFG_SCRATCH_3 :
137 				ASIC_CFG_SCRATCH_2);
138 
139 	ppd->tx_preset_eq = (temp_scratch & TX_EQ_SMASK) >> TX_EQ_SHIFT;
140 	ppd->tx_preset_noeq = (temp_scratch & TX_NO_EQ_SMASK) >> TX_NO_EQ_SHIFT;
141 	ppd->rx_preset = (temp_scratch & RX_SMASK) >> RX_SHIFT;
142 
143 	ppd->max_power_class = (temp_scratch & QSFP_MAX_POWER_SMASK) >>
144 				QSFP_MAX_POWER_SHIFT;
145 
146 	ppd->config_from_scratch = true;
147 }
148 
149 void get_platform_config(struct hfi1_devdata *dd)
150 {
151 	int ret = 0;
152 	u8 *temp_platform_config = NULL;
153 	u32 esize;
154 	const struct firmware *platform_config_file = NULL;
155 
156 	if (is_integrated(dd)) {
157 		if (validate_scratch_checksum(dd)) {
158 			save_platform_config_fields(dd);
159 			return;
160 		}
161 	} else {
162 		ret = eprom_read_platform_config(dd,
163 						 (void **)&temp_platform_config,
164 						 &esize);
165 		if (!ret) {
166 			/* success */
167 			dd->platform_config.data = temp_platform_config;
168 			dd->platform_config.size = esize;
169 			return;
170 		}
171 	}
172 	dd_dev_err(dd,
173 		   "%s: Failed to get platform config, falling back to sub-optimal default file\n",
174 		   __func__);
175 
176 	ret = request_firmware(&platform_config_file,
177 			       DEFAULT_PLATFORM_CONFIG_NAME,
178 			       &dd->pcidev->dev);
179 	if (ret) {
180 		dd_dev_err(dd,
181 			   "%s: No default platform config file found\n",
182 			   __func__);
183 		return;
184 	}
185 
186 	/*
187 	 * Allocate separate memory block to store data and free firmware
188 	 * structure. This allows free_platform_config to treat EPROM and
189 	 * fallback configs in the same manner.
190 	 */
191 	dd->platform_config.data = kmemdup(platform_config_file->data,
192 					   platform_config_file->size,
193 					   GFP_KERNEL);
194 	dd->platform_config.size = platform_config_file->size;
195 	release_firmware(platform_config_file);
196 }
197 
198 void free_platform_config(struct hfi1_devdata *dd)
199 {
200 	/* Release memory allocated for eprom or fallback file read. */
201 	kfree(dd->platform_config.data);
202 	dd->platform_config.data = NULL;
203 }
204 
205 void get_port_type(struct hfi1_pportdata *ppd)
206 {
207 	int ret;
208 	u32 temp;
209 
210 	ret = get_platform_config_field(ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
211 					PORT_TABLE_PORT_TYPE, &temp,
212 					4);
213 	if (ret) {
214 		ppd->port_type = PORT_TYPE_UNKNOWN;
215 		return;
216 	}
217 	ppd->port_type = temp;
218 }
219 
220 int set_qsfp_tx(struct hfi1_pportdata *ppd, int on)
221 {
222 	u8 tx_ctrl_byte = on ? 0x0 : 0xF;
223 	int ret = 0;
224 
225 	ret = qsfp_write(ppd, ppd->dd->hfi1_id, QSFP_TX_CTRL_BYTE_OFFS,
226 			 &tx_ctrl_byte, 1);
227 	/* we expected 1, so consider 0 an error */
228 	if (ret == 0)
229 		ret = -EIO;
230 	else if (ret == 1)
231 		ret = 0;
232 	return ret;
233 }
234 
235 static int qual_power(struct hfi1_pportdata *ppd)
236 {
237 	u32 cable_power_class = 0, power_class_max = 0;
238 	u8 *cache = ppd->qsfp_info.cache;
239 	int ret = 0;
240 
241 	ret = get_platform_config_field(
242 		ppd->dd, PLATFORM_CONFIG_SYSTEM_TABLE, 0,
243 		SYSTEM_TABLE_QSFP_POWER_CLASS_MAX, &power_class_max, 4);
244 	if (ret)
245 		return ret;
246 
247 	cable_power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]);
248 
249 	if (cable_power_class > power_class_max)
250 		ppd->offline_disabled_reason =
251 			HFI1_ODR_MASK(OPA_LINKDOWN_REASON_POWER_POLICY);
252 
253 	if (ppd->offline_disabled_reason ==
254 			HFI1_ODR_MASK(OPA_LINKDOWN_REASON_POWER_POLICY)) {
255 		dd_dev_err(
256 			ppd->dd,
257 			"%s: Port disabled due to system power restrictions\n",
258 			__func__);
259 		ret = -EPERM;
260 	}
261 	return ret;
262 }
263 
264 static int qual_bitrate(struct hfi1_pportdata *ppd)
265 {
266 	u16 lss = ppd->link_speed_supported, lse = ppd->link_speed_enabled;
267 	u8 *cache = ppd->qsfp_info.cache;
268 
269 	if ((lss & OPA_LINK_SPEED_25G) && (lse & OPA_LINK_SPEED_25G) &&
270 	    cache[QSFP_NOM_BIT_RATE_250_OFFS] < 0x64)
271 		ppd->offline_disabled_reason =
272 			   HFI1_ODR_MASK(OPA_LINKDOWN_REASON_LINKSPEED_POLICY);
273 
274 	if ((lss & OPA_LINK_SPEED_12_5G) && (lse & OPA_LINK_SPEED_12_5G) &&
275 	    cache[QSFP_NOM_BIT_RATE_100_OFFS] < 0x7D)
276 		ppd->offline_disabled_reason =
277 			   HFI1_ODR_MASK(OPA_LINKDOWN_REASON_LINKSPEED_POLICY);
278 
279 	if (ppd->offline_disabled_reason ==
280 			HFI1_ODR_MASK(OPA_LINKDOWN_REASON_LINKSPEED_POLICY)) {
281 		dd_dev_err(
282 			ppd->dd,
283 			"%s: Cable failed bitrate check, disabling port\n",
284 			__func__);
285 		return -EPERM;
286 	}
287 	return 0;
288 }
289 
290 static int set_qsfp_high_power(struct hfi1_pportdata *ppd)
291 {
292 	u8 cable_power_class = 0, power_ctrl_byte = 0;
293 	u8 *cache = ppd->qsfp_info.cache;
294 	int ret;
295 
296 	cable_power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]);
297 
298 	if (cable_power_class > QSFP_POWER_CLASS_1) {
299 		power_ctrl_byte = cache[QSFP_PWR_CTRL_BYTE_OFFS];
300 
301 		power_ctrl_byte |= 1;
302 		power_ctrl_byte &= ~(0x2);
303 
304 		ret = qsfp_write(ppd, ppd->dd->hfi1_id,
305 				 QSFP_PWR_CTRL_BYTE_OFFS,
306 				 &power_ctrl_byte, 1);
307 		if (ret != 1)
308 			return -EIO;
309 
310 		if (cable_power_class > QSFP_POWER_CLASS_4) {
311 			power_ctrl_byte |= (1 << 2);
312 			ret = qsfp_write(ppd, ppd->dd->hfi1_id,
313 					 QSFP_PWR_CTRL_BYTE_OFFS,
314 					 &power_ctrl_byte, 1);
315 			if (ret != 1)
316 				return -EIO;
317 		}
318 
319 		/* SFF 8679 rev 1.7 LPMode Deassert time */
320 		msleep(300);
321 	}
322 	return 0;
323 }
324 
325 static void apply_rx_cdr(struct hfi1_pportdata *ppd,
326 			 u32 rx_preset_index,
327 			 u8 *cdr_ctrl_byte)
328 {
329 	u32 rx_preset;
330 	u8 *cache = ppd->qsfp_info.cache;
331 	int cable_power_class;
332 
333 	if (!((cache[QSFP_MOD_PWR_OFFS] & 0x4) &&
334 	      (cache[QSFP_CDR_INFO_OFFS] & 0x40)))
335 		return;
336 
337 	/* RX CDR present, bypass supported */
338 	cable_power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]);
339 
340 	if (cable_power_class <= QSFP_POWER_CLASS_3) {
341 		/* Power class <= 3, ignore config & turn RX CDR on */
342 		*cdr_ctrl_byte |= 0xF;
343 		return;
344 	}
345 
346 	get_platform_config_field(
347 		ppd->dd, PLATFORM_CONFIG_RX_PRESET_TABLE,
348 		rx_preset_index, RX_PRESET_TABLE_QSFP_RX_CDR_APPLY,
349 		&rx_preset, 4);
350 
351 	if (!rx_preset) {
352 		dd_dev_info(
353 			ppd->dd,
354 			"%s: RX_CDR_APPLY is set to disabled\n",
355 			__func__);
356 		return;
357 	}
358 	get_platform_config_field(
359 		ppd->dd, PLATFORM_CONFIG_RX_PRESET_TABLE,
360 		rx_preset_index, RX_PRESET_TABLE_QSFP_RX_CDR,
361 		&rx_preset, 4);
362 
363 	/* Expand cdr setting to all 4 lanes */
364 	rx_preset = (rx_preset | (rx_preset << 1) |
365 			(rx_preset << 2) | (rx_preset << 3));
366 
367 	if (rx_preset) {
368 		*cdr_ctrl_byte |= rx_preset;
369 	} else {
370 		*cdr_ctrl_byte &= rx_preset;
371 		/* Preserve current TX CDR status */
372 		*cdr_ctrl_byte |= (cache[QSFP_CDR_CTRL_BYTE_OFFS] & 0xF0);
373 	}
374 }
375 
376 static void apply_tx_cdr(struct hfi1_pportdata *ppd,
377 			 u32 tx_preset_index,
378 			 u8 *cdr_ctrl_byte)
379 {
380 	u32 tx_preset;
381 	u8 *cache = ppd->qsfp_info.cache;
382 	int cable_power_class;
383 
384 	if (!((cache[QSFP_MOD_PWR_OFFS] & 0x8) &&
385 	      (cache[QSFP_CDR_INFO_OFFS] & 0x80)))
386 		return;
387 
388 	/* TX CDR present, bypass supported */
389 	cable_power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]);
390 
391 	if (cable_power_class <= QSFP_POWER_CLASS_3) {
392 		/* Power class <= 3, ignore config & turn TX CDR on */
393 		*cdr_ctrl_byte |= 0xF0;
394 		return;
395 	}
396 
397 	get_platform_config_field(
398 		ppd->dd,
399 		PLATFORM_CONFIG_TX_PRESET_TABLE, tx_preset_index,
400 		TX_PRESET_TABLE_QSFP_TX_CDR_APPLY, &tx_preset, 4);
401 
402 	if (!tx_preset) {
403 		dd_dev_info(
404 			ppd->dd,
405 			"%s: TX_CDR_APPLY is set to disabled\n",
406 			__func__);
407 		return;
408 	}
409 	get_platform_config_field(
410 		ppd->dd,
411 		PLATFORM_CONFIG_TX_PRESET_TABLE,
412 		tx_preset_index,
413 		TX_PRESET_TABLE_QSFP_TX_CDR, &tx_preset, 4);
414 
415 	/* Expand cdr setting to all 4 lanes */
416 	tx_preset = (tx_preset | (tx_preset << 1) |
417 			(tx_preset << 2) | (tx_preset << 3));
418 
419 	if (tx_preset)
420 		*cdr_ctrl_byte |= (tx_preset << 4);
421 	else
422 		/* Preserve current/determined RX CDR status */
423 		*cdr_ctrl_byte &= ((tx_preset << 4) | 0xF);
424 }
425 
426 static void apply_cdr_settings(
427 		struct hfi1_pportdata *ppd, u32 rx_preset_index,
428 		u32 tx_preset_index)
429 {
430 	u8 *cache = ppd->qsfp_info.cache;
431 	u8 cdr_ctrl_byte = cache[QSFP_CDR_CTRL_BYTE_OFFS];
432 
433 	apply_rx_cdr(ppd, rx_preset_index, &cdr_ctrl_byte);
434 
435 	apply_tx_cdr(ppd, tx_preset_index, &cdr_ctrl_byte);
436 
437 	qsfp_write(ppd, ppd->dd->hfi1_id, QSFP_CDR_CTRL_BYTE_OFFS,
438 		   &cdr_ctrl_byte, 1);
439 }
440 
441 static void apply_tx_eq_auto(struct hfi1_pportdata *ppd)
442 {
443 	u8 *cache = ppd->qsfp_info.cache;
444 	u8 tx_eq;
445 
446 	if (!(cache[QSFP_EQ_INFO_OFFS] & 0x8))
447 		return;
448 	/* Disable adaptive TX EQ if present */
449 	tx_eq = cache[(128 * 3) + 241];
450 	tx_eq &= 0xF0;
451 	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 241, &tx_eq, 1);
452 }
453 
454 static void apply_tx_eq_prog(struct hfi1_pportdata *ppd, u32 tx_preset_index)
455 {
456 	u8 *cache = ppd->qsfp_info.cache;
457 	u32 tx_preset;
458 	u8 tx_eq;
459 
460 	if (!(cache[QSFP_EQ_INFO_OFFS] & 0x4))
461 		return;
462 
463 	get_platform_config_field(
464 		ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE,
465 		tx_preset_index, TX_PRESET_TABLE_QSFP_TX_EQ_APPLY,
466 		&tx_preset, 4);
467 	if (!tx_preset) {
468 		dd_dev_info(
469 			ppd->dd,
470 			"%s: TX_EQ_APPLY is set to disabled\n",
471 			__func__);
472 		return;
473 	}
474 	get_platform_config_field(
475 			ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE,
476 			tx_preset_index, TX_PRESET_TABLE_QSFP_TX_EQ,
477 			&tx_preset, 4);
478 
479 	if (((cache[(128 * 3) + 224] & 0xF0) >> 4) < tx_preset) {
480 		dd_dev_info(
481 			ppd->dd,
482 			"%s: TX EQ %x unsupported\n",
483 			__func__, tx_preset);
484 
485 		dd_dev_info(
486 			ppd->dd,
487 			"%s: Applying EQ %x\n",
488 			__func__, cache[608] & 0xF0);
489 
490 		tx_preset = (cache[608] & 0xF0) >> 4;
491 	}
492 
493 	tx_eq = tx_preset | (tx_preset << 4);
494 	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 234, &tx_eq, 1);
495 	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 235, &tx_eq, 1);
496 }
497 
498 static void apply_rx_eq_emp(struct hfi1_pportdata *ppd, u32 rx_preset_index)
499 {
500 	u32 rx_preset;
501 	u8 rx_eq, *cache = ppd->qsfp_info.cache;
502 
503 	if (!(cache[QSFP_EQ_INFO_OFFS] & 0x2))
504 		return;
505 	get_platform_config_field(
506 			ppd->dd, PLATFORM_CONFIG_RX_PRESET_TABLE,
507 			rx_preset_index, RX_PRESET_TABLE_QSFP_RX_EMP_APPLY,
508 			&rx_preset, 4);
509 
510 	if (!rx_preset) {
511 		dd_dev_info(
512 			ppd->dd,
513 			"%s: RX_EMP_APPLY is set to disabled\n",
514 			__func__);
515 		return;
516 	}
517 	get_platform_config_field(
518 		ppd->dd, PLATFORM_CONFIG_RX_PRESET_TABLE,
519 		rx_preset_index, RX_PRESET_TABLE_QSFP_RX_EMP,
520 		&rx_preset, 4);
521 
522 	if ((cache[(128 * 3) + 224] & 0xF) < rx_preset) {
523 		dd_dev_info(
524 			ppd->dd,
525 			"%s: Requested RX EMP %x\n",
526 			__func__, rx_preset);
527 
528 		dd_dev_info(
529 			ppd->dd,
530 			"%s: Applying supported EMP %x\n",
531 			__func__, cache[608] & 0xF);
532 
533 		rx_preset = cache[608] & 0xF;
534 	}
535 
536 	rx_eq = rx_preset | (rx_preset << 4);
537 
538 	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 236, &rx_eq, 1);
539 	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 237, &rx_eq, 1);
540 }
541 
542 static void apply_eq_settings(struct hfi1_pportdata *ppd,
543 			      u32 rx_preset_index, u32 tx_preset_index)
544 {
545 	u8 *cache = ppd->qsfp_info.cache;
546 
547 	/* no point going on w/o a page 3 */
548 	if (cache[2] & 4) {
549 		dd_dev_info(ppd->dd,
550 			    "%s: Upper page 03 not present\n",
551 			    __func__);
552 		return;
553 	}
554 
555 	apply_tx_eq_auto(ppd);
556 
557 	apply_tx_eq_prog(ppd, tx_preset_index);
558 
559 	apply_rx_eq_emp(ppd, rx_preset_index);
560 }
561 
562 static void apply_rx_amplitude_settings(
563 		struct hfi1_pportdata *ppd, u32 rx_preset_index,
564 		u32 tx_preset_index)
565 {
566 	u32 rx_preset;
567 	u8 rx_amp = 0, i = 0, preferred = 0, *cache = ppd->qsfp_info.cache;
568 
569 	/* no point going on w/o a page 3 */
570 	if (cache[2] & 4) {
571 		dd_dev_info(ppd->dd,
572 			    "%s: Upper page 03 not present\n",
573 			    __func__);
574 		return;
575 	}
576 	if (!(cache[QSFP_EQ_INFO_OFFS] & 0x1)) {
577 		dd_dev_info(ppd->dd,
578 			    "%s: RX_AMP_APPLY is set to disabled\n",
579 			    __func__);
580 		return;
581 	}
582 
583 	get_platform_config_field(ppd->dd,
584 				  PLATFORM_CONFIG_RX_PRESET_TABLE,
585 				  rx_preset_index,
586 				  RX_PRESET_TABLE_QSFP_RX_AMP_APPLY,
587 				  &rx_preset, 4);
588 
589 	if (!rx_preset) {
590 		dd_dev_info(ppd->dd,
591 			    "%s: RX_AMP_APPLY is set to disabled\n",
592 			    __func__);
593 		return;
594 	}
595 	get_platform_config_field(ppd->dd,
596 				  PLATFORM_CONFIG_RX_PRESET_TABLE,
597 				  rx_preset_index,
598 				  RX_PRESET_TABLE_QSFP_RX_AMP,
599 				  &rx_preset, 4);
600 
601 	dd_dev_info(ppd->dd,
602 		    "%s: Requested RX AMP %x\n",
603 		    __func__,
604 		    rx_preset);
605 
606 	for (i = 0; i < 4; i++) {
607 		if (cache[(128 * 3) + 225] & (1 << i)) {
608 			preferred = i;
609 			if (preferred == rx_preset)
610 				break;
611 		}
612 	}
613 
614 	/*
615 	 * Verify that preferred RX amplitude is not just a
616 	 * fall through of the default
617 	 */
618 	if (!preferred && !(cache[(128 * 3) + 225] & 0x1)) {
619 		dd_dev_info(ppd->dd, "No supported RX AMP, not applying\n");
620 		return;
621 	}
622 
623 	dd_dev_info(ppd->dd,
624 		    "%s: Applying RX AMP %x\n", __func__, preferred);
625 
626 	rx_amp = preferred | (preferred << 4);
627 	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 238, &rx_amp, 1);
628 	qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 239, &rx_amp, 1);
629 }
630 
631 #define OPA_INVALID_INDEX 0xFFF
632 
633 static void apply_tx_lanes(struct hfi1_pportdata *ppd, u8 field_id,
634 			   u32 config_data, const char *message)
635 {
636 	u8 i;
637 	int ret = HCMD_SUCCESS;
638 
639 	for (i = 0; i < 4; i++) {
640 		ret = load_8051_config(ppd->dd, field_id, i, config_data);
641 		if (ret != HCMD_SUCCESS) {
642 			dd_dev_err(
643 				ppd->dd,
644 				"%s: %s for lane %u failed\n",
645 				message, __func__, i);
646 		}
647 	}
648 }
649 
650 /*
651  * Return a special SerDes setting for low power AOC cables.  The power class
652  * threshold and setting being used were all found by empirical testing.
653  *
654  * Summary of the logic:
655  *
656  * if (QSFP and QSFP_TYPE == AOC and QSFP_POWER_CLASS < 4)
657  *     return 0xe
658  * return 0; // leave at default
659  */
660 static u8 aoc_low_power_setting(struct hfi1_pportdata *ppd)
661 {
662 	u8 *cache = ppd->qsfp_info.cache;
663 	int power_class;
664 
665 	/* QSFP only */
666 	if (ppd->port_type != PORT_TYPE_QSFP)
667 		return 0; /* leave at default */
668 
669 	/* active optical cables only */
670 	switch ((cache[QSFP_MOD_TECH_OFFS] & 0xF0) >> 4) {
671 	case 0x0 ... 0x9: /* fallthrough */
672 	case 0xC: /* fallthrough */
673 	case 0xE:
674 		/* active AOC */
675 		power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]);
676 		if (power_class < QSFP_POWER_CLASS_4)
677 			return 0xe;
678 	}
679 	return 0; /* leave at default */
680 }
681 
682 static void apply_tunings(
683 		struct hfi1_pportdata *ppd, u32 tx_preset_index,
684 		u8 tuning_method, u32 total_atten, u8 limiting_active)
685 {
686 	int ret = 0;
687 	u32 config_data = 0, tx_preset = 0;
688 	u8 precur = 0, attn = 0, postcur = 0, external_device_config = 0;
689 	u8 *cache = ppd->qsfp_info.cache;
690 
691 	/* Pass tuning method to 8051 */
692 	read_8051_config(ppd->dd, LINK_TUNING_PARAMETERS, GENERAL_CONFIG,
693 			 &config_data);
694 	config_data &= ~(0xff << TUNING_METHOD_SHIFT);
695 	config_data |= ((u32)tuning_method << TUNING_METHOD_SHIFT);
696 	ret = load_8051_config(ppd->dd, LINK_TUNING_PARAMETERS, GENERAL_CONFIG,
697 			       config_data);
698 	if (ret != HCMD_SUCCESS)
699 		dd_dev_err(ppd->dd, "%s: Failed to set tuning method\n",
700 			   __func__);
701 
702 	/* Set same channel loss for both TX and RX */
703 	config_data = 0 | (total_atten << 16) | (total_atten << 24);
704 	apply_tx_lanes(ppd, CHANNEL_LOSS_SETTINGS, config_data,
705 		       "Setting channel loss");
706 
707 	/* Inform 8051 of cable capabilities */
708 	if (ppd->qsfp_info.cache_valid) {
709 		external_device_config =
710 			((cache[QSFP_MOD_PWR_OFFS] & 0x4) << 3) |
711 			((cache[QSFP_MOD_PWR_OFFS] & 0x8) << 2) |
712 			((cache[QSFP_EQ_INFO_OFFS] & 0x2) << 1) |
713 			(cache[QSFP_EQ_INFO_OFFS] & 0x4);
714 		ret = read_8051_config(ppd->dd, DC_HOST_COMM_SETTINGS,
715 				       GENERAL_CONFIG, &config_data);
716 		/* Clear, then set the external device config field */
717 		config_data &= ~(u32)0xFF;
718 		config_data |= external_device_config;
719 		ret = load_8051_config(ppd->dd, DC_HOST_COMM_SETTINGS,
720 				       GENERAL_CONFIG, config_data);
721 		if (ret != HCMD_SUCCESS)
722 			dd_dev_err(ppd->dd,
723 				   "%s: Failed set ext device config params\n",
724 				   __func__);
725 	}
726 
727 	if (tx_preset_index == OPA_INVALID_INDEX) {
728 		if (ppd->port_type == PORT_TYPE_QSFP && limiting_active)
729 			dd_dev_err(ppd->dd, "%s: Invalid Tx preset index\n",
730 				   __func__);
731 		return;
732 	}
733 
734 	/* Following for limiting active channels only */
735 	get_platform_config_field(
736 		ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE, tx_preset_index,
737 		TX_PRESET_TABLE_PRECUR, &tx_preset, 4);
738 	precur = tx_preset;
739 
740 	get_platform_config_field(
741 		ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE,
742 		tx_preset_index, TX_PRESET_TABLE_ATTN, &tx_preset, 4);
743 	attn = tx_preset;
744 
745 	get_platform_config_field(
746 		ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE,
747 		tx_preset_index, TX_PRESET_TABLE_POSTCUR, &tx_preset, 4);
748 	postcur = tx_preset;
749 
750 	/*
751 	 * NOTES:
752 	 * o The aoc_low_power_setting is applied to all lanes even
753 	 *   though only lane 0's value is examined by the firmware.
754 	 * o A lingering low power setting after a cable swap does
755 	 *   not occur.  On cable unplug the 8051 is reset and
756 	 *   restarted on cable insert.  This resets all settings to
757 	 *   their default, erasing any previous low power setting.
758 	 */
759 	config_data = precur | (attn << 8) | (postcur << 16) |
760 			(aoc_low_power_setting(ppd) << 24);
761 
762 	apply_tx_lanes(ppd, TX_EQ_SETTINGS, config_data,
763 		       "Applying TX settings");
764 }
765 
766 /* Must be holding the QSFP i2c resource */
767 static int tune_active_qsfp(struct hfi1_pportdata *ppd, u32 *ptr_tx_preset,
768 			    u32 *ptr_rx_preset, u32 *ptr_total_atten)
769 {
770 	int ret;
771 	u16 lss = ppd->link_speed_supported, lse = ppd->link_speed_enabled;
772 	u8 *cache = ppd->qsfp_info.cache;
773 
774 	ppd->qsfp_info.limiting_active = 1;
775 
776 	ret = set_qsfp_tx(ppd, 0);
777 	if (ret)
778 		return ret;
779 
780 	ret = qual_power(ppd);
781 	if (ret)
782 		return ret;
783 
784 	ret = qual_bitrate(ppd);
785 	if (ret)
786 		return ret;
787 
788 	/*
789 	 * We'll change the QSFP memory contents from here on out, thus we set a
790 	 * flag here to remind ourselves to reset the QSFP module. This prevents
791 	 * reuse of stale settings established in our previous pass through.
792 	 */
793 	if (ppd->qsfp_info.reset_needed) {
794 		ret = reset_qsfp(ppd);
795 		if (ret)
796 			return ret;
797 		refresh_qsfp_cache(ppd, &ppd->qsfp_info);
798 	} else {
799 		ppd->qsfp_info.reset_needed = 1;
800 	}
801 
802 	ret = set_qsfp_high_power(ppd);
803 	if (ret)
804 		return ret;
805 
806 	if (cache[QSFP_EQ_INFO_OFFS] & 0x4) {
807 		ret = get_platform_config_field(
808 			ppd->dd,
809 			PLATFORM_CONFIG_PORT_TABLE, 0,
810 			PORT_TABLE_TX_PRESET_IDX_ACTIVE_EQ,
811 			ptr_tx_preset, 4);
812 		if (ret) {
813 			*ptr_tx_preset = OPA_INVALID_INDEX;
814 			return ret;
815 		}
816 	} else {
817 		ret = get_platform_config_field(
818 			ppd->dd,
819 			PLATFORM_CONFIG_PORT_TABLE, 0,
820 			PORT_TABLE_TX_PRESET_IDX_ACTIVE_NO_EQ,
821 			ptr_tx_preset, 4);
822 		if (ret) {
823 			*ptr_tx_preset = OPA_INVALID_INDEX;
824 			return ret;
825 		}
826 	}
827 
828 	ret = get_platform_config_field(
829 		ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
830 		PORT_TABLE_RX_PRESET_IDX, ptr_rx_preset, 4);
831 	if (ret) {
832 		*ptr_rx_preset = OPA_INVALID_INDEX;
833 		return ret;
834 	}
835 
836 	if ((lss & OPA_LINK_SPEED_25G) && (lse & OPA_LINK_SPEED_25G))
837 		get_platform_config_field(
838 			ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
839 			PORT_TABLE_LOCAL_ATTEN_25G, ptr_total_atten, 4);
840 	else if ((lss & OPA_LINK_SPEED_12_5G) && (lse & OPA_LINK_SPEED_12_5G))
841 		get_platform_config_field(
842 			ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
843 			PORT_TABLE_LOCAL_ATTEN_12G, ptr_total_atten, 4);
844 
845 	apply_cdr_settings(ppd, *ptr_rx_preset, *ptr_tx_preset);
846 
847 	apply_eq_settings(ppd, *ptr_rx_preset, *ptr_tx_preset);
848 
849 	apply_rx_amplitude_settings(ppd, *ptr_rx_preset, *ptr_tx_preset);
850 
851 	ret = set_qsfp_tx(ppd, 1);
852 
853 	return ret;
854 }
855 
856 static int tune_qsfp(struct hfi1_pportdata *ppd,
857 		     u32 *ptr_tx_preset, u32 *ptr_rx_preset,
858 		     u8 *ptr_tuning_method, u32 *ptr_total_atten)
859 {
860 	u32 cable_atten = 0, remote_atten = 0, platform_atten = 0;
861 	u16 lss = ppd->link_speed_supported, lse = ppd->link_speed_enabled;
862 	int ret = 0;
863 	u8 *cache = ppd->qsfp_info.cache;
864 
865 	switch ((cache[QSFP_MOD_TECH_OFFS] & 0xF0) >> 4) {
866 	case 0xA ... 0xB:
867 		ret = get_platform_config_field(
868 			ppd->dd,
869 			PLATFORM_CONFIG_PORT_TABLE, 0,
870 			PORT_TABLE_LOCAL_ATTEN_25G,
871 			&platform_atten, 4);
872 		if (ret)
873 			return ret;
874 
875 		if ((lss & OPA_LINK_SPEED_25G) && (lse & OPA_LINK_SPEED_25G))
876 			cable_atten = cache[QSFP_CU_ATTEN_12G_OFFS];
877 		else if ((lss & OPA_LINK_SPEED_12_5G) &&
878 			 (lse & OPA_LINK_SPEED_12_5G))
879 			cable_atten = cache[QSFP_CU_ATTEN_7G_OFFS];
880 
881 		/* Fallback to configured attenuation if cable memory is bad */
882 		if (cable_atten == 0 || cable_atten > 36) {
883 			ret = get_platform_config_field(
884 				ppd->dd,
885 				PLATFORM_CONFIG_SYSTEM_TABLE, 0,
886 				SYSTEM_TABLE_QSFP_ATTENUATION_DEFAULT_25G,
887 				&cable_atten, 4);
888 			if (ret)
889 				return ret;
890 		}
891 
892 		ret = get_platform_config_field(
893 			ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
894 			PORT_TABLE_REMOTE_ATTEN_25G, &remote_atten, 4);
895 		if (ret)
896 			return ret;
897 
898 		*ptr_total_atten = platform_atten + cable_atten + remote_atten;
899 
900 		*ptr_tuning_method = OPA_PASSIVE_TUNING;
901 		break;
902 	case 0x0 ... 0x9: /* fallthrough */
903 	case 0xC: /* fallthrough */
904 	case 0xE:
905 		ret = tune_active_qsfp(ppd, ptr_tx_preset, ptr_rx_preset,
906 				       ptr_total_atten);
907 		if (ret)
908 			return ret;
909 
910 		*ptr_tuning_method = OPA_ACTIVE_TUNING;
911 		break;
912 	case 0xD: /* fallthrough */
913 	case 0xF:
914 	default:
915 		dd_dev_warn(ppd->dd, "%s: Unknown/unsupported cable\n",
916 			    __func__);
917 		break;
918 	}
919 	return ret;
920 }
921 
922 /*
923  * This function communicates its success or failure via ppd->driver_link_ready
924  * Thus, it depends on its association with start_link(...) which checks
925  * driver_link_ready before proceeding with the link negotiation and
926  * initialization process.
927  */
928 void tune_serdes(struct hfi1_pportdata *ppd)
929 {
930 	int ret = 0;
931 	u32 total_atten = 0;
932 	u32 remote_atten = 0, platform_atten = 0;
933 	u32 rx_preset_index, tx_preset_index;
934 	u8 tuning_method = 0, limiting_active = 0;
935 	struct hfi1_devdata *dd = ppd->dd;
936 
937 	rx_preset_index = OPA_INVALID_INDEX;
938 	tx_preset_index = OPA_INVALID_INDEX;
939 
940 	/* the link defaults to enabled */
941 	ppd->link_enabled = 1;
942 	/* the driver link ready state defaults to not ready */
943 	ppd->driver_link_ready = 0;
944 	ppd->offline_disabled_reason = HFI1_ODR_MASK(OPA_LINKDOWN_REASON_NONE);
945 
946 	/* Skip the tuning for testing (loopback != none) and simulations */
947 	if (loopback != LOOPBACK_NONE ||
948 	    ppd->dd->icode == ICODE_FUNCTIONAL_SIMULATOR) {
949 		ppd->driver_link_ready = 1;
950 
951 		if (qsfp_mod_present(ppd)) {
952 			ret = acquire_chip_resource(ppd->dd,
953 						    qsfp_resource(ppd->dd),
954 						    QSFP_WAIT);
955 			if (ret) {
956 				dd_dev_err(ppd->dd, "%s: hfi%d: cannot lock i2c chain\n",
957 					   __func__, (int)ppd->dd->hfi1_id);
958 				goto bail;
959 			}
960 
961 			refresh_qsfp_cache(ppd, &ppd->qsfp_info);
962 			release_chip_resource(ppd->dd, qsfp_resource(ppd->dd));
963 		}
964 
965 		return;
966 	}
967 
968 	switch (ppd->port_type) {
969 	case PORT_TYPE_DISCONNECTED:
970 		ppd->offline_disabled_reason =
971 			HFI1_ODR_MASK(OPA_LINKDOWN_REASON_DISCONNECTED);
972 		dd_dev_warn(dd, "%s: Port disconnected, disabling port\n",
973 			    __func__);
974 		goto bail;
975 	case PORT_TYPE_FIXED:
976 		/* platform_atten, remote_atten pre-zeroed to catch error */
977 		get_platform_config_field(
978 			ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
979 			PORT_TABLE_LOCAL_ATTEN_25G, &platform_atten, 4);
980 
981 		get_platform_config_field(
982 			ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
983 			PORT_TABLE_REMOTE_ATTEN_25G, &remote_atten, 4);
984 
985 		total_atten = platform_atten + remote_atten;
986 
987 		tuning_method = OPA_PASSIVE_TUNING;
988 		break;
989 	case PORT_TYPE_VARIABLE:
990 		if (qsfp_mod_present(ppd)) {
991 			/*
992 			 * platform_atten, remote_atten pre-zeroed to
993 			 * catch error
994 			 */
995 			get_platform_config_field(
996 				ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
997 				PORT_TABLE_LOCAL_ATTEN_25G,
998 				&platform_atten, 4);
999 
1000 			get_platform_config_field(
1001 				ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0,
1002 				PORT_TABLE_REMOTE_ATTEN_25G,
1003 				&remote_atten, 4);
1004 
1005 			total_atten = platform_atten + remote_atten;
1006 
1007 			tuning_method = OPA_PASSIVE_TUNING;
1008 		} else {
1009 			ppd->offline_disabled_reason =
1010 			     HFI1_ODR_MASK(OPA_LINKDOWN_REASON_CHASSIS_CONFIG);
1011 			goto bail;
1012 		}
1013 		break;
1014 	case PORT_TYPE_QSFP:
1015 		if (qsfp_mod_present(ppd)) {
1016 			ret = acquire_chip_resource(ppd->dd,
1017 						    qsfp_resource(ppd->dd),
1018 						    QSFP_WAIT);
1019 			if (ret) {
1020 				dd_dev_err(ppd->dd, "%s: hfi%d: cannot lock i2c chain\n",
1021 					   __func__, (int)ppd->dd->hfi1_id);
1022 				goto bail;
1023 			}
1024 			refresh_qsfp_cache(ppd, &ppd->qsfp_info);
1025 
1026 			if (ppd->qsfp_info.cache_valid) {
1027 				ret = tune_qsfp(ppd,
1028 						&tx_preset_index,
1029 						&rx_preset_index,
1030 						&tuning_method,
1031 						&total_atten);
1032 
1033 				/*
1034 				 * We may have modified the QSFP memory, so
1035 				 * update the cache to reflect the changes
1036 				 */
1037 				refresh_qsfp_cache(ppd, &ppd->qsfp_info);
1038 				limiting_active =
1039 						ppd->qsfp_info.limiting_active;
1040 			} else {
1041 				dd_dev_err(dd,
1042 					   "%s: Reading QSFP memory failed\n",
1043 					   __func__);
1044 				ret = -EINVAL; /* a fail indication */
1045 			}
1046 			release_chip_resource(ppd->dd, qsfp_resource(ppd->dd));
1047 			if (ret)
1048 				goto bail;
1049 		} else {
1050 			ppd->offline_disabled_reason =
1051 			   HFI1_ODR_MASK(
1052 				OPA_LINKDOWN_REASON_LOCAL_MEDIA_NOT_INSTALLED);
1053 			goto bail;
1054 		}
1055 		break;
1056 	default:
1057 		dd_dev_warn(ppd->dd, "%s: Unknown port type\n", __func__);
1058 		ppd->port_type = PORT_TYPE_UNKNOWN;
1059 		tuning_method = OPA_UNKNOWN_TUNING;
1060 		total_atten = 0;
1061 		limiting_active = 0;
1062 		tx_preset_index = OPA_INVALID_INDEX;
1063 		break;
1064 	}
1065 
1066 	if (ppd->offline_disabled_reason ==
1067 			HFI1_ODR_MASK(OPA_LINKDOWN_REASON_NONE))
1068 		apply_tunings(ppd, tx_preset_index, tuning_method,
1069 			      total_atten, limiting_active);
1070 
1071 	if (!ret)
1072 		ppd->driver_link_ready = 1;
1073 
1074 	return;
1075 bail:
1076 	ppd->driver_link_ready = 0;
1077 }
1078