xref: /linux/drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c (revision 34f7c6e7d4396090692a09789db231e12cb4762b)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  *  Copyright (C) 2013-2015 Chelsio Communications.  All rights reserved.
4  */
5 
6 #include <linux/firmware.h>
7 #include <linux/mdio.h>
8 
9 #include "cxgb4.h"
10 #include "t4_regs.h"
11 #include "t4fw_api.h"
12 #include "cxgb4_cudbg.h"
13 #include "cxgb4_filter.h"
14 #include "cxgb4_tc_flower.h"
15 
16 #define EEPROM_MAGIC 0x38E2F10C
17 
18 static u32 get_msglevel(struct net_device *dev)
19 {
20 	return netdev2adap(dev)->msg_enable;
21 }
22 
23 static void set_msglevel(struct net_device *dev, u32 val)
24 {
25 	netdev2adap(dev)->msg_enable = val;
26 }
27 
28 enum cxgb4_ethtool_tests {
29 	CXGB4_ETHTOOL_LB_TEST,
30 	CXGB4_ETHTOOL_MAX_TEST,
31 };
32 
33 static const char cxgb4_selftest_strings[CXGB4_ETHTOOL_MAX_TEST][ETH_GSTRING_LEN] = {
34 	"Loop back test (offline)",
35 };
36 
37 static const char * const flash_region_strings[] = {
38 	"All",
39 	"Firmware",
40 	"PHY Firmware",
41 	"Boot",
42 	"Boot CFG",
43 };
44 
45 static const char stats_strings[][ETH_GSTRING_LEN] = {
46 	"tx_octets_ok           ",
47 	"tx_frames_ok           ",
48 	"tx_broadcast_frames    ",
49 	"tx_multicast_frames    ",
50 	"tx_unicast_frames      ",
51 	"tx_error_frames        ",
52 
53 	"tx_frames_64           ",
54 	"tx_frames_65_to_127    ",
55 	"tx_frames_128_to_255   ",
56 	"tx_frames_256_to_511   ",
57 	"tx_frames_512_to_1023  ",
58 	"tx_frames_1024_to_1518 ",
59 	"tx_frames_1519_to_max  ",
60 
61 	"tx_frames_dropped      ",
62 	"tx_pause_frames        ",
63 	"tx_ppp0_frames         ",
64 	"tx_ppp1_frames         ",
65 	"tx_ppp2_frames         ",
66 	"tx_ppp3_frames         ",
67 	"tx_ppp4_frames         ",
68 	"tx_ppp5_frames         ",
69 	"tx_ppp6_frames         ",
70 	"tx_ppp7_frames         ",
71 
72 	"rx_octets_ok           ",
73 	"rx_frames_ok           ",
74 	"rx_broadcast_frames    ",
75 	"rx_multicast_frames    ",
76 	"rx_unicast_frames      ",
77 
78 	"rx_frames_too_long     ",
79 	"rx_jabber_errors       ",
80 	"rx_fcs_errors          ",
81 	"rx_length_errors       ",
82 	"rx_symbol_errors       ",
83 	"rx_runt_frames         ",
84 
85 	"rx_frames_64           ",
86 	"rx_frames_65_to_127    ",
87 	"rx_frames_128_to_255   ",
88 	"rx_frames_256_to_511   ",
89 	"rx_frames_512_to_1023  ",
90 	"rx_frames_1024_to_1518 ",
91 	"rx_frames_1519_to_max  ",
92 
93 	"rx_pause_frames        ",
94 	"rx_ppp0_frames         ",
95 	"rx_ppp1_frames         ",
96 	"rx_ppp2_frames         ",
97 	"rx_ppp3_frames         ",
98 	"rx_ppp4_frames         ",
99 	"rx_ppp5_frames         ",
100 	"rx_ppp6_frames         ",
101 	"rx_ppp7_frames         ",
102 
103 	"rx_bg0_frames_dropped  ",
104 	"rx_bg1_frames_dropped  ",
105 	"rx_bg2_frames_dropped  ",
106 	"rx_bg3_frames_dropped  ",
107 	"rx_bg0_frames_trunc    ",
108 	"rx_bg1_frames_trunc    ",
109 	"rx_bg2_frames_trunc    ",
110 	"rx_bg3_frames_trunc    ",
111 
112 	"tso                    ",
113 	"uso                    ",
114 	"tx_csum_offload        ",
115 	"rx_csum_good           ",
116 	"vlan_extractions       ",
117 	"vlan_insertions        ",
118 	"gro_packets            ",
119 	"gro_merged             ",
120 #if  IS_ENABLED(CONFIG_CHELSIO_TLS_DEVICE)
121 	"tx_tls_encrypted_packets",
122 	"tx_tls_encrypted_bytes  ",
123 	"tx_tls_ctx              ",
124 	"tx_tls_ooo              ",
125 	"tx_tls_skip_no_sync_data",
126 	"tx_tls_drop_no_sync_data",
127 	"tx_tls_drop_bypass_req  ",
128 #endif
129 };
130 
131 static char adapter_stats_strings[][ETH_GSTRING_LEN] = {
132 	"db_drop                ",
133 	"db_full                ",
134 	"db_empty               ",
135 	"write_coal_success     ",
136 	"write_coal_fail        ",
137 };
138 
139 static char loopback_stats_strings[][ETH_GSTRING_LEN] = {
140 	"-------Loopback----------- ",
141 	"octets_ok              ",
142 	"frames_ok              ",
143 	"bcast_frames           ",
144 	"mcast_frames           ",
145 	"ucast_frames           ",
146 	"error_frames           ",
147 	"frames_64              ",
148 	"frames_65_to_127       ",
149 	"frames_128_to_255      ",
150 	"frames_256_to_511      ",
151 	"frames_512_to_1023     ",
152 	"frames_1024_to_1518    ",
153 	"frames_1519_to_max     ",
154 	"frames_dropped         ",
155 	"bg0_frames_dropped     ",
156 	"bg1_frames_dropped     ",
157 	"bg2_frames_dropped     ",
158 	"bg3_frames_dropped     ",
159 	"bg0_frames_trunc       ",
160 	"bg1_frames_trunc       ",
161 	"bg2_frames_trunc       ",
162 	"bg3_frames_trunc       ",
163 };
164 
165 static const char cxgb4_priv_flags_strings[][ETH_GSTRING_LEN] = {
166 	[PRIV_FLAG_PORT_TX_VM_BIT] = "port_tx_vm_wr",
167 };
168 
169 static int get_sset_count(struct net_device *dev, int sset)
170 {
171 	switch (sset) {
172 	case ETH_SS_STATS:
173 		return ARRAY_SIZE(stats_strings) +
174 		       ARRAY_SIZE(adapter_stats_strings) +
175 		       ARRAY_SIZE(loopback_stats_strings);
176 	case ETH_SS_PRIV_FLAGS:
177 		return ARRAY_SIZE(cxgb4_priv_flags_strings);
178 	case ETH_SS_TEST:
179 		return ARRAY_SIZE(cxgb4_selftest_strings);
180 	default:
181 		return -EOPNOTSUPP;
182 	}
183 }
184 
185 static int get_regs_len(struct net_device *dev)
186 {
187 	struct adapter *adap = netdev2adap(dev);
188 
189 	return t4_get_regs_len(adap);
190 }
191 
192 static int get_eeprom_len(struct net_device *dev)
193 {
194 	return EEPROMSIZE;
195 }
196 
197 static void get_drvinfo(struct net_device *dev, struct ethtool_drvinfo *info)
198 {
199 	struct adapter *adapter = netdev2adap(dev);
200 	u32 exprom_vers;
201 
202 	strlcpy(info->driver, cxgb4_driver_name, sizeof(info->driver));
203 	strlcpy(info->bus_info, pci_name(adapter->pdev),
204 		sizeof(info->bus_info));
205 	info->regdump_len = get_regs_len(dev);
206 
207 	if (adapter->params.fw_vers)
208 		snprintf(info->fw_version, sizeof(info->fw_version),
209 			 "%u.%u.%u.%u, TP %u.%u.%u.%u",
210 			 FW_HDR_FW_VER_MAJOR_G(adapter->params.fw_vers),
211 			 FW_HDR_FW_VER_MINOR_G(adapter->params.fw_vers),
212 			 FW_HDR_FW_VER_MICRO_G(adapter->params.fw_vers),
213 			 FW_HDR_FW_VER_BUILD_G(adapter->params.fw_vers),
214 			 FW_HDR_FW_VER_MAJOR_G(adapter->params.tp_vers),
215 			 FW_HDR_FW_VER_MINOR_G(adapter->params.tp_vers),
216 			 FW_HDR_FW_VER_MICRO_G(adapter->params.tp_vers),
217 			 FW_HDR_FW_VER_BUILD_G(adapter->params.tp_vers));
218 
219 	if (!t4_get_exprom_version(adapter, &exprom_vers))
220 		snprintf(info->erom_version, sizeof(info->erom_version),
221 			 "%u.%u.%u.%u",
222 			 FW_HDR_FW_VER_MAJOR_G(exprom_vers),
223 			 FW_HDR_FW_VER_MINOR_G(exprom_vers),
224 			 FW_HDR_FW_VER_MICRO_G(exprom_vers),
225 			 FW_HDR_FW_VER_BUILD_G(exprom_vers));
226 	info->n_priv_flags = ARRAY_SIZE(cxgb4_priv_flags_strings);
227 }
228 
229 static void get_strings(struct net_device *dev, u32 stringset, u8 *data)
230 {
231 	if (stringset == ETH_SS_STATS) {
232 		memcpy(data, stats_strings, sizeof(stats_strings));
233 		data += sizeof(stats_strings);
234 		memcpy(data, adapter_stats_strings,
235 		       sizeof(adapter_stats_strings));
236 		data += sizeof(adapter_stats_strings);
237 		memcpy(data, loopback_stats_strings,
238 		       sizeof(loopback_stats_strings));
239 	} else if (stringset == ETH_SS_PRIV_FLAGS) {
240 		memcpy(data, cxgb4_priv_flags_strings,
241 		       sizeof(cxgb4_priv_flags_strings));
242 	} else if (stringset == ETH_SS_TEST) {
243 		memcpy(data, cxgb4_selftest_strings,
244 		       sizeof(cxgb4_selftest_strings));
245 	}
246 }
247 
248 /* port stats maintained per queue of the port. They should be in the same
249  * order as in stats_strings above.
250  */
251 struct queue_port_stats {
252 	u64 tso;
253 	u64 uso;
254 	u64 tx_csum;
255 	u64 rx_csum;
256 	u64 vlan_ex;
257 	u64 vlan_ins;
258 	u64 gro_pkts;
259 	u64 gro_merged;
260 #if IS_ENABLED(CONFIG_CHELSIO_TLS_DEVICE)
261 	u64 tx_tls_encrypted_packets;
262 	u64 tx_tls_encrypted_bytes;
263 	u64 tx_tls_ctx;
264 	u64 tx_tls_ooo;
265 	u64 tx_tls_skip_no_sync_data;
266 	u64 tx_tls_drop_no_sync_data;
267 	u64 tx_tls_drop_bypass_req;
268 #endif
269 };
270 
271 struct adapter_stats {
272 	u64 db_drop;
273 	u64 db_full;
274 	u64 db_empty;
275 	u64 wc_success;
276 	u64 wc_fail;
277 };
278 
279 static void collect_sge_port_stats(const struct adapter *adap,
280 				   const struct port_info *p,
281 				   struct queue_port_stats *s)
282 {
283 	const struct sge_eth_txq *tx = &adap->sge.ethtxq[p->first_qset];
284 	const struct sge_eth_rxq *rx = &adap->sge.ethrxq[p->first_qset];
285 #if IS_ENABLED(CONFIG_CHELSIO_TLS_DEVICE)
286 	const struct ch_ktls_port_stats_debug *ktls_stats;
287 #endif
288 	struct sge_eohw_txq *eohw_tx;
289 	unsigned int i;
290 
291 	memset(s, 0, sizeof(*s));
292 	for (i = 0; i < p->nqsets; i++, rx++, tx++) {
293 		s->tso += tx->tso;
294 		s->uso += tx->uso;
295 		s->tx_csum += tx->tx_cso;
296 		s->rx_csum += rx->stats.rx_cso;
297 		s->vlan_ex += rx->stats.vlan_ex;
298 		s->vlan_ins += tx->vlan_ins;
299 		s->gro_pkts += rx->stats.lro_pkts;
300 		s->gro_merged += rx->stats.lro_merged;
301 	}
302 
303 	if (adap->sge.eohw_txq) {
304 		eohw_tx = &adap->sge.eohw_txq[p->first_qset];
305 		for (i = 0; i < p->nqsets; i++, eohw_tx++) {
306 			s->tso += eohw_tx->tso;
307 			s->uso += eohw_tx->uso;
308 			s->tx_csum += eohw_tx->tx_cso;
309 			s->vlan_ins += eohw_tx->vlan_ins;
310 		}
311 	}
312 #if IS_ENABLED(CONFIG_CHELSIO_TLS_DEVICE)
313 	ktls_stats = &adap->ch_ktls_stats.ktls_port[p->port_id];
314 	s->tx_tls_encrypted_packets =
315 		atomic64_read(&ktls_stats->ktls_tx_encrypted_packets);
316 	s->tx_tls_encrypted_bytes =
317 		atomic64_read(&ktls_stats->ktls_tx_encrypted_bytes);
318 	s->tx_tls_ctx = atomic64_read(&ktls_stats->ktls_tx_ctx);
319 	s->tx_tls_ooo = atomic64_read(&ktls_stats->ktls_tx_ooo);
320 	s->tx_tls_skip_no_sync_data =
321 		atomic64_read(&ktls_stats->ktls_tx_skip_no_sync_data);
322 	s->tx_tls_drop_no_sync_data =
323 		atomic64_read(&ktls_stats->ktls_tx_drop_no_sync_data);
324 	s->tx_tls_drop_bypass_req =
325 		atomic64_read(&ktls_stats->ktls_tx_drop_bypass_req);
326 #endif
327 }
328 
329 static void collect_adapter_stats(struct adapter *adap, struct adapter_stats *s)
330 {
331 	u64 val1, val2;
332 
333 	memset(s, 0, sizeof(*s));
334 
335 	s->db_drop = adap->db_stats.db_drop;
336 	s->db_full = adap->db_stats.db_full;
337 	s->db_empty = adap->db_stats.db_empty;
338 
339 	if (!is_t4(adap->params.chip)) {
340 		int v;
341 
342 		v = t4_read_reg(adap, SGE_STAT_CFG_A);
343 		if (STATSOURCE_T5_G(v) == 7) {
344 			val2 = t4_read_reg(adap, SGE_STAT_MATCH_A);
345 			val1 = t4_read_reg(adap, SGE_STAT_TOTAL_A);
346 			s->wc_success = val1 - val2;
347 			s->wc_fail = val2;
348 		}
349 	}
350 }
351 
352 static void get_stats(struct net_device *dev, struct ethtool_stats *stats,
353 		      u64 *data)
354 {
355 	struct port_info *pi = netdev_priv(dev);
356 	struct adapter *adapter = pi->adapter;
357 	struct lb_port_stats s;
358 	int i;
359 	u64 *p0;
360 
361 	t4_get_port_stats_offset(adapter, pi->tx_chan,
362 				 (struct port_stats *)data,
363 				 &pi->stats_base);
364 
365 	data += sizeof(struct port_stats) / sizeof(u64);
366 	collect_sge_port_stats(adapter, pi, (struct queue_port_stats *)data);
367 	data += sizeof(struct queue_port_stats) / sizeof(u64);
368 	collect_adapter_stats(adapter, (struct adapter_stats *)data);
369 	data += sizeof(struct adapter_stats) / sizeof(u64);
370 
371 	*data++ = (u64)pi->port_id;
372 	memset(&s, 0, sizeof(s));
373 	t4_get_lb_stats(adapter, pi->port_id, &s);
374 
375 	p0 = &s.octets;
376 	for (i = 0; i < ARRAY_SIZE(loopback_stats_strings) - 1; i++)
377 		*data++ = (unsigned long long)*p0++;
378 }
379 
380 static void get_regs(struct net_device *dev, struct ethtool_regs *regs,
381 		     void *buf)
382 {
383 	struct adapter *adap = netdev2adap(dev);
384 	size_t buf_size;
385 
386 	buf_size = t4_get_regs_len(adap);
387 	regs->version = mk_adap_vers(adap);
388 	t4_get_regs(adap, buf, buf_size);
389 }
390 
391 static int restart_autoneg(struct net_device *dev)
392 {
393 	struct port_info *p = netdev_priv(dev);
394 
395 	if (!netif_running(dev))
396 		return -EAGAIN;
397 	if (p->link_cfg.autoneg != AUTONEG_ENABLE)
398 		return -EINVAL;
399 	t4_restart_aneg(p->adapter, p->adapter->pf, p->tx_chan);
400 	return 0;
401 }
402 
403 static int identify_port(struct net_device *dev,
404 			 enum ethtool_phys_id_state state)
405 {
406 	unsigned int val;
407 	struct adapter *adap = netdev2adap(dev);
408 
409 	if (state == ETHTOOL_ID_ACTIVE)
410 		val = 0xffff;
411 	else if (state == ETHTOOL_ID_INACTIVE)
412 		val = 0;
413 	else
414 		return -EINVAL;
415 
416 	return t4_identify_port(adap, adap->pf, netdev2pinfo(dev)->viid, val);
417 }
418 
419 /**
420  *	from_fw_port_mod_type - translate Firmware Port/Module type to Ethtool
421  *	@port_type: Firmware Port Type
422  *	@mod_type: Firmware Module Type
423  *
424  *	Translate Firmware Port/Module type to Ethtool Port Type.
425  */
426 static int from_fw_port_mod_type(enum fw_port_type port_type,
427 				 enum fw_port_module_type mod_type)
428 {
429 	if (port_type == FW_PORT_TYPE_BT_SGMII ||
430 	    port_type == FW_PORT_TYPE_BT_XFI ||
431 	    port_type == FW_PORT_TYPE_BT_XAUI) {
432 		return PORT_TP;
433 	} else if (port_type == FW_PORT_TYPE_FIBER_XFI ||
434 		   port_type == FW_PORT_TYPE_FIBER_XAUI) {
435 		return PORT_FIBRE;
436 	} else if (port_type == FW_PORT_TYPE_SFP ||
437 		   port_type == FW_PORT_TYPE_QSFP_10G ||
438 		   port_type == FW_PORT_TYPE_QSA ||
439 		   port_type == FW_PORT_TYPE_QSFP ||
440 		   port_type == FW_PORT_TYPE_CR4_QSFP ||
441 		   port_type == FW_PORT_TYPE_CR_QSFP ||
442 		   port_type == FW_PORT_TYPE_CR2_QSFP ||
443 		   port_type == FW_PORT_TYPE_SFP28) {
444 		if (mod_type == FW_PORT_MOD_TYPE_LR ||
445 		    mod_type == FW_PORT_MOD_TYPE_SR ||
446 		    mod_type == FW_PORT_MOD_TYPE_ER ||
447 		    mod_type == FW_PORT_MOD_TYPE_LRM)
448 			return PORT_FIBRE;
449 		else if (mod_type == FW_PORT_MOD_TYPE_TWINAX_PASSIVE ||
450 			 mod_type == FW_PORT_MOD_TYPE_TWINAX_ACTIVE)
451 			return PORT_DA;
452 		else
453 			return PORT_OTHER;
454 	} else if (port_type == FW_PORT_TYPE_KR4_100G ||
455 		   port_type == FW_PORT_TYPE_KR_SFP28 ||
456 		   port_type == FW_PORT_TYPE_KR_XLAUI) {
457 		return PORT_NONE;
458 	}
459 
460 	return PORT_OTHER;
461 }
462 
463 /**
464  *	speed_to_fw_caps - translate Port Speed to Firmware Port Capabilities
465  *	@speed: speed in Kb/s
466  *
467  *	Translates a specific Port Speed into a Firmware Port Capabilities
468  *	value.
469  */
470 static unsigned int speed_to_fw_caps(int speed)
471 {
472 	if (speed == 100)
473 		return FW_PORT_CAP32_SPEED_100M;
474 	if (speed == 1000)
475 		return FW_PORT_CAP32_SPEED_1G;
476 	if (speed == 10000)
477 		return FW_PORT_CAP32_SPEED_10G;
478 	if (speed == 25000)
479 		return FW_PORT_CAP32_SPEED_25G;
480 	if (speed == 40000)
481 		return FW_PORT_CAP32_SPEED_40G;
482 	if (speed == 50000)
483 		return FW_PORT_CAP32_SPEED_50G;
484 	if (speed == 100000)
485 		return FW_PORT_CAP32_SPEED_100G;
486 	if (speed == 200000)
487 		return FW_PORT_CAP32_SPEED_200G;
488 	if (speed == 400000)
489 		return FW_PORT_CAP32_SPEED_400G;
490 	return 0;
491 }
492 
493 /**
494  *	fw_caps_to_lmm - translate Firmware to ethtool Link Mode Mask
495  *	@port_type: Firmware Port Type
496  *	@fw_caps: Firmware Port Capabilities
497  *	@link_mode_mask: ethtool Link Mode Mask
498  *
499  *	Translate a Firmware Port Capabilities specification to an ethtool
500  *	Link Mode Mask.
501  */
502 static void fw_caps_to_lmm(enum fw_port_type port_type,
503 			   fw_port_cap32_t fw_caps,
504 			   unsigned long *link_mode_mask)
505 {
506 	#define SET_LMM(__lmm_name) \
507 		do { \
508 			__set_bit(ETHTOOL_LINK_MODE_ ## __lmm_name ## _BIT, \
509 				  link_mode_mask); \
510 		} while (0)
511 
512 	#define FW_CAPS_TO_LMM(__fw_name, __lmm_name) \
513 		do { \
514 			if (fw_caps & FW_PORT_CAP32_ ## __fw_name) \
515 				SET_LMM(__lmm_name); \
516 		} while (0)
517 
518 	switch (port_type) {
519 	case FW_PORT_TYPE_BT_SGMII:
520 	case FW_PORT_TYPE_BT_XFI:
521 	case FW_PORT_TYPE_BT_XAUI:
522 		SET_LMM(TP);
523 		FW_CAPS_TO_LMM(SPEED_100M, 100baseT_Full);
524 		FW_CAPS_TO_LMM(SPEED_1G, 1000baseT_Full);
525 		FW_CAPS_TO_LMM(SPEED_10G, 10000baseT_Full);
526 		break;
527 
528 	case FW_PORT_TYPE_KX4:
529 	case FW_PORT_TYPE_KX:
530 		SET_LMM(Backplane);
531 		FW_CAPS_TO_LMM(SPEED_1G, 1000baseKX_Full);
532 		FW_CAPS_TO_LMM(SPEED_10G, 10000baseKX4_Full);
533 		break;
534 
535 	case FW_PORT_TYPE_KR:
536 		SET_LMM(Backplane);
537 		FW_CAPS_TO_LMM(SPEED_10G, 10000baseKR_Full);
538 		break;
539 
540 	case FW_PORT_TYPE_BP_AP:
541 		SET_LMM(Backplane);
542 		FW_CAPS_TO_LMM(SPEED_1G, 1000baseKX_Full);
543 		FW_CAPS_TO_LMM(SPEED_10G, 10000baseR_FEC);
544 		FW_CAPS_TO_LMM(SPEED_10G, 10000baseKR_Full);
545 		break;
546 
547 	case FW_PORT_TYPE_BP4_AP:
548 		SET_LMM(Backplane);
549 		FW_CAPS_TO_LMM(SPEED_1G, 1000baseKX_Full);
550 		FW_CAPS_TO_LMM(SPEED_10G, 10000baseR_FEC);
551 		FW_CAPS_TO_LMM(SPEED_10G, 10000baseKR_Full);
552 		FW_CAPS_TO_LMM(SPEED_10G, 10000baseKX4_Full);
553 		break;
554 
555 	case FW_PORT_TYPE_FIBER_XFI:
556 	case FW_PORT_TYPE_FIBER_XAUI:
557 	case FW_PORT_TYPE_SFP:
558 	case FW_PORT_TYPE_QSFP_10G:
559 	case FW_PORT_TYPE_QSA:
560 		SET_LMM(FIBRE);
561 		FW_CAPS_TO_LMM(SPEED_1G, 1000baseT_Full);
562 		FW_CAPS_TO_LMM(SPEED_10G, 10000baseT_Full);
563 		break;
564 
565 	case FW_PORT_TYPE_BP40_BA:
566 	case FW_PORT_TYPE_QSFP:
567 		SET_LMM(FIBRE);
568 		FW_CAPS_TO_LMM(SPEED_1G, 1000baseT_Full);
569 		FW_CAPS_TO_LMM(SPEED_10G, 10000baseT_Full);
570 		FW_CAPS_TO_LMM(SPEED_40G, 40000baseSR4_Full);
571 		break;
572 
573 	case FW_PORT_TYPE_CR_QSFP:
574 	case FW_PORT_TYPE_SFP28:
575 		SET_LMM(FIBRE);
576 		FW_CAPS_TO_LMM(SPEED_1G, 1000baseT_Full);
577 		FW_CAPS_TO_LMM(SPEED_10G, 10000baseT_Full);
578 		FW_CAPS_TO_LMM(SPEED_25G, 25000baseCR_Full);
579 		break;
580 
581 	case FW_PORT_TYPE_KR_SFP28:
582 		SET_LMM(Backplane);
583 		FW_CAPS_TO_LMM(SPEED_1G, 1000baseT_Full);
584 		FW_CAPS_TO_LMM(SPEED_10G, 10000baseKR_Full);
585 		FW_CAPS_TO_LMM(SPEED_25G, 25000baseKR_Full);
586 		break;
587 
588 	case FW_PORT_TYPE_KR_XLAUI:
589 		SET_LMM(Backplane);
590 		FW_CAPS_TO_LMM(SPEED_1G, 1000baseKX_Full);
591 		FW_CAPS_TO_LMM(SPEED_10G, 10000baseKR_Full);
592 		FW_CAPS_TO_LMM(SPEED_40G, 40000baseKR4_Full);
593 		break;
594 
595 	case FW_PORT_TYPE_CR2_QSFP:
596 		SET_LMM(FIBRE);
597 		FW_CAPS_TO_LMM(SPEED_50G, 50000baseSR2_Full);
598 		break;
599 
600 	case FW_PORT_TYPE_KR4_100G:
601 	case FW_PORT_TYPE_CR4_QSFP:
602 		SET_LMM(FIBRE);
603 		FW_CAPS_TO_LMM(SPEED_1G,  1000baseT_Full);
604 		FW_CAPS_TO_LMM(SPEED_10G, 10000baseKR_Full);
605 		FW_CAPS_TO_LMM(SPEED_40G, 40000baseSR4_Full);
606 		FW_CAPS_TO_LMM(SPEED_25G, 25000baseCR_Full);
607 		FW_CAPS_TO_LMM(SPEED_50G, 50000baseCR2_Full);
608 		FW_CAPS_TO_LMM(SPEED_100G, 100000baseCR4_Full);
609 		break;
610 
611 	default:
612 		break;
613 	}
614 
615 	if (fw_caps & FW_PORT_CAP32_FEC_V(FW_PORT_CAP32_FEC_M)) {
616 		FW_CAPS_TO_LMM(FEC_RS, FEC_RS);
617 		FW_CAPS_TO_LMM(FEC_BASER_RS, FEC_BASER);
618 	} else {
619 		SET_LMM(FEC_NONE);
620 	}
621 
622 	FW_CAPS_TO_LMM(ANEG, Autoneg);
623 	FW_CAPS_TO_LMM(802_3_PAUSE, Pause);
624 	FW_CAPS_TO_LMM(802_3_ASM_DIR, Asym_Pause);
625 
626 	#undef FW_CAPS_TO_LMM
627 	#undef SET_LMM
628 }
629 
630 /**
631  *	lmm_to_fw_caps - translate ethtool Link Mode Mask to Firmware
632  *	capabilities
633  *	@link_mode_mask: ethtool Link Mode Mask
634  *
635  *	Translate ethtool Link Mode Mask into a Firmware Port capabilities
636  *	value.
637  */
638 static unsigned int lmm_to_fw_caps(const unsigned long *link_mode_mask)
639 {
640 	unsigned int fw_caps = 0;
641 
642 	#define LMM_TO_FW_CAPS(__lmm_name, __fw_name) \
643 		do { \
644 			if (test_bit(ETHTOOL_LINK_MODE_ ## __lmm_name ## _BIT, \
645 				     link_mode_mask)) \
646 				fw_caps |= FW_PORT_CAP32_ ## __fw_name; \
647 		} while (0)
648 
649 	LMM_TO_FW_CAPS(100baseT_Full, SPEED_100M);
650 	LMM_TO_FW_CAPS(1000baseT_Full, SPEED_1G);
651 	LMM_TO_FW_CAPS(10000baseT_Full, SPEED_10G);
652 	LMM_TO_FW_CAPS(40000baseSR4_Full, SPEED_40G);
653 	LMM_TO_FW_CAPS(25000baseCR_Full, SPEED_25G);
654 	LMM_TO_FW_CAPS(50000baseCR2_Full, SPEED_50G);
655 	LMM_TO_FW_CAPS(100000baseCR4_Full, SPEED_100G);
656 
657 	#undef LMM_TO_FW_CAPS
658 
659 	return fw_caps;
660 }
661 
662 static int get_link_ksettings(struct net_device *dev,
663 			      struct ethtool_link_ksettings *link_ksettings)
664 {
665 	struct port_info *pi = netdev_priv(dev);
666 	struct ethtool_link_settings *base = &link_ksettings->base;
667 
668 	/* For the nonce, the Firmware doesn't send up Port State changes
669 	 * when the Virtual Interface attached to the Port is down.  So
670 	 * if it's down, let's grab any changes.
671 	 */
672 	if (!netif_running(dev))
673 		(void)t4_update_port_info(pi);
674 
675 	ethtool_link_ksettings_zero_link_mode(link_ksettings, supported);
676 	ethtool_link_ksettings_zero_link_mode(link_ksettings, advertising);
677 	ethtool_link_ksettings_zero_link_mode(link_ksettings, lp_advertising);
678 
679 	base->port = from_fw_port_mod_type(pi->port_type, pi->mod_type);
680 
681 	if (pi->mdio_addr >= 0) {
682 		base->phy_address = pi->mdio_addr;
683 		base->mdio_support = (pi->port_type == FW_PORT_TYPE_BT_SGMII
684 				      ? ETH_MDIO_SUPPORTS_C22
685 				      : ETH_MDIO_SUPPORTS_C45);
686 	} else {
687 		base->phy_address = 255;
688 		base->mdio_support = 0;
689 	}
690 
691 	fw_caps_to_lmm(pi->port_type, pi->link_cfg.pcaps,
692 		       link_ksettings->link_modes.supported);
693 	fw_caps_to_lmm(pi->port_type,
694 		       t4_link_acaps(pi->adapter,
695 				     pi->lport,
696 				     &pi->link_cfg),
697 		       link_ksettings->link_modes.advertising);
698 	fw_caps_to_lmm(pi->port_type, pi->link_cfg.lpacaps,
699 		       link_ksettings->link_modes.lp_advertising);
700 
701 	base->speed = (netif_carrier_ok(dev)
702 		       ? pi->link_cfg.speed
703 		       : SPEED_UNKNOWN);
704 	base->duplex = DUPLEX_FULL;
705 
706 	base->autoneg = pi->link_cfg.autoneg;
707 	if (pi->link_cfg.pcaps & FW_PORT_CAP32_ANEG)
708 		ethtool_link_ksettings_add_link_mode(link_ksettings,
709 						     supported, Autoneg);
710 	if (pi->link_cfg.autoneg)
711 		ethtool_link_ksettings_add_link_mode(link_ksettings,
712 						     advertising, Autoneg);
713 
714 	return 0;
715 }
716 
717 static int set_link_ksettings(struct net_device *dev,
718 			    const struct ethtool_link_ksettings *link_ksettings)
719 {
720 	struct port_info *pi = netdev_priv(dev);
721 	struct link_config *lc = &pi->link_cfg;
722 	const struct ethtool_link_settings *base = &link_ksettings->base;
723 	struct link_config old_lc;
724 	unsigned int fw_caps;
725 	int ret = 0;
726 
727 	/* only full-duplex supported */
728 	if (base->duplex != DUPLEX_FULL)
729 		return -EINVAL;
730 
731 	old_lc = *lc;
732 	if (!(lc->pcaps & FW_PORT_CAP32_ANEG) ||
733 	    base->autoneg == AUTONEG_DISABLE) {
734 		fw_caps = speed_to_fw_caps(base->speed);
735 
736 		/* Speed must be supported by Physical Port Capabilities. */
737 		if (!(lc->pcaps & fw_caps))
738 			return -EINVAL;
739 
740 		lc->speed_caps = fw_caps;
741 		lc->acaps = fw_caps;
742 	} else {
743 		fw_caps =
744 			lmm_to_fw_caps(link_ksettings->link_modes.advertising);
745 		if (!(lc->pcaps & fw_caps))
746 			return -EINVAL;
747 		lc->speed_caps = 0;
748 		lc->acaps = fw_caps | FW_PORT_CAP32_ANEG;
749 	}
750 	lc->autoneg = base->autoneg;
751 
752 	/* If the firmware rejects the Link Configuration request, back out
753 	 * the changes and report the error.
754 	 */
755 	ret = t4_link_l1cfg(pi->adapter, pi->adapter->mbox, pi->tx_chan, lc);
756 	if (ret)
757 		*lc = old_lc;
758 
759 	return ret;
760 }
761 
762 /* Translate the Firmware FEC value into the ethtool value. */
763 static inline unsigned int fwcap_to_eth_fec(unsigned int fw_fec)
764 {
765 	unsigned int eth_fec = 0;
766 
767 	if (fw_fec & FW_PORT_CAP32_FEC_RS)
768 		eth_fec |= ETHTOOL_FEC_RS;
769 	if (fw_fec & FW_PORT_CAP32_FEC_BASER_RS)
770 		eth_fec |= ETHTOOL_FEC_BASER;
771 
772 	/* if nothing is set, then FEC is off */
773 	if (!eth_fec)
774 		eth_fec = ETHTOOL_FEC_OFF;
775 
776 	return eth_fec;
777 }
778 
779 /* Translate Common Code FEC value into ethtool value. */
780 static inline unsigned int cc_to_eth_fec(unsigned int cc_fec)
781 {
782 	unsigned int eth_fec = 0;
783 
784 	if (cc_fec & FEC_AUTO)
785 		eth_fec |= ETHTOOL_FEC_AUTO;
786 	if (cc_fec & FEC_RS)
787 		eth_fec |= ETHTOOL_FEC_RS;
788 	if (cc_fec & FEC_BASER_RS)
789 		eth_fec |= ETHTOOL_FEC_BASER;
790 
791 	/* if nothing is set, then FEC is off */
792 	if (!eth_fec)
793 		eth_fec = ETHTOOL_FEC_OFF;
794 
795 	return eth_fec;
796 }
797 
798 /* Translate ethtool FEC value into Common Code value. */
799 static inline unsigned int eth_to_cc_fec(unsigned int eth_fec)
800 {
801 	unsigned int cc_fec = 0;
802 
803 	if (eth_fec & ETHTOOL_FEC_OFF)
804 		return cc_fec;
805 
806 	if (eth_fec & ETHTOOL_FEC_AUTO)
807 		cc_fec |= FEC_AUTO;
808 	if (eth_fec & ETHTOOL_FEC_RS)
809 		cc_fec |= FEC_RS;
810 	if (eth_fec & ETHTOOL_FEC_BASER)
811 		cc_fec |= FEC_BASER_RS;
812 
813 	return cc_fec;
814 }
815 
816 static int get_fecparam(struct net_device *dev, struct ethtool_fecparam *fec)
817 {
818 	const struct port_info *pi = netdev_priv(dev);
819 	const struct link_config *lc = &pi->link_cfg;
820 
821 	/* Translate the Firmware FEC Support into the ethtool value.  We
822 	 * always support IEEE 802.3 "automatic" selection of Link FEC type if
823 	 * any FEC is supported.
824 	 */
825 	fec->fec = fwcap_to_eth_fec(lc->pcaps);
826 	if (fec->fec != ETHTOOL_FEC_OFF)
827 		fec->fec |= ETHTOOL_FEC_AUTO;
828 
829 	/* Translate the current internal FEC parameters into the
830 	 * ethtool values.
831 	 */
832 	fec->active_fec = cc_to_eth_fec(lc->fec);
833 
834 	return 0;
835 }
836 
837 static int set_fecparam(struct net_device *dev, struct ethtool_fecparam *fec)
838 {
839 	struct port_info *pi = netdev_priv(dev);
840 	struct link_config *lc = &pi->link_cfg;
841 	struct link_config old_lc;
842 	int ret;
843 
844 	/* Save old Link Configuration in case the L1 Configure below
845 	 * fails.
846 	 */
847 	old_lc = *lc;
848 
849 	/* Try to perform the L1 Configure and return the result of that
850 	 * effort.  If it fails, revert the attempted change.
851 	 */
852 	lc->requested_fec = eth_to_cc_fec(fec->fec);
853 	ret = t4_link_l1cfg(pi->adapter, pi->adapter->mbox,
854 			    pi->tx_chan, lc);
855 	if (ret)
856 		*lc = old_lc;
857 	return ret;
858 }
859 
860 static void get_pauseparam(struct net_device *dev,
861 			   struct ethtool_pauseparam *epause)
862 {
863 	struct port_info *p = netdev_priv(dev);
864 
865 	epause->autoneg = (p->link_cfg.requested_fc & PAUSE_AUTONEG) != 0;
866 	epause->rx_pause = (p->link_cfg.advertised_fc & PAUSE_RX) != 0;
867 	epause->tx_pause = (p->link_cfg.advertised_fc & PAUSE_TX) != 0;
868 }
869 
870 static int set_pauseparam(struct net_device *dev,
871 			  struct ethtool_pauseparam *epause)
872 {
873 	struct port_info *p = netdev_priv(dev);
874 	struct link_config *lc = &p->link_cfg;
875 
876 	if (epause->autoneg == AUTONEG_DISABLE)
877 		lc->requested_fc = 0;
878 	else if (lc->pcaps & FW_PORT_CAP32_ANEG)
879 		lc->requested_fc = PAUSE_AUTONEG;
880 	else
881 		return -EINVAL;
882 
883 	if (epause->rx_pause)
884 		lc->requested_fc |= PAUSE_RX;
885 	if (epause->tx_pause)
886 		lc->requested_fc |= PAUSE_TX;
887 	if (netif_running(dev))
888 		return t4_link_l1cfg(p->adapter, p->adapter->mbox, p->tx_chan,
889 				     lc);
890 	return 0;
891 }
892 
893 static void get_sge_param(struct net_device *dev, struct ethtool_ringparam *e,
894 			  struct kernel_ethtool_ringparam *kernel_e,
895 			  struct netlink_ext_ack *extack)
896 {
897 	const struct port_info *pi = netdev_priv(dev);
898 	const struct sge *s = &pi->adapter->sge;
899 
900 	e->rx_max_pending = MAX_RX_BUFFERS;
901 	e->rx_mini_max_pending = MAX_RSPQ_ENTRIES;
902 	e->rx_jumbo_max_pending = 0;
903 	e->tx_max_pending = MAX_TXQ_ENTRIES;
904 
905 	e->rx_pending = s->ethrxq[pi->first_qset].fl.size - 8;
906 	e->rx_mini_pending = s->ethrxq[pi->first_qset].rspq.size;
907 	e->rx_jumbo_pending = 0;
908 	e->tx_pending = s->ethtxq[pi->first_qset].q.size;
909 }
910 
911 static int set_sge_param(struct net_device *dev, struct ethtool_ringparam *e,
912 			 struct kernel_ethtool_ringparam *kernel_e,
913 			 struct netlink_ext_ack *extack)
914 {
915 	int i;
916 	const struct port_info *pi = netdev_priv(dev);
917 	struct adapter *adapter = pi->adapter;
918 	struct sge *s = &adapter->sge;
919 
920 	if (e->rx_pending > MAX_RX_BUFFERS || e->rx_jumbo_pending ||
921 	    e->tx_pending > MAX_TXQ_ENTRIES ||
922 	    e->rx_mini_pending > MAX_RSPQ_ENTRIES ||
923 	    e->rx_mini_pending < MIN_RSPQ_ENTRIES ||
924 	    e->rx_pending < MIN_FL_ENTRIES || e->tx_pending < MIN_TXQ_ENTRIES)
925 		return -EINVAL;
926 
927 	if (adapter->flags & CXGB4_FULL_INIT_DONE)
928 		return -EBUSY;
929 
930 	for (i = 0; i < pi->nqsets; ++i) {
931 		s->ethtxq[pi->first_qset + i].q.size = e->tx_pending;
932 		s->ethrxq[pi->first_qset + i].fl.size = e->rx_pending + 8;
933 		s->ethrxq[pi->first_qset + i].rspq.size = e->rx_mini_pending;
934 	}
935 	return 0;
936 }
937 
938 /**
939  * set_rx_intr_params - set a net devices's RX interrupt holdoff paramete!
940  * @dev: the network device
941  * @us: the hold-off time in us, or 0 to disable timer
942  * @cnt: the hold-off packet count, or 0 to disable counter
943  *
944  * Set the RX interrupt hold-off parameters for a network device.
945  */
946 static int set_rx_intr_params(struct net_device *dev,
947 			      unsigned int us, unsigned int cnt)
948 {
949 	int i, err;
950 	struct port_info *pi = netdev_priv(dev);
951 	struct adapter *adap = pi->adapter;
952 	struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset];
953 
954 	for (i = 0; i < pi->nqsets; i++, q++) {
955 		err = cxgb4_set_rspq_intr_params(&q->rspq, us, cnt);
956 		if (err)
957 			return err;
958 	}
959 	return 0;
960 }
961 
962 static int set_adaptive_rx_setting(struct net_device *dev, int adaptive_rx)
963 {
964 	int i;
965 	struct port_info *pi = netdev_priv(dev);
966 	struct adapter *adap = pi->adapter;
967 	struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset];
968 
969 	for (i = 0; i < pi->nqsets; i++, q++)
970 		q->rspq.adaptive_rx = adaptive_rx;
971 
972 	return 0;
973 }
974 
975 static int get_adaptive_rx_setting(struct net_device *dev)
976 {
977 	struct port_info *pi = netdev_priv(dev);
978 	struct adapter *adap = pi->adapter;
979 	struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset];
980 
981 	return q->rspq.adaptive_rx;
982 }
983 
984 /* Return the current global Adapter SGE Doorbell Queue Timer Tick for all
985  * Ethernet TX Queues.
986  */
987 static int get_dbqtimer_tick(struct net_device *dev)
988 {
989 	struct port_info *pi = netdev_priv(dev);
990 	struct adapter *adap = pi->adapter;
991 
992 	if (!(adap->flags & CXGB4_SGE_DBQ_TIMER))
993 		return 0;
994 
995 	return adap->sge.dbqtimer_tick;
996 }
997 
998 /* Return the SGE Doorbell Queue Timer Value for the Ethernet TX Queues
999  * associated with a Network Device.
1000  */
1001 static int get_dbqtimer(struct net_device *dev)
1002 {
1003 	struct port_info *pi = netdev_priv(dev);
1004 	struct adapter *adap = pi->adapter;
1005 	struct sge_eth_txq *txq;
1006 
1007 	txq = &adap->sge.ethtxq[pi->first_qset];
1008 
1009 	if (!(adap->flags & CXGB4_SGE_DBQ_TIMER))
1010 		return 0;
1011 
1012 	/* all of the TX Queues use the same Timer Index */
1013 	return adap->sge.dbqtimer_val[txq->dbqtimerix];
1014 }
1015 
1016 /* Set the global Adapter SGE Doorbell Queue Timer Tick for all Ethernet TX
1017  * Queues.  This is the fundamental "Tick" that sets the scale of values which
1018  * can be used.  Individual Ethernet TX Queues index into a relatively small
1019  * array of Tick Multipliers.  Changing the base Tick will thus change all of
1020  * the resulting Timer Values associated with those multipliers for all
1021  * Ethernet TX Queues.
1022  */
1023 static int set_dbqtimer_tick(struct net_device *dev, int usecs)
1024 {
1025 	struct port_info *pi = netdev_priv(dev);
1026 	struct adapter *adap = pi->adapter;
1027 	struct sge *s = &adap->sge;
1028 	u32 param, val;
1029 	int ret;
1030 
1031 	if (!(adap->flags & CXGB4_SGE_DBQ_TIMER))
1032 		return 0;
1033 
1034 	/* return early if it's the same Timer Tick we're already using */
1035 	if (s->dbqtimer_tick == usecs)
1036 		return 0;
1037 
1038 	/* attempt to set the new Timer Tick value */
1039 	param = (FW_PARAMS_MNEM_V(FW_PARAMS_MNEM_DEV) |
1040 		 FW_PARAMS_PARAM_X_V(FW_PARAMS_PARAM_DEV_DBQ_TIMERTICK));
1041 	val = usecs;
1042 	ret = t4_set_params(adap, adap->mbox, adap->pf, 0, 1, &param, &val);
1043 	if (ret)
1044 		return ret;
1045 	s->dbqtimer_tick = usecs;
1046 
1047 	/* if successful, reread resulting dependent Timer values */
1048 	ret = t4_read_sge_dbqtimers(adap, ARRAY_SIZE(s->dbqtimer_val),
1049 				    s->dbqtimer_val);
1050 	return ret;
1051 }
1052 
1053 /* Set the SGE Doorbell Queue Timer Value for the Ethernet TX Queues
1054  * associated with a Network Device.  There is a relatively small array of
1055  * possible Timer Values so we need to pick the closest value available.
1056  */
1057 static int set_dbqtimer(struct net_device *dev, int usecs)
1058 {
1059 	int qix, timerix, min_timerix, delta, min_delta;
1060 	struct port_info *pi = netdev_priv(dev);
1061 	struct adapter *adap = pi->adapter;
1062 	struct sge *s = &adap->sge;
1063 	struct sge_eth_txq *txq;
1064 	u32 param, val;
1065 	int ret;
1066 
1067 	if (!(adap->flags & CXGB4_SGE_DBQ_TIMER))
1068 		return 0;
1069 
1070 	/* Find the SGE Doorbell Timer Value that's closest to the requested
1071 	 * value.
1072 	 */
1073 	min_delta = INT_MAX;
1074 	min_timerix = 0;
1075 	for (timerix = 0; timerix < ARRAY_SIZE(s->dbqtimer_val); timerix++) {
1076 		delta = s->dbqtimer_val[timerix] - usecs;
1077 		if (delta < 0)
1078 			delta = -delta;
1079 		if (delta < min_delta) {
1080 			min_delta = delta;
1081 			min_timerix = timerix;
1082 		}
1083 	}
1084 
1085 	/* Return early if it's the same Timer Index we're already using.
1086 	 * We use the same Timer Index for all of the TX Queues for an
1087 	 * interface so it's only necessary to check the first one.
1088 	 */
1089 	txq = &s->ethtxq[pi->first_qset];
1090 	if (txq->dbqtimerix == min_timerix)
1091 		return 0;
1092 
1093 	for (qix = 0; qix < pi->nqsets; qix++, txq++) {
1094 		if (adap->flags & CXGB4_FULL_INIT_DONE) {
1095 			param =
1096 			 (FW_PARAMS_MNEM_V(FW_PARAMS_MNEM_DMAQ) |
1097 			  FW_PARAMS_PARAM_X_V(FW_PARAMS_PARAM_DMAQ_EQ_TIMERIX) |
1098 			  FW_PARAMS_PARAM_YZ_V(txq->q.cntxt_id));
1099 			val = min_timerix;
1100 			ret = t4_set_params(adap, adap->mbox, adap->pf, 0,
1101 					    1, &param, &val);
1102 			if (ret)
1103 				return ret;
1104 		}
1105 		txq->dbqtimerix = min_timerix;
1106 	}
1107 	return 0;
1108 }
1109 
1110 /* Set the global Adapter SGE Doorbell Queue Timer Tick for all Ethernet TX
1111  * Queues and the Timer Value for the Ethernet TX Queues associated with a
1112  * Network Device.  Since changing the global Tick changes all of the
1113  * available Timer Values, we need to do this first before selecting the
1114  * resulting closest Timer Value.  Moreover, since the Tick is global,
1115  * changing it affects the Timer Values for all Network Devices on the
1116  * adapter.  So, before changing the Tick, we grab all of the current Timer
1117  * Values for other Network Devices on this Adapter and then attempt to select
1118  * new Timer Values which are close to the old values ...
1119  */
1120 static int set_dbqtimer_tickval(struct net_device *dev,
1121 				int tick_usecs, int timer_usecs)
1122 {
1123 	struct port_info *pi = netdev_priv(dev);
1124 	struct adapter *adap = pi->adapter;
1125 	int timer[MAX_NPORTS];
1126 	unsigned int port;
1127 	int ret;
1128 
1129 	/* Grab the other adapter Network Interface current timers and fill in
1130 	 * the new one for this Network Interface.
1131 	 */
1132 	for_each_port(adap, port)
1133 		if (port == pi->port_id)
1134 			timer[port] = timer_usecs;
1135 		else
1136 			timer[port] = get_dbqtimer(adap->port[port]);
1137 
1138 	/* Change the global Tick first ... */
1139 	ret = set_dbqtimer_tick(dev, tick_usecs);
1140 	if (ret)
1141 		return ret;
1142 
1143 	/* ... and then set all of the Network Interface Timer Values ... */
1144 	for_each_port(adap, port) {
1145 		ret = set_dbqtimer(adap->port[port], timer[port]);
1146 		if (ret)
1147 			return ret;
1148 	}
1149 
1150 	return 0;
1151 }
1152 
1153 static int set_coalesce(struct net_device *dev,
1154 			struct ethtool_coalesce *coalesce,
1155 			struct kernel_ethtool_coalesce *kernel_coal,
1156 			struct netlink_ext_ack *extack)
1157 {
1158 	int ret;
1159 
1160 	set_adaptive_rx_setting(dev, coalesce->use_adaptive_rx_coalesce);
1161 
1162 	ret = set_rx_intr_params(dev, coalesce->rx_coalesce_usecs,
1163 				 coalesce->rx_max_coalesced_frames);
1164 	if (ret)
1165 		return ret;
1166 
1167 	return set_dbqtimer_tickval(dev,
1168 				    coalesce->tx_coalesce_usecs_irq,
1169 				    coalesce->tx_coalesce_usecs);
1170 }
1171 
1172 static int get_coalesce(struct net_device *dev, struct ethtool_coalesce *c,
1173 			struct kernel_ethtool_coalesce *kernel_coal,
1174 			struct netlink_ext_ack *extack)
1175 {
1176 	const struct port_info *pi = netdev_priv(dev);
1177 	const struct adapter *adap = pi->adapter;
1178 	const struct sge_rspq *rq = &adap->sge.ethrxq[pi->first_qset].rspq;
1179 
1180 	c->rx_coalesce_usecs = qtimer_val(adap, rq);
1181 	c->rx_max_coalesced_frames = (rq->intr_params & QINTR_CNT_EN_F) ?
1182 		adap->sge.counter_val[rq->pktcnt_idx] : 0;
1183 	c->use_adaptive_rx_coalesce = get_adaptive_rx_setting(dev);
1184 	c->tx_coalesce_usecs_irq = get_dbqtimer_tick(dev);
1185 	c->tx_coalesce_usecs = get_dbqtimer(dev);
1186 	return 0;
1187 }
1188 
1189 /* The next two routines implement eeprom read/write from physical addresses.
1190  */
1191 static int eeprom_rd_phys(struct adapter *adap, unsigned int phys_addr, u32 *v)
1192 {
1193 	int vaddr = t4_eeprom_ptov(phys_addr, adap->pf, EEPROMPFSIZE);
1194 
1195 	if (vaddr >= 0)
1196 		vaddr = pci_read_vpd(adap->pdev, vaddr, sizeof(u32), v);
1197 	return vaddr < 0 ? vaddr : 0;
1198 }
1199 
1200 static int eeprom_wr_phys(struct adapter *adap, unsigned int phys_addr, u32 v)
1201 {
1202 	int vaddr = t4_eeprom_ptov(phys_addr, adap->pf, EEPROMPFSIZE);
1203 
1204 	if (vaddr >= 0)
1205 		vaddr = pci_write_vpd(adap->pdev, vaddr, sizeof(u32), &v);
1206 	return vaddr < 0 ? vaddr : 0;
1207 }
1208 
1209 #define EEPROM_MAGIC 0x38E2F10C
1210 
1211 static int get_eeprom(struct net_device *dev, struct ethtool_eeprom *e,
1212 		      u8 *data)
1213 {
1214 	int i, err = 0;
1215 	struct adapter *adapter = netdev2adap(dev);
1216 	u8 *buf = kvzalloc(EEPROMSIZE, GFP_KERNEL);
1217 
1218 	if (!buf)
1219 		return -ENOMEM;
1220 
1221 	e->magic = EEPROM_MAGIC;
1222 	for (i = e->offset & ~3; !err && i < e->offset + e->len; i += 4)
1223 		err = eeprom_rd_phys(adapter, i, (u32 *)&buf[i]);
1224 
1225 	if (!err)
1226 		memcpy(data, buf + e->offset, e->len);
1227 	kvfree(buf);
1228 	return err;
1229 }
1230 
1231 static int set_eeprom(struct net_device *dev, struct ethtool_eeprom *eeprom,
1232 		      u8 *data)
1233 {
1234 	u8 *buf;
1235 	int err = 0;
1236 	u32 aligned_offset, aligned_len, *p;
1237 	struct adapter *adapter = netdev2adap(dev);
1238 
1239 	if (eeprom->magic != EEPROM_MAGIC)
1240 		return -EINVAL;
1241 
1242 	aligned_offset = eeprom->offset & ~3;
1243 	aligned_len = (eeprom->len + (eeprom->offset & 3) + 3) & ~3;
1244 
1245 	if (adapter->pf > 0) {
1246 		u32 start = 1024 + adapter->pf * EEPROMPFSIZE;
1247 
1248 		if (aligned_offset < start ||
1249 		    aligned_offset + aligned_len > start + EEPROMPFSIZE)
1250 			return -EPERM;
1251 	}
1252 
1253 	if (aligned_offset != eeprom->offset || aligned_len != eeprom->len) {
1254 		/* RMW possibly needed for first or last words.
1255 		 */
1256 		buf = kvzalloc(aligned_len, GFP_KERNEL);
1257 		if (!buf)
1258 			return -ENOMEM;
1259 		err = eeprom_rd_phys(adapter, aligned_offset, (u32 *)buf);
1260 		if (!err && aligned_len > 4)
1261 			err = eeprom_rd_phys(adapter,
1262 					     aligned_offset + aligned_len - 4,
1263 					     (u32 *)&buf[aligned_len - 4]);
1264 		if (err)
1265 			goto out;
1266 		memcpy(buf + (eeprom->offset & 3), data, eeprom->len);
1267 	} else {
1268 		buf = data;
1269 	}
1270 
1271 	err = t4_seeprom_wp(adapter, false);
1272 	if (err)
1273 		goto out;
1274 
1275 	for (p = (u32 *)buf; !err && aligned_len; aligned_len -= 4, p++) {
1276 		err = eeprom_wr_phys(adapter, aligned_offset, *p);
1277 		aligned_offset += 4;
1278 	}
1279 
1280 	if (!err)
1281 		err = t4_seeprom_wp(adapter, true);
1282 out:
1283 	if (buf != data)
1284 		kvfree(buf);
1285 	return err;
1286 }
1287 
1288 static int cxgb4_ethtool_flash_bootcfg(struct net_device *netdev,
1289 				       const u8 *data, u32 size)
1290 {
1291 	struct adapter *adap = netdev2adap(netdev);
1292 	int ret;
1293 
1294 	ret = t4_load_bootcfg(adap, data, size);
1295 	if (ret)
1296 		dev_err(adap->pdev_dev, "Failed to load boot cfg image\n");
1297 
1298 	return ret;
1299 }
1300 
1301 static int cxgb4_ethtool_flash_boot(struct net_device *netdev,
1302 				    const u8 *bdata, u32 size)
1303 {
1304 	struct adapter *adap = netdev2adap(netdev);
1305 	unsigned int offset;
1306 	u8 *data;
1307 	int ret;
1308 
1309 	data = kmemdup(bdata, size, GFP_KERNEL);
1310 	if (!data)
1311 		return -ENOMEM;
1312 
1313 	offset = OFFSET_G(t4_read_reg(adap, PF_REG(0, PCIE_PF_EXPROM_OFST_A)));
1314 
1315 	ret = t4_load_boot(adap, data, offset, size);
1316 	if (ret)
1317 		dev_err(adap->pdev_dev, "Failed to load boot image\n");
1318 
1319 	kfree(data);
1320 	return ret;
1321 }
1322 
1323 #define CXGB4_PHY_SIG 0x130000ea
1324 
1325 static int cxgb4_validate_phy_image(const u8 *data, u32 *size)
1326 {
1327 	struct cxgb4_fw_data *header;
1328 
1329 	header = (struct cxgb4_fw_data *)data;
1330 	if (be32_to_cpu(header->signature) != CXGB4_PHY_SIG)
1331 		return -EINVAL;
1332 
1333 	return 0;
1334 }
1335 
1336 static int cxgb4_ethtool_flash_phy(struct net_device *netdev,
1337 				   const u8 *data, u32 size)
1338 {
1339 	struct adapter *adap = netdev2adap(netdev);
1340 	int ret;
1341 
1342 	ret = cxgb4_validate_phy_image(data, NULL);
1343 	if (ret) {
1344 		dev_err(adap->pdev_dev, "PHY signature mismatch\n");
1345 		return ret;
1346 	}
1347 
1348 	/* We have to RESET the chip/firmware because we need the
1349 	 * chip in uninitialized state for loading new PHY image.
1350 	 * Otherwise, the running firmware will only store the PHY
1351 	 * image in local RAM which will be lost after next reset.
1352 	 */
1353 	ret = t4_fw_reset(adap, adap->mbox, PIORSTMODE_F | PIORST_F);
1354 	if (ret < 0) {
1355 		dev_err(adap->pdev_dev,
1356 			"Set FW to RESET for flashing PHY FW failed. ret: %d\n",
1357 			ret);
1358 		return ret;
1359 	}
1360 
1361 	ret = t4_load_phy_fw(adap, MEMWIN_NIC, NULL, data, size);
1362 	if (ret < 0) {
1363 		dev_err(adap->pdev_dev, "Failed to load PHY FW. ret: %d\n",
1364 			ret);
1365 		return ret;
1366 	}
1367 
1368 	return 0;
1369 }
1370 
1371 static int cxgb4_ethtool_flash_fw(struct net_device *netdev,
1372 				  const u8 *data, u32 size)
1373 {
1374 	struct adapter *adap = netdev2adap(netdev);
1375 	unsigned int mbox = PCIE_FW_MASTER_M + 1;
1376 	int ret;
1377 
1378 	/* If the adapter has been fully initialized then we'll go ahead and
1379 	 * try to get the firmware's cooperation in upgrading to the new
1380 	 * firmware image otherwise we'll try to do the entire job from the
1381 	 * host ... and we always "force" the operation in this path.
1382 	 */
1383 	if (adap->flags & CXGB4_FULL_INIT_DONE)
1384 		mbox = adap->mbox;
1385 
1386 	ret = t4_fw_upgrade(adap, mbox, data, size, 1);
1387 	if (ret)
1388 		dev_err(adap->pdev_dev,
1389 			"Failed to flash firmware\n");
1390 
1391 	return ret;
1392 }
1393 
1394 static int cxgb4_ethtool_flash_region(struct net_device *netdev,
1395 				      const u8 *data, u32 size, u32 region)
1396 {
1397 	struct adapter *adap = netdev2adap(netdev);
1398 	int ret;
1399 
1400 	switch (region) {
1401 	case CXGB4_ETHTOOL_FLASH_FW:
1402 		ret = cxgb4_ethtool_flash_fw(netdev, data, size);
1403 		break;
1404 	case CXGB4_ETHTOOL_FLASH_PHY:
1405 		ret = cxgb4_ethtool_flash_phy(netdev, data, size);
1406 		break;
1407 	case CXGB4_ETHTOOL_FLASH_BOOT:
1408 		ret = cxgb4_ethtool_flash_boot(netdev, data, size);
1409 		break;
1410 	case CXGB4_ETHTOOL_FLASH_BOOTCFG:
1411 		ret = cxgb4_ethtool_flash_bootcfg(netdev, data, size);
1412 		break;
1413 	default:
1414 		ret = -EOPNOTSUPP;
1415 		break;
1416 	}
1417 
1418 	if (!ret)
1419 		dev_info(adap->pdev_dev,
1420 			 "loading %s successful, reload cxgb4 driver\n",
1421 			 flash_region_strings[region]);
1422 	return ret;
1423 }
1424 
1425 #define CXGB4_FW_SIG 0x4368656c
1426 #define CXGB4_FW_SIG_OFFSET 0x160
1427 
1428 static int cxgb4_validate_fw_image(const u8 *data, u32 *size)
1429 {
1430 	struct cxgb4_fw_data *header;
1431 
1432 	header = (struct cxgb4_fw_data *)&data[CXGB4_FW_SIG_OFFSET];
1433 	if (be32_to_cpu(header->signature) != CXGB4_FW_SIG)
1434 		return -EINVAL;
1435 
1436 	if (size)
1437 		*size = be16_to_cpu(((struct fw_hdr *)data)->len512) * 512;
1438 
1439 	return 0;
1440 }
1441 
1442 static int cxgb4_validate_bootcfg_image(const u8 *data, u32 *size)
1443 {
1444 	struct cxgb4_bootcfg_data *header;
1445 
1446 	header = (struct cxgb4_bootcfg_data *)data;
1447 	if (le16_to_cpu(header->signature) != BOOT_CFG_SIG)
1448 		return -EINVAL;
1449 
1450 	return 0;
1451 }
1452 
1453 static int cxgb4_validate_boot_image(const u8 *data, u32 *size)
1454 {
1455 	struct cxgb4_pci_exp_rom_header *exp_header;
1456 	struct cxgb4_pcir_data *pcir_header;
1457 	struct legacy_pci_rom_hdr *header;
1458 	const u8 *cur_header = data;
1459 	u16 pcir_offset;
1460 
1461 	exp_header = (struct cxgb4_pci_exp_rom_header *)data;
1462 
1463 	if (le16_to_cpu(exp_header->signature) != BOOT_SIGNATURE)
1464 		return -EINVAL;
1465 
1466 	if (size) {
1467 		do {
1468 			header = (struct legacy_pci_rom_hdr *)cur_header;
1469 			pcir_offset = le16_to_cpu(header->pcir_offset);
1470 			pcir_header = (struct cxgb4_pcir_data *)(cur_header +
1471 				      pcir_offset);
1472 
1473 			*size += header->size512 * 512;
1474 			cur_header += header->size512 * 512;
1475 		} while (!(pcir_header->indicator & CXGB4_HDR_INDI));
1476 	}
1477 
1478 	return 0;
1479 }
1480 
1481 static int cxgb4_ethtool_get_flash_region(const u8 *data, u32 *size)
1482 {
1483 	if (!cxgb4_validate_fw_image(data, size))
1484 		return CXGB4_ETHTOOL_FLASH_FW;
1485 	if (!cxgb4_validate_boot_image(data, size))
1486 		return CXGB4_ETHTOOL_FLASH_BOOT;
1487 	if (!cxgb4_validate_phy_image(data, size))
1488 		return CXGB4_ETHTOOL_FLASH_PHY;
1489 	if (!cxgb4_validate_bootcfg_image(data, size))
1490 		return CXGB4_ETHTOOL_FLASH_BOOTCFG;
1491 
1492 	return -EOPNOTSUPP;
1493 }
1494 
1495 static int set_flash(struct net_device *netdev, struct ethtool_flash *ef)
1496 {
1497 	struct adapter *adap = netdev2adap(netdev);
1498 	const struct firmware *fw;
1499 	unsigned int master;
1500 	u8 master_vld = 0;
1501 	const u8 *fw_data;
1502 	size_t fw_size;
1503 	u32 size = 0;
1504 	u32 pcie_fw;
1505 	int region;
1506 	int ret;
1507 
1508 	pcie_fw = t4_read_reg(adap, PCIE_FW_A);
1509 	master = PCIE_FW_MASTER_G(pcie_fw);
1510 	if (pcie_fw & PCIE_FW_MASTER_VLD_F)
1511 		master_vld = 1;
1512 	/* if csiostor is the master return */
1513 	if (master_vld && (master != adap->pf)) {
1514 		dev_warn(adap->pdev_dev,
1515 			 "cxgb4 driver needs to be loaded as MASTER to support FW flash\n");
1516 		return -EOPNOTSUPP;
1517 	}
1518 
1519 	ef->data[sizeof(ef->data) - 1] = '\0';
1520 	ret = request_firmware(&fw, ef->data, adap->pdev_dev);
1521 	if (ret < 0)
1522 		return ret;
1523 
1524 	fw_data = fw->data;
1525 	fw_size = fw->size;
1526 	if (ef->region == ETHTOOL_FLASH_ALL_REGIONS) {
1527 		while (fw_size > 0) {
1528 			size = 0;
1529 			region = cxgb4_ethtool_get_flash_region(fw_data, &size);
1530 			if (region < 0 || !size) {
1531 				ret = region;
1532 				goto out_free_fw;
1533 			}
1534 
1535 			ret = cxgb4_ethtool_flash_region(netdev, fw_data, size,
1536 							 region);
1537 			if (ret)
1538 				goto out_free_fw;
1539 
1540 			fw_data += size;
1541 			fw_size -= size;
1542 		}
1543 	} else {
1544 		ret = cxgb4_ethtool_flash_region(netdev, fw_data, fw_size,
1545 						 ef->region);
1546 	}
1547 
1548 out_free_fw:
1549 	release_firmware(fw);
1550 	return ret;
1551 }
1552 
1553 static int get_ts_info(struct net_device *dev, struct ethtool_ts_info *ts_info)
1554 {
1555 	struct port_info *pi = netdev_priv(dev);
1556 	struct  adapter *adapter = pi->adapter;
1557 
1558 	ts_info->so_timestamping = SOF_TIMESTAMPING_TX_SOFTWARE |
1559 				   SOF_TIMESTAMPING_RX_SOFTWARE |
1560 				   SOF_TIMESTAMPING_SOFTWARE;
1561 
1562 	ts_info->so_timestamping |= SOF_TIMESTAMPING_RX_HARDWARE |
1563 				    SOF_TIMESTAMPING_TX_HARDWARE |
1564 				    SOF_TIMESTAMPING_RAW_HARDWARE;
1565 
1566 	ts_info->tx_types = (1 << HWTSTAMP_TX_OFF) |
1567 			    (1 << HWTSTAMP_TX_ON);
1568 
1569 	ts_info->rx_filters = (1 << HWTSTAMP_FILTER_NONE) |
1570 			      (1 << HWTSTAMP_FILTER_PTP_V2_L4_EVENT) |
1571 			      (1 << HWTSTAMP_FILTER_PTP_V1_L4_SYNC) |
1572 			      (1 << HWTSTAMP_FILTER_PTP_V1_L4_DELAY_REQ) |
1573 			      (1 << HWTSTAMP_FILTER_PTP_V2_L4_SYNC) |
1574 			      (1 << HWTSTAMP_FILTER_PTP_V2_L4_DELAY_REQ);
1575 
1576 	if (adapter->ptp_clock)
1577 		ts_info->phc_index = ptp_clock_index(adapter->ptp_clock);
1578 	else
1579 		ts_info->phc_index = -1;
1580 
1581 	return 0;
1582 }
1583 
1584 static u32 get_rss_table_size(struct net_device *dev)
1585 {
1586 	const struct port_info *pi = netdev_priv(dev);
1587 
1588 	return pi->rss_size;
1589 }
1590 
1591 static int get_rss_table(struct net_device *dev, u32 *p, u8 *key, u8 *hfunc)
1592 {
1593 	const struct port_info *pi = netdev_priv(dev);
1594 	unsigned int n = pi->rss_size;
1595 
1596 	if (hfunc)
1597 		*hfunc = ETH_RSS_HASH_TOP;
1598 	if (!p)
1599 		return 0;
1600 	while (n--)
1601 		p[n] = pi->rss[n];
1602 	return 0;
1603 }
1604 
1605 static int set_rss_table(struct net_device *dev, const u32 *p, const u8 *key,
1606 			 const u8 hfunc)
1607 {
1608 	unsigned int i;
1609 	struct port_info *pi = netdev_priv(dev);
1610 
1611 	/* We require at least one supported parameter to be changed and no
1612 	 * change in any of the unsupported parameters
1613 	 */
1614 	if (key ||
1615 	    (hfunc != ETH_RSS_HASH_NO_CHANGE && hfunc != ETH_RSS_HASH_TOP))
1616 		return -EOPNOTSUPP;
1617 	if (!p)
1618 		return 0;
1619 
1620 	/* Interface must be brought up atleast once */
1621 	if (pi->adapter->flags & CXGB4_FULL_INIT_DONE) {
1622 		for (i = 0; i < pi->rss_size; i++)
1623 			pi->rss[i] = p[i];
1624 
1625 		return cxgb4_write_rss(pi, pi->rss);
1626 	}
1627 
1628 	return -EPERM;
1629 }
1630 
1631 static struct filter_entry *cxgb4_get_filter_entry(struct adapter *adap,
1632 						   u32 ftid)
1633 {
1634 	struct tid_info *t = &adap->tids;
1635 
1636 	if (ftid >= t->hpftid_base && ftid < t->hpftid_base + t->nhpftids)
1637 		return &t->hpftid_tab[ftid - t->hpftid_base];
1638 
1639 	if (ftid >= t->ftid_base && ftid < t->ftid_base + t->nftids)
1640 		return &t->ftid_tab[ftid - t->ftid_base];
1641 
1642 	return lookup_tid(t, ftid);
1643 }
1644 
1645 static void cxgb4_fill_filter_rule(struct ethtool_rx_flow_spec *fs,
1646 				   struct ch_filter_specification *dfs)
1647 {
1648 	switch (dfs->val.proto) {
1649 	case IPPROTO_TCP:
1650 		if (dfs->type)
1651 			fs->flow_type = TCP_V6_FLOW;
1652 		else
1653 			fs->flow_type = TCP_V4_FLOW;
1654 		break;
1655 	case IPPROTO_UDP:
1656 		if (dfs->type)
1657 			fs->flow_type = UDP_V6_FLOW;
1658 		else
1659 			fs->flow_type = UDP_V4_FLOW;
1660 		break;
1661 	}
1662 
1663 	if (dfs->type) {
1664 		fs->h_u.tcp_ip6_spec.psrc = cpu_to_be16(dfs->val.fport);
1665 		fs->m_u.tcp_ip6_spec.psrc = cpu_to_be16(dfs->mask.fport);
1666 		fs->h_u.tcp_ip6_spec.pdst = cpu_to_be16(dfs->val.lport);
1667 		fs->m_u.tcp_ip6_spec.pdst = cpu_to_be16(dfs->mask.lport);
1668 		memcpy(&fs->h_u.tcp_ip6_spec.ip6src, &dfs->val.fip[0],
1669 		       sizeof(fs->h_u.tcp_ip6_spec.ip6src));
1670 		memcpy(&fs->m_u.tcp_ip6_spec.ip6src, &dfs->mask.fip[0],
1671 		       sizeof(fs->m_u.tcp_ip6_spec.ip6src));
1672 		memcpy(&fs->h_u.tcp_ip6_spec.ip6dst, &dfs->val.lip[0],
1673 		       sizeof(fs->h_u.tcp_ip6_spec.ip6dst));
1674 		memcpy(&fs->m_u.tcp_ip6_spec.ip6dst, &dfs->mask.lip[0],
1675 		       sizeof(fs->m_u.tcp_ip6_spec.ip6dst));
1676 		fs->h_u.tcp_ip6_spec.tclass = dfs->val.tos;
1677 		fs->m_u.tcp_ip6_spec.tclass = dfs->mask.tos;
1678 	} else {
1679 		fs->h_u.tcp_ip4_spec.psrc = cpu_to_be16(dfs->val.fport);
1680 		fs->m_u.tcp_ip4_spec.psrc = cpu_to_be16(dfs->mask.fport);
1681 		fs->h_u.tcp_ip4_spec.pdst = cpu_to_be16(dfs->val.lport);
1682 		fs->m_u.tcp_ip4_spec.pdst = cpu_to_be16(dfs->mask.lport);
1683 		memcpy(&fs->h_u.tcp_ip4_spec.ip4src, &dfs->val.fip[0],
1684 		       sizeof(fs->h_u.tcp_ip4_spec.ip4src));
1685 		memcpy(&fs->m_u.tcp_ip4_spec.ip4src, &dfs->mask.fip[0],
1686 		       sizeof(fs->m_u.tcp_ip4_spec.ip4src));
1687 		memcpy(&fs->h_u.tcp_ip4_spec.ip4dst, &dfs->val.lip[0],
1688 		       sizeof(fs->h_u.tcp_ip4_spec.ip4dst));
1689 		memcpy(&fs->m_u.tcp_ip4_spec.ip4dst, &dfs->mask.lip[0],
1690 		       sizeof(fs->m_u.tcp_ip4_spec.ip4dst));
1691 		fs->h_u.tcp_ip4_spec.tos = dfs->val.tos;
1692 		fs->m_u.tcp_ip4_spec.tos = dfs->mask.tos;
1693 	}
1694 	fs->h_ext.vlan_tci = cpu_to_be16(dfs->val.ivlan);
1695 	fs->m_ext.vlan_tci = cpu_to_be16(dfs->mask.ivlan);
1696 	fs->flow_type |= FLOW_EXT;
1697 
1698 	if (dfs->action == FILTER_DROP)
1699 		fs->ring_cookie = RX_CLS_FLOW_DISC;
1700 	else
1701 		fs->ring_cookie = dfs->iq;
1702 }
1703 
1704 static int cxgb4_ntuple_get_filter(struct net_device *dev,
1705 				   struct ethtool_rxnfc *cmd,
1706 				   unsigned int loc)
1707 {
1708 	const struct port_info *pi = netdev_priv(dev);
1709 	struct adapter *adap = netdev2adap(dev);
1710 	struct filter_entry *f;
1711 	int ftid;
1712 
1713 	if (!(adap->flags & CXGB4_FULL_INIT_DONE))
1714 		return -EAGAIN;
1715 
1716 	/* Check for maximum filter range */
1717 	if (!adap->ethtool_filters)
1718 		return -EOPNOTSUPP;
1719 
1720 	if (loc >= adap->ethtool_filters->nentries)
1721 		return -ERANGE;
1722 
1723 	if (!test_bit(loc, adap->ethtool_filters->port[pi->port_id].bmap))
1724 		return -ENOENT;
1725 
1726 	ftid = adap->ethtool_filters->port[pi->port_id].loc_array[loc];
1727 
1728 	/* Fetch filter_entry */
1729 	f = cxgb4_get_filter_entry(adap, ftid);
1730 
1731 	cxgb4_fill_filter_rule(&cmd->fs, &f->fs);
1732 
1733 	return 0;
1734 }
1735 
1736 static int get_rxnfc(struct net_device *dev, struct ethtool_rxnfc *info,
1737 		     u32 *rules)
1738 {
1739 	const struct port_info *pi = netdev_priv(dev);
1740 	struct adapter *adap = netdev2adap(dev);
1741 	unsigned int count = 0, index = 0;
1742 	int ret = 0;
1743 
1744 	switch (info->cmd) {
1745 	case ETHTOOL_GRXFH: {
1746 		unsigned int v = pi->rss_mode;
1747 
1748 		info->data = 0;
1749 		switch (info->flow_type) {
1750 		case TCP_V4_FLOW:
1751 			if (v & FW_RSS_VI_CONFIG_CMD_IP4FOURTUPEN_F)
1752 				info->data = RXH_IP_SRC | RXH_IP_DST |
1753 					     RXH_L4_B_0_1 | RXH_L4_B_2_3;
1754 			else if (v & FW_RSS_VI_CONFIG_CMD_IP4TWOTUPEN_F)
1755 				info->data = RXH_IP_SRC | RXH_IP_DST;
1756 			break;
1757 		case UDP_V4_FLOW:
1758 			if ((v & FW_RSS_VI_CONFIG_CMD_IP4FOURTUPEN_F) &&
1759 			    (v & FW_RSS_VI_CONFIG_CMD_UDPEN_F))
1760 				info->data = RXH_IP_SRC | RXH_IP_DST |
1761 					     RXH_L4_B_0_1 | RXH_L4_B_2_3;
1762 			else if (v & FW_RSS_VI_CONFIG_CMD_IP4TWOTUPEN_F)
1763 				info->data = RXH_IP_SRC | RXH_IP_DST;
1764 			break;
1765 		case SCTP_V4_FLOW:
1766 		case AH_ESP_V4_FLOW:
1767 		case IPV4_FLOW:
1768 			if (v & FW_RSS_VI_CONFIG_CMD_IP4TWOTUPEN_F)
1769 				info->data = RXH_IP_SRC | RXH_IP_DST;
1770 			break;
1771 		case TCP_V6_FLOW:
1772 			if (v & FW_RSS_VI_CONFIG_CMD_IP6FOURTUPEN_F)
1773 				info->data = RXH_IP_SRC | RXH_IP_DST |
1774 					     RXH_L4_B_0_1 | RXH_L4_B_2_3;
1775 			else if (v & FW_RSS_VI_CONFIG_CMD_IP6TWOTUPEN_F)
1776 				info->data = RXH_IP_SRC | RXH_IP_DST;
1777 			break;
1778 		case UDP_V6_FLOW:
1779 			if ((v & FW_RSS_VI_CONFIG_CMD_IP6FOURTUPEN_F) &&
1780 			    (v & FW_RSS_VI_CONFIG_CMD_UDPEN_F))
1781 				info->data = RXH_IP_SRC | RXH_IP_DST |
1782 					     RXH_L4_B_0_1 | RXH_L4_B_2_3;
1783 			else if (v & FW_RSS_VI_CONFIG_CMD_IP6TWOTUPEN_F)
1784 				info->data = RXH_IP_SRC | RXH_IP_DST;
1785 			break;
1786 		case SCTP_V6_FLOW:
1787 		case AH_ESP_V6_FLOW:
1788 		case IPV6_FLOW:
1789 			if (v & FW_RSS_VI_CONFIG_CMD_IP6TWOTUPEN_F)
1790 				info->data = RXH_IP_SRC | RXH_IP_DST;
1791 			break;
1792 		}
1793 		return 0;
1794 	}
1795 	case ETHTOOL_GRXRINGS:
1796 		info->data = pi->nqsets;
1797 		return 0;
1798 	case ETHTOOL_GRXCLSRLCNT:
1799 		info->rule_cnt =
1800 		       adap->ethtool_filters->port[pi->port_id].in_use;
1801 		return 0;
1802 	case ETHTOOL_GRXCLSRULE:
1803 		return cxgb4_ntuple_get_filter(dev, info, info->fs.location);
1804 	case ETHTOOL_GRXCLSRLALL:
1805 		info->data = adap->ethtool_filters->nentries;
1806 		while (count < info->rule_cnt) {
1807 			ret = cxgb4_ntuple_get_filter(dev, info, index);
1808 			if (!ret)
1809 				rules[count++] = index;
1810 			index++;
1811 		}
1812 		return 0;
1813 	}
1814 
1815 	return -EOPNOTSUPP;
1816 }
1817 
1818 static int cxgb4_ntuple_del_filter(struct net_device *dev,
1819 				   struct ethtool_rxnfc *cmd)
1820 {
1821 	struct cxgb4_ethtool_filter_info *filter_info;
1822 	struct adapter *adapter = netdev2adap(dev);
1823 	struct port_info *pi = netdev_priv(dev);
1824 	struct filter_entry *f;
1825 	u32 filter_id;
1826 	int ret;
1827 
1828 	if (!(adapter->flags & CXGB4_FULL_INIT_DONE))
1829 		return -EAGAIN;  /* can still change nfilters */
1830 
1831 	if (!adapter->ethtool_filters)
1832 		return -EOPNOTSUPP;
1833 
1834 	if (cmd->fs.location >= adapter->ethtool_filters->nentries) {
1835 		dev_err(adapter->pdev_dev,
1836 			"Location must be < %u",
1837 			adapter->ethtool_filters->nentries);
1838 		return -ERANGE;
1839 	}
1840 
1841 	filter_info = &adapter->ethtool_filters->port[pi->port_id];
1842 
1843 	if (!test_bit(cmd->fs.location, filter_info->bmap))
1844 		return -ENOENT;
1845 
1846 	filter_id = filter_info->loc_array[cmd->fs.location];
1847 	f = cxgb4_get_filter_entry(adapter, filter_id);
1848 
1849 	if (f->fs.prio)
1850 		filter_id -= adapter->tids.hpftid_base;
1851 	else if (!f->fs.hash)
1852 		filter_id -= (adapter->tids.ftid_base - adapter->tids.nhpftids);
1853 
1854 	ret = cxgb4_flow_rule_destroy(dev, f->fs.tc_prio, &f->fs, filter_id);
1855 	if (ret)
1856 		goto err;
1857 
1858 	clear_bit(cmd->fs.location, filter_info->bmap);
1859 	filter_info->in_use--;
1860 
1861 err:
1862 	return ret;
1863 }
1864 
1865 /* Add Ethtool n-tuple filters. */
1866 static int cxgb4_ntuple_set_filter(struct net_device *netdev,
1867 				   struct ethtool_rxnfc *cmd)
1868 {
1869 	struct ethtool_rx_flow_spec_input input = {};
1870 	struct cxgb4_ethtool_filter_info *filter_info;
1871 	struct adapter *adapter = netdev2adap(netdev);
1872 	struct port_info *pi = netdev_priv(netdev);
1873 	struct ch_filter_specification fs;
1874 	struct ethtool_rx_flow_rule *flow;
1875 	u32 tid;
1876 	int ret;
1877 
1878 	if (!(adapter->flags & CXGB4_FULL_INIT_DONE))
1879 		return -EAGAIN;  /* can still change nfilters */
1880 
1881 	if (!adapter->ethtool_filters)
1882 		return -EOPNOTSUPP;
1883 
1884 	if (cmd->fs.location >= adapter->ethtool_filters->nentries) {
1885 		dev_err(adapter->pdev_dev,
1886 			"Location must be < %u",
1887 			adapter->ethtool_filters->nentries);
1888 		return -ERANGE;
1889 	}
1890 
1891 	if (test_bit(cmd->fs.location,
1892 		     adapter->ethtool_filters->port[pi->port_id].bmap))
1893 		return -EEXIST;
1894 
1895 	memset(&fs, 0, sizeof(fs));
1896 
1897 	input.fs = &cmd->fs;
1898 	flow = ethtool_rx_flow_rule_create(&input);
1899 	if (IS_ERR(flow)) {
1900 		ret = PTR_ERR(flow);
1901 		goto exit;
1902 	}
1903 
1904 	fs.hitcnts = 1;
1905 
1906 	ret = cxgb4_flow_rule_replace(netdev, flow->rule, cmd->fs.location,
1907 				      NULL, &fs, &tid);
1908 	if (ret)
1909 		goto free;
1910 
1911 	filter_info = &adapter->ethtool_filters->port[pi->port_id];
1912 
1913 	if (fs.prio)
1914 		tid += adapter->tids.hpftid_base;
1915 	else if (!fs.hash)
1916 		tid += (adapter->tids.ftid_base - adapter->tids.nhpftids);
1917 
1918 	filter_info->loc_array[cmd->fs.location] = tid;
1919 	set_bit(cmd->fs.location, filter_info->bmap);
1920 	filter_info->in_use++;
1921 
1922 free:
1923 	ethtool_rx_flow_rule_destroy(flow);
1924 exit:
1925 	return ret;
1926 }
1927 
1928 static int set_rxnfc(struct net_device *dev, struct ethtool_rxnfc *cmd)
1929 {
1930 	int ret = -EOPNOTSUPP;
1931 
1932 	switch (cmd->cmd) {
1933 	case ETHTOOL_SRXCLSRLINS:
1934 		ret = cxgb4_ntuple_set_filter(dev, cmd);
1935 		break;
1936 	case ETHTOOL_SRXCLSRLDEL:
1937 		ret = cxgb4_ntuple_del_filter(dev, cmd);
1938 		break;
1939 	default:
1940 		break;
1941 	}
1942 
1943 	return ret;
1944 }
1945 
1946 static int set_dump(struct net_device *dev, struct ethtool_dump *eth_dump)
1947 {
1948 	struct adapter *adapter = netdev2adap(dev);
1949 	u32 len = 0;
1950 
1951 	len = sizeof(struct cudbg_hdr) +
1952 	      sizeof(struct cudbg_entity_hdr) * CUDBG_MAX_ENTITY;
1953 	len += cxgb4_get_dump_length(adapter, eth_dump->flag);
1954 
1955 	adapter->eth_dump.flag = eth_dump->flag;
1956 	adapter->eth_dump.len = len;
1957 	return 0;
1958 }
1959 
1960 static int get_dump_flag(struct net_device *dev, struct ethtool_dump *eth_dump)
1961 {
1962 	struct adapter *adapter = netdev2adap(dev);
1963 
1964 	eth_dump->flag = adapter->eth_dump.flag;
1965 	eth_dump->len = adapter->eth_dump.len;
1966 	eth_dump->version = adapter->eth_dump.version;
1967 	return 0;
1968 }
1969 
1970 static int get_dump_data(struct net_device *dev, struct ethtool_dump *eth_dump,
1971 			 void *buf)
1972 {
1973 	struct adapter *adapter = netdev2adap(dev);
1974 	u32 len = 0;
1975 	int ret = 0;
1976 
1977 	if (adapter->eth_dump.flag == CXGB4_ETH_DUMP_NONE)
1978 		return -ENOENT;
1979 
1980 	len = sizeof(struct cudbg_hdr) +
1981 	      sizeof(struct cudbg_entity_hdr) * CUDBG_MAX_ENTITY;
1982 	len += cxgb4_get_dump_length(adapter, adapter->eth_dump.flag);
1983 	if (eth_dump->len < len)
1984 		return -ENOMEM;
1985 
1986 	ret = cxgb4_cudbg_collect(adapter, buf, &len, adapter->eth_dump.flag);
1987 	if (ret)
1988 		return ret;
1989 
1990 	eth_dump->flag = adapter->eth_dump.flag;
1991 	eth_dump->len = len;
1992 	eth_dump->version = adapter->eth_dump.version;
1993 	return 0;
1994 }
1995 
1996 static bool cxgb4_fw_mod_type_info_available(unsigned int fw_mod_type)
1997 {
1998 	/* Read port module EEPROM as long as it is plugged-in and
1999 	 * safe to read.
2000 	 */
2001 	return (fw_mod_type != FW_PORT_MOD_TYPE_NONE &&
2002 		fw_mod_type != FW_PORT_MOD_TYPE_ERROR);
2003 }
2004 
2005 static int cxgb4_get_module_info(struct net_device *dev,
2006 				 struct ethtool_modinfo *modinfo)
2007 {
2008 	struct port_info *pi = netdev_priv(dev);
2009 	u8 sff8472_comp, sff_diag_type, sff_rev;
2010 	struct adapter *adapter = pi->adapter;
2011 	int ret;
2012 
2013 	if (!cxgb4_fw_mod_type_info_available(pi->mod_type))
2014 		return -EINVAL;
2015 
2016 	switch (pi->port_type) {
2017 	case FW_PORT_TYPE_SFP:
2018 	case FW_PORT_TYPE_QSA:
2019 	case FW_PORT_TYPE_SFP28:
2020 		ret = t4_i2c_rd(adapter, adapter->mbox, pi->tx_chan,
2021 				I2C_DEV_ADDR_A0, SFF_8472_COMP_ADDR,
2022 				SFF_8472_COMP_LEN, &sff8472_comp);
2023 		if (ret)
2024 			return ret;
2025 		ret = t4_i2c_rd(adapter, adapter->mbox, pi->tx_chan,
2026 				I2C_DEV_ADDR_A0, SFP_DIAG_TYPE_ADDR,
2027 				SFP_DIAG_TYPE_LEN, &sff_diag_type);
2028 		if (ret)
2029 			return ret;
2030 
2031 		if (!sff8472_comp || (sff_diag_type & SFP_DIAG_ADDRMODE)) {
2032 			modinfo->type = ETH_MODULE_SFF_8079;
2033 			modinfo->eeprom_len = ETH_MODULE_SFF_8079_LEN;
2034 		} else {
2035 			modinfo->type = ETH_MODULE_SFF_8472;
2036 			if (sff_diag_type & SFP_DIAG_IMPLEMENTED)
2037 				modinfo->eeprom_len = ETH_MODULE_SFF_8472_LEN;
2038 			else
2039 				modinfo->eeprom_len = ETH_MODULE_SFF_8472_LEN / 2;
2040 		}
2041 		break;
2042 
2043 	case FW_PORT_TYPE_QSFP:
2044 	case FW_PORT_TYPE_QSFP_10G:
2045 	case FW_PORT_TYPE_CR_QSFP:
2046 	case FW_PORT_TYPE_CR2_QSFP:
2047 	case FW_PORT_TYPE_CR4_QSFP:
2048 		ret = t4_i2c_rd(adapter, adapter->mbox, pi->tx_chan,
2049 				I2C_DEV_ADDR_A0, SFF_REV_ADDR,
2050 				SFF_REV_LEN, &sff_rev);
2051 		/* For QSFP type ports, revision value >= 3
2052 		 * means the SFP is 8636 compliant.
2053 		 */
2054 		if (ret)
2055 			return ret;
2056 		if (sff_rev >= 0x3) {
2057 			modinfo->type = ETH_MODULE_SFF_8636;
2058 			modinfo->eeprom_len = ETH_MODULE_SFF_8636_LEN;
2059 		} else {
2060 			modinfo->type = ETH_MODULE_SFF_8436;
2061 			modinfo->eeprom_len = ETH_MODULE_SFF_8436_LEN;
2062 		}
2063 		break;
2064 
2065 	default:
2066 		return -EINVAL;
2067 	}
2068 
2069 	return 0;
2070 }
2071 
2072 static int cxgb4_get_module_eeprom(struct net_device *dev,
2073 				   struct ethtool_eeprom *eprom, u8 *data)
2074 {
2075 	int ret = 0, offset = eprom->offset, len = eprom->len;
2076 	struct port_info *pi = netdev_priv(dev);
2077 	struct adapter *adapter = pi->adapter;
2078 
2079 	memset(data, 0, eprom->len);
2080 	if (offset + len <= I2C_PAGE_SIZE)
2081 		return t4_i2c_rd(adapter, adapter->mbox, pi->tx_chan,
2082 				 I2C_DEV_ADDR_A0, offset, len, data);
2083 
2084 	/* offset + len spans 0xa0 and 0xa1 pages */
2085 	if (offset <= I2C_PAGE_SIZE) {
2086 		/* read 0xa0 page */
2087 		len = I2C_PAGE_SIZE - offset;
2088 		ret =  t4_i2c_rd(adapter, adapter->mbox, pi->tx_chan,
2089 				 I2C_DEV_ADDR_A0, offset, len, data);
2090 		if (ret)
2091 			return ret;
2092 		offset = I2C_PAGE_SIZE;
2093 		/* Remaining bytes to be read from second page =
2094 		 * Total length - bytes read from first page
2095 		 */
2096 		len = eprom->len - len;
2097 	}
2098 	/* Read additional optical diagnostics from page 0xa2 if supported */
2099 	return t4_i2c_rd(adapter, adapter->mbox, pi->tx_chan, I2C_DEV_ADDR_A2,
2100 			 offset, len, &data[eprom->len - len]);
2101 }
2102 
2103 static u32 cxgb4_get_priv_flags(struct net_device *netdev)
2104 {
2105 	struct port_info *pi = netdev_priv(netdev);
2106 	struct adapter *adapter = pi->adapter;
2107 
2108 	return (adapter->eth_flags | pi->eth_flags);
2109 }
2110 
2111 /**
2112  *	set_flags - set/unset specified flags if passed in new_flags
2113  *	@cur_flags: pointer to current flags
2114  *	@new_flags: new incoming flags
2115  *	@flags: set of flags to set/unset
2116  */
2117 static inline void set_flags(u32 *cur_flags, u32 new_flags, u32 flags)
2118 {
2119 	*cur_flags = (*cur_flags & ~flags) | (new_flags & flags);
2120 }
2121 
2122 static int cxgb4_set_priv_flags(struct net_device *netdev, u32 flags)
2123 {
2124 	struct port_info *pi = netdev_priv(netdev);
2125 	struct adapter *adapter = pi->adapter;
2126 
2127 	set_flags(&adapter->eth_flags, flags, PRIV_FLAGS_ADAP);
2128 	set_flags(&pi->eth_flags, flags, PRIV_FLAGS_PORT);
2129 
2130 	return 0;
2131 }
2132 
2133 static void cxgb4_lb_test(struct net_device *netdev, u64 *lb_status)
2134 {
2135 	int dev_state = netif_running(netdev);
2136 
2137 	if (dev_state) {
2138 		netif_tx_stop_all_queues(netdev);
2139 		netif_carrier_off(netdev);
2140 	}
2141 
2142 	*lb_status = cxgb4_selftest_lb_pkt(netdev);
2143 
2144 	if (dev_state) {
2145 		netif_tx_start_all_queues(netdev);
2146 		netif_carrier_on(netdev);
2147 	}
2148 }
2149 
2150 static void cxgb4_self_test(struct net_device *netdev,
2151 			    struct ethtool_test *eth_test, u64 *data)
2152 {
2153 	struct port_info *pi = netdev_priv(netdev);
2154 	struct adapter *adap = pi->adapter;
2155 
2156 	memset(data, 0, sizeof(u64) * CXGB4_ETHTOOL_MAX_TEST);
2157 
2158 	if (!(adap->flags & CXGB4_FULL_INIT_DONE) ||
2159 	    !(adap->flags & CXGB4_FW_OK)) {
2160 		eth_test->flags |= ETH_TEST_FL_FAILED;
2161 		return;
2162 	}
2163 
2164 	if (eth_test->flags & ETH_TEST_FL_OFFLINE)
2165 		cxgb4_lb_test(netdev, &data[CXGB4_ETHTOOL_LB_TEST]);
2166 
2167 	if (data[CXGB4_ETHTOOL_LB_TEST])
2168 		eth_test->flags |= ETH_TEST_FL_FAILED;
2169 }
2170 
2171 static const struct ethtool_ops cxgb_ethtool_ops = {
2172 	.supported_coalesce_params = ETHTOOL_COALESCE_USECS |
2173 				     ETHTOOL_COALESCE_RX_MAX_FRAMES |
2174 				     ETHTOOL_COALESCE_TX_USECS_IRQ |
2175 				     ETHTOOL_COALESCE_USE_ADAPTIVE_RX,
2176 	.get_link_ksettings = get_link_ksettings,
2177 	.set_link_ksettings = set_link_ksettings,
2178 	.get_fecparam      = get_fecparam,
2179 	.set_fecparam      = set_fecparam,
2180 	.get_drvinfo       = get_drvinfo,
2181 	.get_msglevel      = get_msglevel,
2182 	.set_msglevel      = set_msglevel,
2183 	.get_ringparam     = get_sge_param,
2184 	.set_ringparam     = set_sge_param,
2185 	.get_coalesce      = get_coalesce,
2186 	.set_coalesce      = set_coalesce,
2187 	.get_eeprom_len    = get_eeprom_len,
2188 	.get_eeprom        = get_eeprom,
2189 	.set_eeprom        = set_eeprom,
2190 	.get_pauseparam    = get_pauseparam,
2191 	.set_pauseparam    = set_pauseparam,
2192 	.get_link          = ethtool_op_get_link,
2193 	.get_strings       = get_strings,
2194 	.set_phys_id       = identify_port,
2195 	.nway_reset        = restart_autoneg,
2196 	.get_sset_count    = get_sset_count,
2197 	.get_ethtool_stats = get_stats,
2198 	.get_regs_len      = get_regs_len,
2199 	.get_regs          = get_regs,
2200 	.get_rxnfc         = get_rxnfc,
2201 	.set_rxnfc         = set_rxnfc,
2202 	.get_rxfh_indir_size = get_rss_table_size,
2203 	.get_rxfh	   = get_rss_table,
2204 	.set_rxfh	   = set_rss_table,
2205 	.self_test	   = cxgb4_self_test,
2206 	.flash_device      = set_flash,
2207 	.get_ts_info       = get_ts_info,
2208 	.set_dump          = set_dump,
2209 	.get_dump_flag     = get_dump_flag,
2210 	.get_dump_data     = get_dump_data,
2211 	.get_module_info   = cxgb4_get_module_info,
2212 	.get_module_eeprom = cxgb4_get_module_eeprom,
2213 	.get_priv_flags    = cxgb4_get_priv_flags,
2214 	.set_priv_flags    = cxgb4_set_priv_flags,
2215 };
2216 
2217 void cxgb4_cleanup_ethtool_filters(struct adapter *adap)
2218 {
2219 	struct cxgb4_ethtool_filter_info *eth_filter_info;
2220 	u8 i;
2221 
2222 	if (!adap->ethtool_filters)
2223 		return;
2224 
2225 	eth_filter_info = adap->ethtool_filters->port;
2226 
2227 	if (eth_filter_info) {
2228 		for (i = 0; i < adap->params.nports; i++) {
2229 			kvfree(eth_filter_info[i].loc_array);
2230 			kfree(eth_filter_info[i].bmap);
2231 		}
2232 		kfree(eth_filter_info);
2233 	}
2234 
2235 	kfree(adap->ethtool_filters);
2236 }
2237 
2238 int cxgb4_init_ethtool_filters(struct adapter *adap)
2239 {
2240 	struct cxgb4_ethtool_filter_info *eth_filter_info;
2241 	struct cxgb4_ethtool_filter *eth_filter;
2242 	struct tid_info *tids = &adap->tids;
2243 	u32 nentries, i;
2244 	int ret;
2245 
2246 	eth_filter = kzalloc(sizeof(*eth_filter), GFP_KERNEL);
2247 	if (!eth_filter)
2248 		return -ENOMEM;
2249 
2250 	eth_filter_info = kcalloc(adap->params.nports,
2251 				  sizeof(*eth_filter_info),
2252 				  GFP_KERNEL);
2253 	if (!eth_filter_info) {
2254 		ret = -ENOMEM;
2255 		goto free_eth_filter;
2256 	}
2257 
2258 	eth_filter->port = eth_filter_info;
2259 
2260 	nentries = tids->nhpftids + tids->nftids;
2261 	if (is_hashfilter(adap))
2262 		nentries += tids->nhash +
2263 			    (adap->tids.stid_base - adap->tids.tid_base);
2264 	eth_filter->nentries = nentries;
2265 
2266 	for (i = 0; i < adap->params.nports; i++) {
2267 		eth_filter->port[i].loc_array = kvzalloc(nentries, GFP_KERNEL);
2268 		if (!eth_filter->port[i].loc_array) {
2269 			ret = -ENOMEM;
2270 			goto free_eth_finfo;
2271 		}
2272 
2273 		eth_filter->port[i].bmap = kcalloc(BITS_TO_LONGS(nentries),
2274 						   sizeof(unsigned long),
2275 						   GFP_KERNEL);
2276 		if (!eth_filter->port[i].bmap) {
2277 			ret = -ENOMEM;
2278 			goto free_eth_finfo;
2279 		}
2280 	}
2281 
2282 	adap->ethtool_filters = eth_filter;
2283 	return 0;
2284 
2285 free_eth_finfo:
2286 	while (i-- > 0) {
2287 		kfree(eth_filter->port[i].bmap);
2288 		kvfree(eth_filter->port[i].loc_array);
2289 	}
2290 	kfree(eth_filter_info);
2291 
2292 free_eth_filter:
2293 	kfree(eth_filter);
2294 
2295 	return ret;
2296 }
2297 
2298 void cxgb4_set_ethtool_ops(struct net_device *netdev)
2299 {
2300 	netdev->ethtool_ops = &cxgb_ethtool_ops;
2301 }
2302