xref: /linux/drivers/net/ethernet/mellanox/mlx4/en_dcb_nl.c (revision 827634added7f38b7d724cab1dccdb2b004c13c3)
1 /*
2  * Copyright (c) 2011 Mellanox Technologies. All rights reserved.
3  *
4  * This software is available to you under a choice of one of two
5  * licenses.  You may choose to be licensed under the terms of the GNU
6  * General Public License (GPL) Version 2, available from the file
7  * COPYING in the main directory of this source tree, or the
8  * OpenIB.org BSD license below:
9  *
10  *     Redistribution and use in source and binary forms, with or
11  *     without modification, are permitted provided that the following
12  *     conditions are met:
13  *
14  *      - Redistributions of source code must retain the above
15  *        copyright notice, this list of conditions and the following
16  *        disclaimer.
17  *
18  *      - Redistributions in binary form must reproduce the above
19  *        copyright notice, this list of conditions and the following
20  *        disclaimer in the documentation and/or other materials
21  *        provided with the distribution.
22  *
23  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
24  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
25  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
26  * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
27  * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
28  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
29  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
30  * SOFTWARE.
31  *
32  */
33 
34 #include <linux/dcbnl.h>
35 #include <linux/math64.h>
36 
37 #include "mlx4_en.h"
38 #include "fw_qos.h"
39 
40 /* Definitions for QCN
41  */
42 
43 struct mlx4_congestion_control_mb_prio_802_1_qau_params {
44 	__be32 modify_enable_high;
45 	__be32 modify_enable_low;
46 	__be32 reserved1;
47 	__be32 extended_enable;
48 	__be32 rppp_max_rps;
49 	__be32 rpg_time_reset;
50 	__be32 rpg_byte_reset;
51 	__be32 rpg_threshold;
52 	__be32 rpg_max_rate;
53 	__be32 rpg_ai_rate;
54 	__be32 rpg_hai_rate;
55 	__be32 rpg_gd;
56 	__be32 rpg_min_dec_fac;
57 	__be32 rpg_min_rate;
58 	__be32 max_time_rise;
59 	__be32 max_byte_rise;
60 	__be32 max_qdelta;
61 	__be32 min_qoffset;
62 	__be32 gd_coefficient;
63 	__be32 reserved2[5];
64 	__be32 cp_sample_base;
65 	__be32 reserved3[39];
66 };
67 
68 struct mlx4_congestion_control_mb_prio_802_1_qau_statistics {
69 	__be64 rppp_rp_centiseconds;
70 	__be32 reserved1;
71 	__be32 ignored_cnm;
72 	__be32 rppp_created_rps;
73 	__be32 estimated_total_rate;
74 	__be32 max_active_rate_limiter_index;
75 	__be32 dropped_cnms_busy_fw;
76 	__be32 reserved2;
77 	__be32 cnms_handled_successfully;
78 	__be32 min_total_limiters_rate;
79 	__be32 max_total_limiters_rate;
80 	__be32 reserved3[4];
81 };
82 
83 static int mlx4_en_dcbnl_ieee_getets(struct net_device *dev,
84 				   struct ieee_ets *ets)
85 {
86 	struct mlx4_en_priv *priv = netdev_priv(dev);
87 	struct ieee_ets *my_ets = &priv->ets;
88 
89 	/* No IEEE PFC settings available */
90 	if (!my_ets)
91 		return -EINVAL;
92 
93 	ets->ets_cap = IEEE_8021QAZ_MAX_TCS;
94 	ets->cbs = my_ets->cbs;
95 	memcpy(ets->tc_tx_bw, my_ets->tc_tx_bw, sizeof(ets->tc_tx_bw));
96 	memcpy(ets->tc_tsa, my_ets->tc_tsa, sizeof(ets->tc_tsa));
97 	memcpy(ets->prio_tc, my_ets->prio_tc, sizeof(ets->prio_tc));
98 
99 	return 0;
100 }
101 
102 static int mlx4_en_ets_validate(struct mlx4_en_priv *priv, struct ieee_ets *ets)
103 {
104 	int i;
105 	int total_ets_bw = 0;
106 	int has_ets_tc = 0;
107 
108 	for (i = 0; i < IEEE_8021QAZ_MAX_TCS; i++) {
109 		if (ets->prio_tc[i] >= MLX4_EN_NUM_UP) {
110 			en_err(priv, "Bad priority in UP <=> TC mapping. TC: %d, UP: %d\n",
111 					i, ets->prio_tc[i]);
112 			return -EINVAL;
113 		}
114 
115 		switch (ets->tc_tsa[i]) {
116 		case IEEE_8021QAZ_TSA_STRICT:
117 			break;
118 		case IEEE_8021QAZ_TSA_ETS:
119 			has_ets_tc = 1;
120 			total_ets_bw += ets->tc_tx_bw[i];
121 			break;
122 		default:
123 			en_err(priv, "TC[%d]: Not supported TSA: %d\n",
124 					i, ets->tc_tsa[i]);
125 			return -ENOTSUPP;
126 		}
127 	}
128 
129 	if (has_ets_tc && total_ets_bw != MLX4_EN_BW_MAX) {
130 		en_err(priv, "Bad ETS BW sum: %d. Should be exactly 100%%\n",
131 				total_ets_bw);
132 		return -EINVAL;
133 	}
134 
135 	return 0;
136 }
137 
138 static int mlx4_en_config_port_scheduler(struct mlx4_en_priv *priv,
139 		struct ieee_ets *ets, u16 *ratelimit)
140 {
141 	struct mlx4_en_dev *mdev = priv->mdev;
142 	int num_strict = 0;
143 	int i;
144 	__u8 tc_tx_bw[IEEE_8021QAZ_MAX_TCS] = { 0 };
145 	__u8 pg[IEEE_8021QAZ_MAX_TCS] = { 0 };
146 
147 	ets = ets ?: &priv->ets;
148 	ratelimit = ratelimit ?: priv->maxrate;
149 
150 	/* higher TC means higher priority => lower pg */
151 	for (i = IEEE_8021QAZ_MAX_TCS - 1; i >= 0; i--) {
152 		switch (ets->tc_tsa[i]) {
153 		case IEEE_8021QAZ_TSA_STRICT:
154 			pg[i] = num_strict++;
155 			tc_tx_bw[i] = MLX4_EN_BW_MAX;
156 			break;
157 		case IEEE_8021QAZ_TSA_ETS:
158 			pg[i] = MLX4_EN_TC_ETS;
159 			tc_tx_bw[i] = ets->tc_tx_bw[i] ?: MLX4_EN_BW_MIN;
160 			break;
161 		}
162 	}
163 
164 	return mlx4_SET_PORT_SCHEDULER(mdev->dev, priv->port, tc_tx_bw, pg,
165 			ratelimit);
166 }
167 
168 static int
169 mlx4_en_dcbnl_ieee_setets(struct net_device *dev, struct ieee_ets *ets)
170 {
171 	struct mlx4_en_priv *priv = netdev_priv(dev);
172 	struct mlx4_en_dev *mdev = priv->mdev;
173 	int err;
174 
175 	err = mlx4_en_ets_validate(priv, ets);
176 	if (err)
177 		return err;
178 
179 	err = mlx4_SET_PORT_PRIO2TC(mdev->dev, priv->port, ets->prio_tc);
180 	if (err)
181 		return err;
182 
183 	err = mlx4_en_config_port_scheduler(priv, ets, NULL);
184 	if (err)
185 		return err;
186 
187 	memcpy(&priv->ets, ets, sizeof(priv->ets));
188 
189 	return 0;
190 }
191 
192 static int mlx4_en_dcbnl_ieee_getpfc(struct net_device *dev,
193 		struct ieee_pfc *pfc)
194 {
195 	struct mlx4_en_priv *priv = netdev_priv(dev);
196 
197 	pfc->pfc_cap = IEEE_8021QAZ_MAX_TCS;
198 	pfc->pfc_en = priv->prof->tx_ppp;
199 
200 	return 0;
201 }
202 
203 static int mlx4_en_dcbnl_ieee_setpfc(struct net_device *dev,
204 		struct ieee_pfc *pfc)
205 {
206 	struct mlx4_en_priv *priv = netdev_priv(dev);
207 	struct mlx4_en_port_profile *prof = priv->prof;
208 	struct mlx4_en_dev *mdev = priv->mdev;
209 	int err;
210 
211 	en_dbg(DRV, priv, "cap: 0x%x en: 0x%x mbc: 0x%x delay: %d\n",
212 			pfc->pfc_cap,
213 			pfc->pfc_en,
214 			pfc->mbc,
215 			pfc->delay);
216 
217 	prof->rx_pause = !pfc->pfc_en;
218 	prof->tx_pause = !pfc->pfc_en;
219 	prof->rx_ppp = pfc->pfc_en;
220 	prof->tx_ppp = pfc->pfc_en;
221 
222 	err = mlx4_SET_PORT_general(mdev->dev, priv->port,
223 				    priv->rx_skb_size + ETH_FCS_LEN,
224 				    prof->tx_pause,
225 				    prof->tx_ppp,
226 				    prof->rx_pause,
227 				    prof->rx_ppp);
228 	if (err)
229 		en_err(priv, "Failed setting pause params\n");
230 	else
231 		mlx4_en_update_pfc_stats_bitmap(mdev->dev, &priv->stats_bitmap,
232 						prof->rx_ppp, prof->rx_pause,
233 						prof->tx_ppp, prof->tx_pause);
234 
235 	return err;
236 }
237 
238 static u8 mlx4_en_dcbnl_getdcbx(struct net_device *dev)
239 {
240 	return DCB_CAP_DCBX_HOST | DCB_CAP_DCBX_VER_IEEE;
241 }
242 
243 static u8 mlx4_en_dcbnl_setdcbx(struct net_device *dev, u8 mode)
244 {
245 	if ((mode & DCB_CAP_DCBX_LLD_MANAGED) ||
246 	    (mode & DCB_CAP_DCBX_VER_CEE) ||
247 	    !(mode & DCB_CAP_DCBX_VER_IEEE) ||
248 	    !(mode & DCB_CAP_DCBX_HOST))
249 		return 1;
250 
251 	return 0;
252 }
253 
254 #define MLX4_RATELIMIT_UNITS_IN_KB 100000 /* rate-limit HW unit in Kbps */
255 static int mlx4_en_dcbnl_ieee_getmaxrate(struct net_device *dev,
256 				   struct ieee_maxrate *maxrate)
257 {
258 	struct mlx4_en_priv *priv = netdev_priv(dev);
259 	int i;
260 
261 	for (i = 0; i < IEEE_8021QAZ_MAX_TCS; i++)
262 		maxrate->tc_maxrate[i] =
263 			priv->maxrate[i] * MLX4_RATELIMIT_UNITS_IN_KB;
264 
265 	return 0;
266 }
267 
268 static int mlx4_en_dcbnl_ieee_setmaxrate(struct net_device *dev,
269 		struct ieee_maxrate *maxrate)
270 {
271 	struct mlx4_en_priv *priv = netdev_priv(dev);
272 	u16 tmp[IEEE_8021QAZ_MAX_TCS];
273 	int i, err;
274 
275 	for (i = 0; i < IEEE_8021QAZ_MAX_TCS; i++) {
276 		/* Convert from Kbps into HW units, rounding result up.
277 		 * Setting to 0, means unlimited BW.
278 		 */
279 		tmp[i] = div_u64(maxrate->tc_maxrate[i] +
280 				 MLX4_RATELIMIT_UNITS_IN_KB - 1,
281 				 MLX4_RATELIMIT_UNITS_IN_KB);
282 	}
283 
284 	err = mlx4_en_config_port_scheduler(priv, NULL, tmp);
285 	if (err)
286 		return err;
287 
288 	memcpy(priv->maxrate, tmp, sizeof(priv->maxrate));
289 
290 	return 0;
291 }
292 
293 #define RPG_ENABLE_BIT	31
294 #define CN_TAG_BIT	30
295 
296 static int mlx4_en_dcbnl_ieee_getqcn(struct net_device *dev,
297 				     struct ieee_qcn *qcn)
298 {
299 	struct mlx4_en_priv *priv = netdev_priv(dev);
300 	struct mlx4_congestion_control_mb_prio_802_1_qau_params *hw_qcn;
301 	struct mlx4_cmd_mailbox *mailbox_out = NULL;
302 	u64 mailbox_in_dma = 0;
303 	u32 inmod = 0;
304 	int i, err;
305 
306 	if (!(priv->mdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_QCN))
307 		return -EOPNOTSUPP;
308 
309 	mailbox_out = mlx4_alloc_cmd_mailbox(priv->mdev->dev);
310 	if (IS_ERR(mailbox_out))
311 		return -ENOMEM;
312 	hw_qcn =
313 	(struct mlx4_congestion_control_mb_prio_802_1_qau_params *)
314 	mailbox_out->buf;
315 
316 	for (i = 0; i < IEEE_8021QAZ_MAX_TCS; i++) {
317 		inmod = priv->port | ((1 << i) << 8) |
318 			 (MLX4_CTRL_ALGO_802_1_QAU_REACTION_POINT << 16);
319 		err = mlx4_cmd_box(priv->mdev->dev, mailbox_in_dma,
320 				   mailbox_out->dma,
321 				   inmod, MLX4_CONGESTION_CONTROL_GET_PARAMS,
322 				   MLX4_CMD_CONGESTION_CTRL_OPCODE,
323 				   MLX4_CMD_TIME_CLASS_C,
324 				   MLX4_CMD_NATIVE);
325 		if (err) {
326 			mlx4_free_cmd_mailbox(priv->mdev->dev, mailbox_out);
327 			return err;
328 		}
329 
330 		qcn->rpg_enable[i] =
331 			be32_to_cpu(hw_qcn->extended_enable) >> RPG_ENABLE_BIT;
332 		qcn->rppp_max_rps[i] =
333 			be32_to_cpu(hw_qcn->rppp_max_rps);
334 		qcn->rpg_time_reset[i] =
335 			be32_to_cpu(hw_qcn->rpg_time_reset);
336 		qcn->rpg_byte_reset[i] =
337 			be32_to_cpu(hw_qcn->rpg_byte_reset);
338 		qcn->rpg_threshold[i] =
339 			be32_to_cpu(hw_qcn->rpg_threshold);
340 		qcn->rpg_max_rate[i] =
341 			be32_to_cpu(hw_qcn->rpg_max_rate);
342 		qcn->rpg_ai_rate[i] =
343 			be32_to_cpu(hw_qcn->rpg_ai_rate);
344 		qcn->rpg_hai_rate[i] =
345 			be32_to_cpu(hw_qcn->rpg_hai_rate);
346 		qcn->rpg_gd[i] =
347 			be32_to_cpu(hw_qcn->rpg_gd);
348 		qcn->rpg_min_dec_fac[i] =
349 			be32_to_cpu(hw_qcn->rpg_min_dec_fac);
350 		qcn->rpg_min_rate[i] =
351 			be32_to_cpu(hw_qcn->rpg_min_rate);
352 		qcn->cndd_state_machine[i] =
353 			priv->cndd_state[i];
354 	}
355 	mlx4_free_cmd_mailbox(priv->mdev->dev, mailbox_out);
356 	return 0;
357 }
358 
359 static int mlx4_en_dcbnl_ieee_setqcn(struct net_device *dev,
360 				     struct ieee_qcn *qcn)
361 {
362 	struct mlx4_en_priv *priv = netdev_priv(dev);
363 	struct mlx4_congestion_control_mb_prio_802_1_qau_params *hw_qcn;
364 	struct mlx4_cmd_mailbox *mailbox_in = NULL;
365 	u64 mailbox_in_dma = 0;
366 	u32 inmod = 0;
367 	int i, err;
368 #define MODIFY_ENABLE_HIGH_MASK 0xc0000000
369 #define MODIFY_ENABLE_LOW_MASK 0xffc00000
370 
371 	if (!(priv->mdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_QCN))
372 		return -EOPNOTSUPP;
373 
374 	mailbox_in = mlx4_alloc_cmd_mailbox(priv->mdev->dev);
375 	if (IS_ERR(mailbox_in))
376 		return -ENOMEM;
377 
378 	mailbox_in_dma = mailbox_in->dma;
379 	hw_qcn =
380 	(struct mlx4_congestion_control_mb_prio_802_1_qau_params *)mailbox_in->buf;
381 	for (i = 0; i < IEEE_8021QAZ_MAX_TCS; i++) {
382 		inmod = priv->port | ((1 << i) << 8) |
383 			 (MLX4_CTRL_ALGO_802_1_QAU_REACTION_POINT << 16);
384 
385 		/* Before updating QCN parameter,
386 		 * need to set it's modify enable bit to 1
387 		 */
388 
389 		hw_qcn->modify_enable_high = cpu_to_be32(
390 						MODIFY_ENABLE_HIGH_MASK);
391 		hw_qcn->modify_enable_low = cpu_to_be32(MODIFY_ENABLE_LOW_MASK);
392 
393 		hw_qcn->extended_enable = cpu_to_be32(qcn->rpg_enable[i] << RPG_ENABLE_BIT);
394 		hw_qcn->rppp_max_rps = cpu_to_be32(qcn->rppp_max_rps[i]);
395 		hw_qcn->rpg_time_reset = cpu_to_be32(qcn->rpg_time_reset[i]);
396 		hw_qcn->rpg_byte_reset = cpu_to_be32(qcn->rpg_byte_reset[i]);
397 		hw_qcn->rpg_threshold = cpu_to_be32(qcn->rpg_threshold[i]);
398 		hw_qcn->rpg_max_rate = cpu_to_be32(qcn->rpg_max_rate[i]);
399 		hw_qcn->rpg_ai_rate = cpu_to_be32(qcn->rpg_ai_rate[i]);
400 		hw_qcn->rpg_hai_rate = cpu_to_be32(qcn->rpg_hai_rate[i]);
401 		hw_qcn->rpg_gd = cpu_to_be32(qcn->rpg_gd[i]);
402 		hw_qcn->rpg_min_dec_fac = cpu_to_be32(qcn->rpg_min_dec_fac[i]);
403 		hw_qcn->rpg_min_rate = cpu_to_be32(qcn->rpg_min_rate[i]);
404 		priv->cndd_state[i] = qcn->cndd_state_machine[i];
405 		if (qcn->cndd_state_machine[i] == DCB_CNDD_INTERIOR_READY)
406 			hw_qcn->extended_enable |= cpu_to_be32(1 << CN_TAG_BIT);
407 
408 		err = mlx4_cmd(priv->mdev->dev, mailbox_in_dma, inmod,
409 			       MLX4_CONGESTION_CONTROL_SET_PARAMS,
410 			       MLX4_CMD_CONGESTION_CTRL_OPCODE,
411 			       MLX4_CMD_TIME_CLASS_C,
412 			       MLX4_CMD_NATIVE);
413 		if (err) {
414 			mlx4_free_cmd_mailbox(priv->mdev->dev, mailbox_in);
415 			return err;
416 		}
417 	}
418 	mlx4_free_cmd_mailbox(priv->mdev->dev, mailbox_in);
419 	return 0;
420 }
421 
422 static int mlx4_en_dcbnl_ieee_getqcnstats(struct net_device *dev,
423 					  struct ieee_qcn_stats *qcn_stats)
424 {
425 	struct mlx4_en_priv *priv = netdev_priv(dev);
426 	struct mlx4_congestion_control_mb_prio_802_1_qau_statistics *hw_qcn_stats;
427 	struct mlx4_cmd_mailbox *mailbox_out = NULL;
428 	u64 mailbox_in_dma = 0;
429 	u32 inmod = 0;
430 	int i, err;
431 
432 	if (!(priv->mdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_QCN))
433 		return -EOPNOTSUPP;
434 
435 	mailbox_out = mlx4_alloc_cmd_mailbox(priv->mdev->dev);
436 	if (IS_ERR(mailbox_out))
437 		return -ENOMEM;
438 
439 	hw_qcn_stats =
440 	(struct mlx4_congestion_control_mb_prio_802_1_qau_statistics *)
441 	mailbox_out->buf;
442 
443 	for (i = 0; i < IEEE_8021QAZ_MAX_TCS; i++) {
444 		inmod = priv->port | ((1 << i) << 8) |
445 			 (MLX4_CTRL_ALGO_802_1_QAU_REACTION_POINT << 16);
446 		err = mlx4_cmd_box(priv->mdev->dev, mailbox_in_dma,
447 				   mailbox_out->dma, inmod,
448 				   MLX4_CONGESTION_CONTROL_GET_STATISTICS,
449 				   MLX4_CMD_CONGESTION_CTRL_OPCODE,
450 				   MLX4_CMD_TIME_CLASS_C,
451 				   MLX4_CMD_NATIVE);
452 		if (err) {
453 			mlx4_free_cmd_mailbox(priv->mdev->dev, mailbox_out);
454 			return err;
455 		}
456 		qcn_stats->rppp_rp_centiseconds[i] =
457 			be64_to_cpu(hw_qcn_stats->rppp_rp_centiseconds);
458 		qcn_stats->rppp_created_rps[i] =
459 			be32_to_cpu(hw_qcn_stats->rppp_created_rps);
460 	}
461 	mlx4_free_cmd_mailbox(priv->mdev->dev, mailbox_out);
462 	return 0;
463 }
464 
465 const struct dcbnl_rtnl_ops mlx4_en_dcbnl_ops = {
466 	.ieee_getets	= mlx4_en_dcbnl_ieee_getets,
467 	.ieee_setets	= mlx4_en_dcbnl_ieee_setets,
468 	.ieee_getmaxrate = mlx4_en_dcbnl_ieee_getmaxrate,
469 	.ieee_setmaxrate = mlx4_en_dcbnl_ieee_setmaxrate,
470 	.ieee_getpfc	= mlx4_en_dcbnl_ieee_getpfc,
471 	.ieee_setpfc	= mlx4_en_dcbnl_ieee_setpfc,
472 
473 	.getdcbx	= mlx4_en_dcbnl_getdcbx,
474 	.setdcbx	= mlx4_en_dcbnl_setdcbx,
475 	.ieee_getqcn	= mlx4_en_dcbnl_ieee_getqcn,
476 	.ieee_setqcn	= mlx4_en_dcbnl_ieee_setqcn,
477 	.ieee_getqcnstats = mlx4_en_dcbnl_ieee_getqcnstats,
478 };
479 
480 const struct dcbnl_rtnl_ops mlx4_en_dcbnl_pfc_ops = {
481 	.ieee_getpfc	= mlx4_en_dcbnl_ieee_getpfc,
482 	.ieee_setpfc	= mlx4_en_dcbnl_ieee_setpfc,
483 
484 	.getdcbx	= mlx4_en_dcbnl_getdcbx,
485 	.setdcbx	= mlx4_en_dcbnl_setdcbx,
486 };
487