xref: /linux/drivers/net/ethernet/mellanox/mlx5/core/en/reporter_tx.c (revision 8be4d31cb8aaeea27bde4b7ddb26e28a89062ebf)
1 /* SPDX-License-Identifier: GPL-2.0 */
2 /* Copyright (c) 2019 Mellanox Technologies. */
3 
4 #include "health.h"
5 #include "en/ptp.h"
6 #include "en/devlink.h"
7 #include "lib/tout.h"
8 
9 /* Keep this string array consistent with the MLX5E_SQ_STATE_* enums in en.h */
10 static const char * const sq_sw_state_type_name[] = {
11 	[MLX5E_SQ_STATE_ENABLED] = "enabled",
12 	[MLX5E_SQ_STATE_MPWQE] = "mpwqe",
13 	[MLX5E_SQ_STATE_RECOVERING] = "recovering",
14 	[MLX5E_SQ_STATE_IPSEC] = "ipsec",
15 	[MLX5E_SQ_STATE_DIM] = "dim",
16 	[MLX5E_SQ_STATE_PENDING_XSK_TX] = "pending_xsk_tx",
17 	[MLX5E_SQ_STATE_PENDING_TLS_RX_RESYNC] = "pending_tls_rx_resync",
18 };
19 
mlx5e_wait_for_sq_flush(struct mlx5e_txqsq * sq)20 static int mlx5e_wait_for_sq_flush(struct mlx5e_txqsq *sq)
21 {
22 	struct mlx5_core_dev *dev = sq->mdev;
23 	unsigned long exp_time;
24 
25 	exp_time = jiffies + msecs_to_jiffies(mlx5_tout_ms(dev, FLUSH_ON_ERROR));
26 
27 	while (time_before(jiffies, exp_time)) {
28 		if (sq->cc == sq->pc)
29 			return 0;
30 
31 		msleep(20);
32 	}
33 
34 	netdev_err(sq->netdev,
35 		   "Wait for SQ 0x%x flush timeout (sq cc = 0x%x, sq pc = 0x%x)\n",
36 		   sq->sqn, sq->cc, sq->pc);
37 
38 	return -ETIMEDOUT;
39 }
40 
mlx5e_reset_txqsq_cc_pc(struct mlx5e_txqsq * sq)41 static void mlx5e_reset_txqsq_cc_pc(struct mlx5e_txqsq *sq)
42 {
43 	WARN_ONCE(sq->cc != sq->pc,
44 		  "SQ 0x%x: cc (0x%x) != pc (0x%x)\n",
45 		  sq->sqn, sq->cc, sq->pc);
46 	sq->cc = 0;
47 	sq->dma_fifo_cc = 0;
48 	sq->pc = 0;
49 }
50 
mlx5e_health_sq_put_sw_state(struct devlink_fmsg * fmsg,struct mlx5e_txqsq * sq)51 static void mlx5e_health_sq_put_sw_state(struct devlink_fmsg *fmsg, struct mlx5e_txqsq *sq)
52 {
53 	int i;
54 
55 	BUILD_BUG_ON_MSG(ARRAY_SIZE(sq_sw_state_type_name) != MLX5E_NUM_SQ_STATES,
56 			 "sq_sw_state_type_name string array must be consistent with MLX5E_SQ_STATE_* enum in en.h");
57 	mlx5e_health_fmsg_named_obj_nest_start(fmsg, "SW State");
58 
59 	for (i = 0; i < ARRAY_SIZE(sq_sw_state_type_name); ++i)
60 		devlink_fmsg_u32_pair_put(fmsg, sq_sw_state_type_name[i],
61 					  test_bit(i, &sq->state));
62 
63 	mlx5e_health_fmsg_named_obj_nest_end(fmsg);
64 }
65 
mlx5e_tx_reporter_err_cqe_recover(void * ctx)66 static int mlx5e_tx_reporter_err_cqe_recover(void *ctx)
67 {
68 	struct mlx5_core_dev *mdev;
69 	struct net_device *dev;
70 	struct mlx5e_txqsq *sq;
71 	u8 state;
72 	int err;
73 
74 	sq = ctx;
75 	mdev = sq->mdev;
76 	dev = sq->netdev;
77 
78 	if (!test_bit(MLX5E_SQ_STATE_RECOVERING, &sq->state))
79 		return 0;
80 
81 	err = mlx5_core_query_sq_state(mdev, sq->sqn, &state);
82 	if (err) {
83 		netdev_err(dev, "Failed to query SQ 0x%x state. err = %d\n",
84 			   sq->sqn, err);
85 		goto out;
86 	}
87 
88 	if (state != MLX5_SQC_STATE_ERR)
89 		goto out;
90 
91 	mlx5e_tx_disable_queue(sq->txq);
92 
93 	err = mlx5e_wait_for_sq_flush(sq);
94 	if (err)
95 		goto out;
96 
97 	/* At this point, no new packets will arrive from the stack as TXQ is
98 	 * marked with QUEUE_STATE_DRV_XOFF. In addition, NAPI cleared all
99 	 * pending WQEs. SQ can safely reset the SQ.
100 	 */
101 
102 	err = mlx5e_health_sq_to_ready(mdev, dev, sq->sqn);
103 	if (err)
104 		goto out;
105 
106 	mlx5e_reset_txqsq_cc_pc(sq);
107 	sq->stats->recover++;
108 	clear_bit(MLX5E_SQ_STATE_RECOVERING, &sq->state);
109 	mlx5e_activate_txqsq(sq);
110 
111 	if (sq->channel)
112 		mlx5e_trigger_napi_icosq(sq->channel);
113 	else
114 		mlx5e_trigger_napi_sched(sq->cq.napi);
115 
116 	return 0;
117 out:
118 	clear_bit(MLX5E_SQ_STATE_RECOVERING, &sq->state);
119 	return err;
120 }
121 
122 struct mlx5e_tx_timeout_ctx {
123 	struct mlx5e_txqsq *sq;
124 	signed int status;
125 };
126 
mlx5e_tx_reporter_timeout_recover(void * ctx)127 static int mlx5e_tx_reporter_timeout_recover(void *ctx)
128 {
129 	struct mlx5e_tx_timeout_ctx *to_ctx;
130 	struct mlx5e_priv *priv;
131 	struct mlx5_eq_comp *eq;
132 	struct mlx5e_txqsq *sq;
133 	int err;
134 
135 	to_ctx = ctx;
136 	sq = to_ctx->sq;
137 	eq = sq->cq.mcq.eq;
138 	priv = sq->priv;
139 	err = mlx5e_health_channel_eq_recover(sq->netdev, eq, sq->cq.ch_stats);
140 	if (!err) {
141 		to_ctx->status = 0; /* this sq recovered */
142 		return err;
143 	}
144 
145 	mutex_lock(&priv->state_lock);
146 	err = mlx5e_safe_reopen_channels(priv);
147 	mutex_unlock(&priv->state_lock);
148 	if (!err) {
149 		to_ctx->status = 1; /* all channels recovered */
150 		return err;
151 	}
152 
153 	to_ctx->status = err;
154 	clear_bit(MLX5E_SQ_STATE_ENABLED, &sq->state);
155 	netdev_err(priv->netdev,
156 		   "mlx5e_safe_reopen_channels failed recovering from a tx_timeout, err(%d).\n",
157 		   err);
158 
159 	return err;
160 }
161 
mlx5e_tx_reporter_ptpsq_unhealthy_recover(void * ctx)162 static int mlx5e_tx_reporter_ptpsq_unhealthy_recover(void *ctx)
163 {
164 	struct mlx5e_ptpsq *ptpsq = ctx;
165 	struct mlx5e_channels *chs;
166 	struct net_device *netdev;
167 	struct mlx5e_priv *priv;
168 	int carrier_ok;
169 	int err;
170 
171 	if (!test_bit(MLX5E_SQ_STATE_RECOVERING, &ptpsq->txqsq.state))
172 		return 0;
173 
174 	priv = ptpsq->txqsq.priv;
175 
176 	mutex_lock(&priv->state_lock);
177 	chs = &priv->channels;
178 	netdev = priv->netdev;
179 
180 	carrier_ok = netif_carrier_ok(netdev);
181 	netif_carrier_off(netdev);
182 
183 	mlx5e_deactivate_priv_channels(priv);
184 
185 	mlx5e_ptp_close(chs->ptp);
186 	err = mlx5e_ptp_open(priv, &chs->params, chs->c[0]->lag_port, &chs->ptp);
187 
188 	mlx5e_activate_priv_channels(priv);
189 
190 	/* return carrier back if needed */
191 	if (carrier_ok)
192 		netif_carrier_on(netdev);
193 
194 	mutex_unlock(&priv->state_lock);
195 
196 	return err;
197 }
198 
199 /* state lock cannot be grabbed within this function.
200  * It can cause a dead lock or a read-after-free.
201  */
mlx5e_tx_reporter_recover_from_ctx(struct mlx5e_err_ctx * err_ctx)202 static int mlx5e_tx_reporter_recover_from_ctx(struct mlx5e_err_ctx *err_ctx)
203 {
204 	return err_ctx->recover(err_ctx->ctx);
205 }
206 
mlx5e_tx_reporter_recover(struct devlink_health_reporter * reporter,void * context,struct netlink_ext_ack * extack)207 static int mlx5e_tx_reporter_recover(struct devlink_health_reporter *reporter,
208 				     void *context,
209 				     struct netlink_ext_ack *extack)
210 {
211 	struct mlx5e_priv *priv = devlink_health_reporter_priv(reporter);
212 	struct mlx5e_err_ctx *err_ctx = context;
213 
214 	return err_ctx ? mlx5e_tx_reporter_recover_from_ctx(err_ctx) :
215 			 mlx5e_health_recover_channels(priv);
216 }
217 
218 static void
mlx5e_tx_reporter_build_diagnose_output_sq_common(struct devlink_fmsg * fmsg,struct mlx5e_txqsq * sq,int tc)219 mlx5e_tx_reporter_build_diagnose_output_sq_common(struct devlink_fmsg *fmsg,
220 						  struct mlx5e_txqsq *sq, int tc)
221 {
222 	bool stopped = netif_xmit_stopped(sq->txq);
223 	u8 state;
224 	int err;
225 
226 	devlink_fmsg_u32_pair_put(fmsg, "tc", tc);
227 	devlink_fmsg_u32_pair_put(fmsg, "txq ix", sq->txq_ix);
228 	devlink_fmsg_u32_pair_put(fmsg, "sqn", sq->sqn);
229 
230 	err = mlx5_core_query_sq_state(sq->mdev, sq->sqn, &state);
231 	if (!err)
232 		devlink_fmsg_u8_pair_put(fmsg, "HW state", state);
233 
234 	devlink_fmsg_bool_pair_put(fmsg, "stopped", stopped);
235 	devlink_fmsg_u32_pair_put(fmsg, "cc", sq->cc);
236 	devlink_fmsg_u32_pair_put(fmsg, "pc", sq->pc);
237 	mlx5e_health_sq_put_sw_state(fmsg, sq);
238 	mlx5e_health_cq_diag_fmsg(&sq->cq, fmsg);
239 	mlx5e_health_eq_diag_fmsg(sq->cq.mcq.eq, fmsg);
240 }
241 
242 static void
mlx5e_tx_reporter_build_diagnose_output(struct devlink_fmsg * fmsg,struct mlx5e_txqsq * sq,int tc)243 mlx5e_tx_reporter_build_diagnose_output(struct devlink_fmsg *fmsg,
244 					struct mlx5e_txqsq *sq, int tc)
245 {
246 	devlink_fmsg_obj_nest_start(fmsg);
247 	devlink_fmsg_u32_pair_put(fmsg, "channel ix", sq->ch_ix);
248 	mlx5e_tx_reporter_build_diagnose_output_sq_common(fmsg, sq, tc);
249 	devlink_fmsg_obj_nest_end(fmsg);
250 }
251 
252 static void
mlx5e_tx_reporter_build_diagnose_output_ptpsq(struct devlink_fmsg * fmsg,struct mlx5e_ptpsq * ptpsq,int tc)253 mlx5e_tx_reporter_build_diagnose_output_ptpsq(struct devlink_fmsg *fmsg,
254 					      struct mlx5e_ptpsq *ptpsq, int tc)
255 {
256 	devlink_fmsg_obj_nest_start(fmsg);
257 	devlink_fmsg_string_pair_put(fmsg, "channel", "ptp");
258 	mlx5e_tx_reporter_build_diagnose_output_sq_common(fmsg, &ptpsq->txqsq, tc);
259 	mlx5e_health_fmsg_named_obj_nest_start(fmsg, "Port TS");
260 	mlx5e_health_cq_diag_fmsg(&ptpsq->ts_cq, fmsg);
261 	mlx5e_health_fmsg_named_obj_nest_end(fmsg);
262 	devlink_fmsg_obj_nest_end(fmsg);
263 }
264 
265 static void
mlx5e_tx_reporter_diagnose_generic_txqsq(struct devlink_fmsg * fmsg,struct mlx5e_txqsq * txqsq)266 mlx5e_tx_reporter_diagnose_generic_txqsq(struct devlink_fmsg *fmsg,
267 					 struct mlx5e_txqsq *txqsq)
268 {
269 	bool real_time =  mlx5_is_real_time_sq(txqsq->mdev);
270 	u32 sq_sz = mlx5_wq_cyc_get_size(&txqsq->wq);
271 	u32 sq_stride = MLX5_SEND_WQE_BB;
272 
273 	mlx5e_health_fmsg_named_obj_nest_start(fmsg, "SQ");
274 	devlink_fmsg_u64_pair_put(fmsg, "stride size", sq_stride);
275 	devlink_fmsg_u32_pair_put(fmsg, "size", sq_sz);
276 	devlink_fmsg_string_pair_put(fmsg, "ts_format", real_time ? "RT" : "FRC");
277 	mlx5e_health_cq_common_diag_fmsg(&txqsq->cq, fmsg);
278 	mlx5e_health_fmsg_named_obj_nest_end(fmsg);
279 }
280 
281 static void
mlx5e_tx_reporter_diagnose_generic_tx_port_ts(struct devlink_fmsg * fmsg,struct mlx5e_ptpsq * ptpsq)282 mlx5e_tx_reporter_diagnose_generic_tx_port_ts(struct devlink_fmsg *fmsg,
283 					      struct mlx5e_ptpsq *ptpsq)
284 {
285 	mlx5e_health_fmsg_named_obj_nest_start(fmsg, "Port TS");
286 	mlx5e_health_cq_common_diag_fmsg(&ptpsq->ts_cq, fmsg);
287 	mlx5e_health_fmsg_named_obj_nest_end(fmsg);
288 }
289 
290 static void
mlx5e_tx_reporter_diagnose_common_config(struct devlink_health_reporter * reporter,struct devlink_fmsg * fmsg)291 mlx5e_tx_reporter_diagnose_common_config(struct devlink_health_reporter *reporter,
292 					 struct devlink_fmsg *fmsg)
293 {
294 	struct mlx5e_priv *priv = devlink_health_reporter_priv(reporter);
295 	struct mlx5e_txqsq *generic_sq = priv->txq2sq[0];
296 	struct mlx5e_ptp *ptp_ch = priv->channels.ptp;
297 	struct mlx5e_ptpsq *generic_ptpsq;
298 
299 	mlx5e_health_fmsg_named_obj_nest_start(fmsg, "Common Config");
300 	mlx5e_tx_reporter_diagnose_generic_txqsq(fmsg, generic_sq);
301 
302 	if (!ptp_ch || !test_bit(MLX5E_PTP_STATE_TX, ptp_ch->state))
303 		goto out;
304 
305 	generic_ptpsq = &ptp_ch->ptpsq[0];
306 	mlx5e_health_fmsg_named_obj_nest_start(fmsg, "PTP");
307 	mlx5e_tx_reporter_diagnose_generic_txqsq(fmsg, &generic_ptpsq->txqsq);
308 	mlx5e_tx_reporter_diagnose_generic_tx_port_ts(fmsg, generic_ptpsq);
309 	mlx5e_health_fmsg_named_obj_nest_end(fmsg);
310 out:
311 	mlx5e_health_fmsg_named_obj_nest_end(fmsg);
312 }
313 
314 static void
mlx5e_tx_reporter_diagnose_tis_config(struct devlink_health_reporter * reporter,struct devlink_fmsg * fmsg)315 mlx5e_tx_reporter_diagnose_tis_config(struct devlink_health_reporter *reporter,
316 				      struct devlink_fmsg *fmsg)
317 {
318 	struct mlx5e_priv *priv = devlink_health_reporter_priv(reporter);
319 	u8 num_tc = mlx5e_get_dcb_num_tc(&priv->channels.params);
320 	u32 tc, i, tisn;
321 
322 	devlink_fmsg_arr_pair_nest_start(fmsg, "TIS Config");
323 	for (i = 0; i < mlx5e_get_num_lag_ports(priv->mdev); i++) {
324 		for (tc = 0; tc < num_tc; tc++) {
325 			tisn = mlx5e_profile_get_tisn(priv->mdev, priv,
326 						      priv->profile, i, tc);
327 
328 			devlink_fmsg_obj_nest_start(fmsg);
329 			devlink_fmsg_u32_pair_put(fmsg, "lag port", i);
330 			devlink_fmsg_u32_pair_put(fmsg, "tc", tc);
331 			devlink_fmsg_u32_pair_put(fmsg, "tisn", tisn);
332 			devlink_fmsg_obj_nest_end(fmsg);
333 		}
334 	}
335 	devlink_fmsg_arr_pair_nest_end(fmsg);
336 }
337 
mlx5e_tx_reporter_diagnose(struct devlink_health_reporter * reporter,struct devlink_fmsg * fmsg,struct netlink_ext_ack * extack)338 static int mlx5e_tx_reporter_diagnose(struct devlink_health_reporter *reporter,
339 				      struct devlink_fmsg *fmsg,
340 				      struct netlink_ext_ack *extack)
341 {
342 	struct mlx5e_priv *priv = devlink_health_reporter_priv(reporter);
343 	struct mlx5e_ptp *ptp_ch = priv->channels.ptp;
344 
345 	int i, tc;
346 
347 	mutex_lock(&priv->state_lock);
348 
349 	if (!test_bit(MLX5E_STATE_OPENED, &priv->state))
350 		goto unlock;
351 
352 	mlx5e_tx_reporter_diagnose_common_config(reporter, fmsg);
353 	mlx5e_tx_reporter_diagnose_tis_config(reporter, fmsg);
354 	devlink_fmsg_arr_pair_nest_start(fmsg, "SQs");
355 
356 	for (i = 0; i < priv->channels.num; i++) {
357 		struct mlx5e_channel *c = priv->channels.c[i];
358 
359 		for (tc = 0; tc < mlx5e_get_dcb_num_tc(&priv->channels.params); tc++) {
360 			struct mlx5e_txqsq *sq = &c->sq[tc];
361 
362 			mlx5e_tx_reporter_build_diagnose_output(fmsg, sq, tc);
363 		}
364 	}
365 
366 	if (!ptp_ch || !test_bit(MLX5E_PTP_STATE_TX, ptp_ch->state))
367 		goto close_sqs_nest;
368 
369 	for (tc = 0; tc < mlx5e_get_dcb_num_tc(&priv->channels.params); tc++)
370 		mlx5e_tx_reporter_build_diagnose_output_ptpsq(fmsg,
371 							      &ptp_ch->ptpsq[tc],
372 							      tc);
373 
374 close_sqs_nest:
375 	devlink_fmsg_arr_pair_nest_end(fmsg);
376 unlock:
377 	mutex_unlock(&priv->state_lock);
378 	return 0;
379 }
380 
mlx5e_tx_reporter_dump_sq(struct mlx5e_priv * priv,struct devlink_fmsg * fmsg,void * ctx)381 static int mlx5e_tx_reporter_dump_sq(struct mlx5e_priv *priv, struct devlink_fmsg *fmsg,
382 				     void *ctx)
383 {
384 	struct mlx5_rsc_key key = {};
385 	struct mlx5e_txqsq *sq = ctx;
386 
387 	if (!test_bit(MLX5E_STATE_OPENED, &priv->state))
388 		return 0;
389 
390 	mlx5e_health_fmsg_named_obj_nest_start(fmsg, "SX Slice");
391 	key.size = PAGE_SIZE;
392 	key.rsc = MLX5_SGMT_TYPE_SX_SLICE_ALL;
393 	mlx5e_health_rsc_fmsg_dump(priv, &key, fmsg);
394 	mlx5e_health_fmsg_named_obj_nest_end(fmsg);
395 
396 	mlx5e_health_fmsg_named_obj_nest_start(fmsg, "SQ");
397 	mlx5e_health_fmsg_named_obj_nest_start(fmsg, "QPC");
398 	key.rsc = MLX5_SGMT_TYPE_FULL_QPC;
399 	key.index1 = sq->sqn;
400 	key.num_of_obj1 = 1;
401 	mlx5e_health_rsc_fmsg_dump(priv, &key, fmsg);
402 	mlx5e_health_fmsg_named_obj_nest_end(fmsg);
403 
404 	mlx5e_health_fmsg_named_obj_nest_start(fmsg, "send_buff");
405 	key.rsc = MLX5_SGMT_TYPE_SND_BUFF;
406 	key.num_of_obj2 = MLX5_RSC_DUMP_ALL;
407 	mlx5e_health_rsc_fmsg_dump(priv, &key, fmsg);
408 	mlx5e_health_fmsg_named_obj_nest_end(fmsg);
409 
410 	mlx5e_health_fmsg_named_obj_nest_end(fmsg);
411 
412 	return 0;
413 }
414 
mlx5e_tx_reporter_timeout_dump(struct mlx5e_priv * priv,struct devlink_fmsg * fmsg,void * ctx)415 static int mlx5e_tx_reporter_timeout_dump(struct mlx5e_priv *priv, struct devlink_fmsg *fmsg,
416 					  void *ctx)
417 {
418 	struct mlx5e_tx_timeout_ctx *to_ctx = ctx;
419 
420 	return mlx5e_tx_reporter_dump_sq(priv, fmsg, to_ctx->sq);
421 }
422 
mlx5e_tx_reporter_ptpsq_unhealthy_dump(struct mlx5e_priv * priv,struct devlink_fmsg * fmsg,void * ctx)423 static int mlx5e_tx_reporter_ptpsq_unhealthy_dump(struct mlx5e_priv *priv,
424 						  struct devlink_fmsg *fmsg,
425 						  void *ctx)
426 {
427 	struct mlx5e_ptpsq *ptpsq = ctx;
428 
429 	return mlx5e_tx_reporter_dump_sq(priv, fmsg, &ptpsq->txqsq);
430 }
431 
mlx5e_tx_reporter_dump_all_sqs(struct mlx5e_priv * priv,struct devlink_fmsg * fmsg)432 static int mlx5e_tx_reporter_dump_all_sqs(struct mlx5e_priv *priv,
433 					  struct devlink_fmsg *fmsg)
434 {
435 	struct mlx5e_ptp *ptp_ch = priv->channels.ptp;
436 	struct mlx5_rsc_key key = {};
437 	int i, tc;
438 
439 	if (!test_bit(MLX5E_STATE_OPENED, &priv->state))
440 		return 0;
441 
442 	mlx5e_health_fmsg_named_obj_nest_start(fmsg, "SX Slice");
443 	key.size = PAGE_SIZE;
444 	key.rsc = MLX5_SGMT_TYPE_SX_SLICE_ALL;
445 	mlx5e_health_rsc_fmsg_dump(priv, &key, fmsg);
446 	mlx5e_health_fmsg_named_obj_nest_end(fmsg);
447 	devlink_fmsg_arr_pair_nest_start(fmsg, "SQs");
448 
449 	for (i = 0; i < priv->channels.num; i++) {
450 		struct mlx5e_channel *c = priv->channels.c[i];
451 
452 		for (tc = 0; tc < mlx5e_get_dcb_num_tc(&priv->channels.params); tc++) {
453 			struct mlx5e_txqsq *sq = &c->sq[tc];
454 
455 			mlx5e_health_queue_dump(priv, fmsg, sq->sqn, "SQ");
456 		}
457 	}
458 
459 	if (ptp_ch && test_bit(MLX5E_PTP_STATE_TX, ptp_ch->state)) {
460 		for (tc = 0; tc < mlx5e_get_dcb_num_tc(&priv->channels.params); tc++) {
461 			struct mlx5e_txqsq *sq = &ptp_ch->ptpsq[tc].txqsq;
462 
463 			mlx5e_health_queue_dump(priv, fmsg, sq->sqn, "PTP SQ");
464 		}
465 	}
466 
467 	devlink_fmsg_arr_pair_nest_end(fmsg);
468 	return 0;
469 }
470 
mlx5e_tx_reporter_dump_from_ctx(struct mlx5e_priv * priv,struct mlx5e_err_ctx * err_ctx,struct devlink_fmsg * fmsg)471 static int mlx5e_tx_reporter_dump_from_ctx(struct mlx5e_priv *priv,
472 					   struct mlx5e_err_ctx *err_ctx,
473 					   struct devlink_fmsg *fmsg)
474 {
475 	return err_ctx->dump(priv, fmsg, err_ctx->ctx);
476 }
477 
mlx5e_tx_reporter_dump(struct devlink_health_reporter * reporter,struct devlink_fmsg * fmsg,void * context,struct netlink_ext_ack * extack)478 static int mlx5e_tx_reporter_dump(struct devlink_health_reporter *reporter,
479 				  struct devlink_fmsg *fmsg, void *context,
480 				  struct netlink_ext_ack *extack)
481 {
482 	struct mlx5e_priv *priv = devlink_health_reporter_priv(reporter);
483 	struct mlx5e_err_ctx *err_ctx = context;
484 
485 	return err_ctx ? mlx5e_tx_reporter_dump_from_ctx(priv, err_ctx, fmsg) :
486 			 mlx5e_tx_reporter_dump_all_sqs(priv, fmsg);
487 }
488 
mlx5e_reporter_tx_err_cqe(struct mlx5e_txqsq * sq)489 void mlx5e_reporter_tx_err_cqe(struct mlx5e_txqsq *sq)
490 {
491 	char err_str[MLX5E_REPORTER_PER_Q_MAX_LEN];
492 	struct mlx5e_priv *priv = sq->priv;
493 	struct mlx5e_err_ctx err_ctx = {};
494 
495 	err_ctx.ctx = sq;
496 	err_ctx.recover = mlx5e_tx_reporter_err_cqe_recover;
497 	err_ctx.dump = mlx5e_tx_reporter_dump_sq;
498 	snprintf(err_str, sizeof(err_str), "ERR CQE on SQ: 0x%x", sq->sqn);
499 
500 	mlx5e_health_report(priv, priv->tx_reporter, err_str, &err_ctx);
501 }
502 
mlx5e_reporter_tx_timeout(struct mlx5e_txqsq * sq)503 int mlx5e_reporter_tx_timeout(struct mlx5e_txqsq *sq)
504 {
505 	char err_str[MLX5E_REPORTER_PER_Q_MAX_LEN];
506 	struct mlx5e_tx_timeout_ctx to_ctx = {};
507 	struct mlx5e_priv *priv = sq->priv;
508 	struct mlx5e_err_ctx err_ctx = {};
509 
510 	to_ctx.sq = sq;
511 	err_ctx.ctx = &to_ctx;
512 	err_ctx.recover = mlx5e_tx_reporter_timeout_recover;
513 	err_ctx.dump = mlx5e_tx_reporter_timeout_dump;
514 	snprintf(err_str, sizeof(err_str),
515 		 "TX timeout on queue: %d, SQ: 0x%x, CQ: 0x%x, SQ Cons: 0x%x SQ Prod: 0x%x, usecs since last trans: %u",
516 		 sq->ch_ix, sq->sqn, sq->cq.mcq.cqn, sq->cc, sq->pc,
517 		 jiffies_to_usecs(jiffies - READ_ONCE(sq->txq->trans_start)));
518 
519 	mlx5e_health_report(priv, priv->tx_reporter, err_str, &err_ctx);
520 	return to_ctx.status;
521 }
522 
mlx5e_reporter_tx_ptpsq_unhealthy(struct mlx5e_ptpsq * ptpsq)523 void mlx5e_reporter_tx_ptpsq_unhealthy(struct mlx5e_ptpsq *ptpsq)
524 {
525 	struct mlx5e_ptp_metadata_map *map = &ptpsq->metadata_map;
526 	char err_str[MLX5E_REPORTER_PER_Q_MAX_LEN];
527 	struct mlx5e_txqsq *txqsq = &ptpsq->txqsq;
528 	struct mlx5e_cq *ts_cq = &ptpsq->ts_cq;
529 	struct mlx5e_priv *priv = txqsq->priv;
530 	struct mlx5e_err_ctx err_ctx = {};
531 
532 	err_ctx.ctx = ptpsq;
533 	err_ctx.recover = mlx5e_tx_reporter_ptpsq_unhealthy_recover;
534 	err_ctx.dump = mlx5e_tx_reporter_ptpsq_unhealthy_dump;
535 	snprintf(err_str, sizeof(err_str),
536 		 "Unhealthy TX port TS queue: %d, SQ: 0x%x, CQ: 0x%x, Undelivered CQEs: %u Map Capacity: %u",
537 		 txqsq->ch_ix, txqsq->sqn, ts_cq->mcq.cqn, map->undelivered_counter, map->capacity);
538 
539 	mlx5e_health_report(priv, priv->tx_reporter, err_str, &err_ctx);
540 }
541 
542 static const struct devlink_health_reporter_ops mlx5_tx_reporter_ops = {
543 		.name = "tx",
544 		.recover = mlx5e_tx_reporter_recover,
545 		.diagnose = mlx5e_tx_reporter_diagnose,
546 		.dump = mlx5e_tx_reporter_dump,
547 };
548 
549 #define MLX5_REPORTER_TX_GRACEFUL_PERIOD 500
550 
mlx5e_reporter_tx_create(struct mlx5e_priv * priv)551 void mlx5e_reporter_tx_create(struct mlx5e_priv *priv)
552 {
553 	struct devlink_health_reporter *reporter;
554 
555 	reporter = devlink_port_health_reporter_create(priv->netdev->devlink_port,
556 						       &mlx5_tx_reporter_ops,
557 						       MLX5_REPORTER_TX_GRACEFUL_PERIOD, priv);
558 	if (IS_ERR(reporter)) {
559 		netdev_warn(priv->netdev,
560 			    "Failed to create tx reporter, err = %ld\n",
561 			    PTR_ERR(reporter));
562 		return;
563 	}
564 	priv->tx_reporter = reporter;
565 }
566 
mlx5e_reporter_tx_destroy(struct mlx5e_priv * priv)567 void mlx5e_reporter_tx_destroy(struct mlx5e_priv *priv)
568 {
569 	if (!priv->tx_reporter)
570 		return;
571 
572 	devlink_health_reporter_destroy(priv->tx_reporter);
573 	priv->tx_reporter = NULL;
574 }
575