1 /* SPDX-License-Identifier: GPL-2.0 */
2 /* Copyright (c) 2019 Mellanox Technologies. */
3
4 #include "health.h"
5 #include "en/ptp.h"
6 #include "en/devlink.h"
7 #include "lib/tout.h"
8
9 /* Keep this string array consistent with the MLX5E_SQ_STATE_* enums in en.h */
10 static const char * const sq_sw_state_type_name[] = {
11 [MLX5E_SQ_STATE_ENABLED] = "enabled",
12 [MLX5E_SQ_STATE_MPWQE] = "mpwqe",
13 [MLX5E_SQ_STATE_RECOVERING] = "recovering",
14 [MLX5E_SQ_STATE_IPSEC] = "ipsec",
15 [MLX5E_SQ_STATE_DIM] = "dim",
16 [MLX5E_SQ_STATE_VLAN_NEED_L2_INLINE] = "vlan_need_l2_inline",
17 [MLX5E_SQ_STATE_PENDING_XSK_TX] = "pending_xsk_tx",
18 [MLX5E_SQ_STATE_PENDING_TLS_RX_RESYNC] = "pending_tls_rx_resync",
19 };
20
mlx5e_wait_for_sq_flush(struct mlx5e_txqsq * sq)21 static int mlx5e_wait_for_sq_flush(struct mlx5e_txqsq *sq)
22 {
23 struct mlx5_core_dev *dev = sq->mdev;
24 unsigned long exp_time;
25
26 exp_time = jiffies + msecs_to_jiffies(mlx5_tout_ms(dev, FLUSH_ON_ERROR));
27
28 while (time_before(jiffies, exp_time)) {
29 if (sq->cc == sq->pc)
30 return 0;
31
32 msleep(20);
33 }
34
35 netdev_err(sq->netdev,
36 "Wait for SQ 0x%x flush timeout (sq cc = 0x%x, sq pc = 0x%x)\n",
37 sq->sqn, sq->cc, sq->pc);
38
39 return -ETIMEDOUT;
40 }
41
mlx5e_reset_txqsq_cc_pc(struct mlx5e_txqsq * sq)42 static void mlx5e_reset_txqsq_cc_pc(struct mlx5e_txqsq *sq)
43 {
44 WARN_ONCE(sq->cc != sq->pc,
45 "SQ 0x%x: cc (0x%x) != pc (0x%x)\n",
46 sq->sqn, sq->cc, sq->pc);
47 sq->cc = 0;
48 sq->dma_fifo_cc = 0;
49 sq->pc = 0;
50 }
51
mlx5e_health_sq_put_sw_state(struct devlink_fmsg * fmsg,struct mlx5e_txqsq * sq)52 static void mlx5e_health_sq_put_sw_state(struct devlink_fmsg *fmsg, struct mlx5e_txqsq *sq)
53 {
54 int i;
55
56 BUILD_BUG_ON_MSG(ARRAY_SIZE(sq_sw_state_type_name) != MLX5E_NUM_SQ_STATES,
57 "sq_sw_state_type_name string array must be consistent with MLX5E_SQ_STATE_* enum in en.h");
58 mlx5e_health_fmsg_named_obj_nest_start(fmsg, "SW State");
59
60 for (i = 0; i < ARRAY_SIZE(sq_sw_state_type_name); ++i)
61 devlink_fmsg_u32_pair_put(fmsg, sq_sw_state_type_name[i],
62 test_bit(i, &sq->state));
63
64 mlx5e_health_fmsg_named_obj_nest_end(fmsg);
65 }
66
mlx5e_tx_reporter_err_cqe_recover(void * ctx)67 static int mlx5e_tx_reporter_err_cqe_recover(void *ctx)
68 {
69 struct mlx5_core_dev *mdev;
70 struct net_device *dev;
71 struct mlx5e_txqsq *sq;
72 u8 state;
73 int err;
74
75 sq = ctx;
76 mdev = sq->mdev;
77 dev = sq->netdev;
78
79 if (!test_bit(MLX5E_SQ_STATE_RECOVERING, &sq->state))
80 return 0;
81
82 err = mlx5_core_query_sq_state(mdev, sq->sqn, &state);
83 if (err) {
84 netdev_err(dev, "Failed to query SQ 0x%x state. err = %d\n",
85 sq->sqn, err);
86 goto out;
87 }
88
89 if (state != MLX5_SQC_STATE_ERR)
90 goto out;
91
92 mlx5e_tx_disable_queue(sq->txq);
93
94 err = mlx5e_wait_for_sq_flush(sq);
95 if (err)
96 goto out;
97
98 /* At this point, no new packets will arrive from the stack as TXQ is
99 * marked with QUEUE_STATE_DRV_XOFF. In addition, NAPI cleared all
100 * pending WQEs. SQ can safely reset the SQ.
101 */
102
103 err = mlx5e_health_sq_to_ready(mdev, dev, sq->sqn);
104 if (err)
105 goto out;
106
107 mlx5e_reset_txqsq_cc_pc(sq);
108 sq->stats->recover++;
109 clear_bit(MLX5E_SQ_STATE_RECOVERING, &sq->state);
110 rtnl_lock();
111 mlx5e_activate_txqsq(sq);
112 rtnl_unlock();
113
114 if (sq->channel)
115 mlx5e_trigger_napi_icosq(sq->channel);
116 else
117 mlx5e_trigger_napi_sched(sq->cq.napi);
118
119 return 0;
120 out:
121 clear_bit(MLX5E_SQ_STATE_RECOVERING, &sq->state);
122 return err;
123 }
124
125 struct mlx5e_tx_timeout_ctx {
126 struct mlx5e_txqsq *sq;
127 signed int status;
128 };
129
mlx5e_tx_reporter_timeout_recover(void * ctx)130 static int mlx5e_tx_reporter_timeout_recover(void *ctx)
131 {
132 struct mlx5e_tx_timeout_ctx *to_ctx;
133 struct mlx5e_priv *priv;
134 struct mlx5_eq_comp *eq;
135 struct mlx5e_txqsq *sq;
136 int err;
137
138 to_ctx = ctx;
139 sq = to_ctx->sq;
140 eq = sq->cq.mcq.eq;
141 priv = sq->priv;
142 err = mlx5e_health_channel_eq_recover(sq->netdev, eq, sq->cq.ch_stats);
143 if (!err) {
144 to_ctx->status = 0; /* this sq recovered */
145 return err;
146 }
147
148 mutex_lock(&priv->state_lock);
149 err = mlx5e_safe_reopen_channels(priv);
150 mutex_unlock(&priv->state_lock);
151 if (!err) {
152 to_ctx->status = 1; /* all channels recovered */
153 return err;
154 }
155
156 to_ctx->status = err;
157 clear_bit(MLX5E_SQ_STATE_ENABLED, &sq->state);
158 netdev_err(priv->netdev,
159 "mlx5e_safe_reopen_channels failed recovering from a tx_timeout, err(%d).\n",
160 err);
161
162 return err;
163 }
164
mlx5e_tx_reporter_ptpsq_unhealthy_recover(void * ctx)165 static int mlx5e_tx_reporter_ptpsq_unhealthy_recover(void *ctx)
166 {
167 struct mlx5e_ptpsq *ptpsq = ctx;
168 struct mlx5e_channels *chs;
169 struct net_device *netdev;
170 struct mlx5e_priv *priv;
171 int carrier_ok;
172 int err;
173
174 if (!test_bit(MLX5E_SQ_STATE_RECOVERING, &ptpsq->txqsq.state))
175 return 0;
176
177 priv = ptpsq->txqsq.priv;
178
179 rtnl_lock();
180 mutex_lock(&priv->state_lock);
181 chs = &priv->channels;
182 netdev = priv->netdev;
183
184 carrier_ok = netif_carrier_ok(netdev);
185 netif_carrier_off(netdev);
186
187 mlx5e_deactivate_priv_channels(priv);
188
189 mlx5e_ptp_close(chs->ptp);
190 err = mlx5e_ptp_open(priv, &chs->params, chs->c[0]->lag_port, &chs->ptp);
191
192 mlx5e_activate_priv_channels(priv);
193
194 /* return carrier back if needed */
195 if (carrier_ok)
196 netif_carrier_on(netdev);
197
198 mutex_unlock(&priv->state_lock);
199 rtnl_unlock();
200
201 return err;
202 }
203
204 /* state lock cannot be grabbed within this function.
205 * It can cause a dead lock or a read-after-free.
206 */
mlx5e_tx_reporter_recover_from_ctx(struct mlx5e_err_ctx * err_ctx)207 static int mlx5e_tx_reporter_recover_from_ctx(struct mlx5e_err_ctx *err_ctx)
208 {
209 return err_ctx->recover(err_ctx->ctx);
210 }
211
mlx5e_tx_reporter_recover(struct devlink_health_reporter * reporter,void * context,struct netlink_ext_ack * extack)212 static int mlx5e_tx_reporter_recover(struct devlink_health_reporter *reporter,
213 void *context,
214 struct netlink_ext_ack *extack)
215 {
216 struct mlx5e_priv *priv = devlink_health_reporter_priv(reporter);
217 struct mlx5e_err_ctx *err_ctx = context;
218
219 return err_ctx ? mlx5e_tx_reporter_recover_from_ctx(err_ctx) :
220 mlx5e_health_recover_channels(priv);
221 }
222
223 static void
mlx5e_tx_reporter_build_diagnose_output_sq_common(struct devlink_fmsg * fmsg,struct mlx5e_txqsq * sq,int tc)224 mlx5e_tx_reporter_build_diagnose_output_sq_common(struct devlink_fmsg *fmsg,
225 struct mlx5e_txqsq *sq, int tc)
226 {
227 bool stopped = netif_xmit_stopped(sq->txq);
228 u8 state;
229 int err;
230
231 devlink_fmsg_u32_pair_put(fmsg, "tc", tc);
232 devlink_fmsg_u32_pair_put(fmsg, "txq ix", sq->txq_ix);
233 devlink_fmsg_u32_pair_put(fmsg, "sqn", sq->sqn);
234
235 err = mlx5_core_query_sq_state(sq->mdev, sq->sqn, &state);
236 if (!err)
237 devlink_fmsg_u8_pair_put(fmsg, "HW state", state);
238
239 devlink_fmsg_bool_pair_put(fmsg, "stopped", stopped);
240 devlink_fmsg_u32_pair_put(fmsg, "cc", sq->cc);
241 devlink_fmsg_u32_pair_put(fmsg, "pc", sq->pc);
242 mlx5e_health_sq_put_sw_state(fmsg, sq);
243 mlx5e_health_cq_diag_fmsg(&sq->cq, fmsg);
244 mlx5e_health_eq_diag_fmsg(sq->cq.mcq.eq, fmsg);
245 }
246
247 static void
mlx5e_tx_reporter_build_diagnose_output(struct devlink_fmsg * fmsg,struct mlx5e_txqsq * sq,int tc)248 mlx5e_tx_reporter_build_diagnose_output(struct devlink_fmsg *fmsg,
249 struct mlx5e_txqsq *sq, int tc)
250 {
251 devlink_fmsg_obj_nest_start(fmsg);
252 devlink_fmsg_u32_pair_put(fmsg, "channel ix", sq->ch_ix);
253 mlx5e_tx_reporter_build_diagnose_output_sq_common(fmsg, sq, tc);
254 devlink_fmsg_obj_nest_end(fmsg);
255 }
256
257 static void
mlx5e_tx_reporter_build_diagnose_output_ptpsq(struct devlink_fmsg * fmsg,struct mlx5e_ptpsq * ptpsq,int tc)258 mlx5e_tx_reporter_build_diagnose_output_ptpsq(struct devlink_fmsg *fmsg,
259 struct mlx5e_ptpsq *ptpsq, int tc)
260 {
261 devlink_fmsg_obj_nest_start(fmsg);
262 devlink_fmsg_string_pair_put(fmsg, "channel", "ptp");
263 mlx5e_tx_reporter_build_diagnose_output_sq_common(fmsg, &ptpsq->txqsq, tc);
264 mlx5e_health_fmsg_named_obj_nest_start(fmsg, "Port TS");
265 mlx5e_health_cq_diag_fmsg(&ptpsq->ts_cq, fmsg);
266 mlx5e_health_fmsg_named_obj_nest_end(fmsg);
267 devlink_fmsg_obj_nest_end(fmsg);
268 }
269
270 static void
mlx5e_tx_reporter_diagnose_generic_txqsq(struct devlink_fmsg * fmsg,struct mlx5e_txqsq * txqsq)271 mlx5e_tx_reporter_diagnose_generic_txqsq(struct devlink_fmsg *fmsg,
272 struct mlx5e_txqsq *txqsq)
273 {
274 bool real_time = mlx5_is_real_time_sq(txqsq->mdev);
275 u32 sq_sz = mlx5_wq_cyc_get_size(&txqsq->wq);
276 u32 sq_stride = MLX5_SEND_WQE_BB;
277
278 mlx5e_health_fmsg_named_obj_nest_start(fmsg, "SQ");
279 devlink_fmsg_u64_pair_put(fmsg, "stride size", sq_stride);
280 devlink_fmsg_u32_pair_put(fmsg, "size", sq_sz);
281 devlink_fmsg_string_pair_put(fmsg, "ts_format", real_time ? "RT" : "FRC");
282 mlx5e_health_cq_common_diag_fmsg(&txqsq->cq, fmsg);
283 mlx5e_health_fmsg_named_obj_nest_end(fmsg);
284 }
285
286 static void
mlx5e_tx_reporter_diagnose_generic_tx_port_ts(struct devlink_fmsg * fmsg,struct mlx5e_ptpsq * ptpsq)287 mlx5e_tx_reporter_diagnose_generic_tx_port_ts(struct devlink_fmsg *fmsg,
288 struct mlx5e_ptpsq *ptpsq)
289 {
290 mlx5e_health_fmsg_named_obj_nest_start(fmsg, "Port TS");
291 mlx5e_health_cq_common_diag_fmsg(&ptpsq->ts_cq, fmsg);
292 mlx5e_health_fmsg_named_obj_nest_end(fmsg);
293 }
294
295 static void
mlx5e_tx_reporter_diagnose_common_config(struct devlink_health_reporter * reporter,struct devlink_fmsg * fmsg)296 mlx5e_tx_reporter_diagnose_common_config(struct devlink_health_reporter *reporter,
297 struct devlink_fmsg *fmsg)
298 {
299 struct mlx5e_priv *priv = devlink_health_reporter_priv(reporter);
300 struct mlx5e_txqsq *generic_sq = priv->txq2sq[0];
301 struct mlx5e_ptp *ptp_ch = priv->channels.ptp;
302 struct mlx5e_ptpsq *generic_ptpsq;
303
304 mlx5e_health_fmsg_named_obj_nest_start(fmsg, "Common Config");
305 mlx5e_tx_reporter_diagnose_generic_txqsq(fmsg, generic_sq);
306
307 if (!ptp_ch || !test_bit(MLX5E_PTP_STATE_TX, ptp_ch->state))
308 goto out;
309
310 generic_ptpsq = &ptp_ch->ptpsq[0];
311 mlx5e_health_fmsg_named_obj_nest_start(fmsg, "PTP");
312 mlx5e_tx_reporter_diagnose_generic_txqsq(fmsg, &generic_ptpsq->txqsq);
313 mlx5e_tx_reporter_diagnose_generic_tx_port_ts(fmsg, generic_ptpsq);
314 mlx5e_health_fmsg_named_obj_nest_end(fmsg);
315 out:
316 mlx5e_health_fmsg_named_obj_nest_end(fmsg);
317 }
318
mlx5e_tx_reporter_diagnose(struct devlink_health_reporter * reporter,struct devlink_fmsg * fmsg,struct netlink_ext_ack * extack)319 static int mlx5e_tx_reporter_diagnose(struct devlink_health_reporter *reporter,
320 struct devlink_fmsg *fmsg,
321 struct netlink_ext_ack *extack)
322 {
323 struct mlx5e_priv *priv = devlink_health_reporter_priv(reporter);
324 struct mlx5e_ptp *ptp_ch = priv->channels.ptp;
325
326 int i, tc;
327
328 mutex_lock(&priv->state_lock);
329
330 if (!test_bit(MLX5E_STATE_OPENED, &priv->state))
331 goto unlock;
332
333 mlx5e_tx_reporter_diagnose_common_config(reporter, fmsg);
334 devlink_fmsg_arr_pair_nest_start(fmsg, "SQs");
335
336 for (i = 0; i < priv->channels.num; i++) {
337 struct mlx5e_channel *c = priv->channels.c[i];
338
339 for (tc = 0; tc < mlx5e_get_dcb_num_tc(&priv->channels.params); tc++) {
340 struct mlx5e_txqsq *sq = &c->sq[tc];
341
342 mlx5e_tx_reporter_build_diagnose_output(fmsg, sq, tc);
343 }
344 }
345
346 if (!ptp_ch || !test_bit(MLX5E_PTP_STATE_TX, ptp_ch->state))
347 goto close_sqs_nest;
348
349 for (tc = 0; tc < mlx5e_get_dcb_num_tc(&priv->channels.params); tc++)
350 mlx5e_tx_reporter_build_diagnose_output_ptpsq(fmsg,
351 &ptp_ch->ptpsq[tc],
352 tc);
353
354 close_sqs_nest:
355 devlink_fmsg_arr_pair_nest_end(fmsg);
356 unlock:
357 mutex_unlock(&priv->state_lock);
358 return 0;
359 }
360
mlx5e_tx_reporter_dump_sq(struct mlx5e_priv * priv,struct devlink_fmsg * fmsg,void * ctx)361 static int mlx5e_tx_reporter_dump_sq(struct mlx5e_priv *priv, struct devlink_fmsg *fmsg,
362 void *ctx)
363 {
364 struct mlx5_rsc_key key = {};
365 struct mlx5e_txqsq *sq = ctx;
366
367 if (!test_bit(MLX5E_STATE_OPENED, &priv->state))
368 return 0;
369
370 mlx5e_health_fmsg_named_obj_nest_start(fmsg, "SX Slice");
371 key.size = PAGE_SIZE;
372 key.rsc = MLX5_SGMT_TYPE_SX_SLICE_ALL;
373 mlx5e_health_rsc_fmsg_dump(priv, &key, fmsg);
374 mlx5e_health_fmsg_named_obj_nest_end(fmsg);
375
376 mlx5e_health_fmsg_named_obj_nest_start(fmsg, "SQ");
377 mlx5e_health_fmsg_named_obj_nest_start(fmsg, "QPC");
378 key.rsc = MLX5_SGMT_TYPE_FULL_QPC;
379 key.index1 = sq->sqn;
380 key.num_of_obj1 = 1;
381 mlx5e_health_rsc_fmsg_dump(priv, &key, fmsg);
382 mlx5e_health_fmsg_named_obj_nest_end(fmsg);
383
384 mlx5e_health_fmsg_named_obj_nest_start(fmsg, "send_buff");
385 key.rsc = MLX5_SGMT_TYPE_SND_BUFF;
386 key.num_of_obj2 = MLX5_RSC_DUMP_ALL;
387 mlx5e_health_rsc_fmsg_dump(priv, &key, fmsg);
388 mlx5e_health_fmsg_named_obj_nest_end(fmsg);
389
390 mlx5e_health_fmsg_named_obj_nest_end(fmsg);
391
392 return 0;
393 }
394
mlx5e_tx_reporter_timeout_dump(struct mlx5e_priv * priv,struct devlink_fmsg * fmsg,void * ctx)395 static int mlx5e_tx_reporter_timeout_dump(struct mlx5e_priv *priv, struct devlink_fmsg *fmsg,
396 void *ctx)
397 {
398 struct mlx5e_tx_timeout_ctx *to_ctx = ctx;
399
400 return mlx5e_tx_reporter_dump_sq(priv, fmsg, to_ctx->sq);
401 }
402
mlx5e_tx_reporter_ptpsq_unhealthy_dump(struct mlx5e_priv * priv,struct devlink_fmsg * fmsg,void * ctx)403 static int mlx5e_tx_reporter_ptpsq_unhealthy_dump(struct mlx5e_priv *priv,
404 struct devlink_fmsg *fmsg,
405 void *ctx)
406 {
407 struct mlx5e_ptpsq *ptpsq = ctx;
408
409 return mlx5e_tx_reporter_dump_sq(priv, fmsg, &ptpsq->txqsq);
410 }
411
mlx5e_tx_reporter_dump_all_sqs(struct mlx5e_priv * priv,struct devlink_fmsg * fmsg)412 static int mlx5e_tx_reporter_dump_all_sqs(struct mlx5e_priv *priv,
413 struct devlink_fmsg *fmsg)
414 {
415 struct mlx5e_ptp *ptp_ch = priv->channels.ptp;
416 struct mlx5_rsc_key key = {};
417 int i, tc;
418
419 if (!test_bit(MLX5E_STATE_OPENED, &priv->state))
420 return 0;
421
422 mlx5e_health_fmsg_named_obj_nest_start(fmsg, "SX Slice");
423 key.size = PAGE_SIZE;
424 key.rsc = MLX5_SGMT_TYPE_SX_SLICE_ALL;
425 mlx5e_health_rsc_fmsg_dump(priv, &key, fmsg);
426 mlx5e_health_fmsg_named_obj_nest_end(fmsg);
427 devlink_fmsg_arr_pair_nest_start(fmsg, "SQs");
428
429 for (i = 0; i < priv->channels.num; i++) {
430 struct mlx5e_channel *c = priv->channels.c[i];
431
432 for (tc = 0; tc < mlx5e_get_dcb_num_tc(&priv->channels.params); tc++) {
433 struct mlx5e_txqsq *sq = &c->sq[tc];
434
435 mlx5e_health_queue_dump(priv, fmsg, sq->sqn, "SQ");
436 }
437 }
438
439 if (ptp_ch && test_bit(MLX5E_PTP_STATE_TX, ptp_ch->state)) {
440 for (tc = 0; tc < mlx5e_get_dcb_num_tc(&priv->channels.params); tc++) {
441 struct mlx5e_txqsq *sq = &ptp_ch->ptpsq[tc].txqsq;
442
443 mlx5e_health_queue_dump(priv, fmsg, sq->sqn, "PTP SQ");
444 }
445 }
446
447 devlink_fmsg_arr_pair_nest_end(fmsg);
448 return 0;
449 }
450
mlx5e_tx_reporter_dump_from_ctx(struct mlx5e_priv * priv,struct mlx5e_err_ctx * err_ctx,struct devlink_fmsg * fmsg)451 static int mlx5e_tx_reporter_dump_from_ctx(struct mlx5e_priv *priv,
452 struct mlx5e_err_ctx *err_ctx,
453 struct devlink_fmsg *fmsg)
454 {
455 return err_ctx->dump(priv, fmsg, err_ctx->ctx);
456 }
457
mlx5e_tx_reporter_dump(struct devlink_health_reporter * reporter,struct devlink_fmsg * fmsg,void * context,struct netlink_ext_ack * extack)458 static int mlx5e_tx_reporter_dump(struct devlink_health_reporter *reporter,
459 struct devlink_fmsg *fmsg, void *context,
460 struct netlink_ext_ack *extack)
461 {
462 struct mlx5e_priv *priv = devlink_health_reporter_priv(reporter);
463 struct mlx5e_err_ctx *err_ctx = context;
464
465 return err_ctx ? mlx5e_tx_reporter_dump_from_ctx(priv, err_ctx, fmsg) :
466 mlx5e_tx_reporter_dump_all_sqs(priv, fmsg);
467 }
468
mlx5e_reporter_tx_err_cqe(struct mlx5e_txqsq * sq)469 void mlx5e_reporter_tx_err_cqe(struct mlx5e_txqsq *sq)
470 {
471 char err_str[MLX5E_REPORTER_PER_Q_MAX_LEN];
472 struct mlx5e_priv *priv = sq->priv;
473 struct mlx5e_err_ctx err_ctx = {};
474
475 err_ctx.ctx = sq;
476 err_ctx.recover = mlx5e_tx_reporter_err_cqe_recover;
477 err_ctx.dump = mlx5e_tx_reporter_dump_sq;
478 snprintf(err_str, sizeof(err_str), "ERR CQE on SQ: 0x%x", sq->sqn);
479
480 mlx5e_health_report(priv, priv->tx_reporter, err_str, &err_ctx);
481 }
482
mlx5e_reporter_tx_timeout(struct mlx5e_txqsq * sq)483 int mlx5e_reporter_tx_timeout(struct mlx5e_txqsq *sq)
484 {
485 char err_str[MLX5E_REPORTER_PER_Q_MAX_LEN];
486 struct mlx5e_tx_timeout_ctx to_ctx = {};
487 struct mlx5e_priv *priv = sq->priv;
488 struct mlx5e_err_ctx err_ctx = {};
489
490 to_ctx.sq = sq;
491 err_ctx.ctx = &to_ctx;
492 err_ctx.recover = mlx5e_tx_reporter_timeout_recover;
493 err_ctx.dump = mlx5e_tx_reporter_timeout_dump;
494 snprintf(err_str, sizeof(err_str),
495 "TX timeout on queue: %d, SQ: 0x%x, CQ: 0x%x, SQ Cons: 0x%x SQ Prod: 0x%x, usecs since last trans: %u",
496 sq->ch_ix, sq->sqn, sq->cq.mcq.cqn, sq->cc, sq->pc,
497 jiffies_to_usecs(jiffies - READ_ONCE(sq->txq->trans_start)));
498
499 mlx5e_health_report(priv, priv->tx_reporter, err_str, &err_ctx);
500 return to_ctx.status;
501 }
502
mlx5e_reporter_tx_ptpsq_unhealthy(struct mlx5e_ptpsq * ptpsq)503 void mlx5e_reporter_tx_ptpsq_unhealthy(struct mlx5e_ptpsq *ptpsq)
504 {
505 struct mlx5e_ptp_metadata_map *map = &ptpsq->metadata_map;
506 char err_str[MLX5E_REPORTER_PER_Q_MAX_LEN];
507 struct mlx5e_txqsq *txqsq = &ptpsq->txqsq;
508 struct mlx5e_cq *ts_cq = &ptpsq->ts_cq;
509 struct mlx5e_priv *priv = txqsq->priv;
510 struct mlx5e_err_ctx err_ctx = {};
511
512 err_ctx.ctx = ptpsq;
513 err_ctx.recover = mlx5e_tx_reporter_ptpsq_unhealthy_recover;
514 err_ctx.dump = mlx5e_tx_reporter_ptpsq_unhealthy_dump;
515 snprintf(err_str, sizeof(err_str),
516 "Unhealthy TX port TS queue: %d, SQ: 0x%x, CQ: 0x%x, Undelivered CQEs: %u Map Capacity: %u",
517 txqsq->ch_ix, txqsq->sqn, ts_cq->mcq.cqn, map->undelivered_counter, map->capacity);
518
519 mlx5e_health_report(priv, priv->tx_reporter, err_str, &err_ctx);
520 }
521
522 static const struct devlink_health_reporter_ops mlx5_tx_reporter_ops = {
523 .name = "tx",
524 .recover = mlx5e_tx_reporter_recover,
525 .diagnose = mlx5e_tx_reporter_diagnose,
526 .dump = mlx5e_tx_reporter_dump,
527 };
528
529 #define MLX5_REPORTER_TX_GRACEFUL_PERIOD 500
530
mlx5e_reporter_tx_create(struct mlx5e_priv * priv)531 void mlx5e_reporter_tx_create(struct mlx5e_priv *priv)
532 {
533 struct devlink_health_reporter *reporter;
534
535 reporter = devlink_port_health_reporter_create(priv->netdev->devlink_port,
536 &mlx5_tx_reporter_ops,
537 MLX5_REPORTER_TX_GRACEFUL_PERIOD, priv);
538 if (IS_ERR(reporter)) {
539 netdev_warn(priv->netdev,
540 "Failed to create tx reporter, err = %ld\n",
541 PTR_ERR(reporter));
542 return;
543 }
544 priv->tx_reporter = reporter;
545 }
546
mlx5e_reporter_tx_destroy(struct mlx5e_priv * priv)547 void mlx5e_reporter_tx_destroy(struct mlx5e_priv *priv)
548 {
549 if (!priv->tx_reporter)
550 return;
551
552 devlink_health_reporter_destroy(priv->tx_reporter);
553 priv->tx_reporter = NULL;
554 }
555