| /linux/drivers/infiniband/hw/cxgb4/ |
| H A D | ev.c | 62 static void dump_err_cqe(struct c4iw_dev *dev, struct t4_cqe *err_cqe) in dump_err_cqe() argument 64 __be64 *p = (void *)err_cqe; in dump_err_cqe() 69 CQE_QPID(err_cqe), CQE_OPCODE(err_cqe), in dump_err_cqe() 70 CQE_STATUS(err_cqe), CQE_TYPE(err_cqe), ntohl(err_cqe->len), in dump_err_cqe() 71 CQE_WRID_HI(err_cqe), CQE_WRID_LOW(err_cqe)); in dump_err_cqe() 82 if (RQ_TYPE(err_cqe) && (CQE_OPCODE(err_cqe) == FW_RI_RDMA_WRITE || in dump_err_cqe() 83 CQE_OPCODE(err_cqe) == FW_RI_READ_RESP)) in dump_err_cqe() 84 print_tpte(dev, CQE_WRID_STAG(err_cqe)); in dump_err_cqe() 89 struct t4_cqe *err_cqe, in post_qp_event() argument 96 dump_err_cqe(dev, err_cqe); in post_qp_event() [all …]
|
| H A D | qp.c | 1421 static inline void build_term_codes(struct t4_cqe *err_cqe, u8 *layer_type, in build_term_codes() argument 1430 if (!err_cqe) { in build_term_codes() 1436 status = CQE_STATUS(err_cqe); in build_term_codes() 1437 opcode = CQE_OPCODE(err_cqe); in build_term_codes() 1438 rqtype = RQ_TYPE(err_cqe); in build_term_codes() 1557 static void post_terminate(struct c4iw_qp *qhp, struct t4_cqe *err_cqe, in post_terminate() argument 1586 build_term_codes(err_cqe, &term->layer_etype, &term->ecode); in post_terminate()
|
| /linux/include/linux/mlx5/ |
| H A D | cq.h | 200 struct mlx5_err_cqe *err_cqe) in mlx5_dump_err_cqe() argument 202 print_hex_dump(KERN_WARNING, "", DUMP_PREFIX_OFFSET, 16, 1, err_cqe, in mlx5_dump_err_cqe() 203 sizeof(*err_cqe), false); in mlx5_dump_err_cqe()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/lib/ |
| H A D | aso.c | 406 struct mlx5_err_cqe *err_cqe; in mlx5_aso_poll_cq() local 411 err_cqe = (struct mlx5_err_cqe *)cqe; in mlx5_aso_poll_cq() 413 err_cqe->vendor_err_synd); in mlx5_aso_poll_cq() 415 err_cqe->syndrome); in mlx5_aso_poll_cq() 417 16, 1, err_cqe, in mlx5_aso_poll_cq() 418 sizeof(*err_cqe), false); in mlx5_aso_poll_cq()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| H A D | send.c | 408 struct mlx5_err_cqe *err_cqe = (struct mlx5_err_cqe *)cqe; in hws_send_engine_dump_error_cqe() local 413 err_cqe->rsvd1[16]); in hws_send_engine_dump_error_cqe() 417 err_cqe->rsvd1[17] >> 4); in hws_send_engine_dump_error_cqe() 421 err_cqe->vendor_err_synd); in hws_send_engine_dump_error_cqe() 425 err_cqe->syndrome); in hws_send_engine_dump_error_cqe() 560 struct mlx5_err_cqe *err_cqe = (struct mlx5_err_cqe *)cqe64; in mlx5hws_parse_cqe() local 563 mlx5_core_err(cq->mdev, "vendor_err_synd=%x\n", err_cqe->vendor_err_synd); in mlx5hws_parse_cqe() 564 mlx5_core_err(cq->mdev, "syndrome=%x\n", err_cqe->syndrome); in mlx5hws_parse_cqe() 566 16, 1, err_cqe, in mlx5hws_parse_cqe() 567 sizeof(*err_cqe), false); in mlx5hws_parse_cqe()
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | cq.c | 455 struct mlx5_err_cqe *err_cqe; in mlx5_poll_one() local 525 err_cqe = (struct mlx5_err_cqe *)cqe64; in mlx5_poll_one() 526 mlx5_handle_error_cqe(dev, err_cqe, wc); in mlx5_poll_one() 531 err_cqe->syndrome, err_cqe->vendor_err_synd); in mlx5_poll_one()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | en_rx.c | 1679 struct mlx5_err_cqe *err_cqe = (struct mlx5_err_cqe *)cqe; in trigger_report() local 1682 if (cqe_syndrome_needs_recover(err_cqe->syndrome) && in trigger_report() 1684 mlx5e_dump_error_cqe(&rq->cq, rq->rqn, err_cqe); in trigger_report()
|
| H A D | en_stats.c | 1192 ts_stats->err += stats->abort + stats->err_cqe + in mlx5e_stats_ts_get() 2384 { MLX5E_DECLARE_PTP_CQ_STAT(struct mlx5e_ptp_cq_stats, err_cqe) },
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
| H A D | ptp.c | 215 ptpsq->cq_stats->err_cqe++; in mlx5e_ptp_handle_ts_cqe()
|