Home
last modified time | relevance | path

Searched refs:err_cqe (Results 1 – 12 of 12) sorted by relevance

/linux/drivers/infiniband/hw/cxgb4/
H A Dev.c62 static void dump_err_cqe(struct c4iw_dev *dev, struct t4_cqe *err_cqe) in dump_err_cqe() argument
64 __be64 *p = (void *)err_cqe; in dump_err_cqe()
69 CQE_QPID(err_cqe), CQE_OPCODE(err_cqe), in dump_err_cqe()
70 CQE_STATUS(err_cqe), CQE_TYPE(err_cqe), ntohl(err_cqe->len), in dump_err_cqe()
71 CQE_WRID_HI(err_cqe), CQE_WRID_LOW(err_cqe)); in dump_err_cqe()
82 if (RQ_TYPE(err_cqe) && (CQE_OPCODE(err_cqe) == FW_RI_RDMA_WRITE || in dump_err_cqe()
83 CQE_OPCODE(err_cqe) == FW_RI_READ_RESP)) in dump_err_cqe()
84 print_tpte(dev, CQE_WRID_STAG(err_cqe)); in dump_err_cqe()
89 struct t4_cqe *err_cqe, in post_qp_event() argument
96 dump_err_cqe(dev, err_cqe); in post_qp_event()
[all …]
H A Diw_cxgb4.h1049 void c4iw_ev_dispatch(struct c4iw_dev *dev, struct t4_cqe *err_cqe);
/linux/include/linux/mlx5/
H A Dcq.h200 struct mlx5_err_cqe *err_cqe) in mlx5_dump_err_cqe() argument
202 print_hex_dump(KERN_WARNING, "", DUMP_PREFIX_OFFSET, 16, 1, err_cqe, in mlx5_dump_err_cqe()
203 sizeof(*err_cqe), false); in mlx5_dump_err_cqe()
/linux/drivers/net/ethernet/mellanox/mlx5/core/lib/
H A Daso.c406 struct mlx5_err_cqe *err_cqe; in mlx5_aso_poll_cq() local
411 err_cqe = (struct mlx5_err_cqe *)cqe; in mlx5_aso_poll_cq()
413 err_cqe->vendor_err_synd); in mlx5_aso_poll_cq()
415 err_cqe->syndrome); in mlx5_aso_poll_cq()
417 16, 1, err_cqe, in mlx5_aso_poll_cq()
418 sizeof(*err_cqe), false); in mlx5_aso_poll_cq()
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dtxrx.h377 struct mlx5_err_cqe *err_cqe) in mlx5e_dump_error_cqe() argument
387 get_cqe_opcode((struct mlx5_cqe64 *)err_cqe), in mlx5e_dump_error_cqe()
388 err_cqe->syndrome, err_cqe->vendor_err_synd); in mlx5e_dump_error_cqe()
389 mlx5_dump_err_cqe(cq->mdev, err_cqe); in mlx5e_dump_error_cqe()
H A Dptp.c205 ptpsq->cq_stats->err_cqe++; in mlx5e_ptp_handle_ts_cqe()
/linux/drivers/net/ethernet/mellanox/mlx4/
H A Den_tx.c400 static void mlx4_en_handle_err_cqe(struct mlx4_en_priv *priv, struct mlx4_err_cqe *err_cqe, in mlx4_en_handle_err_cqe() argument
410 ring->sp_cqn, cqe_index, err_cqe->vendor_err_syndrome, err_cqe->syndrome); in mlx4_en_handle_err_cqe()
411 print_hex_dump(KERN_WARNING, "", DUMP_PREFIX_OFFSET, 16, 1, err_cqe, sizeof(*err_cqe), in mlx4_en_handle_err_cqe()
414 wqe_index = be16_to_cpu(err_cqe->wqe_index) & ring->size_mask; in mlx4_en_handle_err_cqe()
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/
H A Dsend.c444 struct mlx5_err_cqe *err_cqe = (struct mlx5_err_cqe *)cqe64; in mlx5hws_parse_cqe() local
447 mlx5_core_err(cq->mdev, "vendor_err_synd=%x\n", err_cqe->vendor_err_synd); in mlx5hws_parse_cqe()
448 mlx5_core_err(cq->mdev, "syndrome=%x\n", err_cqe->syndrome); in mlx5hws_parse_cqe()
450 16, 1, err_cqe, in mlx5hws_parse_cqe()
451 sizeof(*err_cqe), false); in mlx5hws_parse_cqe()
/linux/drivers/infiniband/hw/mlx5/
H A Dcq.c455 struct mlx5_err_cqe *err_cqe; in mlx5_poll_one() local
525 err_cqe = (struct mlx5_err_cqe *)cqe64; in mlx5_poll_one()
526 mlx5_handle_error_cqe(dev, err_cqe, wc); in mlx5_poll_one()
531 err_cqe->syndrome, err_cqe->vendor_err_synd); in mlx5_poll_one()
/linux/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_stats.h471 u64 err_cqe; member
H A Den_rx.c1798 struct mlx5_err_cqe *err_cqe = (struct mlx5_err_cqe *)cqe; in trigger_report() local
1801 if (cqe_syndrome_needs_recover(err_cqe->syndrome) && in trigger_report()
1803 mlx5e_dump_error_cqe(&rq->cq, rq->rqn, err_cqe); in trigger_report()
H A Den_stats.c1204 ts_stats->err += stats->abort + stats->err_cqe + in mlx5e_stats_ts_get()
2236 { MLX5E_DECLARE_PTP_CQ_STAT(struct mlx5e_ptp_cq_stats, err_cqe) },