1 /* 2 * Broadcom NetXtreme-E RoCE driver. 3 * 4 * Copyright (c) 2016 - 2017, Broadcom. All rights reserved. The term 5 * Broadcom refers to Broadcom Limited and/or its subsidiaries. 6 * 7 * This software is available to you under a choice of one of two 8 * licenses. You may choose to be licensed under the terms of the GNU 9 * General Public License (GPL) Version 2, available from the file 10 * COPYING in the main directory of this source tree, or the 11 * BSD license below: 12 * 13 * Redistribution and use in source and binary forms, with or without 14 * modification, are permitted provided that the following conditions 15 * are met: 16 * 17 * 1. Redistributions of source code must retain the above copyright 18 * notice, this list of conditions and the following disclaimer. 19 * 2. Redistributions in binary form must reproduce the above copyright 20 * notice, this list of conditions and the following disclaimer in 21 * the documentation and/or other materials provided with the 22 * distribution. 23 * 24 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' 25 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, 26 * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR 27 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS 28 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 29 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 30 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR 31 * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, 32 * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE 33 * OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN 34 * IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 35 * 36 * Description: Statistics 37 * 38 */ 39 40 #include <linux/types.h> 41 #include <linux/pci.h> 42 #include <rdma/ib_mad.h> 43 #include <rdma/ib_pma.h> 44 45 #include "roce_hsi.h" 46 #include "qplib_res.h" 47 #include "qplib_sp.h" 48 #include "qplib_fp.h" 49 #include "qplib_rcfw.h" 50 #include "bnxt_re.h" 51 #include "hw_counters.h" 52 53 static const struct rdma_stat_desc bnxt_re_stat_descs[] = { 54 [BNXT_RE_RX_PKTS].name = "rx_pkts", 55 [BNXT_RE_RX_BYTES].name = "rx_bytes", 56 [BNXT_RE_TX_PKTS].name = "tx_pkts", 57 [BNXT_RE_TX_BYTES].name = "tx_bytes", 58 [BNXT_RE_RECOVERABLE_ERRORS].name = "recoverable_errors", 59 [BNXT_RE_TX_ERRORS].name = "tx_roce_errors", 60 [BNXT_RE_TX_DISCARDS].name = "tx_roce_discards", 61 [BNXT_RE_RX_ERRORS].name = "rx_roce_errors", 62 [BNXT_RE_RX_DISCARDS].name = "rx_roce_discards", 63 [BNXT_RE_TO_RETRANSMITS].name = "local_ack_timeout_err", 64 [BNXT_RE_SEQ_ERR_NAKS_RCVD].name = "packet_seq_err", 65 [BNXT_RE_MAX_RETRY_EXCEEDED].name = "max_retry_exceeded", 66 [BNXT_RE_RNR_NAKS_RCVD].name = "rnr_nak_retry_err", 67 [BNXT_RE_MISSING_RESP].name = "implied_nak_seq_err", 68 [BNXT_RE_UNRECOVERABLE_ERR].name = "unrecoverable_err", 69 [BNXT_RE_BAD_RESP_ERR].name = "bad_resp_err", 70 [BNXT_RE_LOCAL_QP_OP_ERR].name = "local_qp_op_err", 71 [BNXT_RE_LOCAL_PROTECTION_ERR].name = "local_protection_err", 72 [BNXT_RE_MEM_MGMT_OP_ERR].name = "mem_mgmt_op_err", 73 [BNXT_RE_REMOTE_INVALID_REQ_ERR].name = "req_remote_invalid_request", 74 [BNXT_RE_REMOTE_ACCESS_ERR].name = "req_remote_access_errors", 75 [BNXT_RE_REMOTE_OP_ERR].name = "remote_op_err", 76 [BNXT_RE_DUP_REQ].name = "duplicate_request", 77 [BNXT_RE_RES_EXCEED_MAX].name = "res_exceed_max", 78 [BNXT_RE_RES_LENGTH_MISMATCH].name = "resp_local_length_error", 79 [BNXT_RE_RES_EXCEEDS_WQE].name = "res_exceeds_wqe", 80 [BNXT_RE_RES_OPCODE_ERR].name = "res_opcode_err", 81 [BNXT_RE_RES_RX_INVALID_RKEY].name = "res_rx_invalid_rkey", 82 [BNXT_RE_RES_RX_DOMAIN_ERR].name = "res_rx_domain_err", 83 [BNXT_RE_RES_RX_NO_PERM].name = "res_rx_no_perm", 84 [BNXT_RE_RES_RX_RANGE_ERR].name = "res_rx_range_err", 85 [BNXT_RE_RES_TX_INVALID_RKEY].name = "res_tx_invalid_rkey", 86 [BNXT_RE_RES_TX_DOMAIN_ERR].name = "res_tx_domain_err", 87 [BNXT_RE_RES_TX_NO_PERM].name = "res_tx_no_perm", 88 [BNXT_RE_RES_TX_RANGE_ERR].name = "res_tx_range_err", 89 [BNXT_RE_RES_IRRQ_OFLOW].name = "res_irrq_oflow", 90 [BNXT_RE_RES_UNSUP_OPCODE].name = "res_unsup_opcode", 91 [BNXT_RE_RES_UNALIGNED_ATOMIC].name = "res_unaligned_atomic", 92 [BNXT_RE_RES_REM_INV_ERR].name = "res_rem_inv_err", 93 [BNXT_RE_RES_MEM_ERROR].name = "res_mem_err", 94 [BNXT_RE_RES_SRQ_ERR].name = "res_srq_err", 95 [BNXT_RE_RES_CMP_ERR].name = "res_cmp_err", 96 [BNXT_RE_RES_INVALID_DUP_RKEY].name = "res_invalid_dup_rkey", 97 [BNXT_RE_RES_WQE_FORMAT_ERR].name = "res_wqe_format_err", 98 [BNXT_RE_RES_CQ_LOAD_ERR].name = "res_cq_load_err", 99 [BNXT_RE_RES_SRQ_LOAD_ERR].name = "res_srq_load_err", 100 [BNXT_RE_RES_TX_PCI_ERR].name = "res_tx_pci_err", 101 [BNXT_RE_RES_RX_PCI_ERR].name = "res_rx_pci_err", 102 [BNXT_RE_OUT_OF_SEQ_ERR].name = "out_of_sequence", 103 [BNXT_RE_TX_ATOMIC_REQ].name = "tx_atomic_req", 104 [BNXT_RE_TX_READ_REQ].name = "tx_read_req", 105 [BNXT_RE_TX_READ_RES].name = "tx_read_resp", 106 [BNXT_RE_TX_WRITE_REQ].name = "tx_write_req", 107 [BNXT_RE_TX_SEND_REQ].name = "tx_send_req", 108 [BNXT_RE_TX_ROCE_PKTS].name = "tx_roce_only_pkts", 109 [BNXT_RE_TX_ROCE_BYTES].name = "tx_roce_only_bytes", 110 [BNXT_RE_RX_ATOMIC_REQ].name = "rx_atomic_requests", 111 [BNXT_RE_RX_READ_REQ].name = "rx_read_requests", 112 [BNXT_RE_RX_READ_RESP].name = "rx_read_resp", 113 [BNXT_RE_RX_WRITE_REQ].name = "rx_write_requests", 114 [BNXT_RE_RX_SEND_REQ].name = "rx_send_req", 115 [BNXT_RE_RX_ROCE_PKTS].name = "rx_roce_only_pkts", 116 [BNXT_RE_RX_ROCE_BYTES].name = "rx_roce_only_bytes", 117 [BNXT_RE_RX_ROCE_GOOD_PKTS].name = "rx_roce_good_pkts", 118 [BNXT_RE_RX_ROCE_GOOD_BYTES].name = "rx_roce_good_bytes", 119 [BNXT_RE_OOB].name = "out_of_buffer", 120 [BNXT_RE_TX_CNP].name = "np_cnp_pkts", 121 [BNXT_RE_RX_CNP].name = "rp_cnp_handled", 122 [BNXT_RE_RX_ECN].name = "np_ecn_marked_roce_packets", 123 [BNXT_RE_REQ_CQE_ERROR].name = "req_cqe_error", 124 [BNXT_RE_RESP_CQE_ERROR].name = "resp_cqe_error", 125 [BNXT_RE_RESP_REMOTE_ACCESS_ERRS].name = "resp_remote_access_errors", 126 }; 127 128 static void bnxt_re_copy_ext_stats(struct bnxt_re_dev *rdev, 129 struct rdma_hw_stats *stats, 130 struct bnxt_qplib_ext_stat *s) 131 { 132 stats->value[BNXT_RE_TX_ATOMIC_REQ] = s->tx_atomic_req; 133 stats->value[BNXT_RE_TX_READ_REQ] = s->tx_read_req; 134 stats->value[BNXT_RE_TX_READ_RES] = s->tx_read_res; 135 stats->value[BNXT_RE_TX_WRITE_REQ] = s->tx_write_req; 136 stats->value[BNXT_RE_TX_SEND_REQ] = s->tx_send_req; 137 stats->value[BNXT_RE_TX_ROCE_PKTS] = s->tx_roce_pkts; 138 stats->value[BNXT_RE_TX_ROCE_BYTES] = s->tx_roce_bytes; 139 stats->value[BNXT_RE_RX_ATOMIC_REQ] = s->rx_atomic_req; 140 stats->value[BNXT_RE_RX_READ_REQ] = s->rx_read_req; 141 stats->value[BNXT_RE_RX_READ_RESP] = s->rx_read_res; 142 stats->value[BNXT_RE_RX_WRITE_REQ] = s->rx_write_req; 143 stats->value[BNXT_RE_RX_SEND_REQ] = s->rx_send_req; 144 stats->value[BNXT_RE_RX_ROCE_PKTS] = s->rx_roce_pkts; 145 stats->value[BNXT_RE_RX_ROCE_BYTES] = s->rx_roce_bytes; 146 stats->value[BNXT_RE_RX_ROCE_GOOD_PKTS] = s->rx_roce_good_pkts; 147 stats->value[BNXT_RE_RX_ROCE_GOOD_BYTES] = s->rx_roce_good_bytes; 148 stats->value[BNXT_RE_OOB] = s->rx_out_of_buffer; 149 stats->value[BNXT_RE_TX_CNP] = s->tx_cnp; 150 stats->value[BNXT_RE_RX_CNP] = s->rx_cnp; 151 stats->value[BNXT_RE_RX_ECN] = s->rx_ecn_marked; 152 stats->value[BNXT_RE_OUT_OF_SEQ_ERR] = s->rx_out_of_sequence; 153 } 154 155 static int bnxt_re_get_ext_stat(struct bnxt_re_dev *rdev, 156 struct rdma_hw_stats *stats) 157 { 158 struct bnxt_qplib_ext_stat *estat = &rdev->stats.rstat.ext_stat; 159 u32 fid; 160 int rc; 161 162 fid = PCI_FUNC(rdev->en_dev->pdev->devfn); 163 rc = bnxt_qplib_qext_stat(&rdev->rcfw, fid, estat); 164 if (rc) 165 goto done; 166 bnxt_re_copy_ext_stats(rdev, stats, estat); 167 168 done: 169 return rc; 170 } 171 172 static void bnxt_re_copy_err_stats(struct bnxt_re_dev *rdev, 173 struct rdma_hw_stats *stats, 174 struct bnxt_qplib_roce_stats *err_s) 175 { 176 stats->value[BNXT_RE_TO_RETRANSMITS] = 177 err_s->to_retransmits; 178 stats->value[BNXT_RE_SEQ_ERR_NAKS_RCVD] = 179 err_s->seq_err_naks_rcvd; 180 stats->value[BNXT_RE_MAX_RETRY_EXCEEDED] = 181 err_s->max_retry_exceeded; 182 stats->value[BNXT_RE_RNR_NAKS_RCVD] = 183 err_s->rnr_naks_rcvd; 184 stats->value[BNXT_RE_MISSING_RESP] = 185 err_s->missing_resp; 186 stats->value[BNXT_RE_UNRECOVERABLE_ERR] = 187 err_s->unrecoverable_err; 188 stats->value[BNXT_RE_BAD_RESP_ERR] = 189 err_s->bad_resp_err; 190 stats->value[BNXT_RE_LOCAL_QP_OP_ERR] = 191 err_s->local_qp_op_err; 192 stats->value[BNXT_RE_LOCAL_PROTECTION_ERR] = 193 err_s->local_protection_err; 194 stats->value[BNXT_RE_MEM_MGMT_OP_ERR] = 195 err_s->mem_mgmt_op_err; 196 stats->value[BNXT_RE_REMOTE_INVALID_REQ_ERR] = 197 err_s->remote_invalid_req_err; 198 stats->value[BNXT_RE_REMOTE_ACCESS_ERR] = 199 err_s->remote_access_err; 200 stats->value[BNXT_RE_REMOTE_OP_ERR] = 201 err_s->remote_op_err; 202 stats->value[BNXT_RE_DUP_REQ] = 203 err_s->dup_req; 204 stats->value[BNXT_RE_RES_EXCEED_MAX] = 205 err_s->res_exceed_max; 206 stats->value[BNXT_RE_RES_LENGTH_MISMATCH] = 207 err_s->res_length_mismatch; 208 stats->value[BNXT_RE_RES_EXCEEDS_WQE] = 209 err_s->res_exceeds_wqe; 210 stats->value[BNXT_RE_RES_OPCODE_ERR] = 211 err_s->res_opcode_err; 212 stats->value[BNXT_RE_RES_RX_INVALID_RKEY] = 213 err_s->res_rx_invalid_rkey; 214 stats->value[BNXT_RE_RES_RX_DOMAIN_ERR] = 215 err_s->res_rx_domain_err; 216 stats->value[BNXT_RE_RES_RX_NO_PERM] = 217 err_s->res_rx_no_perm; 218 stats->value[BNXT_RE_RES_RX_RANGE_ERR] = 219 err_s->res_rx_range_err; 220 stats->value[BNXT_RE_RES_TX_INVALID_RKEY] = 221 err_s->res_tx_invalid_rkey; 222 stats->value[BNXT_RE_RES_TX_DOMAIN_ERR] = 223 err_s->res_tx_domain_err; 224 stats->value[BNXT_RE_RES_TX_NO_PERM] = 225 err_s->res_tx_no_perm; 226 stats->value[BNXT_RE_RES_TX_RANGE_ERR] = 227 err_s->res_tx_range_err; 228 stats->value[BNXT_RE_RES_IRRQ_OFLOW] = 229 err_s->res_irrq_oflow; 230 stats->value[BNXT_RE_RES_UNSUP_OPCODE] = 231 err_s->res_unsup_opcode; 232 stats->value[BNXT_RE_RES_UNALIGNED_ATOMIC] = 233 err_s->res_unaligned_atomic; 234 stats->value[BNXT_RE_RES_REM_INV_ERR] = 235 err_s->res_rem_inv_err; 236 stats->value[BNXT_RE_RES_MEM_ERROR] = 237 err_s->res_mem_error; 238 stats->value[BNXT_RE_RES_SRQ_ERR] = 239 err_s->res_srq_err; 240 stats->value[BNXT_RE_RES_CMP_ERR] = 241 err_s->res_cmp_err; 242 stats->value[BNXT_RE_RES_INVALID_DUP_RKEY] = 243 err_s->res_invalid_dup_rkey; 244 stats->value[BNXT_RE_RES_WQE_FORMAT_ERR] = 245 err_s->res_wqe_format_err; 246 stats->value[BNXT_RE_RES_CQ_LOAD_ERR] = 247 err_s->res_cq_load_err; 248 stats->value[BNXT_RE_RES_SRQ_LOAD_ERR] = 249 err_s->res_srq_load_err; 250 stats->value[BNXT_RE_RES_TX_PCI_ERR] = 251 err_s->res_tx_pci_err; 252 stats->value[BNXT_RE_RES_RX_PCI_ERR] = 253 err_s->res_rx_pci_err; 254 stats->value[BNXT_RE_OUT_OF_SEQ_ERR] = 255 err_s->res_oos_drop_count; 256 stats->value[BNXT_RE_REQ_CQE_ERROR] = 257 err_s->bad_resp_err + 258 err_s->local_qp_op_err + 259 err_s->local_protection_err + 260 err_s->mem_mgmt_op_err + 261 err_s->remote_invalid_req_err + 262 err_s->remote_access_err + 263 err_s->remote_op_err; 264 stats->value[BNXT_RE_RESP_CQE_ERROR] = 265 err_s->res_cmp_err + 266 err_s->res_cq_load_err; 267 stats->value[BNXT_RE_RESP_REMOTE_ACCESS_ERRS] = 268 err_s->res_rx_no_perm + 269 err_s->res_tx_no_perm; 270 } 271 272 int bnxt_re_assign_pma_port_ext_counters(struct bnxt_re_dev *rdev, struct ib_mad *out_mad) 273 { 274 struct ib_pma_portcounters_ext *pma_cnt_ext; 275 struct bnxt_qplib_ext_stat *estat = &rdev->stats.rstat.ext_stat; 276 struct ctx_hw_stats *hw_stats = NULL; 277 int rc; 278 279 hw_stats = rdev->qplib_ctx.stats.dma; 280 281 pma_cnt_ext = (struct ib_pma_portcounters_ext *)(out_mad->data + 40); 282 if (_is_ext_stats_supported(rdev->dev_attr->dev_cap_flags)) { 283 u32 fid = PCI_FUNC(rdev->en_dev->pdev->devfn); 284 285 rc = bnxt_qplib_qext_stat(&rdev->rcfw, fid, estat); 286 if (rc) 287 return rc; 288 } 289 290 pma_cnt_ext = (struct ib_pma_portcounters_ext *)(out_mad->data + 40); 291 if ((bnxt_qplib_is_chip_gen_p5(rdev->chip_ctx) && rdev->is_virtfn) || 292 !bnxt_qplib_is_chip_gen_p5(rdev->chip_ctx)) { 293 pma_cnt_ext->port_xmit_data = 294 cpu_to_be64(le64_to_cpu(hw_stats->tx_ucast_bytes) / 4); 295 pma_cnt_ext->port_rcv_data = 296 cpu_to_be64(le64_to_cpu(hw_stats->rx_ucast_bytes) / 4); 297 pma_cnt_ext->port_xmit_packets = 298 cpu_to_be64(le64_to_cpu(hw_stats->tx_ucast_pkts)); 299 pma_cnt_ext->port_rcv_packets = 300 cpu_to_be64(le64_to_cpu(hw_stats->rx_ucast_pkts)); 301 pma_cnt_ext->port_unicast_rcv_packets = 302 cpu_to_be64(le64_to_cpu(hw_stats->rx_ucast_pkts)); 303 pma_cnt_ext->port_unicast_xmit_packets = 304 cpu_to_be64(le64_to_cpu(hw_stats->tx_ucast_pkts)); 305 306 } else { 307 pma_cnt_ext->port_rcv_packets = cpu_to_be64(estat->rx_roce_good_pkts); 308 pma_cnt_ext->port_rcv_data = cpu_to_be64(estat->rx_roce_good_bytes / 4); 309 pma_cnt_ext->port_xmit_packets = cpu_to_be64(estat->tx_roce_pkts); 310 pma_cnt_ext->port_xmit_data = cpu_to_be64(estat->tx_roce_bytes / 4); 311 pma_cnt_ext->port_unicast_rcv_packets = cpu_to_be64(estat->rx_roce_good_pkts); 312 pma_cnt_ext->port_unicast_xmit_packets = cpu_to_be64(estat->tx_roce_pkts); 313 } 314 return 0; 315 } 316 317 int bnxt_re_assign_pma_port_counters(struct bnxt_re_dev *rdev, struct ib_mad *out_mad) 318 { 319 struct bnxt_qplib_ext_stat *estat = &rdev->stats.rstat.ext_stat; 320 struct ib_pma_portcounters *pma_cnt; 321 struct ctx_hw_stats *hw_stats = NULL; 322 int rc; 323 324 hw_stats = rdev->qplib_ctx.stats.dma; 325 326 pma_cnt = (struct ib_pma_portcounters *)(out_mad->data + 40); 327 if (_is_ext_stats_supported(rdev->dev_attr->dev_cap_flags)) { 328 u32 fid = PCI_FUNC(rdev->en_dev->pdev->devfn); 329 330 rc = bnxt_qplib_qext_stat(&rdev->rcfw, fid, estat); 331 if (rc) 332 return rc; 333 } 334 if ((bnxt_qplib_is_chip_gen_p5(rdev->chip_ctx) && rdev->is_virtfn) || 335 !bnxt_qplib_is_chip_gen_p5(rdev->chip_ctx)) { 336 pma_cnt->port_rcv_packets = 337 cpu_to_be32((u32)(le64_to_cpu(hw_stats->rx_ucast_pkts)) & 0xFFFFFFFF); 338 pma_cnt->port_rcv_data = 339 cpu_to_be32((u32)((le64_to_cpu(hw_stats->rx_ucast_bytes) & 340 0xFFFFFFFF) / 4)); 341 pma_cnt->port_xmit_packets = 342 cpu_to_be32((u32)(le64_to_cpu(hw_stats->tx_ucast_pkts)) & 0xFFFFFFFF); 343 pma_cnt->port_xmit_data = 344 cpu_to_be32((u32)((le64_to_cpu(hw_stats->tx_ucast_bytes) 345 & 0xFFFFFFFF) / 4)); 346 } else { 347 pma_cnt->port_rcv_packets = cpu_to_be32(estat->rx_roce_good_pkts); 348 pma_cnt->port_rcv_data = cpu_to_be32((estat->rx_roce_good_bytes / 4)); 349 pma_cnt->port_xmit_packets = cpu_to_be32(estat->tx_roce_pkts); 350 pma_cnt->port_xmit_data = cpu_to_be32((estat->tx_roce_bytes / 4)); 351 } 352 pma_cnt->port_rcv_constraint_errors = (u8)(le64_to_cpu(hw_stats->rx_discard_pkts) & 0xFF); 353 pma_cnt->port_rcv_errors = cpu_to_be16((u16)(le64_to_cpu(hw_stats->rx_error_pkts) 354 & 0xFFFF)); 355 pma_cnt->port_xmit_constraint_errors = (u8)(le64_to_cpu(hw_stats->tx_error_pkts) & 0xFF); 356 pma_cnt->port_xmit_discards = cpu_to_be16((u16)(le64_to_cpu(hw_stats->tx_discard_pkts) 357 & 0xFFFF)); 358 359 return 0; 360 } 361 362 int bnxt_re_ib_get_hw_stats(struct ib_device *ibdev, 363 struct rdma_hw_stats *stats, 364 u32 port, int index) 365 { 366 struct bnxt_re_dev *rdev = to_bnxt_re_dev(ibdev, ibdev); 367 struct bnxt_qplib_roce_stats *err_s = NULL; 368 struct ctx_hw_stats *hw_stats = NULL; 369 int rc = 0; 370 371 hw_stats = rdev->qplib_ctx.stats.dma; 372 if (!port || !stats) 373 return -EINVAL; 374 375 if (hw_stats) { 376 stats->value[BNXT_RE_RECOVERABLE_ERRORS] = 377 le64_to_cpu(hw_stats->tx_bcast_pkts); 378 stats->value[BNXT_RE_TX_DISCARDS] = 379 le64_to_cpu(hw_stats->tx_discard_pkts); 380 stats->value[BNXT_RE_TX_ERRORS] = 381 le64_to_cpu(hw_stats->tx_error_pkts); 382 stats->value[BNXT_RE_RX_ERRORS] = 383 le64_to_cpu(hw_stats->rx_error_pkts); 384 stats->value[BNXT_RE_RX_DISCARDS] = 385 le64_to_cpu(hw_stats->rx_discard_pkts); 386 stats->value[BNXT_RE_RX_PKTS] = 387 le64_to_cpu(hw_stats->rx_ucast_pkts); 388 stats->value[BNXT_RE_RX_BYTES] = 389 le64_to_cpu(hw_stats->rx_ucast_bytes); 390 stats->value[BNXT_RE_TX_PKTS] = 391 le64_to_cpu(hw_stats->tx_ucast_pkts); 392 stats->value[BNXT_RE_TX_BYTES] = 393 le64_to_cpu(hw_stats->tx_ucast_bytes); 394 } 395 err_s = &rdev->stats.rstat.errs; 396 if (test_bit(BNXT_RE_FLAG_ISSUE_ROCE_STATS, &rdev->flags)) { 397 rc = bnxt_qplib_get_roce_stats(&rdev->rcfw, err_s); 398 if (rc) { 399 clear_bit(BNXT_RE_FLAG_ISSUE_ROCE_STATS, 400 &rdev->flags); 401 goto done; 402 } 403 bnxt_re_copy_err_stats(rdev, stats, err_s); 404 if (bnxt_ext_stats_supported(rdev->chip_ctx, rdev->dev_attr->dev_cap_flags, 405 rdev->is_virtfn)) { 406 rc = bnxt_re_get_ext_stat(rdev, stats); 407 if (rc) { 408 clear_bit(BNXT_RE_FLAG_ISSUE_ROCE_STATS, 409 &rdev->flags); 410 goto done; 411 } 412 } 413 } 414 415 done: 416 return bnxt_qplib_is_chip_gen_p5_p7(rdev->chip_ctx) ? 417 BNXT_RE_NUM_EXT_COUNTERS : BNXT_RE_NUM_STD_COUNTERS; 418 } 419 420 struct rdma_hw_stats *bnxt_re_ib_alloc_hw_port_stats(struct ib_device *ibdev, 421 u32 port_num) 422 { 423 struct bnxt_re_dev *rdev = to_bnxt_re_dev(ibdev, ibdev); 424 int num_counters = 0; 425 426 if (bnxt_qplib_is_chip_gen_p5_p7(rdev->chip_ctx)) 427 num_counters = BNXT_RE_NUM_EXT_COUNTERS; 428 else 429 num_counters = BNXT_RE_NUM_STD_COUNTERS; 430 431 return rdma_alloc_hw_stats_struct(bnxt_re_stat_descs, num_counters, 432 RDMA_HW_STATS_DEFAULT_LIFESPAN); 433 } 434