1 /*- 2 * Copyright (c) 2007-2014 QLogic Corporation. All rights reserved. 3 * 4 * Redistribution and use in source and binary forms, with or without 5 * modification, are permitted provided that the following conditions 6 * are met: 7 * 8 * 1. Redistributions of source code must retain the above copyright 9 * notice, this list of conditions and the following disclaimer. 10 * 2. Redistributions in binary form must reproduce the above copyright 11 * notice, this list of conditions and the following disclaimer in the 12 * documentation and/or other materials provided with the distribution. 13 * 14 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS' 15 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 17 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS 18 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 19 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 20 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 21 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 22 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 23 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF 24 * THE POSSIBILITY OF SUCH DAMAGE. 25 */ 26 27 #include <sys/cdefs.h> 28 __FBSDID("$FreeBSD$"); 29 30 #include "bxe.h" 31 #include "bxe_stats.h" 32 33 #ifdef __i386__ 34 #define BITS_PER_LONG 32 35 #else 36 #define BITS_PER_LONG 64 37 #endif 38 39 static inline long 40 bxe_hilo(uint32_t *hiref) 41 { 42 uint32_t lo = *(hiref + 1); 43 #if (BITS_PER_LONG == 64) 44 uint32_t hi = *hiref; 45 return (HILO_U64(hi, lo)); 46 #else 47 return (lo); 48 #endif 49 } 50 51 static inline uint16_t 52 bxe_get_port_stats_dma_len(struct bxe_softc *sc) 53 { 54 uint16_t res = 0; 55 uint32_t size; 56 57 /* 'newest' convention - shmem2 contains the size of the port stats */ 58 if (SHMEM2_HAS(sc, sizeof_port_stats)) { 59 size = SHMEM2_RD(sc, sizeof_port_stats); 60 if (size) { 61 res = size; 62 } 63 64 /* prevent newer BC from causing buffer overflow */ 65 if (res > sizeof(struct host_port_stats)) { 66 res = sizeof(struct host_port_stats); 67 } 68 } 69 70 /* 71 * Older convention - all BCs support the port stats fields up until 72 * the 'not_used' field 73 */ 74 if (!res) { 75 res = (offsetof(struct host_port_stats, not_used) + 4); 76 77 /* if PFC stats are supported by the MFW, DMA them as well */ 78 if (sc->devinfo.bc_ver >= REQ_BC_VER_4_PFC_STATS_SUPPORTED) { 79 res += (offsetof(struct host_port_stats, pfc_frames_rx_lo) - 80 offsetof(struct host_port_stats, pfc_frames_tx_hi) + 4); 81 } 82 } 83 84 res >>= 2; 85 86 DBASSERT(sc, !(res > 2 * DMAE_LEN32_RD_MAX), ("big stats dmae length\n")); 87 return (res); 88 } 89 90 /* 91 * Init service functions 92 */ 93 94 static void 95 bxe_dp_stats(struct bxe_softc *sc) 96 { 97 int i; 98 99 BLOGD(sc, DBG_STATS, 100 "dumping stats:\n" 101 " fw_stats_req\n" 102 " hdr\n" 103 " cmd_num %d\n" 104 " reserved0 %d\n" 105 " drv_stats_counter %d\n" 106 " reserved1 %d\n" 107 " stats_counters_addrs %x %x\n", 108 sc->fw_stats_req->hdr.cmd_num, 109 sc->fw_stats_req->hdr.reserved0, 110 sc->fw_stats_req->hdr.drv_stats_counter, 111 sc->fw_stats_req->hdr.reserved1, 112 sc->fw_stats_req->hdr.stats_counters_addrs.hi, 113 sc->fw_stats_req->hdr.stats_counters_addrs.lo); 114 115 for (i = 0; i < sc->fw_stats_req->hdr.cmd_num; i++) { 116 BLOGD(sc, DBG_STATS, 117 "query[%d]\n" 118 " kind %d\n" 119 " index %d\n" 120 " funcID %d\n" 121 " reserved %d\n" 122 " address %x %x\n", 123 i, 124 sc->fw_stats_req->query[i].kind, 125 sc->fw_stats_req->query[i].index, 126 sc->fw_stats_req->query[i].funcID, 127 sc->fw_stats_req->query[i].reserved, 128 sc->fw_stats_req->query[i].address.hi, 129 sc->fw_stats_req->query[i].address.lo); 130 } 131 } 132 133 /* 134 * Post the next statistics ramrod. Protect it with the lock in 135 * order to ensure the strict order between statistics ramrods 136 * (each ramrod has a sequence number passed in a 137 * sc->fw_stats_req->hdr.drv_stats_counter and ramrods must be 138 * sent in order). 139 */ 140 static void 141 bxe_storm_stats_post(struct bxe_softc *sc) 142 { 143 int rc; 144 145 if (!sc->stats_pending) { 146 BXE_STATS_LOCK(sc); 147 148 if (sc->stats_pending) { 149 BXE_STATS_UNLOCK(sc); 150 return; 151 } 152 153 sc->fw_stats_req->hdr.drv_stats_counter = 154 htole16(sc->stats_counter++); 155 156 BLOGD(sc, DBG_STATS, 157 "sending statistics ramrod %d\n", 158 le16toh(sc->fw_stats_req->hdr.drv_stats_counter)); 159 160 /* adjust the ramrod to include VF queues statistics */ 161 // XXX bxe_iov_adjust_stats_req(sc); 162 163 bxe_dp_stats(sc); 164 165 /* send FW stats ramrod */ 166 rc = bxe_sp_post(sc, RAMROD_CMD_ID_COMMON_STAT_QUERY, 0, 167 U64_HI(sc->fw_stats_req_mapping), 168 U64_LO(sc->fw_stats_req_mapping), 169 NONE_CONNECTION_TYPE); 170 if (rc == 0) { 171 sc->stats_pending = 1; 172 } 173 174 BXE_STATS_UNLOCK(sc); 175 } 176 } 177 178 static void 179 bxe_hw_stats_post(struct bxe_softc *sc) 180 { 181 struct dmae_command *dmae = &sc->stats_dmae; 182 uint32_t *stats_comp = BXE_SP(sc, stats_comp); 183 int loader_idx; 184 uint32_t opcode; 185 186 *stats_comp = DMAE_COMP_VAL; 187 if (CHIP_REV_IS_SLOW(sc)) { 188 return; 189 } 190 191 /* Update MCP's statistics if possible */ 192 if (sc->func_stx) { 193 memcpy(BXE_SP(sc, func_stats), &sc->func_stats, 194 sizeof(sc->func_stats)); 195 } 196 197 /* loader */ 198 if (sc->executer_idx) { 199 loader_idx = PMF_DMAE_C(sc); 200 opcode = bxe_dmae_opcode(sc, DMAE_SRC_PCI, DMAE_DST_GRC, 201 TRUE, DMAE_COMP_GRC); 202 opcode = bxe_dmae_opcode_clr_src_reset(opcode); 203 204 memset(dmae, 0, sizeof(struct dmae_command)); 205 dmae->opcode = opcode; 206 dmae->src_addr_lo = U64_LO(BXE_SP_MAPPING(sc, dmae[0])); 207 dmae->src_addr_hi = U64_HI(BXE_SP_MAPPING(sc, dmae[0])); 208 dmae->dst_addr_lo = ((DMAE_REG_CMD_MEM + 209 sizeof(struct dmae_command) * 210 (loader_idx + 1)) >> 2); 211 dmae->dst_addr_hi = 0; 212 dmae->len = sizeof(struct dmae_command) >> 2; 213 if (CHIP_IS_E1(sc)) { 214 dmae->len--; 215 } 216 dmae->comp_addr_lo = (dmae_reg_go_c[loader_idx + 1] >> 2); 217 dmae->comp_addr_hi = 0; 218 dmae->comp_val = 1; 219 220 *stats_comp = 0; 221 bxe_post_dmae(sc, dmae, loader_idx); 222 } else if (sc->func_stx) { 223 *stats_comp = 0; 224 bxe_post_dmae(sc, dmae, INIT_DMAE_C(sc)); 225 } 226 } 227 228 static int 229 bxe_stats_comp(struct bxe_softc *sc) 230 { 231 uint32_t *stats_comp = BXE_SP(sc, stats_comp); 232 int cnt = 10; 233 234 while (*stats_comp != DMAE_COMP_VAL) { 235 if (!cnt) { 236 BLOGE(sc, "Timeout waiting for stats finished\n"); 237 break; 238 } 239 240 cnt--; 241 DELAY(1000); 242 } 243 244 return (1); 245 } 246 247 /* 248 * Statistics service functions 249 */ 250 251 static void 252 bxe_stats_pmf_update(struct bxe_softc *sc) 253 { 254 struct dmae_command *dmae; 255 uint32_t opcode; 256 int loader_idx = PMF_DMAE_C(sc); 257 uint32_t *stats_comp = BXE_SP(sc, stats_comp); 258 259 if (sc->devinfo.bc_ver <= 0x06001400) { 260 /* 261 * Bootcode v6.0.21 fixed a GRC timeout that occurs when accessing 262 * BRB registers while the BRB block is in reset. The DMA transfer 263 * below triggers this issue resulting in the DMAE to stop 264 * functioning. Skip this initial stats transfer for old bootcode 265 * versions <= 6.0.20. 266 */ 267 return; 268 } 269 270 /* sanity */ 271 if (!sc->port.pmf || !sc->port.port_stx) { 272 BLOGE(sc, "BUG!\n"); 273 return; 274 } 275 276 sc->executer_idx = 0; 277 278 opcode = bxe_dmae_opcode(sc, DMAE_SRC_GRC, DMAE_DST_PCI, FALSE, 0); 279 280 dmae = BXE_SP(sc, dmae[sc->executer_idx++]); 281 dmae->opcode = bxe_dmae_opcode_add_comp(opcode, DMAE_COMP_GRC); 282 dmae->src_addr_lo = (sc->port.port_stx >> 2); 283 dmae->src_addr_hi = 0; 284 dmae->dst_addr_lo = U64_LO(BXE_SP_MAPPING(sc, port_stats)); 285 dmae->dst_addr_hi = U64_HI(BXE_SP_MAPPING(sc, port_stats)); 286 dmae->len = DMAE_LEN32_RD_MAX; 287 dmae->comp_addr_lo = (dmae_reg_go_c[loader_idx] >> 2); 288 dmae->comp_addr_hi = 0; 289 dmae->comp_val = 1; 290 291 dmae = BXE_SP(sc, dmae[sc->executer_idx++]); 292 dmae->opcode = bxe_dmae_opcode_add_comp(opcode, DMAE_COMP_PCI); 293 dmae->src_addr_lo = ((sc->port.port_stx >> 2) + DMAE_LEN32_RD_MAX); 294 dmae->src_addr_hi = 0; 295 dmae->dst_addr_lo = U64_LO(BXE_SP_MAPPING(sc, port_stats) + 296 DMAE_LEN32_RD_MAX * 4); 297 dmae->dst_addr_hi = U64_HI(BXE_SP_MAPPING(sc, port_stats) + 298 DMAE_LEN32_RD_MAX * 4); 299 dmae->len = (bxe_get_port_stats_dma_len(sc) - DMAE_LEN32_RD_MAX); 300 301 dmae->comp_addr_lo = U64_LO(BXE_SP_MAPPING(sc, stats_comp)); 302 dmae->comp_addr_hi = U64_HI(BXE_SP_MAPPING(sc, stats_comp)); 303 dmae->comp_val = DMAE_COMP_VAL; 304 305 *stats_comp = 0; 306 bxe_hw_stats_post(sc); 307 bxe_stats_comp(sc); 308 } 309 310 static void 311 bxe_port_stats_init(struct bxe_softc *sc) 312 { 313 struct dmae_command *dmae; 314 int port = SC_PORT(sc); 315 uint32_t opcode; 316 int loader_idx = PMF_DMAE_C(sc); 317 uint32_t mac_addr; 318 uint32_t *stats_comp = BXE_SP(sc, stats_comp); 319 320 /* sanity */ 321 if (!sc->link_vars.link_up || !sc->port.pmf) { 322 BLOGE(sc, "BUG!\n"); 323 return; 324 } 325 326 sc->executer_idx = 0; 327 328 /* MCP */ 329 opcode = bxe_dmae_opcode(sc, DMAE_SRC_PCI, DMAE_DST_GRC, 330 TRUE, DMAE_COMP_GRC); 331 332 if (sc->port.port_stx) { 333 dmae = BXE_SP(sc, dmae[sc->executer_idx++]); 334 dmae->opcode = opcode; 335 dmae->src_addr_lo = U64_LO(BXE_SP_MAPPING(sc, port_stats)); 336 dmae->src_addr_hi = U64_HI(BXE_SP_MAPPING(sc, port_stats)); 337 dmae->dst_addr_lo = sc->port.port_stx >> 2; 338 dmae->dst_addr_hi = 0; 339 dmae->len = bxe_get_port_stats_dma_len(sc); 340 dmae->comp_addr_lo = dmae_reg_go_c[loader_idx] >> 2; 341 dmae->comp_addr_hi = 0; 342 dmae->comp_val = 1; 343 } 344 345 if (sc->func_stx) { 346 dmae = BXE_SP(sc, dmae[sc->executer_idx++]); 347 dmae->opcode = opcode; 348 dmae->src_addr_lo = U64_LO(BXE_SP_MAPPING(sc, func_stats)); 349 dmae->src_addr_hi = U64_HI(BXE_SP_MAPPING(sc, func_stats)); 350 dmae->dst_addr_lo = (sc->func_stx >> 2); 351 dmae->dst_addr_hi = 0; 352 dmae->len = (sizeof(struct host_func_stats) >> 2); 353 dmae->comp_addr_lo = (dmae_reg_go_c[loader_idx] >> 2); 354 dmae->comp_addr_hi = 0; 355 dmae->comp_val = 1; 356 } 357 358 /* MAC */ 359 opcode = bxe_dmae_opcode(sc, DMAE_SRC_GRC, DMAE_DST_PCI, 360 TRUE, DMAE_COMP_GRC); 361 362 /* EMAC is special */ 363 if (sc->link_vars.mac_type == ELINK_MAC_TYPE_EMAC) { 364 mac_addr = (port ? GRCBASE_EMAC1 : GRCBASE_EMAC0); 365 366 /* EMAC_REG_EMAC_RX_STAT_AC (EMAC_REG_EMAC_RX_STAT_AC_COUNT)*/ 367 dmae = BXE_SP(sc, dmae[sc->executer_idx++]); 368 dmae->opcode = opcode; 369 dmae->src_addr_lo = (mac_addr + EMAC_REG_EMAC_RX_STAT_AC) >> 2; 370 dmae->src_addr_hi = 0; 371 dmae->dst_addr_lo = U64_LO(BXE_SP_MAPPING(sc, mac_stats)); 372 dmae->dst_addr_hi = U64_HI(BXE_SP_MAPPING(sc, mac_stats)); 373 dmae->len = EMAC_REG_EMAC_RX_STAT_AC_COUNT; 374 dmae->comp_addr_lo = (dmae_reg_go_c[loader_idx] >> 2); 375 dmae->comp_addr_hi = 0; 376 dmae->comp_val = 1; 377 378 /* EMAC_REG_EMAC_RX_STAT_AC_28 */ 379 dmae = BXE_SP(sc, dmae[sc->executer_idx++]); 380 dmae->opcode = opcode; 381 dmae->src_addr_lo = ((mac_addr + EMAC_REG_EMAC_RX_STAT_AC_28) >> 2); 382 dmae->src_addr_hi = 0; 383 dmae->dst_addr_lo = U64_LO(BXE_SP_MAPPING(sc, mac_stats) + 384 offsetof(struct emac_stats, 385 rx_stat_falsecarriererrors)); 386 dmae->dst_addr_hi = U64_HI(BXE_SP_MAPPING(sc, mac_stats) + 387 offsetof(struct emac_stats, 388 rx_stat_falsecarriererrors)); 389 dmae->len = 1; 390 dmae->comp_addr_lo = (dmae_reg_go_c[loader_idx] >> 2); 391 dmae->comp_addr_hi = 0; 392 dmae->comp_val = 1; 393 394 /* EMAC_REG_EMAC_TX_STAT_AC (EMAC_REG_EMAC_TX_STAT_AC_COUNT)*/ 395 dmae = BXE_SP(sc, dmae[sc->executer_idx++]); 396 dmae->opcode = opcode; 397 dmae->src_addr_lo = ((mac_addr + EMAC_REG_EMAC_TX_STAT_AC) >> 2); 398 dmae->src_addr_hi = 0; 399 dmae->dst_addr_lo = U64_LO(BXE_SP_MAPPING(sc, mac_stats) + 400 offsetof(struct emac_stats, 401 tx_stat_ifhcoutoctets)); 402 dmae->dst_addr_hi = U64_HI(BXE_SP_MAPPING(sc, mac_stats) + 403 offsetof(struct emac_stats, 404 tx_stat_ifhcoutoctets)); 405 dmae->len = EMAC_REG_EMAC_TX_STAT_AC_COUNT; 406 dmae->comp_addr_lo = (dmae_reg_go_c[loader_idx] >> 2); 407 dmae->comp_addr_hi = 0; 408 dmae->comp_val = 1; 409 } else { 410 uint32_t tx_src_addr_lo, rx_src_addr_lo; 411 uint16_t rx_len, tx_len; 412 413 /* configure the params according to MAC type */ 414 switch (sc->link_vars.mac_type) { 415 case ELINK_MAC_TYPE_BMAC: 416 mac_addr = (port) ? NIG_REG_INGRESS_BMAC1_MEM : 417 NIG_REG_INGRESS_BMAC0_MEM; 418 419 /* BIGMAC_REGISTER_TX_STAT_GTPKT .. 420 BIGMAC_REGISTER_TX_STAT_GTBYT */ 421 if (CHIP_IS_E1x(sc)) { 422 tx_src_addr_lo = 423 ((mac_addr + BIGMAC_REGISTER_TX_STAT_GTPKT) >> 2); 424 tx_len = ((8 + BIGMAC_REGISTER_TX_STAT_GTBYT - 425 BIGMAC_REGISTER_TX_STAT_GTPKT) >> 2); 426 rx_src_addr_lo = 427 ((mac_addr + BIGMAC_REGISTER_RX_STAT_GR64) >> 2); 428 rx_len = ((8 + BIGMAC_REGISTER_RX_STAT_GRIPJ - 429 BIGMAC_REGISTER_RX_STAT_GR64) >> 2); 430 } else { 431 tx_src_addr_lo = 432 ((mac_addr + BIGMAC2_REGISTER_TX_STAT_GTPOK) >> 2); 433 tx_len = ((8 + BIGMAC2_REGISTER_TX_STAT_GTBYT - 434 BIGMAC2_REGISTER_TX_STAT_GTPOK) >> 2); 435 rx_src_addr_lo = 436 ((mac_addr + BIGMAC2_REGISTER_RX_STAT_GR64) >> 2); 437 rx_len = ((8 + BIGMAC2_REGISTER_RX_STAT_GRIPJ - 438 BIGMAC2_REGISTER_RX_STAT_GR64) >> 2); 439 } 440 441 break; 442 443 case ELINK_MAC_TYPE_UMAC: /* handled by MSTAT */ 444 case ELINK_MAC_TYPE_XMAC: /* handled by MSTAT */ 445 default: 446 mac_addr = (port) ? GRCBASE_MSTAT1 : GRCBASE_MSTAT0; 447 tx_src_addr_lo = ((mac_addr + MSTAT_REG_TX_STAT_GTXPOK_LO) >> 2); 448 rx_src_addr_lo = ((mac_addr + MSTAT_REG_RX_STAT_GR64_LO) >> 2); 449 tx_len = 450 (sizeof(sc->sp->mac_stats.mstat_stats.stats_tx) >> 2); 451 rx_len = 452 (sizeof(sc->sp->mac_stats.mstat_stats.stats_rx) >> 2); 453 break; 454 } 455 456 /* TX stats */ 457 dmae = BXE_SP(sc, dmae[sc->executer_idx++]); 458 dmae->opcode = opcode; 459 dmae->src_addr_lo = tx_src_addr_lo; 460 dmae->src_addr_hi = 0; 461 dmae->len = tx_len; 462 dmae->dst_addr_lo = U64_LO(BXE_SP_MAPPING(sc, mac_stats)); 463 dmae->dst_addr_hi = U64_HI(BXE_SP_MAPPING(sc, mac_stats)); 464 dmae->comp_addr_lo = dmae_reg_go_c[loader_idx] >> 2; 465 dmae->comp_addr_hi = 0; 466 dmae->comp_val = 1; 467 468 /* RX stats */ 469 dmae = BXE_SP(sc, dmae[sc->executer_idx++]); 470 dmae->opcode = opcode; 471 dmae->src_addr_hi = 0; 472 dmae->src_addr_lo = rx_src_addr_lo; 473 dmae->dst_addr_lo = 474 U64_LO(BXE_SP_MAPPING(sc, mac_stats) + (tx_len << 2)); 475 dmae->dst_addr_hi = 476 U64_HI(BXE_SP_MAPPING(sc, mac_stats) + (tx_len << 2)); 477 dmae->len = rx_len; 478 dmae->comp_addr_lo = dmae_reg_go_c[loader_idx] >> 2; 479 dmae->comp_addr_hi = 0; 480 dmae->comp_val = 1; 481 } 482 483 /* NIG */ 484 if (!CHIP_IS_E3(sc)) { 485 dmae = BXE_SP(sc, dmae[sc->executer_idx++]); 486 dmae->opcode = opcode; 487 dmae->src_addr_lo = 488 (port ? NIG_REG_STAT1_EGRESS_MAC_PKT0 : 489 NIG_REG_STAT0_EGRESS_MAC_PKT0) >> 2; 490 dmae->src_addr_hi = 0; 491 dmae->dst_addr_lo = U64_LO(BXE_SP_MAPPING(sc, nig_stats) + 492 offsetof(struct nig_stats, 493 egress_mac_pkt0_lo)); 494 dmae->dst_addr_hi = U64_HI(BXE_SP_MAPPING(sc, nig_stats) + 495 offsetof(struct nig_stats, 496 egress_mac_pkt0_lo)); 497 dmae->len = ((2 * sizeof(uint32_t)) >> 2); 498 dmae->comp_addr_lo = (dmae_reg_go_c[loader_idx] >> 2); 499 dmae->comp_addr_hi = 0; 500 dmae->comp_val = 1; 501 502 dmae = BXE_SP(sc, dmae[sc->executer_idx++]); 503 dmae->opcode = opcode; 504 dmae->src_addr_lo = 505 (port ? NIG_REG_STAT1_EGRESS_MAC_PKT1 : 506 NIG_REG_STAT0_EGRESS_MAC_PKT1) >> 2; 507 dmae->src_addr_hi = 0; 508 dmae->dst_addr_lo = U64_LO(BXE_SP_MAPPING(sc, nig_stats) + 509 offsetof(struct nig_stats, 510 egress_mac_pkt1_lo)); 511 dmae->dst_addr_hi = U64_HI(BXE_SP_MAPPING(sc, nig_stats) + 512 offsetof(struct nig_stats, 513 egress_mac_pkt1_lo)); 514 dmae->len = ((2 * sizeof(uint32_t)) >> 2); 515 dmae->comp_addr_lo = (dmae_reg_go_c[loader_idx] >> 2); 516 dmae->comp_addr_hi = 0; 517 dmae->comp_val = 1; 518 } 519 520 dmae = BXE_SP(sc, dmae[sc->executer_idx++]); 521 dmae->opcode = bxe_dmae_opcode(sc, DMAE_SRC_GRC, DMAE_DST_PCI, 522 TRUE, DMAE_COMP_PCI); 523 dmae->src_addr_lo = 524 (port ? NIG_REG_STAT1_BRB_DISCARD : 525 NIG_REG_STAT0_BRB_DISCARD) >> 2; 526 dmae->src_addr_hi = 0; 527 dmae->dst_addr_lo = U64_LO(BXE_SP_MAPPING(sc, nig_stats)); 528 dmae->dst_addr_hi = U64_HI(BXE_SP_MAPPING(sc, nig_stats)); 529 dmae->len = (sizeof(struct nig_stats) - 4*sizeof(uint32_t)) >> 2; 530 531 dmae->comp_addr_lo = U64_LO(BXE_SP_MAPPING(sc, stats_comp)); 532 dmae->comp_addr_hi = U64_HI(BXE_SP_MAPPING(sc, stats_comp)); 533 dmae->comp_val = DMAE_COMP_VAL; 534 535 *stats_comp = 0; 536 } 537 538 static void 539 bxe_func_stats_init(struct bxe_softc *sc) 540 { 541 struct dmae_command *dmae = &sc->stats_dmae; 542 uint32_t *stats_comp = BXE_SP(sc, stats_comp); 543 544 /* sanity */ 545 if (!sc->func_stx) { 546 BLOGE(sc, "BUG!\n"); 547 return; 548 } 549 550 sc->executer_idx = 0; 551 memset(dmae, 0, sizeof(struct dmae_command)); 552 553 dmae->opcode = bxe_dmae_opcode(sc, DMAE_SRC_PCI, DMAE_DST_GRC, 554 TRUE, DMAE_COMP_PCI); 555 dmae->src_addr_lo = U64_LO(BXE_SP_MAPPING(sc, func_stats)); 556 dmae->src_addr_hi = U64_HI(BXE_SP_MAPPING(sc, func_stats)); 557 dmae->dst_addr_lo = (sc->func_stx >> 2); 558 dmae->dst_addr_hi = 0; 559 dmae->len = (sizeof(struct host_func_stats) >> 2); 560 dmae->comp_addr_lo = U64_LO(BXE_SP_MAPPING(sc, stats_comp)); 561 dmae->comp_addr_hi = U64_HI(BXE_SP_MAPPING(sc, stats_comp)); 562 dmae->comp_val = DMAE_COMP_VAL; 563 564 *stats_comp = 0; 565 } 566 567 static void 568 bxe_stats_start(struct bxe_softc *sc) 569 { 570 /* 571 * VFs travel through here as part of the statistics FSM, but no action 572 * is required 573 */ 574 if (IS_VF(sc)) { 575 return; 576 } 577 578 if (sc->port.pmf) { 579 bxe_port_stats_init(sc); 580 } 581 582 else if (sc->func_stx) { 583 bxe_func_stats_init(sc); 584 } 585 586 bxe_hw_stats_post(sc); 587 bxe_storm_stats_post(sc); 588 } 589 590 static void 591 bxe_stats_pmf_start(struct bxe_softc *sc) 592 { 593 bxe_stats_comp(sc); 594 bxe_stats_pmf_update(sc); 595 bxe_stats_start(sc); 596 } 597 598 static void 599 bxe_stats_restart(struct bxe_softc *sc) 600 { 601 /* 602 * VFs travel through here as part of the statistics FSM, but no action 603 * is required 604 */ 605 if (IS_VF(sc)) { 606 return; 607 } 608 609 bxe_stats_comp(sc); 610 bxe_stats_start(sc); 611 } 612 613 static void 614 bxe_bmac_stats_update(struct bxe_softc *sc) 615 { 616 struct host_port_stats *pstats = BXE_SP(sc, port_stats); 617 struct bxe_eth_stats *estats = &sc->eth_stats; 618 struct { 619 uint32_t lo; 620 uint32_t hi; 621 } diff; 622 623 if (CHIP_IS_E1x(sc)) { 624 struct bmac1_stats *new = BXE_SP(sc, mac_stats.bmac1_stats); 625 626 /* the macros below will use "bmac1_stats" type */ 627 UPDATE_STAT64(rx_stat_grerb, rx_stat_ifhcinbadoctets); 628 UPDATE_STAT64(rx_stat_grfcs, rx_stat_dot3statsfcserrors); 629 UPDATE_STAT64(rx_stat_grund, rx_stat_etherstatsundersizepkts); 630 UPDATE_STAT64(rx_stat_grovr, rx_stat_dot3statsframestoolong); 631 UPDATE_STAT64(rx_stat_grfrg, rx_stat_etherstatsfragments); 632 UPDATE_STAT64(rx_stat_grjbr, rx_stat_etherstatsjabbers); 633 UPDATE_STAT64(rx_stat_grxcf, rx_stat_maccontrolframesreceived); 634 UPDATE_STAT64(rx_stat_grxpf, rx_stat_xoffstateentered); 635 UPDATE_STAT64(rx_stat_grxpf, rx_stat_mac_xpf); 636 637 UPDATE_STAT64(tx_stat_gtxpf, tx_stat_outxoffsent); 638 UPDATE_STAT64(tx_stat_gtxpf, tx_stat_flowcontroldone); 639 UPDATE_STAT64(tx_stat_gt64, tx_stat_etherstatspkts64octets); 640 UPDATE_STAT64(tx_stat_gt127, 641 tx_stat_etherstatspkts65octetsto127octets); 642 UPDATE_STAT64(tx_stat_gt255, 643 tx_stat_etherstatspkts128octetsto255octets); 644 UPDATE_STAT64(tx_stat_gt511, 645 tx_stat_etherstatspkts256octetsto511octets); 646 UPDATE_STAT64(tx_stat_gt1023, 647 tx_stat_etherstatspkts512octetsto1023octets); 648 UPDATE_STAT64(tx_stat_gt1518, 649 tx_stat_etherstatspkts1024octetsto1522octets); 650 UPDATE_STAT64(tx_stat_gt2047, tx_stat_mac_2047); 651 UPDATE_STAT64(tx_stat_gt4095, tx_stat_mac_4095); 652 UPDATE_STAT64(tx_stat_gt9216, tx_stat_mac_9216); 653 UPDATE_STAT64(tx_stat_gt16383, tx_stat_mac_16383); 654 UPDATE_STAT64(tx_stat_gterr, 655 tx_stat_dot3statsinternalmactransmiterrors); 656 UPDATE_STAT64(tx_stat_gtufl, tx_stat_mac_ufl); 657 } else { 658 struct bmac2_stats *new = BXE_SP(sc, mac_stats.bmac2_stats); 659 struct bxe_fw_port_stats_old *fwstats = &sc->fw_stats_old; 660 661 /* the macros below will use "bmac2_stats" type */ 662 UPDATE_STAT64(rx_stat_grerb, rx_stat_ifhcinbadoctets); 663 UPDATE_STAT64(rx_stat_grfcs, rx_stat_dot3statsfcserrors); 664 UPDATE_STAT64(rx_stat_grund, rx_stat_etherstatsundersizepkts); 665 UPDATE_STAT64(rx_stat_grovr, rx_stat_dot3statsframestoolong); 666 UPDATE_STAT64(rx_stat_grfrg, rx_stat_etherstatsfragments); 667 UPDATE_STAT64(rx_stat_grjbr, rx_stat_etherstatsjabbers); 668 UPDATE_STAT64(rx_stat_grxcf, rx_stat_maccontrolframesreceived); 669 UPDATE_STAT64(rx_stat_grxpf, rx_stat_xoffstateentered); 670 UPDATE_STAT64(rx_stat_grxpf, rx_stat_mac_xpf); 671 UPDATE_STAT64(tx_stat_gtxpf, tx_stat_outxoffsent); 672 UPDATE_STAT64(tx_stat_gtxpf, tx_stat_flowcontroldone); 673 UPDATE_STAT64(tx_stat_gt64, tx_stat_etherstatspkts64octets); 674 UPDATE_STAT64(tx_stat_gt127, 675 tx_stat_etherstatspkts65octetsto127octets); 676 UPDATE_STAT64(tx_stat_gt255, 677 tx_stat_etherstatspkts128octetsto255octets); 678 UPDATE_STAT64(tx_stat_gt511, 679 tx_stat_etherstatspkts256octetsto511octets); 680 UPDATE_STAT64(tx_stat_gt1023, 681 tx_stat_etherstatspkts512octetsto1023octets); 682 UPDATE_STAT64(tx_stat_gt1518, 683 tx_stat_etherstatspkts1024octetsto1522octets); 684 UPDATE_STAT64(tx_stat_gt2047, tx_stat_mac_2047); 685 UPDATE_STAT64(tx_stat_gt4095, tx_stat_mac_4095); 686 UPDATE_STAT64(tx_stat_gt9216, tx_stat_mac_9216); 687 UPDATE_STAT64(tx_stat_gt16383, tx_stat_mac_16383); 688 UPDATE_STAT64(tx_stat_gterr, 689 tx_stat_dot3statsinternalmactransmiterrors); 690 UPDATE_STAT64(tx_stat_gtufl, tx_stat_mac_ufl); 691 692 /* collect PFC stats */ 693 pstats->pfc_frames_tx_hi = new->tx_stat_gtpp_hi; 694 pstats->pfc_frames_tx_lo = new->tx_stat_gtpp_lo; 695 ADD_64(pstats->pfc_frames_tx_hi, fwstats->pfc_frames_tx_hi, 696 pstats->pfc_frames_tx_lo, fwstats->pfc_frames_tx_lo); 697 698 pstats->pfc_frames_rx_hi = new->rx_stat_grpp_hi; 699 pstats->pfc_frames_rx_lo = new->rx_stat_grpp_lo; 700 ADD_64(pstats->pfc_frames_rx_hi, fwstats->pfc_frames_rx_hi, 701 pstats->pfc_frames_rx_lo, fwstats->pfc_frames_rx_lo); 702 } 703 704 estats->pause_frames_received_hi = pstats->mac_stx[1].rx_stat_mac_xpf_hi; 705 estats->pause_frames_received_lo = pstats->mac_stx[1].rx_stat_mac_xpf_lo; 706 707 estats->pause_frames_sent_hi = pstats->mac_stx[1].tx_stat_outxoffsent_hi; 708 estats->pause_frames_sent_lo = pstats->mac_stx[1].tx_stat_outxoffsent_lo; 709 710 estats->pfc_frames_received_hi = pstats->pfc_frames_rx_hi; 711 estats->pfc_frames_received_lo = pstats->pfc_frames_rx_lo; 712 estats->pfc_frames_sent_hi = pstats->pfc_frames_tx_hi; 713 estats->pfc_frames_sent_lo = pstats->pfc_frames_tx_lo; 714 } 715 716 static void 717 bxe_mstat_stats_update(struct bxe_softc *sc) 718 { 719 struct host_port_stats *pstats = BXE_SP(sc, port_stats); 720 struct bxe_eth_stats *estats = &sc->eth_stats; 721 struct mstat_stats *new = BXE_SP(sc, mac_stats.mstat_stats); 722 723 ADD_STAT64(stats_rx.rx_grerb, rx_stat_ifhcinbadoctets); 724 ADD_STAT64(stats_rx.rx_grfcs, rx_stat_dot3statsfcserrors); 725 ADD_STAT64(stats_rx.rx_grund, rx_stat_etherstatsundersizepkts); 726 ADD_STAT64(stats_rx.rx_grovr, rx_stat_dot3statsframestoolong); 727 ADD_STAT64(stats_rx.rx_grfrg, rx_stat_etherstatsfragments); 728 ADD_STAT64(stats_rx.rx_grxcf, rx_stat_maccontrolframesreceived); 729 ADD_STAT64(stats_rx.rx_grxpf, rx_stat_xoffstateentered); 730 ADD_STAT64(stats_rx.rx_grxpf, rx_stat_mac_xpf); 731 ADD_STAT64(stats_tx.tx_gtxpf, tx_stat_outxoffsent); 732 ADD_STAT64(stats_tx.tx_gtxpf, tx_stat_flowcontroldone); 733 734 /* collect pfc stats */ 735 ADD_64(pstats->pfc_frames_tx_hi, new->stats_tx.tx_gtxpp_hi, 736 pstats->pfc_frames_tx_lo, new->stats_tx.tx_gtxpp_lo); 737 ADD_64(pstats->pfc_frames_rx_hi, new->stats_rx.rx_grxpp_hi, 738 pstats->pfc_frames_rx_lo, new->stats_rx.rx_grxpp_lo); 739 740 ADD_STAT64(stats_tx.tx_gt64, tx_stat_etherstatspkts64octets); 741 ADD_STAT64(stats_tx.tx_gt127, tx_stat_etherstatspkts65octetsto127octets); 742 ADD_STAT64(stats_tx.tx_gt255, tx_stat_etherstatspkts128octetsto255octets); 743 ADD_STAT64(stats_tx.tx_gt511, tx_stat_etherstatspkts256octetsto511octets); 744 ADD_STAT64(stats_tx.tx_gt1023, 745 tx_stat_etherstatspkts512octetsto1023octets); 746 ADD_STAT64(stats_tx.tx_gt1518, 747 tx_stat_etherstatspkts1024octetsto1522octets); 748 ADD_STAT64(stats_tx.tx_gt2047, tx_stat_mac_2047); 749 750 ADD_STAT64(stats_tx.tx_gt4095, tx_stat_mac_4095); 751 ADD_STAT64(stats_tx.tx_gt9216, tx_stat_mac_9216); 752 ADD_STAT64(stats_tx.tx_gt16383, tx_stat_mac_16383); 753 754 ADD_STAT64(stats_tx.tx_gterr, tx_stat_dot3statsinternalmactransmiterrors); 755 ADD_STAT64(stats_tx.tx_gtufl, tx_stat_mac_ufl); 756 757 estats->etherstatspkts1024octetsto1522octets_hi = 758 pstats->mac_stx[1].tx_stat_etherstatspkts1024octetsto1522octets_hi; 759 estats->etherstatspkts1024octetsto1522octets_lo = 760 pstats->mac_stx[1].tx_stat_etherstatspkts1024octetsto1522octets_lo; 761 762 estats->etherstatspktsover1522octets_hi = 763 pstats->mac_stx[1].tx_stat_mac_2047_hi; 764 estats->etherstatspktsover1522octets_lo = 765 pstats->mac_stx[1].tx_stat_mac_2047_lo; 766 767 ADD_64(estats->etherstatspktsover1522octets_hi, 768 pstats->mac_stx[1].tx_stat_mac_4095_hi, 769 estats->etherstatspktsover1522octets_lo, 770 pstats->mac_stx[1].tx_stat_mac_4095_lo); 771 772 ADD_64(estats->etherstatspktsover1522octets_hi, 773 pstats->mac_stx[1].tx_stat_mac_9216_hi, 774 estats->etherstatspktsover1522octets_lo, 775 pstats->mac_stx[1].tx_stat_mac_9216_lo); 776 777 ADD_64(estats->etherstatspktsover1522octets_hi, 778 pstats->mac_stx[1].tx_stat_mac_16383_hi, 779 estats->etherstatspktsover1522octets_lo, 780 pstats->mac_stx[1].tx_stat_mac_16383_lo); 781 782 estats->pause_frames_received_hi = pstats->mac_stx[1].rx_stat_mac_xpf_hi; 783 estats->pause_frames_received_lo = pstats->mac_stx[1].rx_stat_mac_xpf_lo; 784 785 estats->pause_frames_sent_hi = pstats->mac_stx[1].tx_stat_outxoffsent_hi; 786 estats->pause_frames_sent_lo = pstats->mac_stx[1].tx_stat_outxoffsent_lo; 787 788 estats->pfc_frames_received_hi = pstats->pfc_frames_rx_hi; 789 estats->pfc_frames_received_lo = pstats->pfc_frames_rx_lo; 790 estats->pfc_frames_sent_hi = pstats->pfc_frames_tx_hi; 791 estats->pfc_frames_sent_lo = pstats->pfc_frames_tx_lo; 792 } 793 794 static void 795 bxe_emac_stats_update(struct bxe_softc *sc) 796 { 797 struct emac_stats *new = BXE_SP(sc, mac_stats.emac_stats); 798 struct host_port_stats *pstats = BXE_SP(sc, port_stats); 799 struct bxe_eth_stats *estats = &sc->eth_stats; 800 801 UPDATE_EXTEND_STAT(rx_stat_ifhcinbadoctets); 802 UPDATE_EXTEND_STAT(tx_stat_ifhcoutbadoctets); 803 UPDATE_EXTEND_STAT(rx_stat_dot3statsfcserrors); 804 UPDATE_EXTEND_STAT(rx_stat_dot3statsalignmenterrors); 805 UPDATE_EXTEND_STAT(rx_stat_dot3statscarriersenseerrors); 806 UPDATE_EXTEND_STAT(rx_stat_falsecarriererrors); 807 UPDATE_EXTEND_STAT(rx_stat_etherstatsundersizepkts); 808 UPDATE_EXTEND_STAT(rx_stat_dot3statsframestoolong); 809 UPDATE_EXTEND_STAT(rx_stat_etherstatsfragments); 810 UPDATE_EXTEND_STAT(rx_stat_etherstatsjabbers); 811 UPDATE_EXTEND_STAT(rx_stat_maccontrolframesreceived); 812 UPDATE_EXTEND_STAT(rx_stat_xoffstateentered); 813 UPDATE_EXTEND_STAT(rx_stat_xonpauseframesreceived); 814 UPDATE_EXTEND_STAT(rx_stat_xoffpauseframesreceived); 815 UPDATE_EXTEND_STAT(tx_stat_outxonsent); 816 UPDATE_EXTEND_STAT(tx_stat_outxoffsent); 817 UPDATE_EXTEND_STAT(tx_stat_flowcontroldone); 818 UPDATE_EXTEND_STAT(tx_stat_etherstatscollisions); 819 UPDATE_EXTEND_STAT(tx_stat_dot3statssinglecollisionframes); 820 UPDATE_EXTEND_STAT(tx_stat_dot3statsmultiplecollisionframes); 821 UPDATE_EXTEND_STAT(tx_stat_dot3statsdeferredtransmissions); 822 UPDATE_EXTEND_STAT(tx_stat_dot3statsexcessivecollisions); 823 UPDATE_EXTEND_STAT(tx_stat_dot3statslatecollisions); 824 UPDATE_EXTEND_STAT(tx_stat_etherstatspkts64octets); 825 UPDATE_EXTEND_STAT(tx_stat_etherstatspkts65octetsto127octets); 826 UPDATE_EXTEND_STAT(tx_stat_etherstatspkts128octetsto255octets); 827 UPDATE_EXTEND_STAT(tx_stat_etherstatspkts256octetsto511octets); 828 UPDATE_EXTEND_STAT(tx_stat_etherstatspkts512octetsto1023octets); 829 UPDATE_EXTEND_STAT(tx_stat_etherstatspkts1024octetsto1522octets); 830 UPDATE_EXTEND_STAT(tx_stat_etherstatspktsover1522octets); 831 UPDATE_EXTEND_STAT(tx_stat_dot3statsinternalmactransmiterrors); 832 833 estats->pause_frames_received_hi = 834 pstats->mac_stx[1].rx_stat_xonpauseframesreceived_hi; 835 estats->pause_frames_received_lo = 836 pstats->mac_stx[1].rx_stat_xonpauseframesreceived_lo; 837 ADD_64(estats->pause_frames_received_hi, 838 pstats->mac_stx[1].rx_stat_xoffpauseframesreceived_hi, 839 estats->pause_frames_received_lo, 840 pstats->mac_stx[1].rx_stat_xoffpauseframesreceived_lo); 841 842 estats->pause_frames_sent_hi = 843 pstats->mac_stx[1].tx_stat_outxonsent_hi; 844 estats->pause_frames_sent_lo = 845 pstats->mac_stx[1].tx_stat_outxonsent_lo; 846 ADD_64(estats->pause_frames_sent_hi, 847 pstats->mac_stx[1].tx_stat_outxoffsent_hi, 848 estats->pause_frames_sent_lo, 849 pstats->mac_stx[1].tx_stat_outxoffsent_lo); 850 } 851 852 static int 853 bxe_hw_stats_update(struct bxe_softc *sc) 854 { 855 struct nig_stats *new = BXE_SP(sc, nig_stats); 856 struct nig_stats *old = &(sc->port.old_nig_stats); 857 struct host_port_stats *pstats = BXE_SP(sc, port_stats); 858 struct bxe_eth_stats *estats = &sc->eth_stats; 859 uint32_t lpi_reg, nig_timer_max; 860 struct { 861 uint32_t lo; 862 uint32_t hi; 863 } diff; 864 865 switch (sc->link_vars.mac_type) { 866 case ELINK_MAC_TYPE_BMAC: 867 bxe_bmac_stats_update(sc); 868 break; 869 870 case ELINK_MAC_TYPE_EMAC: 871 bxe_emac_stats_update(sc); 872 break; 873 874 case ELINK_MAC_TYPE_UMAC: 875 case ELINK_MAC_TYPE_XMAC: 876 bxe_mstat_stats_update(sc); 877 break; 878 879 case ELINK_MAC_TYPE_NONE: /* unreached */ 880 BLOGD(sc, DBG_STATS, 881 "stats updated by DMAE but no MAC active\n"); 882 return (-1); 883 884 default: /* unreached */ 885 BLOGE(sc, "stats update failed, unknown MAC type\n"); 886 } 887 888 ADD_EXTEND_64(pstats->brb_drop_hi, pstats->brb_drop_lo, 889 new->brb_discard - old->brb_discard); 890 ADD_EXTEND_64(estats->brb_truncate_hi, estats->brb_truncate_lo, 891 new->brb_truncate - old->brb_truncate); 892 893 if (!CHIP_IS_E3(sc)) { 894 UPDATE_STAT64_NIG(egress_mac_pkt0, 895 etherstatspkts1024octetsto1522octets); 896 UPDATE_STAT64_NIG(egress_mac_pkt1, 897 etherstatspktsover1522octets); 898 } 899 900 memcpy(old, new, sizeof(struct nig_stats)); 901 902 memcpy(&(estats->rx_stat_ifhcinbadoctets_hi), &(pstats->mac_stx[1]), 903 sizeof(struct mac_stx)); 904 estats->brb_drop_hi = pstats->brb_drop_hi; 905 estats->brb_drop_lo = pstats->brb_drop_lo; 906 907 pstats->host_port_stats_counter++; 908 909 if (CHIP_IS_E3(sc)) { 910 lpi_reg = (SC_PORT(sc)) ? 911 MISC_REG_CPMU_LP_SM_ENT_CNT_P1 : 912 MISC_REG_CPMU_LP_SM_ENT_CNT_P0; 913 estats->eee_tx_lpi += REG_RD(sc, lpi_reg); 914 } 915 916 if (!BXE_NOMCP(sc)) { 917 nig_timer_max = SHMEM_RD(sc, port_mb[SC_PORT(sc)].stat_nig_timer); 918 if (nig_timer_max != estats->nig_timer_max) { 919 estats->nig_timer_max = nig_timer_max; 920 BLOGE(sc, "invalid NIG timer max (%u)\n", 921 estats->nig_timer_max); 922 } 923 } 924 925 return (0); 926 } 927 928 static int 929 bxe_storm_stats_validate_counters(struct bxe_softc *sc) 930 { 931 struct stats_counter *counters = &sc->fw_stats_data->storm_counters; 932 uint16_t cur_stats_counter; 933 934 /* 935 * Make sure we use the value of the counter 936 * used for sending the last stats ramrod. 937 */ 938 BXE_STATS_LOCK(sc); 939 cur_stats_counter = (sc->stats_counter - 1); 940 BXE_STATS_UNLOCK(sc); 941 942 /* are storm stats valid? */ 943 if (le16toh(counters->xstats_counter) != cur_stats_counter) { 944 BLOGD(sc, DBG_STATS, 945 "stats not updated by xstorm, " 946 "counter 0x%x != stats_counter 0x%x\n", 947 le16toh(counters->xstats_counter), sc->stats_counter); 948 return (-EAGAIN); 949 } 950 951 if (le16toh(counters->ustats_counter) != cur_stats_counter) { 952 BLOGD(sc, DBG_STATS, 953 "stats not updated by ustorm, " 954 "counter 0x%x != stats_counter 0x%x\n", 955 le16toh(counters->ustats_counter), sc->stats_counter); 956 return (-EAGAIN); 957 } 958 959 if (le16toh(counters->cstats_counter) != cur_stats_counter) { 960 BLOGD(sc, DBG_STATS, 961 "stats not updated by cstorm, " 962 "counter 0x%x != stats_counter 0x%x\n", 963 le16toh(counters->cstats_counter), sc->stats_counter); 964 return (-EAGAIN); 965 } 966 967 if (le16toh(counters->tstats_counter) != cur_stats_counter) { 968 BLOGD(sc, DBG_STATS, 969 "stats not updated by tstorm, " 970 "counter 0x%x != stats_counter 0x%x\n", 971 le16toh(counters->tstats_counter), sc->stats_counter); 972 return (-EAGAIN); 973 } 974 975 return (0); 976 } 977 978 static int 979 bxe_storm_stats_update(struct bxe_softc *sc) 980 { 981 struct tstorm_per_port_stats *tport = 982 &sc->fw_stats_data->port.tstorm_port_statistics; 983 struct tstorm_per_pf_stats *tfunc = 984 &sc->fw_stats_data->pf.tstorm_pf_statistics; 985 struct host_func_stats *fstats = &sc->func_stats; 986 struct bxe_eth_stats *estats = &sc->eth_stats; 987 struct bxe_eth_stats_old *estats_old = &sc->eth_stats_old; 988 int i; 989 990 /* vfs stat counter is managed by pf */ 991 if (IS_PF(sc) && bxe_storm_stats_validate_counters(sc)) { 992 return (-EAGAIN); 993 } 994 995 estats->error_bytes_received_hi = 0; 996 estats->error_bytes_received_lo = 0; 997 998 for (i = 0; i < sc->num_queues; i++) { 999 struct bxe_fastpath *fp = &sc->fp[i]; 1000 struct tstorm_per_queue_stats *tclient = 1001 &sc->fw_stats_data->queue_stats[i].tstorm_queue_statistics; 1002 struct tstorm_per_queue_stats *old_tclient = &fp->old_tclient; 1003 struct ustorm_per_queue_stats *uclient = 1004 &sc->fw_stats_data->queue_stats[i].ustorm_queue_statistics; 1005 struct ustorm_per_queue_stats *old_uclient = &fp->old_uclient; 1006 struct xstorm_per_queue_stats *xclient = 1007 &sc->fw_stats_data->queue_stats[i].xstorm_queue_statistics; 1008 struct xstorm_per_queue_stats *old_xclient = &fp->old_xclient; 1009 struct bxe_eth_q_stats *qstats = &fp->eth_q_stats; 1010 struct bxe_eth_q_stats_old *qstats_old = &fp->eth_q_stats_old; 1011 1012 uint32_t diff; 1013 1014 BLOGD(sc, DBG_STATS, 1015 "queue[%d]: ucast_sent 0x%x bcast_sent 0x%x mcast_sent 0x%x\n", 1016 i, xclient->ucast_pkts_sent, xclient->bcast_pkts_sent, 1017 xclient->mcast_pkts_sent); 1018 1019 BLOGD(sc, DBG_STATS, "---------------\n"); 1020 1021 UPDATE_QSTAT(tclient->rcv_bcast_bytes, 1022 total_broadcast_bytes_received); 1023 UPDATE_QSTAT(tclient->rcv_mcast_bytes, 1024 total_multicast_bytes_received); 1025 UPDATE_QSTAT(tclient->rcv_ucast_bytes, 1026 total_unicast_bytes_received); 1027 1028 /* 1029 * sum to total_bytes_received all 1030 * unicast/multicast/broadcast 1031 */ 1032 qstats->total_bytes_received_hi = 1033 qstats->total_broadcast_bytes_received_hi; 1034 qstats->total_bytes_received_lo = 1035 qstats->total_broadcast_bytes_received_lo; 1036 1037 ADD_64(qstats->total_bytes_received_hi, 1038 qstats->total_multicast_bytes_received_hi, 1039 qstats->total_bytes_received_lo, 1040 qstats->total_multicast_bytes_received_lo); 1041 1042 ADD_64(qstats->total_bytes_received_hi, 1043 qstats->total_unicast_bytes_received_hi, 1044 qstats->total_bytes_received_lo, 1045 qstats->total_unicast_bytes_received_lo); 1046 1047 qstats->valid_bytes_received_hi = qstats->total_bytes_received_hi; 1048 qstats->valid_bytes_received_lo = qstats->total_bytes_received_lo; 1049 1050 UPDATE_EXTEND_TSTAT(rcv_ucast_pkts, total_unicast_packets_received); 1051 UPDATE_EXTEND_TSTAT(rcv_mcast_pkts, total_multicast_packets_received); 1052 UPDATE_EXTEND_TSTAT(rcv_bcast_pkts, total_broadcast_packets_received); 1053 UPDATE_EXTEND_E_TSTAT(pkts_too_big_discard, 1054 etherstatsoverrsizepkts, 32); 1055 UPDATE_EXTEND_E_TSTAT(no_buff_discard, no_buff_discard, 16); 1056 1057 SUB_EXTEND_USTAT(ucast_no_buff_pkts, total_unicast_packets_received); 1058 SUB_EXTEND_USTAT(mcast_no_buff_pkts, 1059 total_multicast_packets_received); 1060 SUB_EXTEND_USTAT(bcast_no_buff_pkts, 1061 total_broadcast_packets_received); 1062 UPDATE_EXTEND_E_USTAT(ucast_no_buff_pkts, no_buff_discard); 1063 UPDATE_EXTEND_E_USTAT(mcast_no_buff_pkts, no_buff_discard); 1064 UPDATE_EXTEND_E_USTAT(bcast_no_buff_pkts, no_buff_discard); 1065 1066 UPDATE_QSTAT(xclient->bcast_bytes_sent, 1067 total_broadcast_bytes_transmitted); 1068 UPDATE_QSTAT(xclient->mcast_bytes_sent, 1069 total_multicast_bytes_transmitted); 1070 UPDATE_QSTAT(xclient->ucast_bytes_sent, 1071 total_unicast_bytes_transmitted); 1072 1073 /* 1074 * sum to total_bytes_transmitted all 1075 * unicast/multicast/broadcast 1076 */ 1077 qstats->total_bytes_transmitted_hi = 1078 qstats->total_unicast_bytes_transmitted_hi; 1079 qstats->total_bytes_transmitted_lo = 1080 qstats->total_unicast_bytes_transmitted_lo; 1081 1082 ADD_64(qstats->total_bytes_transmitted_hi, 1083 qstats->total_broadcast_bytes_transmitted_hi, 1084 qstats->total_bytes_transmitted_lo, 1085 qstats->total_broadcast_bytes_transmitted_lo); 1086 1087 ADD_64(qstats->total_bytes_transmitted_hi, 1088 qstats->total_multicast_bytes_transmitted_hi, 1089 qstats->total_bytes_transmitted_lo, 1090 qstats->total_multicast_bytes_transmitted_lo); 1091 1092 UPDATE_EXTEND_XSTAT(ucast_pkts_sent, 1093 total_unicast_packets_transmitted); 1094 UPDATE_EXTEND_XSTAT(mcast_pkts_sent, 1095 total_multicast_packets_transmitted); 1096 UPDATE_EXTEND_XSTAT(bcast_pkts_sent, 1097 total_broadcast_packets_transmitted); 1098 1099 UPDATE_EXTEND_TSTAT(checksum_discard, 1100 total_packets_received_checksum_discarded); 1101 UPDATE_EXTEND_TSTAT(ttl0_discard, 1102 total_packets_received_ttl0_discarded); 1103 1104 UPDATE_EXTEND_XSTAT(error_drop_pkts, 1105 total_transmitted_dropped_packets_error); 1106 1107 /* TPA aggregations completed */ 1108 UPDATE_EXTEND_E_USTAT(coalesced_events, total_tpa_aggregations); 1109 /* Number of network frames aggregated by TPA */ 1110 UPDATE_EXTEND_E_USTAT(coalesced_pkts, total_tpa_aggregated_frames); 1111 /* Total number of bytes in completed TPA aggregations */ 1112 UPDATE_QSTAT(uclient->coalesced_bytes, total_tpa_bytes); 1113 1114 UPDATE_ESTAT_QSTAT_64(total_tpa_bytes); 1115 1116 UPDATE_FSTAT_QSTAT(total_bytes_received); 1117 UPDATE_FSTAT_QSTAT(total_bytes_transmitted); 1118 UPDATE_FSTAT_QSTAT(total_unicast_packets_received); 1119 UPDATE_FSTAT_QSTAT(total_multicast_packets_received); 1120 UPDATE_FSTAT_QSTAT(total_broadcast_packets_received); 1121 UPDATE_FSTAT_QSTAT(total_unicast_packets_transmitted); 1122 UPDATE_FSTAT_QSTAT(total_multicast_packets_transmitted); 1123 UPDATE_FSTAT_QSTAT(total_broadcast_packets_transmitted); 1124 UPDATE_FSTAT_QSTAT(valid_bytes_received); 1125 } 1126 1127 ADD_64(estats->total_bytes_received_hi, 1128 estats->rx_stat_ifhcinbadoctets_hi, 1129 estats->total_bytes_received_lo, 1130 estats->rx_stat_ifhcinbadoctets_lo); 1131 1132 ADD_64_LE(estats->total_bytes_received_hi, 1133 tfunc->rcv_error_bytes.hi, 1134 estats->total_bytes_received_lo, 1135 tfunc->rcv_error_bytes.lo); 1136 1137 ADD_64_LE(estats->error_bytes_received_hi, 1138 tfunc->rcv_error_bytes.hi, 1139 estats->error_bytes_received_lo, 1140 tfunc->rcv_error_bytes.lo); 1141 1142 UPDATE_ESTAT(etherstatsoverrsizepkts, rx_stat_dot3statsframestoolong); 1143 1144 ADD_64(estats->error_bytes_received_hi, 1145 estats->rx_stat_ifhcinbadoctets_hi, 1146 estats->error_bytes_received_lo, 1147 estats->rx_stat_ifhcinbadoctets_lo); 1148 1149 if (sc->port.pmf) { 1150 struct bxe_fw_port_stats_old *fwstats = &sc->fw_stats_old; 1151 UPDATE_FW_STAT(mac_filter_discard); 1152 UPDATE_FW_STAT(mf_tag_discard); 1153 UPDATE_FW_STAT(brb_truncate_discard); 1154 UPDATE_FW_STAT(mac_discard); 1155 } 1156 1157 fstats->host_func_stats_start = ++fstats->host_func_stats_end; 1158 1159 sc->stats_pending = 0; 1160 1161 return (0); 1162 } 1163 1164 static void 1165 bxe_net_stats_update(struct bxe_softc *sc) 1166 { 1167 struct bxe_eth_stats *estats = &sc->eth_stats; 1168 struct ifnet *ifnet = sc->ifnet; 1169 unsigned long tmp; 1170 int i; 1171 1172 ifnet->if_data.ifi_ipackets = 1173 bxe_hilo(&estats->total_unicast_packets_received_hi) + 1174 bxe_hilo(&estats->total_multicast_packets_received_hi) + 1175 bxe_hilo(&estats->total_broadcast_packets_received_hi); 1176 1177 ifnet->if_data.ifi_opackets = 1178 bxe_hilo(&estats->total_unicast_packets_transmitted_hi) + 1179 bxe_hilo(&estats->total_multicast_packets_transmitted_hi) + 1180 bxe_hilo(&estats->total_broadcast_packets_transmitted_hi); 1181 1182 ifnet->if_data.ifi_ibytes = bxe_hilo(&estats->total_bytes_received_hi); 1183 1184 ifnet->if_data.ifi_obytes = bxe_hilo(&estats->total_bytes_transmitted_hi); 1185 1186 tmp = 0; 1187 for (i = 0; i < sc->num_queues; i++) { 1188 struct tstorm_per_queue_stats *old_tclient = 1189 &sc->fp[i].old_tclient; 1190 tmp += le32toh(old_tclient->checksum_discard); 1191 } 1192 1193 ifnet->if_data.ifi_iqdrops = tmp; 1194 1195 ifnet->if_data.ifi_ierrors = 1196 bxe_hilo(&estats->rx_stat_etherstatsundersizepkts_hi) + 1197 bxe_hilo(&estats->etherstatsoverrsizepkts_hi) + 1198 bxe_hilo(&estats->brb_drop_hi) + 1199 bxe_hilo(&estats->brb_truncate_hi) + 1200 bxe_hilo(&estats->rx_stat_dot3statsfcserrors_hi) + 1201 bxe_hilo(&estats->rx_stat_dot3statsalignmenterrors_hi) + 1202 bxe_hilo(&estats->no_buff_discard_hi); 1203 1204 ifnet->if_data.ifi_oerrors = 1205 bxe_hilo(&estats->rx_stat_dot3statscarriersenseerrors_hi) + 1206 bxe_hilo(&estats->tx_stat_dot3statsinternalmactransmiterrors_hi); 1207 1208 ifnet->if_data.ifi_imcasts = 1209 bxe_hilo(&estats->total_multicast_packets_received_hi); 1210 1211 ifnet->if_data.ifi_collisions = 1212 bxe_hilo(&estats->tx_stat_etherstatscollisions_hi) + 1213 bxe_hilo(&estats->tx_stat_dot3statslatecollisions_hi) + 1214 bxe_hilo(&estats->tx_stat_dot3statsexcessivecollisions_hi); 1215 } 1216 1217 static void 1218 bxe_drv_stats_update(struct bxe_softc *sc) 1219 { 1220 struct bxe_eth_stats *estats = &sc->eth_stats; 1221 int i; 1222 1223 for (i = 0; i < sc->num_queues; i++) { 1224 struct bxe_eth_q_stats *qstats = &sc->fp[i].eth_q_stats; 1225 struct bxe_eth_q_stats_old *qstats_old = &sc->fp[i].eth_q_stats_old; 1226 1227 UPDATE_ESTAT_QSTAT(rx_calls); 1228 UPDATE_ESTAT_QSTAT(rx_pkts); 1229 UPDATE_ESTAT_QSTAT(rx_tpa_pkts); 1230 UPDATE_ESTAT_QSTAT(rx_soft_errors); 1231 UPDATE_ESTAT_QSTAT(rx_hw_csum_errors); 1232 UPDATE_ESTAT_QSTAT(rx_ofld_frames_csum_ip); 1233 UPDATE_ESTAT_QSTAT(rx_ofld_frames_csum_tcp_udp); 1234 UPDATE_ESTAT_QSTAT(rx_budget_reached); 1235 UPDATE_ESTAT_QSTAT(tx_pkts); 1236 UPDATE_ESTAT_QSTAT(tx_soft_errors); 1237 UPDATE_ESTAT_QSTAT(tx_ofld_frames_csum_ip); 1238 UPDATE_ESTAT_QSTAT(tx_ofld_frames_csum_tcp); 1239 UPDATE_ESTAT_QSTAT(tx_ofld_frames_csum_udp); 1240 UPDATE_ESTAT_QSTAT(tx_ofld_frames_lso); 1241 UPDATE_ESTAT_QSTAT(tx_ofld_frames_lso_hdr_splits); 1242 UPDATE_ESTAT_QSTAT(tx_encap_failures); 1243 UPDATE_ESTAT_QSTAT(tx_hw_queue_full); 1244 UPDATE_ESTAT_QSTAT(tx_hw_max_queue_depth); 1245 UPDATE_ESTAT_QSTAT(tx_dma_mapping_failure); 1246 UPDATE_ESTAT_QSTAT(tx_max_drbr_queue_depth); 1247 UPDATE_ESTAT_QSTAT(tx_window_violation_std); 1248 UPDATE_ESTAT_QSTAT(tx_window_violation_tso); 1249 //UPDATE_ESTAT_QSTAT(tx_unsupported_tso_request_ipv6); 1250 //UPDATE_ESTAT_QSTAT(tx_unsupported_tso_request_not_tcp); 1251 UPDATE_ESTAT_QSTAT(tx_chain_lost_mbuf); 1252 UPDATE_ESTAT_QSTAT(tx_frames_deferred); 1253 UPDATE_ESTAT_QSTAT(tx_queue_xoff); 1254 1255 /* mbuf driver statistics */ 1256 UPDATE_ESTAT_QSTAT(mbuf_defrag_attempts); 1257 UPDATE_ESTAT_QSTAT(mbuf_defrag_failures); 1258 UPDATE_ESTAT_QSTAT(mbuf_rx_bd_alloc_failed); 1259 UPDATE_ESTAT_QSTAT(mbuf_rx_bd_mapping_failed); 1260 UPDATE_ESTAT_QSTAT(mbuf_rx_tpa_alloc_failed); 1261 UPDATE_ESTAT_QSTAT(mbuf_rx_tpa_mapping_failed); 1262 UPDATE_ESTAT_QSTAT(mbuf_rx_sge_alloc_failed); 1263 UPDATE_ESTAT_QSTAT(mbuf_rx_sge_mapping_failed); 1264 1265 /* track the number of allocated mbufs */ 1266 UPDATE_ESTAT_QSTAT(mbuf_alloc_tx); 1267 UPDATE_ESTAT_QSTAT(mbuf_alloc_rx); 1268 UPDATE_ESTAT_QSTAT(mbuf_alloc_sge); 1269 UPDATE_ESTAT_QSTAT(mbuf_alloc_tpa); 1270 } 1271 } 1272 1273 static uint8_t 1274 bxe_edebug_stats_stopped(struct bxe_softc *sc) 1275 { 1276 uint32_t val; 1277 1278 if (SHMEM2_HAS(sc, edebug_driver_if[1])) { 1279 val = SHMEM2_RD(sc, edebug_driver_if[1]); 1280 1281 if (val == EDEBUG_DRIVER_IF_OP_CODE_DISABLE_STAT) { 1282 return (TRUE); 1283 } 1284 } 1285 1286 return (FALSE); 1287 } 1288 1289 static void 1290 bxe_stats_update(struct bxe_softc *sc) 1291 { 1292 uint32_t *stats_comp = BXE_SP(sc, stats_comp); 1293 1294 if (bxe_edebug_stats_stopped(sc)) { 1295 return; 1296 } 1297 1298 if (IS_PF(sc)) { 1299 if (*stats_comp != DMAE_COMP_VAL) { 1300 return; 1301 } 1302 1303 if (sc->port.pmf) { 1304 bxe_hw_stats_update(sc); 1305 } 1306 1307 if (bxe_storm_stats_update(sc)) { 1308 if (sc->stats_pending++ == 3) { 1309 bxe_panic(sc, ("storm stats not updated for 3 times\n")); 1310 } 1311 return; 1312 } 1313 } else { 1314 /* 1315 * VF doesn't collect HW statistics, and doesn't get completions, 1316 * performs only update. 1317 */ 1318 bxe_storm_stats_update(sc); 1319 } 1320 1321 bxe_net_stats_update(sc); 1322 bxe_drv_stats_update(sc); 1323 1324 /* vf is done */ 1325 if (IS_VF(sc)) { 1326 return; 1327 } 1328 1329 bxe_hw_stats_post(sc); 1330 bxe_storm_stats_post(sc); 1331 } 1332 1333 static void 1334 bxe_port_stats_stop(struct bxe_softc *sc) 1335 { 1336 struct dmae_command *dmae; 1337 uint32_t opcode; 1338 int loader_idx = PMF_DMAE_C(sc); 1339 uint32_t *stats_comp = BXE_SP(sc, stats_comp); 1340 1341 sc->executer_idx = 0; 1342 1343 opcode = bxe_dmae_opcode(sc, DMAE_SRC_PCI, DMAE_DST_GRC, FALSE, 0); 1344 1345 if (sc->port.port_stx) { 1346 dmae = BXE_SP(sc, dmae[sc->executer_idx++]); 1347 1348 if (sc->func_stx) { 1349 dmae->opcode = bxe_dmae_opcode_add_comp(opcode, DMAE_COMP_GRC); 1350 } else { 1351 dmae->opcode = bxe_dmae_opcode_add_comp(opcode, DMAE_COMP_PCI); 1352 } 1353 1354 dmae->src_addr_lo = U64_LO(BXE_SP_MAPPING(sc, port_stats)); 1355 dmae->src_addr_hi = U64_HI(BXE_SP_MAPPING(sc, port_stats)); 1356 dmae->dst_addr_lo = sc->port.port_stx >> 2; 1357 dmae->dst_addr_hi = 0; 1358 dmae->len = bxe_get_port_stats_dma_len(sc); 1359 if (sc->func_stx) { 1360 dmae->comp_addr_lo = (dmae_reg_go_c[loader_idx] >> 2); 1361 dmae->comp_addr_hi = 0; 1362 dmae->comp_val = 1; 1363 } else { 1364 dmae->comp_addr_lo = U64_LO(BXE_SP_MAPPING(sc, stats_comp)); 1365 dmae->comp_addr_hi = U64_HI(BXE_SP_MAPPING(sc, stats_comp)); 1366 dmae->comp_val = DMAE_COMP_VAL; 1367 1368 *stats_comp = 0; 1369 } 1370 } 1371 1372 if (sc->func_stx) { 1373 dmae = BXE_SP(sc, dmae[sc->executer_idx++]); 1374 dmae->opcode = bxe_dmae_opcode_add_comp(opcode, DMAE_COMP_PCI); 1375 dmae->src_addr_lo = U64_LO(BXE_SP_MAPPING(sc, func_stats)); 1376 dmae->src_addr_hi = U64_HI(BXE_SP_MAPPING(sc, func_stats)); 1377 dmae->dst_addr_lo = (sc->func_stx >> 2); 1378 dmae->dst_addr_hi = 0; 1379 dmae->len = (sizeof(struct host_func_stats) >> 2); 1380 dmae->comp_addr_lo = U64_LO(BXE_SP_MAPPING(sc, stats_comp)); 1381 dmae->comp_addr_hi = U64_HI(BXE_SP_MAPPING(sc, stats_comp)); 1382 dmae->comp_val = DMAE_COMP_VAL; 1383 1384 *stats_comp = 0; 1385 } 1386 } 1387 1388 static void 1389 bxe_stats_stop(struct bxe_softc *sc) 1390 { 1391 uint8_t update = FALSE; 1392 1393 bxe_stats_comp(sc); 1394 1395 if (sc->port.pmf) { 1396 update = bxe_hw_stats_update(sc) == 0; 1397 } 1398 1399 update |= bxe_storm_stats_update(sc) == 0; 1400 1401 if (update) { 1402 bxe_net_stats_update(sc); 1403 1404 if (sc->port.pmf) { 1405 bxe_port_stats_stop(sc); 1406 } 1407 1408 bxe_hw_stats_post(sc); 1409 bxe_stats_comp(sc); 1410 } 1411 } 1412 1413 static void 1414 bxe_stats_do_nothing(struct bxe_softc *sc) 1415 { 1416 return; 1417 } 1418 1419 static const struct { 1420 void (*action)(struct bxe_softc *sc); 1421 enum bxe_stats_state next_state; 1422 } bxe_stats_stm[STATS_STATE_MAX][STATS_EVENT_MAX] = { 1423 { 1424 /* DISABLED PMF */ { bxe_stats_pmf_update, STATS_STATE_DISABLED }, 1425 /* LINK_UP */ { bxe_stats_start, STATS_STATE_ENABLED }, 1426 /* UPDATE */ { bxe_stats_do_nothing, STATS_STATE_DISABLED }, 1427 /* STOP */ { bxe_stats_do_nothing, STATS_STATE_DISABLED } 1428 }, 1429 { 1430 /* ENABLED PMF */ { bxe_stats_pmf_start, STATS_STATE_ENABLED }, 1431 /* LINK_UP */ { bxe_stats_restart, STATS_STATE_ENABLED }, 1432 /* UPDATE */ { bxe_stats_update, STATS_STATE_ENABLED }, 1433 /* STOP */ { bxe_stats_stop, STATS_STATE_DISABLED } 1434 } 1435 }; 1436 1437 void bxe_stats_handle(struct bxe_softc *sc, 1438 enum bxe_stats_event event) 1439 { 1440 enum bxe_stats_state state; 1441 1442 if (__predict_false(sc->panic)) { 1443 return; 1444 } 1445 1446 BXE_STATS_LOCK(sc); 1447 state = sc->stats_state; 1448 sc->stats_state = bxe_stats_stm[state][event].next_state; 1449 BXE_STATS_UNLOCK(sc); 1450 1451 bxe_stats_stm[state][event].action(sc); 1452 1453 if (event != STATS_EVENT_UPDATE) { 1454 BLOGD(sc, DBG_STATS, 1455 "state %d -> event %d -> state %d\n", 1456 state, event, sc->stats_state); 1457 } 1458 } 1459 1460 static void 1461 bxe_port_stats_base_init(struct bxe_softc *sc) 1462 { 1463 struct dmae_command *dmae; 1464 uint32_t *stats_comp = BXE_SP(sc, stats_comp); 1465 1466 /* sanity */ 1467 if (!sc->port.pmf || !sc->port.port_stx) { 1468 BLOGE(sc, "BUG!\n"); 1469 return; 1470 } 1471 1472 sc->executer_idx = 0; 1473 1474 dmae = BXE_SP(sc, dmae[sc->executer_idx++]); 1475 dmae->opcode = bxe_dmae_opcode(sc, DMAE_SRC_PCI, DMAE_DST_GRC, 1476 TRUE, DMAE_COMP_PCI); 1477 dmae->src_addr_lo = U64_LO(BXE_SP_MAPPING(sc, port_stats)); 1478 dmae->src_addr_hi = U64_HI(BXE_SP_MAPPING(sc, port_stats)); 1479 dmae->dst_addr_lo = (sc->port.port_stx >> 2); 1480 dmae->dst_addr_hi = 0; 1481 dmae->len = bxe_get_port_stats_dma_len(sc); 1482 dmae->comp_addr_lo = U64_LO(BXE_SP_MAPPING(sc, stats_comp)); 1483 dmae->comp_addr_hi = U64_HI(BXE_SP_MAPPING(sc, stats_comp)); 1484 dmae->comp_val = DMAE_COMP_VAL; 1485 1486 *stats_comp = 0; 1487 bxe_hw_stats_post(sc); 1488 bxe_stats_comp(sc); 1489 } 1490 1491 /* 1492 * This function will prepare the statistics ramrod data the way 1493 * we will only have to increment the statistics counter and 1494 * send the ramrod each time we have to. 1495 */ 1496 static void 1497 bxe_prep_fw_stats_req(struct bxe_softc *sc) 1498 { 1499 int i; 1500 int first_queue_query_index; 1501 struct stats_query_header *stats_hdr = &sc->fw_stats_req->hdr; 1502 bus_addr_t cur_data_offset; 1503 struct stats_query_entry *cur_query_entry; 1504 1505 stats_hdr->cmd_num = sc->fw_stats_num; 1506 stats_hdr->drv_stats_counter = 0; 1507 1508 /* 1509 * The storm_counters struct contains the counters of completed 1510 * statistics requests per storm which are incremented by FW 1511 * each time it completes hadning a statistics ramrod. We will 1512 * check these counters in the timer handler and discard a 1513 * (statistics) ramrod completion. 1514 */ 1515 cur_data_offset = (sc->fw_stats_data_mapping + 1516 offsetof(struct bxe_fw_stats_data, storm_counters)); 1517 1518 stats_hdr->stats_counters_addrs.hi = htole32(U64_HI(cur_data_offset)); 1519 stats_hdr->stats_counters_addrs.lo = htole32(U64_LO(cur_data_offset)); 1520 1521 /* 1522 * Prepare the first stats ramrod (will be completed with 1523 * the counters equal to zero) - init counters to somethig different. 1524 */ 1525 memset(&sc->fw_stats_data->storm_counters, 0xff, 1526 sizeof(struct stats_counter)); 1527 1528 /**** Port FW statistics data ****/ 1529 cur_data_offset = (sc->fw_stats_data_mapping + 1530 offsetof(struct bxe_fw_stats_data, port)); 1531 1532 cur_query_entry = &sc->fw_stats_req->query[BXE_PORT_QUERY_IDX]; 1533 1534 cur_query_entry->kind = STATS_TYPE_PORT; 1535 /* For port query index is a DONT CARE */ 1536 cur_query_entry->index = SC_PORT(sc); 1537 /* For port query funcID is a DONT CARE */ 1538 cur_query_entry->funcID = htole16(SC_FUNC(sc)); 1539 cur_query_entry->address.hi = htole32(U64_HI(cur_data_offset)); 1540 cur_query_entry->address.lo = htole32(U64_LO(cur_data_offset)); 1541 1542 /**** PF FW statistics data ****/ 1543 cur_data_offset = (sc->fw_stats_data_mapping + 1544 offsetof(struct bxe_fw_stats_data, pf)); 1545 1546 cur_query_entry = &sc->fw_stats_req->query[BXE_PF_QUERY_IDX]; 1547 1548 cur_query_entry->kind = STATS_TYPE_PF; 1549 /* For PF query index is a DONT CARE */ 1550 cur_query_entry->index = SC_PORT(sc); 1551 cur_query_entry->funcID = htole16(SC_FUNC(sc)); 1552 cur_query_entry->address.hi = htole32(U64_HI(cur_data_offset)); 1553 cur_query_entry->address.lo = htole32(U64_LO(cur_data_offset)); 1554 1555 #if 0 1556 /**** FCoE FW statistics data ****/ 1557 if (!NO_FCOE(sc)) { 1558 cur_data_offset = (sc->fw_stats_data_mapping + 1559 offsetof(struct bxe_fw_stats_data, fcoe)); 1560 1561 cur_query_entry = &sc->fw_stats_req->query[BXE_FCOE_QUERY_IDX]; 1562 1563 cur_query_entry->kind = STATS_TYPE_FCOE; 1564 /* For FCoE query index is a DONT CARE */ 1565 cur_query_entry->index = SC_PORT(sc); 1566 cur_query_entry->funcID = cpu_to_le16(SC_FUNC(sc)); 1567 cur_query_entry->address.hi = htole32(U64_HI(cur_data_offset)); 1568 cur_query_entry->address.lo = htole32(U64_LO(cur_data_offset)); 1569 } 1570 #endif 1571 1572 /**** Clients' queries ****/ 1573 cur_data_offset = (sc->fw_stats_data_mapping + 1574 offsetof(struct bxe_fw_stats_data, queue_stats)); 1575 1576 /* 1577 * First queue query index depends whether FCoE offloaded request will 1578 * be included in the ramrod 1579 */ 1580 #if 0 1581 if (!NO_FCOE(sc)) 1582 first_queue_query_index = BXE_FIRST_QUEUE_QUERY_IDX; 1583 else 1584 #endif 1585 first_queue_query_index = (BXE_FIRST_QUEUE_QUERY_IDX - 1); 1586 1587 for (i = 0; i < sc->num_queues; i++) { 1588 cur_query_entry = 1589 &sc->fw_stats_req->query[first_queue_query_index + i]; 1590 1591 cur_query_entry->kind = STATS_TYPE_QUEUE; 1592 cur_query_entry->index = bxe_stats_id(&sc->fp[i]); 1593 cur_query_entry->funcID = htole16(SC_FUNC(sc)); 1594 cur_query_entry->address.hi = htole32(U64_HI(cur_data_offset)); 1595 cur_query_entry->address.lo = htole32(U64_LO(cur_data_offset)); 1596 1597 cur_data_offset += sizeof(struct per_queue_stats); 1598 } 1599 1600 #if 0 1601 /* add FCoE queue query if needed */ 1602 if (!NO_FCOE(sc)) { 1603 cur_query_entry = 1604 &sc->fw_stats_req->query[first_queue_query_index + i]; 1605 1606 cur_query_entry->kind = STATS_TYPE_QUEUE; 1607 cur_query_entry->index = bxe_stats_id(&sc->fp[FCOE_IDX(sc)]); 1608 cur_query_entry->funcID = htole16(SC_FUNC(sc)); 1609 cur_query_entry->address.hi = htole32(U64_HI(cur_data_offset)); 1610 cur_query_entry->address.lo = htole32(U64_LO(cur_data_offset)); 1611 } 1612 #endif 1613 } 1614 1615 void 1616 bxe_stats_init(struct bxe_softc *sc) 1617 { 1618 int /*abs*/port = SC_PORT(sc); 1619 int mb_idx = SC_FW_MB_IDX(sc); 1620 int i; 1621 1622 sc->stats_pending = 0; 1623 sc->executer_idx = 0; 1624 sc->stats_counter = 0; 1625 1626 /* port and func stats for management */ 1627 if (!BXE_NOMCP(sc)) { 1628 sc->port.port_stx = SHMEM_RD(sc, port_mb[port].port_stx); 1629 sc->func_stx = SHMEM_RD(sc, func_mb[mb_idx].fw_mb_param); 1630 } else { 1631 sc->port.port_stx = 0; 1632 sc->func_stx = 0; 1633 } 1634 1635 BLOGD(sc, DBG_STATS, "port_stx 0x%x func_stx 0x%x\n", 1636 sc->port.port_stx, sc->func_stx); 1637 1638 /* pmf should retrieve port statistics from SP on a non-init*/ 1639 if (!sc->stats_init && sc->port.pmf && sc->port.port_stx) { 1640 bxe_stats_handle(sc, STATS_EVENT_PMF); 1641 } 1642 1643 port = SC_PORT(sc); 1644 /* port stats */ 1645 memset(&(sc->port.old_nig_stats), 0, sizeof(struct nig_stats)); 1646 sc->port.old_nig_stats.brb_discard = 1647 REG_RD(sc, NIG_REG_STAT0_BRB_DISCARD + port*0x38); 1648 sc->port.old_nig_stats.brb_truncate = 1649 REG_RD(sc, NIG_REG_STAT0_BRB_TRUNCATE + port*0x38); 1650 if (!CHIP_IS_E3(sc)) { 1651 REG_RD_DMAE(sc, NIG_REG_STAT0_EGRESS_MAC_PKT0 + port*0x50, 1652 &(sc->port.old_nig_stats.egress_mac_pkt0_lo), 2); 1653 REG_RD_DMAE(sc, NIG_REG_STAT0_EGRESS_MAC_PKT1 + port*0x50, 1654 &(sc->port.old_nig_stats.egress_mac_pkt1_lo), 2); 1655 } 1656 1657 /* function stats */ 1658 for (i = 0; i < sc->num_queues; i++) { 1659 memset(&sc->fp[i].old_tclient, 0, sizeof(sc->fp[i].old_tclient)); 1660 memset(&sc->fp[i].old_uclient, 0, sizeof(sc->fp[i].old_uclient)); 1661 memset(&sc->fp[i].old_xclient, 0, sizeof(sc->fp[i].old_xclient)); 1662 if (sc->stats_init) { 1663 memset(&sc->fp[i].eth_q_stats, 0, 1664 sizeof(sc->fp[i].eth_q_stats)); 1665 memset(&sc->fp[i].eth_q_stats_old, 0, 1666 sizeof(sc->fp[i].eth_q_stats_old)); 1667 } 1668 } 1669 1670 /* prepare statistics ramrod data */ 1671 bxe_prep_fw_stats_req(sc); 1672 1673 sc->ifnet->if_data.ifi_ipackets = 0; 1674 sc->ifnet->if_data.ifi_opackets = 0; 1675 sc->ifnet->if_data.ifi_ibytes = 0; 1676 sc->ifnet->if_data.ifi_obytes = 0; 1677 sc->ifnet->if_data.ifi_ierrors = 0; 1678 sc->ifnet->if_data.ifi_oerrors = 0; 1679 sc->ifnet->if_data.ifi_imcasts = 0; 1680 sc->ifnet->if_data.ifi_collisions = 0; 1681 1682 if (sc->stats_init) { 1683 memset(&sc->net_stats_old, 0, sizeof(sc->net_stats_old)); 1684 memset(&sc->fw_stats_old, 0, sizeof(sc->fw_stats_old)); 1685 memset(&sc->eth_stats_old, 0, sizeof(sc->eth_stats_old)); 1686 memset(&sc->eth_stats, 0, sizeof(sc->eth_stats)); 1687 memset(&sc->func_stats, 0, sizeof(sc->func_stats)); 1688 1689 /* Clean SP from previous statistics */ 1690 if (sc->func_stx) { 1691 memset(BXE_SP(sc, func_stats), 0, sizeof(struct host_func_stats)); 1692 bxe_func_stats_init(sc); 1693 bxe_hw_stats_post(sc); 1694 bxe_stats_comp(sc); 1695 } 1696 } 1697 1698 sc->stats_state = STATS_STATE_DISABLED; 1699 1700 if (sc->port.pmf && sc->port.port_stx) { 1701 bxe_port_stats_base_init(sc); 1702 } 1703 1704 /* mark the end of statistics initializiation */ 1705 sc->stats_init = FALSE; 1706 } 1707 1708 void 1709 bxe_save_statistics(struct bxe_softc *sc) 1710 { 1711 int i; 1712 1713 /* save queue statistics */ 1714 for (i = 0; i < sc->num_queues; i++) { 1715 struct bxe_fastpath *fp = &sc->fp[i]; 1716 struct bxe_eth_q_stats *qstats = &fp->eth_q_stats; 1717 struct bxe_eth_q_stats_old *qstats_old = &fp->eth_q_stats_old; 1718 1719 UPDATE_QSTAT_OLD(total_unicast_bytes_received_hi); 1720 UPDATE_QSTAT_OLD(total_unicast_bytes_received_lo); 1721 UPDATE_QSTAT_OLD(total_broadcast_bytes_received_hi); 1722 UPDATE_QSTAT_OLD(total_broadcast_bytes_received_lo); 1723 UPDATE_QSTAT_OLD(total_multicast_bytes_received_hi); 1724 UPDATE_QSTAT_OLD(total_multicast_bytes_received_lo); 1725 UPDATE_QSTAT_OLD(total_unicast_bytes_transmitted_hi); 1726 UPDATE_QSTAT_OLD(total_unicast_bytes_transmitted_lo); 1727 UPDATE_QSTAT_OLD(total_broadcast_bytes_transmitted_hi); 1728 UPDATE_QSTAT_OLD(total_broadcast_bytes_transmitted_lo); 1729 UPDATE_QSTAT_OLD(total_multicast_bytes_transmitted_hi); 1730 UPDATE_QSTAT_OLD(total_multicast_bytes_transmitted_lo); 1731 UPDATE_QSTAT_OLD(total_tpa_bytes_hi); 1732 UPDATE_QSTAT_OLD(total_tpa_bytes_lo); 1733 } 1734 1735 /* save net_device_stats statistics */ 1736 sc->net_stats_old.rx_dropped = sc->ifnet->if_data.ifi_iqdrops; 1737 1738 /* store port firmware statistics */ 1739 if (sc->port.pmf) { 1740 struct bxe_eth_stats *estats = &sc->eth_stats; 1741 struct bxe_fw_port_stats_old *fwstats = &sc->fw_stats_old; 1742 struct host_port_stats *pstats = BXE_SP(sc, port_stats); 1743 1744 fwstats->pfc_frames_rx_hi = pstats->pfc_frames_rx_hi; 1745 fwstats->pfc_frames_rx_lo = pstats->pfc_frames_rx_lo; 1746 fwstats->pfc_frames_tx_hi = pstats->pfc_frames_tx_hi; 1747 fwstats->pfc_frames_tx_lo = pstats->pfc_frames_tx_lo; 1748 1749 if (IS_MF(sc)) { 1750 UPDATE_FW_STAT_OLD(mac_filter_discard); 1751 UPDATE_FW_STAT_OLD(mf_tag_discard); 1752 UPDATE_FW_STAT_OLD(brb_truncate_discard); 1753 UPDATE_FW_STAT_OLD(mac_discard); 1754 } 1755 } 1756 } 1757 1758 void 1759 bxe_afex_collect_stats(struct bxe_softc *sc, 1760 void *void_afex_stats, 1761 uint32_t stats_type) 1762 { 1763 int i; 1764 struct afex_stats *afex_stats = (struct afex_stats *)void_afex_stats; 1765 struct bxe_eth_stats *estats = &sc->eth_stats; 1766 #if 0 1767 struct per_queue_stats *fcoe_q_stats = 1768 &sc->fw_stats_data->queue_stats[FCOE_IDX(sc)]; 1769 1770 struct tstorm_per_queue_stats *fcoe_q_tstorm_stats = 1771 &fcoe_q_stats->tstorm_queue_statistics; 1772 1773 struct ustorm_per_queue_stats *fcoe_q_ustorm_stats = 1774 &fcoe_q_stats->ustorm_queue_statistics; 1775 1776 struct xstorm_per_queue_stats *fcoe_q_xstorm_stats = 1777 &fcoe_q_stats->xstorm_queue_statistics; 1778 1779 struct fcoe_statistics_params *fw_fcoe_stat = 1780 &sc->fw_stats_data->fcoe; 1781 #endif 1782 1783 memset(afex_stats, 0, sizeof(struct afex_stats)); 1784 1785 for (i = 0; i < sc->num_queues; i++) { 1786 struct bxe_eth_q_stats *qstats = &sc->fp[i].eth_q_stats; 1787 1788 ADD_64(afex_stats->rx_unicast_bytes_hi, 1789 qstats->total_unicast_bytes_received_hi, 1790 afex_stats->rx_unicast_bytes_lo, 1791 qstats->total_unicast_bytes_received_lo); 1792 1793 ADD_64(afex_stats->rx_broadcast_bytes_hi, 1794 qstats->total_broadcast_bytes_received_hi, 1795 afex_stats->rx_broadcast_bytes_lo, 1796 qstats->total_broadcast_bytes_received_lo); 1797 1798 ADD_64(afex_stats->rx_multicast_bytes_hi, 1799 qstats->total_multicast_bytes_received_hi, 1800 afex_stats->rx_multicast_bytes_lo, 1801 qstats->total_multicast_bytes_received_lo); 1802 1803 ADD_64(afex_stats->rx_unicast_frames_hi, 1804 qstats->total_unicast_packets_received_hi, 1805 afex_stats->rx_unicast_frames_lo, 1806 qstats->total_unicast_packets_received_lo); 1807 1808 ADD_64(afex_stats->rx_broadcast_frames_hi, 1809 qstats->total_broadcast_packets_received_hi, 1810 afex_stats->rx_broadcast_frames_lo, 1811 qstats->total_broadcast_packets_received_lo); 1812 1813 ADD_64(afex_stats->rx_multicast_frames_hi, 1814 qstats->total_multicast_packets_received_hi, 1815 afex_stats->rx_multicast_frames_lo, 1816 qstats->total_multicast_packets_received_lo); 1817 1818 /* 1819 * sum to rx_frames_discarded all discarded 1820 * packets due to size, ttl0 and checksum 1821 */ 1822 ADD_64(afex_stats->rx_frames_discarded_hi, 1823 qstats->total_packets_received_checksum_discarded_hi, 1824 afex_stats->rx_frames_discarded_lo, 1825 qstats->total_packets_received_checksum_discarded_lo); 1826 1827 ADD_64(afex_stats->rx_frames_discarded_hi, 1828 qstats->total_packets_received_ttl0_discarded_hi, 1829 afex_stats->rx_frames_discarded_lo, 1830 qstats->total_packets_received_ttl0_discarded_lo); 1831 1832 ADD_64(afex_stats->rx_frames_discarded_hi, 1833 qstats->etherstatsoverrsizepkts_hi, 1834 afex_stats->rx_frames_discarded_lo, 1835 qstats->etherstatsoverrsizepkts_lo); 1836 1837 ADD_64(afex_stats->rx_frames_dropped_hi, 1838 qstats->no_buff_discard_hi, 1839 afex_stats->rx_frames_dropped_lo, 1840 qstats->no_buff_discard_lo); 1841 1842 ADD_64(afex_stats->tx_unicast_bytes_hi, 1843 qstats->total_unicast_bytes_transmitted_hi, 1844 afex_stats->tx_unicast_bytes_lo, 1845 qstats->total_unicast_bytes_transmitted_lo); 1846 1847 ADD_64(afex_stats->tx_broadcast_bytes_hi, 1848 qstats->total_broadcast_bytes_transmitted_hi, 1849 afex_stats->tx_broadcast_bytes_lo, 1850 qstats->total_broadcast_bytes_transmitted_lo); 1851 1852 ADD_64(afex_stats->tx_multicast_bytes_hi, 1853 qstats->total_multicast_bytes_transmitted_hi, 1854 afex_stats->tx_multicast_bytes_lo, 1855 qstats->total_multicast_bytes_transmitted_lo); 1856 1857 ADD_64(afex_stats->tx_unicast_frames_hi, 1858 qstats->total_unicast_packets_transmitted_hi, 1859 afex_stats->tx_unicast_frames_lo, 1860 qstats->total_unicast_packets_transmitted_lo); 1861 1862 ADD_64(afex_stats->tx_broadcast_frames_hi, 1863 qstats->total_broadcast_packets_transmitted_hi, 1864 afex_stats->tx_broadcast_frames_lo, 1865 qstats->total_broadcast_packets_transmitted_lo); 1866 1867 ADD_64(afex_stats->tx_multicast_frames_hi, 1868 qstats->total_multicast_packets_transmitted_hi, 1869 afex_stats->tx_multicast_frames_lo, 1870 qstats->total_multicast_packets_transmitted_lo); 1871 1872 ADD_64(afex_stats->tx_frames_dropped_hi, 1873 qstats->total_transmitted_dropped_packets_error_hi, 1874 afex_stats->tx_frames_dropped_lo, 1875 qstats->total_transmitted_dropped_packets_error_lo); 1876 } 1877 1878 #if 0 1879 /* 1880 * Now add FCoE statistics which are collected separately 1881 * (both offloaded and non offloaded) 1882 */ 1883 if (!NO_FCOE(sc)) { 1884 ADD_64_LE(afex_stats->rx_unicast_bytes_hi, 1885 LE32_0, 1886 afex_stats->rx_unicast_bytes_lo, 1887 fw_fcoe_stat->rx_stat0.fcoe_rx_byte_cnt); 1888 1889 ADD_64_LE(afex_stats->rx_unicast_bytes_hi, 1890 fcoe_q_tstorm_stats->rcv_ucast_bytes.hi, 1891 afex_stats->rx_unicast_bytes_lo, 1892 fcoe_q_tstorm_stats->rcv_ucast_bytes.lo); 1893 1894 ADD_64_LE(afex_stats->rx_broadcast_bytes_hi, 1895 fcoe_q_tstorm_stats->rcv_bcast_bytes.hi, 1896 afex_stats->rx_broadcast_bytes_lo, 1897 fcoe_q_tstorm_stats->rcv_bcast_bytes.lo); 1898 1899 ADD_64_LE(afex_stats->rx_multicast_bytes_hi, 1900 fcoe_q_tstorm_stats->rcv_mcast_bytes.hi, 1901 afex_stats->rx_multicast_bytes_lo, 1902 fcoe_q_tstorm_stats->rcv_mcast_bytes.lo); 1903 1904 ADD_64_LE(afex_stats->rx_unicast_frames_hi, 1905 LE32_0, 1906 afex_stats->rx_unicast_frames_lo, 1907 fw_fcoe_stat->rx_stat0.fcoe_rx_pkt_cnt); 1908 1909 ADD_64_LE(afex_stats->rx_unicast_frames_hi, 1910 LE32_0, 1911 afex_stats->rx_unicast_frames_lo, 1912 fcoe_q_tstorm_stats->rcv_ucast_pkts); 1913 1914 ADD_64_LE(afex_stats->rx_broadcast_frames_hi, 1915 LE32_0, 1916 afex_stats->rx_broadcast_frames_lo, 1917 fcoe_q_tstorm_stats->rcv_bcast_pkts); 1918 1919 ADD_64_LE(afex_stats->rx_multicast_frames_hi, 1920 LE32_0, 1921 afex_stats->rx_multicast_frames_lo, 1922 fcoe_q_tstorm_stats->rcv_ucast_pkts); 1923 1924 ADD_64_LE(afex_stats->rx_frames_discarded_hi, 1925 LE32_0, 1926 afex_stats->rx_frames_discarded_lo, 1927 fcoe_q_tstorm_stats->checksum_discard); 1928 1929 ADD_64_LE(afex_stats->rx_frames_discarded_hi, 1930 LE32_0, 1931 afex_stats->rx_frames_discarded_lo, 1932 fcoe_q_tstorm_stats->pkts_too_big_discard); 1933 1934 ADD_64_LE(afex_stats->rx_frames_discarded_hi, 1935 LE32_0, 1936 afex_stats->rx_frames_discarded_lo, 1937 fcoe_q_tstorm_stats->ttl0_discard); 1938 1939 ADD_64_LE16(afex_stats->rx_frames_dropped_hi, 1940 LE16_0, 1941 afex_stats->rx_frames_dropped_lo, 1942 fcoe_q_tstorm_stats->no_buff_discard); 1943 1944 ADD_64_LE(afex_stats->rx_frames_dropped_hi, 1945 LE32_0, 1946 afex_stats->rx_frames_dropped_lo, 1947 fcoe_q_ustorm_stats->ucast_no_buff_pkts); 1948 1949 ADD_64_LE(afex_stats->rx_frames_dropped_hi, 1950 LE32_0, 1951 afex_stats->rx_frames_dropped_lo, 1952 fcoe_q_ustorm_stats->mcast_no_buff_pkts); 1953 1954 ADD_64_LE(afex_stats->rx_frames_dropped_hi, 1955 LE32_0, 1956 afex_stats->rx_frames_dropped_lo, 1957 fcoe_q_ustorm_stats->bcast_no_buff_pkts); 1958 1959 ADD_64_LE(afex_stats->rx_frames_dropped_hi, 1960 LE32_0, 1961 afex_stats->rx_frames_dropped_lo, 1962 fw_fcoe_stat->rx_stat1.fcoe_rx_drop_pkt_cnt); 1963 1964 ADD_64_LE(afex_stats->rx_frames_dropped_hi, 1965 LE32_0, 1966 afex_stats->rx_frames_dropped_lo, 1967 fw_fcoe_stat->rx_stat2.fcoe_rx_drop_pkt_cnt); 1968 1969 ADD_64_LE(afex_stats->tx_unicast_bytes_hi, 1970 LE32_0, 1971 afex_stats->tx_unicast_bytes_lo, 1972 fw_fcoe_stat->tx_stat.fcoe_tx_byte_cnt); 1973 1974 ADD_64_LE(afex_stats->tx_unicast_bytes_hi, 1975 fcoe_q_xstorm_stats->ucast_bytes_sent.hi, 1976 afex_stats->tx_unicast_bytes_lo, 1977 fcoe_q_xstorm_stats->ucast_bytes_sent.lo); 1978 1979 ADD_64_LE(afex_stats->tx_broadcast_bytes_hi, 1980 fcoe_q_xstorm_stats->bcast_bytes_sent.hi, 1981 afex_stats->tx_broadcast_bytes_lo, 1982 fcoe_q_xstorm_stats->bcast_bytes_sent.lo); 1983 1984 ADD_64_LE(afex_stats->tx_multicast_bytes_hi, 1985 fcoe_q_xstorm_stats->mcast_bytes_sent.hi, 1986 afex_stats->tx_multicast_bytes_lo, 1987 fcoe_q_xstorm_stats->mcast_bytes_sent.lo); 1988 1989 ADD_64_LE(afex_stats->tx_unicast_frames_hi, 1990 LE32_0, 1991 afex_stats->tx_unicast_frames_lo, 1992 fw_fcoe_stat->tx_stat.fcoe_tx_pkt_cnt); 1993 1994 ADD_64_LE(afex_stats->tx_unicast_frames_hi, 1995 LE32_0, 1996 afex_stats->tx_unicast_frames_lo, 1997 fcoe_q_xstorm_stats->ucast_pkts_sent); 1998 1999 ADD_64_LE(afex_stats->tx_broadcast_frames_hi, 2000 LE32_0, 2001 afex_stats->tx_broadcast_frames_lo, 2002 fcoe_q_xstorm_stats->bcast_pkts_sent); 2003 2004 ADD_64_LE(afex_stats->tx_multicast_frames_hi, 2005 LE32_0, 2006 afex_stats->tx_multicast_frames_lo, 2007 fcoe_q_xstorm_stats->mcast_pkts_sent); 2008 2009 ADD_64_LE(afex_stats->tx_frames_dropped_hi, 2010 LE32_0, 2011 afex_stats->tx_frames_dropped_lo, 2012 fcoe_q_xstorm_stats->error_drop_pkts); 2013 } 2014 #endif 2015 2016 /* 2017 * If port stats are requested, add them to the PMF 2018 * stats, as anyway they will be accumulated by the 2019 * MCP before sent to the switch 2020 */ 2021 if ((sc->port.pmf) && (stats_type == VICSTATST_UIF_INDEX)) { 2022 ADD_64(afex_stats->rx_frames_dropped_hi, 2023 0, 2024 afex_stats->rx_frames_dropped_lo, 2025 estats->mac_filter_discard); 2026 ADD_64(afex_stats->rx_frames_dropped_hi, 2027 0, 2028 afex_stats->rx_frames_dropped_lo, 2029 estats->brb_truncate_discard); 2030 ADD_64(afex_stats->rx_frames_discarded_hi, 2031 0, 2032 afex_stats->rx_frames_discarded_lo, 2033 estats->mac_discard); 2034 } 2035 } 2036 2037