1 /*- 2 * Copyright (c) 2017 Chelsio Communications, Inc. 3 * All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright 9 * notice, this list of conditions and the following disclaimer. 10 * 2. Redistributions in binary form must reproduce the above copyright 11 * notice, this list of conditions and the following disclaimer in the 12 * documentation and/or other materials provided with the distribution. 13 * 14 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 15 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 17 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 24 * SUCH DAMAGE. 25 */ 26 27 #include <sys/cdefs.h> 28 __FBSDID("$FreeBSD$"); 29 30 #include <sys/types.h> 31 #include <sys/param.h> 32 33 #include "common/common.h" 34 #include "common/t4_regs.h" 35 #include "cudbg.h" 36 #include "cudbg_lib_common.h" 37 #include "cudbg_entity.h" 38 39 int collect_wtp_data(struct cudbg_init *pdbg_init, 40 struct cudbg_buffer *dbg_buff, 41 struct cudbg_error *cudbg_err); 42 /*SGE_DEBUG Registers.*/ 43 #define TP_MIB_SIZE 0x5e 44 45 struct sge_debug_reg_data { 46 /*indx0*/ 47 u32 reserved1:4; 48 u32 reserved2:4; 49 u32 debug_uP_SOP_cnt:4; 50 u32 debug_uP_EOP_cnt:4; 51 u32 debug_CIM_SOP1_cnt:4; 52 u32 debug_CIM_EOP1_cnt:4; 53 u32 debug_CIM_SOP0_cnt:4; 54 u32 debug_CIM_EOP0_cnt:4; 55 56 /*indx1*/ 57 u32 reserved3:32; 58 59 /*indx2*/ 60 u32 debug_T_Rx_SOP1_cnt:4; 61 u32 debug_T_Rx_EOP1_cnt:4; 62 u32 debug_T_Rx_SOP0_cnt:4; 63 u32 debug_T_Rx_EOP0_cnt:4; 64 u32 debug_U_Rx_SOP1_cnt:4; 65 u32 debug_U_Rx_EOP1_cnt:4; 66 u32 debug_U_Rx_SOP0_cnt:4; 67 u32 debug_U_Rx_EOP0_cnt:4; 68 69 /*indx3*/ 70 u32 reserved4:32; 71 72 /*indx4*/ 73 u32 debug_UD_Rx_SOP3_cnt:4; 74 u32 debug_UD_Rx_EOP3_cnt:4; 75 u32 debug_UD_Rx_SOP2_cnt:4; 76 u32 debug_UD_Rx_EOP2_cnt:4; 77 u32 debug_UD_Rx_SOP1_cnt:4; 78 u32 debug_UD_Rx_EOP1_cnt:4; 79 u32 debug_UD_Rx_SOP0_cnt:4; 80 u32 debug_UD_Rx_EOP0_cnt:4; 81 82 /*indx5*/ 83 u32 reserved5:32; 84 85 /*indx6*/ 86 u32 debug_U_Tx_SOP3_cnt:4; 87 u32 debug_U_Tx_EOP3_cnt:4; 88 u32 debug_U_Tx_SOP2_cnt:4; 89 u32 debug_U_Tx_EOP2_cnt:4; 90 u32 debug_U_Tx_SOP1_cnt:4; 91 u32 debug_U_Tx_EOP1_cnt:4; 92 u32 debug_U_Tx_SOP0_cnt:4; 93 u32 debug_U_Tx_EOP0_cnt:4; 94 95 /*indx7*/ 96 u32 reserved6:32; 97 98 /*indx8*/ 99 u32 debug_PC_Rsp_SOP1_cnt:4; 100 u32 debug_PC_Rsp_EOP1_cnt:4; 101 u32 debug_PC_Rsp_SOP0_cnt:4; 102 u32 debug_PC_Rsp_EOP0_cnt:4; 103 u32 debug_PC_Req_SOP1_cnt:4; 104 u32 debug_PC_Req_EOP1_cnt:4; 105 u32 debug_PC_Req_SOP0_cnt:4; 106 u32 debug_PC_Req_EOP0_cnt:4; 107 108 /*indx9*/ 109 u32 reserved7:32; 110 111 /*indx10*/ 112 u32 debug_PD_Req_SOP3_cnt:4; 113 u32 debug_PD_Req_EOP3_cnt:4; 114 u32 debug_PD_Req_SOP2_cnt:4; 115 u32 debug_PD_Req_EOP2_cnt:4; 116 u32 debug_PD_Req_SOP1_cnt:4; 117 u32 debug_PD_Req_EOP1_cnt:4; 118 u32 debug_PD_Req_SOP0_cnt:4; 119 u32 debug_PD_Req_EOP0_cnt:4; 120 121 /*indx11*/ 122 u32 reserved8:32; 123 124 /*indx12*/ 125 u32 debug_PD_Rsp_SOP3_cnt:4; 126 u32 debug_PD_Rsp_EOP3_cnt:4; 127 u32 debug_PD_Rsp_SOP2_cnt:4; 128 u32 debug_PD_Rsp_EOP2_cnt:4; 129 u32 debug_PD_Rsp_SOP1_cnt:4; 130 u32 debug_PD_Rsp_EOP1_cnt:4; 131 u32 debug_PD_Rsp_SOP0_cnt:4; 132 u32 debug_PD_Rsp_EOP0_cnt:4; 133 134 /*indx13*/ 135 u32 reserved9:32; 136 137 /*indx14*/ 138 u32 debug_CPLSW_TP_Rx_SOP1_cnt:4; 139 u32 debug_CPLSW_TP_Rx_EOP1_cnt:4; 140 u32 debug_CPLSW_TP_Rx_SOP0_cnt:4; 141 u32 debug_CPLSW_TP_Rx_EOP0_cnt:4; 142 u32 debug_CPLSW_CIM_SOP1_cnt:4; 143 u32 debug_CPLSW_CIM_EOP1_cnt:4; 144 u32 debug_CPLSW_CIM_SOP0_cnt:4; 145 u32 debug_CPLSW_CIM_EOP0_cnt:4; 146 147 /*indx15*/ 148 u32 reserved10:32; 149 150 /*indx16*/ 151 u32 debug_PD_Req_Rd3_cnt:4; 152 u32 debug_PD_Req_Rd2_cnt:4; 153 u32 debug_PD_Req_Rd1_cnt:4; 154 u32 debug_PD_Req_Rd0_cnt:4; 155 u32 debug_PD_Req_Int3_cnt:4; 156 u32 debug_PD_Req_Int2_cnt:4; 157 u32 debug_PD_Req_Int1_cnt:4; 158 u32 debug_PD_Req_Int0_cnt:4; 159 160 }; 161 162 struct tp_mib_type tp_mib[] = { 163 {"tp_mib_mac_in_err_0", 0x0}, 164 {"tp_mib_mac_in_err_1", 0x1}, 165 {"tp_mib_mac_in_err_2", 0x2}, 166 {"tp_mib_mac_in_err_3", 0x3}, 167 {"tp_mib_hdr_in_err_0", 0x4}, 168 {"tp_mib_hdr_in_err_1", 0x5}, 169 {"tp_mib_hdr_in_err_2", 0x6}, 170 {"tp_mib_hdr_in_err_3", 0x7}, 171 {"tp_mib_tcp_in_err_0", 0x8}, 172 {"tp_mib_tcp_in_err_1", 0x9}, 173 {"tp_mib_tcp_in_err_2", 0xa}, 174 {"tp_mib_tcp_in_err_3", 0xb}, 175 {"tp_mib_tcp_out_rst", 0xc}, 176 {"tp_mib_tcp_in_seg_hi", 0x10}, 177 {"tp_mib_tcp_in_seg_lo", 0x11}, 178 {"tp_mib_tcp_out_seg_hi", 0x12}, 179 {"tp_mib_tcp_out_seg_lo", 0x13}, 180 {"tp_mib_tcp_rxt_seg_hi", 0x14}, 181 {"tp_mib_tcp_rxt_seg_lo", 0x15}, 182 {"tp_mib_tnl_cng_drop_0", 0x18}, 183 {"tp_mib_tnl_cng_drop_1", 0x19}, 184 {"tp_mib_tnl_cng_drop_2", 0x1a}, 185 {"tp_mib_tnl_cng_drop_3", 0x1b}, 186 {"tp_mib_ofd_chn_drop_0", 0x1c}, 187 {"tp_mib_ofd_chn_drop_1", 0x1d}, 188 {"tp_mib_ofd_chn_drop_2", 0x1e}, 189 {"tp_mib_ofd_chn_drop_3", 0x1f}, 190 {"tp_mib_tnl_out_pkt_0", 0x20}, 191 {"tp_mib_tnl_out_pkt_1", 0x21}, 192 {"tp_mib_tnl_out_pkt_2", 0x22}, 193 {"tp_mib_tnl_out_pkt_3", 0x23}, 194 {"tp_mib_tnl_in_pkt_0", 0x24}, 195 {"tp_mib_tnl_in_pkt_1", 0x25}, 196 {"tp_mib_tnl_in_pkt_2", 0x26}, 197 {"tp_mib_tnl_in_pkt_3", 0x27}, 198 {"tp_mib_tcp_v6in_err_0", 0x28}, 199 {"tp_mib_tcp_v6in_err_1", 0x29}, 200 {"tp_mib_tcp_v6in_err_2", 0x2a}, 201 {"tp_mib_tcp_v6in_err_3", 0x2b}, 202 {"tp_mib_tcp_v6out_rst", 0x2c}, 203 {"tp_mib_tcp_v6in_seg_hi", 0x30}, 204 {"tp_mib_tcp_v6in_seg_lo", 0x31}, 205 {"tp_mib_tcp_v6out_seg_hi", 0x32}, 206 {"tp_mib_tcp_v6out_seg_lo", 0x33}, 207 {"tp_mib_tcp_v6rxt_seg_hi", 0x34}, 208 {"tp_mib_tcp_v6rxt_seg_lo", 0x35}, 209 {"tp_mib_ofd_arp_drop", 0x36}, 210 {"tp_mib_ofd_dfr_drop", 0x37}, 211 {"tp_mib_cpl_in_req_0", 0x38}, 212 {"tp_mib_cpl_in_req_1", 0x39}, 213 {"tp_mib_cpl_in_req_2", 0x3a}, 214 {"tp_mib_cpl_in_req_3", 0x3b}, 215 {"tp_mib_cpl_out_rsp_0", 0x3c}, 216 {"tp_mib_cpl_out_rsp_1", 0x3d}, 217 {"tp_mib_cpl_out_rsp_2", 0x3e}, 218 {"tp_mib_cpl_out_rsp_3", 0x3f}, 219 {"tp_mib_tnl_lpbk_0", 0x40}, 220 {"tp_mib_tnl_lpbk_1", 0x41}, 221 {"tp_mib_tnl_lpbk_2", 0x42}, 222 {"tp_mib_tnl_lpbk_3", 0x43}, 223 {"tp_mib_tnl_drop_0", 0x44}, 224 {"tp_mib_tnl_drop_1", 0x45}, 225 {"tp_mib_tnl_drop_2", 0x46}, 226 {"tp_mib_tnl_drop_3", 0x47}, 227 {"tp_mib_fcoe_ddp_0", 0x48}, 228 {"tp_mib_fcoe_ddp_1", 0x49}, 229 {"tp_mib_fcoe_ddp_2", 0x4a}, 230 {"tp_mib_fcoe_ddp_3", 0x4b}, 231 {"tp_mib_fcoe_drop_0", 0x4c}, 232 {"tp_mib_fcoe_drop_1", 0x4d}, 233 {"tp_mib_fcoe_drop_2", 0x4e}, 234 {"tp_mib_fcoe_drop_3", 0x4f}, 235 {"tp_mib_fcoe_byte_0_hi", 0x50}, 236 {"tp_mib_fcoe_byte_0_lo", 0x51}, 237 {"tp_mib_fcoe_byte_1_hi", 0x52}, 238 {"tp_mib_fcoe_byte_1_lo", 0x53}, 239 {"tp_mib_fcoe_byte_2_hi", 0x54}, 240 {"tp_mib_fcoe_byte_2_lo", 0x55}, 241 {"tp_mib_fcoe_byte_3_hi", 0x56}, 242 {"tp_mib_fcoe_byte_3_lo", 0x57}, 243 {"tp_mib_ofd_vln_drop_0", 0x58}, 244 {"tp_mib_ofd_vln_drop_1", 0x59}, 245 {"tp_mib_ofd_vln_drop_2", 0x5a}, 246 {"tp_mib_ofd_vln_drop_3", 0x5b}, 247 {"tp_mib_usm_pkts", 0x5c}, 248 {"tp_mib_usm_drop", 0x5d}, 249 {"tp_mib_usm_bytes_hi", 0x5e}, 250 {"tp_mib_usm_bytes_lo", 0x5f}, 251 {"tp_mib_tid_del", 0x60}, 252 {"tp_mib_tid_inv", 0x61}, 253 {"tp_mib_tid_act", 0x62}, 254 {"tp_mib_tid_pas", 0x63}, 255 {"tp_mib_rqe_dfr_mod", 0x64}, 256 {"tp_mib_rqe_dfr_pkt", 0x65} 257 }; 258 259 static u32 read_sge_debug_data(struct cudbg_init *pdbg_init, u32 *sge_dbg_reg) 260 { 261 struct adapter *padap = pdbg_init->adap; 262 u32 value; 263 int i = 0; 264 265 for (i = 0; i <= 15; i++) { 266 t4_write_reg(padap, A_SGE_DEBUG_INDEX, (u32)i); 267 value = t4_read_reg(padap, A_SGE_DEBUG_DATA_LOW); 268 /*printf("LOW 0x%08x\n", value);*/ 269 sge_dbg_reg[(i << 1) | 1] = HTONL_NIBBLE(value); 270 value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH); 271 /*printf("HIGH 0x%08x\n", value);*/ 272 sge_dbg_reg[(i << 1)] = HTONL_NIBBLE(value); 273 } 274 return 0; 275 } 276 277 static u32 read_tp_mib_data(struct cudbg_init *pdbg_init, 278 struct tp_mib_data **ppTp_Mib) 279 { 280 struct adapter *padap = pdbg_init->adap; 281 u32 i = 0; 282 283 for (i = 0; i < TP_MIB_SIZE; i++) { 284 t4_tp_mib_read(padap, &tp_mib[i].value, 1, 285 (u32)tp_mib[i].addr, true); 286 } 287 *ppTp_Mib = (struct tp_mib_data *)&tp_mib[0]; 288 289 return 0; 290 } 291 292 static int t5_wtp_data(struct cudbg_init *pdbg_init, 293 struct cudbg_buffer *dbg_buff, 294 struct cudbg_error *cudbg_err) 295 { 296 struct adapter *padap = pdbg_init->adap; 297 struct sge_debug_reg_data *sge_dbg_reg = NULL; 298 struct cudbg_buffer scratch_buff; 299 struct tp_mib_data *ptp_mib = NULL; 300 struct wtp_data *wtp; 301 u32 Sge_Dbg[32] = {0}; 302 u32 value = 0; 303 u32 i = 0; 304 u32 drop = 0; 305 u32 err = 0; 306 u32 offset; 307 int rc = 0; 308 309 rc = get_scratch_buff(dbg_buff, sizeof(struct wtp_data), &scratch_buff); 310 311 if (rc) 312 goto err; 313 314 offset = scratch_buff.offset; 315 wtp = (struct wtp_data *)((char *)scratch_buff.data + offset); 316 317 read_sge_debug_data(pdbg_init, Sge_Dbg); 318 read_tp_mib_data(pdbg_init, &ptp_mib); 319 320 sge_dbg_reg = (struct sge_debug_reg_data *) &Sge_Dbg[0]; 321 322 /*#######################################################################*/ 323 /*# TX PATH, starting from pcie*/ 324 /*#######################################################################*/ 325 326 /* Get Reqests of commands from SGE to PCIE*/ 327 328 wtp->sge_pcie_cmd_req.sop[0] = sge_dbg_reg->debug_PC_Req_SOP0_cnt; 329 wtp->sge_pcie_cmd_req.sop[1] = sge_dbg_reg->debug_PC_Req_SOP1_cnt; 330 331 wtp->sge_pcie_cmd_req.eop[0] = sge_dbg_reg->debug_PC_Req_EOP0_cnt; 332 wtp->sge_pcie_cmd_req.eop[1] = sge_dbg_reg->debug_PC_Req_EOP1_cnt; 333 334 /* Get Reqests of commands from PCIE to core*/ 335 value = t4_read_reg(padap, A_PCIE_CMDR_REQ_CNT); 336 337 wtp->pcie_core_cmd_req.sop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/ 338 wtp->pcie_core_cmd_req.sop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/ 339 /* there is no EOP for this, so we fake it.*/ 340 wtp->pcie_core_cmd_req.eop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/ 341 wtp->pcie_core_cmd_req.eop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/ 342 343 /* Get DMA stats*/ 344 for (i = 0; i < 4; i++) { 345 value = t4_read_reg(padap, A_PCIE_T5_DMA_STAT3 + (i * 0x10)); 346 wtp->pcie_t5_dma_stat3.sop[i] = value & 0xFF; 347 wtp->pcie_t5_dma_stat3.eop[i] = ((value >> 16) & 0xFF); 348 } 349 350 /* Get SGE debug data high index 6*/ 351 value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_6); 352 wtp->sge_debug_data_high_index_6.sop[0] = ((value >> 4) & 0x0F); 353 wtp->sge_debug_data_high_index_6.eop[0] = ((value >> 0) & 0x0F); 354 wtp->sge_debug_data_high_index_6.sop[1] = ((value >> 12) & 0x0F); 355 wtp->sge_debug_data_high_index_6.eop[1] = ((value >> 8) & 0x0F); 356 wtp->sge_debug_data_high_index_6.sop[2] = ((value >> 20) & 0x0F); 357 wtp->sge_debug_data_high_index_6.eop[2] = ((value >> 16) & 0x0F); 358 wtp->sge_debug_data_high_index_6.sop[3] = ((value >> 28) & 0x0F); 359 wtp->sge_debug_data_high_index_6.eop[3] = ((value >> 24) & 0x0F); 360 361 /* Get SGE debug data high index 3*/ 362 value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_3); 363 wtp->sge_debug_data_high_index_3.sop[0] = ((value >> 4) & 0x0F); 364 wtp->sge_debug_data_high_index_3.eop[0] = ((value >> 0) & 0x0F); 365 wtp->sge_debug_data_high_index_3.sop[1] = ((value >> 12) & 0x0F); 366 wtp->sge_debug_data_high_index_3.eop[1] = ((value >> 8) & 0x0F); 367 wtp->sge_debug_data_high_index_3.sop[2] = ((value >> 20) & 0x0F); 368 wtp->sge_debug_data_high_index_3.eop[2] = ((value >> 16) & 0x0F); 369 wtp->sge_debug_data_high_index_3.sop[3] = ((value >> 28) & 0x0F); 370 wtp->sge_debug_data_high_index_3.eop[3] = ((value >> 24) & 0x0F); 371 372 /* Get ULP SE CNT CHx*/ 373 for (i = 0; i < 4; i++) { 374 value = t4_read_reg(padap, A_ULP_TX_SE_CNT_CH0 + (i * 4)); 375 wtp->ulp_se_cnt_chx.sop[i] = ((value >> 28) & 0x0F); 376 wtp->ulp_se_cnt_chx.eop[i] = ((value >> 24) & 0x0F); 377 } 378 379 /* Get MAC PORTx PKT COUNT*/ 380 for (i = 0; i < 4; i++) { 381 value = t4_read_reg(padap, 0x3081c + ((i * 4) << 12)); 382 wtp->mac_portx_pkt_count.sop[i] = ((value >> 24) & 0xFF); 383 wtp->mac_portx_pkt_count.eop[i] = ((value >> 16) & 0xFF); 384 wtp->mac_porrx_pkt_count.sop[i] = ((value >> 8) & 0xFF); 385 wtp->mac_porrx_pkt_count.eop[i] = ((value >> 0) & 0xFF); 386 } 387 388 /* Get mac portx aFramesTransmittedok*/ 389 for (i = 0; i < 4; i++) { 390 value = t4_read_reg(padap, 0x30a80 + ((i * 4) << 12)); 391 wtp->mac_portx_aframestra_ok.sop[i] = (value & 0xFF); 392 wtp->mac_portx_aframestra_ok.eop[i] = (value & 0xFF); 393 } 394 395 /* Get command respones from core to PCIE*/ 396 value = t4_read_reg(padap, A_PCIE_CMDR_RSP_CNT); 397 398 wtp->core_pcie_cmd_rsp.sop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/ 399 wtp->core_pcie_cmd_rsp.sop[1] = ((value >> 16) & 0xFF); /*bit 16:23*/ 400 401 wtp->core_pcie_cmd_rsp.eop[0] = ((value >> 8) & 0xFF); /*bit 8:15*/ 402 wtp->core_pcie_cmd_rsp.eop[1] = ((value >> 24) & 0xFF); /*bit 24:31*/ 403 404 /*Get command Resposes from PCIE to SGE*/ 405 wtp->pcie_sge_cmd_rsp.sop[0] = sge_dbg_reg->debug_PC_Rsp_SOP0_cnt; 406 wtp->pcie_sge_cmd_rsp.sop[1] = sge_dbg_reg->debug_PC_Rsp_SOP1_cnt; 407 408 wtp->pcie_sge_cmd_rsp.eop[0] = sge_dbg_reg->debug_PC_Rsp_EOP0_cnt; 409 wtp->pcie_sge_cmd_rsp.eop[1] = sge_dbg_reg->debug_PC_Rsp_EOP1_cnt; 410 411 /* Get commands sent from SGE to CIM/uP*/ 412 wtp->sge_cim.sop[0] = sge_dbg_reg->debug_CIM_SOP0_cnt; 413 wtp->sge_cim.sop[1] = sge_dbg_reg->debug_CIM_SOP1_cnt; 414 415 wtp->sge_cim.eop[0] = sge_dbg_reg->debug_CIM_EOP0_cnt; 416 wtp->sge_cim.eop[1] = sge_dbg_reg->debug_CIM_EOP1_cnt; 417 418 /* Get Reqests of data from PCIE by SGE*/ 419 wtp->utx_sge_dma_req.sop[0] = sge_dbg_reg->debug_UD_Rx_SOP0_cnt; 420 wtp->utx_sge_dma_req.sop[1] = sge_dbg_reg->debug_UD_Rx_SOP1_cnt; 421 wtp->utx_sge_dma_req.sop[2] = sge_dbg_reg->debug_UD_Rx_SOP2_cnt; 422 wtp->utx_sge_dma_req.sop[3] = sge_dbg_reg->debug_UD_Rx_SOP3_cnt; 423 424 wtp->utx_sge_dma_req.eop[0] = sge_dbg_reg->debug_UD_Rx_EOP0_cnt; 425 wtp->utx_sge_dma_req.eop[1] = sge_dbg_reg->debug_UD_Rx_EOP1_cnt; 426 wtp->utx_sge_dma_req.eop[2] = sge_dbg_reg->debug_UD_Rx_EOP2_cnt; 427 wtp->utx_sge_dma_req.eop[3] = sge_dbg_reg->debug_UD_Rx_EOP3_cnt; 428 429 /* Get Reqests of data from PCIE by SGE*/ 430 wtp->sge_pcie_dma_req.sop[0] = sge_dbg_reg->debug_PD_Req_Rd0_cnt; 431 wtp->sge_pcie_dma_req.sop[1] = sge_dbg_reg->debug_PD_Req_Rd1_cnt; 432 wtp->sge_pcie_dma_req.sop[2] = sge_dbg_reg->debug_PD_Req_Rd2_cnt; 433 wtp->sge_pcie_dma_req.sop[3] = sge_dbg_reg->debug_PD_Req_Rd3_cnt; 434 /*no EOP's, so fake it.*/ 435 wtp->sge_pcie_dma_req.eop[0] = sge_dbg_reg->debug_PD_Req_Rd0_cnt; 436 wtp->sge_pcie_dma_req.eop[1] = sge_dbg_reg->debug_PD_Req_Rd1_cnt; 437 wtp->sge_pcie_dma_req.eop[2] = sge_dbg_reg->debug_PD_Req_Rd2_cnt; 438 wtp->sge_pcie_dma_req.eop[3] = sge_dbg_reg->debug_PD_Req_Rd3_cnt; 439 440 /* Get Reqests of data from PCIE to core*/ 441 value = t4_read_reg(padap, A_PCIE_DMAR_REQ_CNT); 442 443 wtp->pcie_core_dma_req.sop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/ 444 wtp->pcie_core_dma_req.sop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/ 445 wtp->pcie_core_dma_req.sop[2] = ((value >> 16) & 0xFF); /*bit 16:23*/ 446 wtp->pcie_core_dma_req.sop[3] = ((value >> 24) & 0xFF); /*bit 24:31*/ 447 /* There is no eop so fake it.*/ 448 wtp->pcie_core_dma_req.eop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/ 449 wtp->pcie_core_dma_req.eop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/ 450 wtp->pcie_core_dma_req.eop[2] = ((value >> 16) & 0xFF); /*bit 16:23*/ 451 wtp->pcie_core_dma_req.eop[3] = ((value >> 24) & 0xFF); /*bit 24:31*/ 452 453 /* Get data responses from core to PCIE*/ 454 value = t4_read_reg(padap, A_PCIE_DMAR_RSP_SOP_CNT); 455 456 wtp->core_pcie_dma_rsp.sop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/ 457 wtp->core_pcie_dma_rsp.sop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/ 458 wtp->core_pcie_dma_rsp.sop[2] = ((value >> 16) & 0xFF); /*bit 16:23*/ 459 wtp->core_pcie_dma_rsp.sop[3] = ((value >> 24) & 0xFF); /*bit 24:31*/ 460 461 value = t4_read_reg(padap, A_PCIE_DMAR_RSP_EOP_CNT); 462 463 wtp->core_pcie_dma_rsp.eop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/ 464 wtp->core_pcie_dma_rsp.eop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/ 465 wtp->core_pcie_dma_rsp.eop[2] = ((value >> 16) & 0xFF); /*bit 16:23*/ 466 wtp->core_pcie_dma_rsp.eop[3] = ((value >> 24) & 0xFF); /*bit 24:31*/ 467 468 /* Get PCIE_DATA to SGE*/ 469 wtp->pcie_sge_dma_rsp.sop[0] = sge_dbg_reg->debug_PD_Rsp_SOP0_cnt; 470 wtp->pcie_sge_dma_rsp.sop[1] = sge_dbg_reg->debug_PD_Rsp_SOP1_cnt; 471 wtp->pcie_sge_dma_rsp.sop[2] = sge_dbg_reg->debug_PD_Rsp_SOP2_cnt; 472 wtp->pcie_sge_dma_rsp.sop[3] = sge_dbg_reg->debug_PD_Rsp_SOP3_cnt; 473 474 wtp->pcie_sge_dma_rsp.eop[0] = sge_dbg_reg->debug_PD_Rsp_EOP0_cnt; 475 wtp->pcie_sge_dma_rsp.eop[1] = sge_dbg_reg->debug_PD_Rsp_EOP1_cnt; 476 wtp->pcie_sge_dma_rsp.eop[2] = sge_dbg_reg->debug_PD_Rsp_EOP2_cnt; 477 wtp->pcie_sge_dma_rsp.eop[3] = sge_dbg_reg->debug_PD_Rsp_EOP3_cnt; 478 479 /*Get SGE to ULP_TX*/ 480 wtp->sge_utx.sop[0] = sge_dbg_reg->debug_U_Tx_SOP0_cnt; 481 wtp->sge_utx.sop[1] = sge_dbg_reg->debug_U_Tx_SOP1_cnt; 482 wtp->sge_utx.sop[2] = sge_dbg_reg->debug_U_Tx_SOP2_cnt; 483 wtp->sge_utx.sop[3] = sge_dbg_reg->debug_U_Tx_SOP3_cnt; 484 485 wtp->sge_utx.eop[0] = sge_dbg_reg->debug_U_Tx_EOP0_cnt; 486 wtp->sge_utx.eop[1] = sge_dbg_reg->debug_U_Tx_EOP1_cnt; 487 wtp->sge_utx.eop[2] = sge_dbg_reg->debug_U_Tx_EOP2_cnt; 488 wtp->sge_utx.eop[3] = sge_dbg_reg->debug_U_Tx_EOP3_cnt; 489 490 /* Get ULP_TX to TP*/ 491 for (i = 0; i < 4; i++) { 492 value = t4_read_reg(padap, (A_ULP_TX_SE_CNT_CH0 + (i*4))); 493 494 wtp->utx_tp.sop[i] = ((value >> 28) & 0xF); /*bits 28:31*/ 495 wtp->utx_tp.eop[i] = ((value >> 24) & 0xF); /*bits 24:27*/ 496 } 497 498 /* Get TP_DBG_CSIDE registers*/ 499 for (i = 0; i < 4; i++) { 500 t4_tp_pio_read(padap, &value, 1, (u32)(A_TP_DBG_CSIDE_RX0 + i), 501 true); 502 503 wtp->utx_tpcside.sop[i] = ((value >> 28) & 0xF);/*bits 28:31*/ 504 wtp->utx_tpcside.eop[i] = ((value >> 24) & 0xF);/*bits 24:27*/ 505 wtp->tpcside_rxpld.sop[i] = ((value >> 20) & 0xF);/*bits 20:23*/ 506 wtp->tpcside_rxpld.eop[i] = ((value >> 16) & 0xF);/*bits 16:19*/ 507 wtp->tpcside_rxarb.sop[i] = ((value >> 12) & 0xF);/*bits 12:15*/ 508 wtp->tpcside_rxarb.eop[i] = ((value >> 8) & 0xF); /*bits 8:11*/ 509 wtp->tpcside_rxcpl.sop[i] = ((value >> 4) & 0xF); /*bits 4:7*/ 510 wtp->tpcside_rxcpl.eop[i] = ((value >> 0) & 0xF); /*bits 0:3*/ 511 } 512 513 /* TP_DBG_ESIDE*/ 514 for (i = 0; i < 4; i++) { 515 t4_tp_pio_read(padap, &value, 1, (u32)(A_TP_DBG_ESIDE_PKT0 + i), 516 true); 517 518 wtp->tpeside_mps.sop[i] = ((value >> 28) & 0xF); /*bits 28:31*/ 519 wtp->tpeside_mps.eop[i] = ((value >> 24) & 0xF); /*bits 24:27*/ 520 wtp->tpeside_pm.sop[i] = ((value >> 20) & 0xF); /*bits 20:23*/ 521 wtp->tpeside_pm.eop[i] = ((value >> 16) & 0xF); /*bits 16:19*/ 522 wtp->mps_tpeside.sop[i] = ((value >> 12) & 0xF); /*bits 12:15*/ 523 wtp->mps_tpeside.eop[i] = ((value >> 8) & 0xF); /*bits 8:11*/ 524 wtp->tpeside_pld.sop[i] = ((value >> 4) & 0xF); /*bits 4:7*/ 525 wtp->tpeside_pld.eop[i] = ((value >> 0) & 0xF); /*bits 0:3*/ 526 527 } 528 529 /*PCIE CMD STAT2*/ 530 for (i = 0; i < 3; i++) { 531 value = t4_read_reg(padap, 0x5988 + (i * 0x10)); 532 wtp->pcie_cmd_stat2.sop[i] = value & 0xFF; 533 wtp->pcie_cmd_stat2.eop[i] = value & 0xFF; 534 } 535 536 /*PCIE cmd stat3*/ 537 for (i = 0; i < 3; i++) { 538 value = t4_read_reg(padap, 0x598c + (i * 0x10)); 539 wtp->pcie_cmd_stat3.sop[i] = value & 0xFF; 540 wtp->pcie_cmd_stat3.eop[i] = value & 0xFF; 541 } 542 543 /* ULP_RX input/output*/ 544 for (i = 0; i < 2; i++) { 545 value = t4_read_reg(padap, (A_ULP_RX_SE_CNT_CH0 + (i*4))); 546 547 wtp->pmrx_ulprx.sop[i] = ((value >> 4) & 0xF); /*bits 4:7*/ 548 wtp->pmrx_ulprx.eop[i] = ((value >> 0) & 0xF); /*bits 0:3*/ 549 wtp->ulprx_tpcside.sop[i] = ((value >> 28) & 0xF);/*bits 28:31*/ 550 wtp->ulprx_tpcside.eop[i] = ((value >> 24) & 0xF);/*bits 24:27*/ 551 } 552 553 /* Get the MPS input from TP*/ 554 drop = 0; 555 for (i = 0; i < 2; i++) { 556 value = t4_read_reg(padap, (A_MPS_TX_SE_CNT_TP01 + (i << 2))); 557 wtp->tp_mps.sop[(i*2)] = ((value >> 8) & 0xFF); /*bit 8:15*/ 558 wtp->tp_mps.eop[(i*2)] = ((value >> 0) & 0xFF); /*bit 0:7*/ 559 wtp->tp_mps.sop[(i*2) + 1] = ((value >> 24) & 0xFF);/*bit 24:31 560 */ 561 wtp->tp_mps.eop[(i*2) + 1] = ((value >> 16) & 0xFF);/*bit 16:23 562 */ 563 } 564 drop = ptp_mib->TP_MIB_OFD_ARP_DROP.value; 565 drop += ptp_mib->TP_MIB_OFD_DFR_DROP.value; 566 567 drop += ptp_mib->TP_MIB_TNL_DROP_0.value; 568 drop += ptp_mib->TP_MIB_TNL_DROP_1.value; 569 drop += ptp_mib->TP_MIB_TNL_DROP_2.value; 570 drop += ptp_mib->TP_MIB_TNL_DROP_3.value; 571 572 wtp->tp_mps.drops = drop; 573 574 /* Get the MPS output to the MAC's*/ 575 drop = 0; 576 for (i = 0; i < 2; i++) { 577 value = t4_read_reg(padap, (A_MPS_TX_SE_CNT_MAC01 + (i << 2))); 578 wtp->mps_xgm.sop[(i*2)] = ((value >> 8) & 0xFF);/*bit 8:15*/ 579 wtp->mps_xgm.eop[(i*2)] = ((value >> 0) & 0xFF);/*bit 0:7*/ 580 wtp->mps_xgm.sop[(i*2) + 1] = ((value >> 24) & 0xFF);/*bit 24:31 581 */ 582 wtp->mps_xgm.eop[(i*2) + 1] = ((value >> 16) & 0xFF);/*bit 16:23 583 */ 584 } 585 for (i = 0; i < 4; i++) { 586 value = t4_read_reg(padap, 587 (T5_PORT0_REG(A_MPS_PORT_STAT_TX_PORT_DROP_L) + 588 (i * T5_PORT_STRIDE))); 589 drop += value; 590 } 591 wtp->mps_xgm.drops = (drop & 0xFF); 592 593 /* Get the SOP/EOP counters into and out of MAC. [JHANEL] I think this 594 * is*/ 595 /* clear on read, so you have to read both TX and RX path at same 596 * time.*/ 597 drop = 0; 598 for (i = 0; i < 4; i++) { 599 value = t4_read_reg(padap, 600 (T5_PORT0_REG(A_MAC_PORT_PKT_COUNT) + 601 (i * T5_PORT_STRIDE))); 602 603 wtp->tx_xgm_xgm.sop[i] = ((value >> 24) & 0xFF); /*bit 24:31*/ 604 wtp->tx_xgm_xgm.eop[i] = ((value >> 16) & 0xFF); /*bit 16:23*/ 605 wtp->rx_xgm_xgm.sop[i] = ((value >> 8) & 0xFF); /*bit 8:15*/ 606 wtp->rx_xgm_xgm.eop[i] = ((value >> 0) & 0xFF); /*bit 0:7*/ 607 } 608 609 /* Get the MAC's output to the wire*/ 610 drop = 0; 611 for (i = 0; i < 4; i++) { 612 value = t4_read_reg(padap, 613 (T5_PORT0_REG(A_MAC_PORT_AFRAMESTRANSMITTEDOK) + 614 (i * T5_PORT_STRIDE))); 615 wtp->xgm_wire.sop[i] = (value); 616 wtp->xgm_wire.eop[i] = (value); /* No EOP for XGMAC, so fake 617 it.*/ 618 } 619 620 /*########################################################################*/ 621 /*# RX PATH, starting from wire*/ 622 /*########################################################################*/ 623 624 /* Add up the wire input to the MAC*/ 625 drop = 0; 626 for (i = 0; i < 4; i++) { 627 value = t4_read_reg(padap, 628 (T5_PORT0_REG(A_MAC_PORT_AFRAMESRECEIVEDOK) + 629 (i * T5_PORT_STRIDE))); 630 631 wtp->wire_xgm.sop[i] = (value); 632 wtp->wire_xgm.eop[i] = (value); /* No EOP for XGMAC, so fake 633 it.*/ 634 } 635 636 /* Already read the rx_xgm_xgm when reading TX path.*/ 637 638 /* Add up SOP/EOP's on all 8 MPS buffer channels*/ 639 drop = 0; 640 for (i = 0; i < 8; i++) { 641 value = t4_read_reg(padap, (A_MPS_RX_SE_CNT_IN0 + (i << 2))); 642 643 wtp->xgm_mps.sop[i] = ((value >> 8) & 0xFF); /*bits 8:15*/ 644 wtp->xgm_mps.eop[i] = ((value >> 0) & 0xFF); /*bits 0:7*/ 645 } 646 for (i = 0; i < 4; i++) { 647 value = t4_read_reg(padap, (A_MPS_RX_CLS_DROP_CNT0 + (i << 2))); 648 /* typo in JHANEL's code.*/ 649 drop += (value & 0xFFFF) + ((value >> 16) & 0xFFFF); 650 } 651 wtp->xgm_mps.cls_drop = drop & 0xFF; 652 653 /* Add up the overflow drops on all 4 ports.*/ 654 drop = 0; 655 for (i = 0; i < 4; i++) { 656 value = t4_read_reg(padap, 657 (A_MPS_STAT_RX_BG_0_MAC_DROP_FRAME_L + 658 (i << 3))); 659 drop += value; 660 value = t4_read_reg(padap, 661 (A_MPS_STAT_RX_BG_0_MAC_DROP_FRAME_L + 662 (i << 2))); 663 value = t4_read_reg(padap, 664 (A_MPS_STAT_RX_BG_0_LB_DROP_FRAME_L + 665 (i << 3))); 666 drop += value; 667 value = t4_read_reg(padap, 668 (A_MPS_STAT_RX_BG_0_MAC_DROP_FRAME_L + 669 (i << 2))); 670 671 value = t4_read_reg(padap, 672 (A_MPS_STAT_RX_BG_0_MAC_TRUNC_FRAME_L + 673 (i << 3))); 674 drop += value; 675 value = t4_read_reg(padap, 676 (A_MPS_STAT_RX_BG_0_MAC_TRUNC_FRAME_L + 677 (i << 3))); 678 value = t4_read_reg(padap, 679 (A_MPS_STAT_RX_BG_0_LB_TRUNC_FRAME_L + 680 (i << 3))); 681 drop += value; 682 value = t4_read_reg(padap, 683 (A_MPS_STAT_RX_BG_0_MAC_TRUNC_FRAME_L + 684 (i << 3))); 685 686 value = t4_read_reg(padap, 687 T5_PORT0_REG(A_MPS_PORT_STAT_LB_PORT_DROP_FRAMES) + 688 (i * T5_PORT_STRIDE)); 689 drop += value; 690 } 691 wtp->xgm_mps.drop = (drop & 0xFF); 692 693 /* Add up the MPS errors that should result in dropped packets*/ 694 err = 0; 695 for (i = 0; i < 4; i++) { 696 697 value = t4_read_reg(padap, 698 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_MTU_ERROR_L) + 699 (i * T5_PORT_STRIDE))); 700 err += value; 701 value = t4_read_reg(padap, 702 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_MTU_ERROR_L) + 703 (i * T5_PORT_STRIDE) + 4)); 704 705 value = t4_read_reg(padap, 706 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_MTU_CRC_ERROR_L) + 707 (i * T5_PORT_STRIDE))); 708 err += value; 709 value = t4_read_reg(padap, 710 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_MTU_CRC_ERROR_L) + 711 (i * T5_PORT_STRIDE) + 4)); 712 713 value = t4_read_reg(padap, 714 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_CRC_ERROR_L) + 715 (i * T5_PORT_STRIDE))); 716 err += value; 717 value = t4_read_reg(padap, 718 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_CRC_ERROR_L) + 719 (i * T5_PORT_STRIDE) + 4)); 720 721 value = t4_read_reg(padap, 722 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_LEN_ERROR_L) + 723 (i * T5_PORT_STRIDE))); 724 err += value; 725 value = t4_read_reg(padap, 726 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_LEN_ERROR_L) + 727 (i * T5_PORT_STRIDE) + 4)); 728 729 value = t4_read_reg(padap, 730 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_SYM_ERROR_L) + 731 (i * T5_PORT_STRIDE))); 732 err += value; 733 value = t4_read_reg(padap, 734 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_SYM_ERROR_L) + 735 (i * T5_PORT_STRIDE) + 4)); 736 737 value = t4_read_reg(padap, 738 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_LESS_64B_L) + 739 (i * T5_PORT_STRIDE))); 740 err += value; 741 value = t4_read_reg(padap, 742 (T5_PORT0_REG((A_MPS_PORT_STAT_RX_PORT_LESS_64B_L) + 743 (i * T5_PORT_STRIDE) + 4))); 744 } 745 wtp->xgm_mps.err = (err & 0xFF); 746 747 drop = 0; 748 for (i = 0; i < 2; i++) { 749 value = t4_read_reg(padap, (A_MPS_RX_SE_CNT_OUT01 + (i << 2))); 750 751 wtp->mps_tp.sop[(i*2)] = ((value >> 8) & 0xFF); /*bit 8:15*/ 752 wtp->mps_tp.eop[(i*2)] = ((value >> 0) & 0xFF); /*bit 0:7*/ 753 wtp->mps_tp.sop[(i*2) + 1] = ((value >> 24) & 0xFF);/*bit 24:31 754 */ 755 wtp->mps_tp.eop[(i*2) + 1] = ((value >> 16) & 0xFF);/*bit 16:23 756 */ 757 } 758 drop = ptp_mib->TP_MIB_TNL_CNG_DROP_0.value; 759 drop += ptp_mib->TP_MIB_TNL_CNG_DROP_1.value; 760 drop += ptp_mib->TP_MIB_TNL_CNG_DROP_2.value; 761 drop += ptp_mib->TP_MIB_TNL_CNG_DROP_3.value; 762 drop += ptp_mib->TP_MIB_OFD_CHN_DROP_0.value; 763 drop += ptp_mib->TP_MIB_OFD_CHN_DROP_1.value; 764 drop += ptp_mib->TP_MIB_OFD_CHN_DROP_2.value; 765 drop += ptp_mib->TP_MIB_OFD_CHN_DROP_3.value; 766 drop += ptp_mib->TP_MIB_FCOE_DROP_0.value; 767 drop += ptp_mib->TP_MIB_FCOE_DROP_1.value; 768 drop += ptp_mib->TP_MIB_FCOE_DROP_2.value; 769 drop += ptp_mib->TP_MIB_FCOE_DROP_3.value; 770 drop += ptp_mib->TP_MIB_OFD_VLN_DROP_0.value; 771 drop += ptp_mib->TP_MIB_OFD_VLN_DROP_1.value; 772 drop += ptp_mib->TP_MIB_OFD_VLN_DROP_2.value; 773 drop += ptp_mib->TP_MIB_OFD_VLN_DROP_3.value; 774 drop += ptp_mib->TP_MIB_USM_DROP.value; 775 776 wtp->mps_tp.drops = drop; 777 778 /* Get TP_DBG_CSIDE_TX registers*/ 779 for (i = 0; i < 4; i++) { 780 t4_tp_pio_read(padap, &value, 1, (u32)(A_TP_DBG_CSIDE_RX0 + i), 781 true); 782 783 wtp->tpcside_csw.sop[i] = ((value >> 28) & 0xF);/*bits 28:31*/ 784 wtp->tpcside_csw.eop[i] = ((value >> 24) & 0xF);/*bits 24:27*/ 785 wtp->tpcside_pm.sop[i] = ((value >> 20) & 0xF);/*bits 20:23*/ 786 wtp->tpcside_pm.eop[i] = ((value >> 16) & 0xF);/*bits 16:19*/ 787 wtp->tpcside_uturn.sop[i] = ((value >> 12) & 0xF);/*bits 12:15*/ 788 wtp->tpcside_uturn.eop[i] = ((value >> 8) & 0xF); /*bits 8:11*/ 789 wtp->tpcside_txcpl.sop[i] = ((value >> 4) & 0xF); /*bits 4:7*/ 790 wtp->tpcside_txcpl.eop[i] = ((value >> 0) & 0xF); /*bits 0:3*/ 791 } 792 793 /* TP to CPL_SWITCH*/ 794 wtp->tp_csw.sop[0] = sge_dbg_reg->debug_CPLSW_TP_Rx_SOP0_cnt; 795 wtp->tp_csw.sop[1] = sge_dbg_reg->debug_CPLSW_TP_Rx_SOP1_cnt; 796 797 wtp->tp_csw.eop[0] = sge_dbg_reg->debug_CPLSW_TP_Rx_EOP0_cnt; 798 wtp->tp_csw.eop[1] = sge_dbg_reg->debug_CPLSW_TP_Rx_EOP1_cnt; 799 800 /* TP/CPL_SWITCH to SGE*/ 801 wtp->csw_sge.sop[0] = sge_dbg_reg->debug_T_Rx_SOP0_cnt; 802 wtp->csw_sge.sop[1] = sge_dbg_reg->debug_T_Rx_SOP1_cnt; 803 804 wtp->csw_sge.eop[0] = sge_dbg_reg->debug_T_Rx_EOP0_cnt; 805 wtp->csw_sge.eop[1] = sge_dbg_reg->debug_T_Rx_EOP1_cnt; 806 807 wtp->sge_pcie.sop[0] = sge_dbg_reg->debug_PD_Req_SOP0_cnt; 808 wtp->sge_pcie.sop[1] = sge_dbg_reg->debug_PD_Req_SOP1_cnt; 809 wtp->sge_pcie.sop[2] = sge_dbg_reg->debug_PD_Req_SOP2_cnt; 810 wtp->sge_pcie.sop[3] = sge_dbg_reg->debug_PD_Req_SOP3_cnt; 811 812 wtp->sge_pcie.eop[0] = sge_dbg_reg->debug_PD_Req_EOP0_cnt; 813 wtp->sge_pcie.eop[1] = sge_dbg_reg->debug_PD_Req_EOP1_cnt; 814 wtp->sge_pcie.eop[2] = sge_dbg_reg->debug_PD_Req_EOP2_cnt; 815 wtp->sge_pcie.eop[3] = sge_dbg_reg->debug_PD_Req_EOP3_cnt; 816 817 wtp->sge_pcie_ints.sop[0] = sge_dbg_reg->debug_PD_Req_Int0_cnt; 818 wtp->sge_pcie_ints.sop[1] = sge_dbg_reg->debug_PD_Req_Int1_cnt; 819 wtp->sge_pcie_ints.sop[2] = sge_dbg_reg->debug_PD_Req_Int2_cnt; 820 wtp->sge_pcie_ints.sop[3] = sge_dbg_reg->debug_PD_Req_Int3_cnt; 821 /* NO EOP, so fake it.*/ 822 wtp->sge_pcie_ints.eop[0] = sge_dbg_reg->debug_PD_Req_Int0_cnt; 823 wtp->sge_pcie_ints.eop[1] = sge_dbg_reg->debug_PD_Req_Int1_cnt; 824 wtp->sge_pcie_ints.eop[2] = sge_dbg_reg->debug_PD_Req_Int2_cnt; 825 wtp->sge_pcie_ints.eop[3] = sge_dbg_reg->debug_PD_Req_Int3_cnt; 826 827 /*Get PCIE DMA1 STAT2*/ 828 for (i = 0; i < 4; i++) { 829 value = t4_read_reg(padap, A_PCIE_T5_DMA_STAT2 + (i * 0x10)); 830 wtp->pcie_dma1_stat2.sop[i] = ((value >> 8) & 0x0F); 831 wtp->pcie_dma1_stat2.eop[i] = ((value >> 8) & 0x0F); 832 wtp->pcie_dma1_stat2_core.sop[i] += value & 0x0F; 833 wtp->pcie_dma1_stat2_core.eop[i] += value & 0x0F; 834 } 835 836 /* Get mac porrx aFramesTransmittedok*/ 837 for (i = 0; i < 4; i++) { 838 value = t4_read_reg(padap, 0x30a88 + ((i * 4) << 12)); 839 wtp->mac_porrx_aframestra_ok.sop[i] = (value & 0xFF); 840 wtp->mac_porrx_aframestra_ok.eop[i] = (value & 0xFF); 841 } 842 843 /*Get SGE debug data high index 7*/ 844 value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_7); 845 wtp->sge_debug_data_high_indx7.sop[0] = ((value >> 4) & 0x0F); 846 wtp->sge_debug_data_high_indx7.eop[0] = ((value >> 0) & 0x0F); 847 wtp->sge_debug_data_high_indx7.sop[1] = ((value >> 12) & 0x0F); 848 wtp->sge_debug_data_high_indx7.eop[1] = ((value >> 8) & 0x0F); 849 wtp->sge_debug_data_high_indx7.sop[2] = ((value >> 20) & 0x0F); 850 wtp->sge_debug_data_high_indx7.eop[2] = ((value >> 16) & 0x0F); 851 wtp->sge_debug_data_high_indx7.sop[3] = ((value >> 28) & 0x0F); 852 wtp->sge_debug_data_high_indx7.eop[3] = ((value >> 24) & 0x0F); 853 854 /*Get SGE debug data high index 1*/ 855 value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_1); 856 wtp->sge_debug_data_high_indx1.sop[0] = ((value >> 20) & 0x0F); 857 wtp->sge_debug_data_high_indx1.eop[0] = ((value >> 16) & 0x0F); 858 wtp->sge_debug_data_high_indx1.sop[1] = ((value >> 28) & 0x0F); 859 wtp->sge_debug_data_high_indx1.eop[1] = ((value >> 24) & 0x0F); 860 861 /*Get TP debug CSIDE Tx registers*/ 862 for (i = 0; i < 2; i++) { 863 t4_tp_pio_read(padap, &value, 1, (u32)(A_TP_DBG_CSIDE_TX0 + i), 864 true); 865 866 wtp->utx_tpcside_tx.sop[i] = ((value >> 28) & 0xF);/*bits 28:31 867 */ 868 wtp->utx_tpcside_tx.eop[i] = ((value >> 24) & 0xF); 869 } 870 871 /*Get SGE debug data high index 9*/ 872 value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_9); 873 wtp->sge_debug_data_high_indx9.sop[0] = ((value >> 20) & 0x0F); 874 wtp->sge_debug_data_high_indx9.sop[1] = ((value >> 28) & 0x0F); 875 wtp->sge_debug_data_high_indx9.eop[0] = ((value >> 16) & 0x0F); 876 wtp->sge_debug_data_high_indx9.eop[1] = ((value >> 24) & 0x0F); 877 wtp->sge_work_req_pkt.sop[0] = ((value >> 4) & 0x0F); 878 wtp->sge_work_req_pkt.sop[1] = ((value >> 12) & 0x0F); 879 880 /*Get LE DB response count*/ 881 value = t4_read_reg(padap, A_LE_DB_REQ_RSP_CNT); 882 wtp->le_db_rsp_cnt.sop = value & 0xF; 883 wtp->le_db_rsp_cnt.eop = (value >> 16) & 0xF; 884 885 /*Get TP debug Eside PKTx*/ 886 for (i = 0; i < 4; i++) { 887 t4_tp_pio_read(padap, &value, 1, (u32)(A_TP_DBG_ESIDE_PKT0 + i), 888 true); 889 890 wtp->tp_dbg_eside_pktx.sop[i] = ((value >> 12) & 0xF); 891 wtp->tp_dbg_eside_pktx.eop[i] = ((value >> 8) & 0xF); 892 } 893 894 /* Get data responses from core to PCIE*/ 895 value = t4_read_reg(padap, A_PCIE_DMAW_SOP_CNT); 896 897 wtp->pcie_core_dmaw.sop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/ 898 wtp->pcie_core_dmaw.sop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/ 899 wtp->pcie_core_dmaw.sop[2] = ((value >> 16) & 0xFF); /*bit 16:23*/ 900 wtp->pcie_core_dmaw.sop[3] = ((value >> 24) & 0xFF); /*bit 24:31*/ 901 902 value = t4_read_reg(padap, A_PCIE_DMAW_EOP_CNT); 903 904 wtp->pcie_core_dmaw.eop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/ 905 wtp->pcie_core_dmaw.eop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/ 906 wtp->pcie_core_dmaw.eop[2] = ((value >> 16) & 0xFF); /*bit 16:23*/ 907 wtp->pcie_core_dmaw.eop[3] = ((value >> 24) & 0xFF); /*bit 24:31*/ 908 909 value = t4_read_reg(padap, A_PCIE_DMAI_CNT); 910 911 wtp->pcie_core_dmai.sop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/ 912 wtp->pcie_core_dmai.sop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/ 913 wtp->pcie_core_dmai.sop[2] = ((value >> 16) & 0xFF); /*bit 16:23*/ 914 wtp->pcie_core_dmai.sop[3] = ((value >> 24) & 0xFF); /*bit 24:31*/ 915 /* no eop for interrups, just fake it.*/ 916 wtp->pcie_core_dmai.eop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/ 917 wtp->pcie_core_dmai.eop[1] = ((value >> 8) & 0xFF); /*bit 8:15*/ 918 wtp->pcie_core_dmai.eop[2] = ((value >> 16) & 0xFF); /*bit 16:23*/ 919 wtp->pcie_core_dmai.eop[3] = ((value >> 24) & 0xFF); /*bit 24:31*/ 920 921 rc = write_compression_hdr(&scratch_buff, dbg_buff); 922 923 if (rc) 924 goto err1; 925 926 rc = compress_buff(&scratch_buff, dbg_buff); 927 928 err1: 929 release_scratch_buff(&scratch_buff, dbg_buff); 930 err: 931 return rc; 932 } 933 934 static int t6_wtp_data(struct cudbg_init *pdbg_init, 935 struct cudbg_buffer *dbg_buff, 936 struct cudbg_error *cudbg_err) 937 { 938 struct adapter *padap = pdbg_init->adap; 939 struct sge_debug_reg_data *sge_dbg_reg = NULL; 940 struct cudbg_buffer scratch_buff; 941 struct tp_mib_data *ptp_mib = NULL; 942 struct wtp_data *wtp; 943 u32 Sge_Dbg[32] = {0}; 944 u32 value = 0; 945 u32 i = 0; 946 u32 drop = 0; 947 u32 err = 0; 948 u32 offset; 949 int rc = 0; 950 951 rc = get_scratch_buff(dbg_buff, sizeof(struct wtp_data), &scratch_buff); 952 953 if (rc) 954 goto err; 955 956 offset = scratch_buff.offset; 957 wtp = (struct wtp_data *)((char *)scratch_buff.data + offset); 958 959 read_sge_debug_data(pdbg_init, Sge_Dbg); 960 read_tp_mib_data(pdbg_init, &ptp_mib); 961 962 sge_dbg_reg = (struct sge_debug_reg_data *) &Sge_Dbg[0]; 963 964 /*# TX PATH*/ 965 966 /*PCIE CMD STAT2*/ 967 value = t4_read_reg(padap, A_PCIE_T5_CMD_STAT2); 968 wtp->pcie_cmd_stat2.sop[0] = value & 0xFF; 969 wtp->pcie_cmd_stat2.eop[0] = value & 0xFF; 970 971 value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_7); 972 wtp->sge_pcie_cmd_req.sop[0] = ((value >> 20) & 0x0F); 973 wtp->sge_pcie_cmd_req.eop[0] = ((value >> 16) & 0x0F); 974 wtp->sge_pcie_cmd_req.sop[1] = ((value >> 28) & 0x0F); 975 wtp->sge_pcie_cmd_req.eop[1] = ((value >> 24) & 0x0F); 976 977 value = t4_read_reg(padap, A_PCIE_T5_CMD_STAT3); 978 wtp->pcie_cmd_stat3.sop[0] = value & 0xFF; 979 wtp->pcie_cmd_stat3.eop[0] = value & 0xFF; 980 981 /*Get command Resposes from PCIE to SGE*/ 982 wtp->pcie_sge_cmd_rsp.sop[0] = sge_dbg_reg->debug_PC_Rsp_SOP0_cnt; 983 wtp->pcie_sge_cmd_rsp.eop[0] = sge_dbg_reg->debug_PC_Rsp_EOP0_cnt; 984 wtp->pcie_sge_cmd_rsp.sop[1] = sge_dbg_reg->debug_PC_Rsp_SOP1_cnt; 985 wtp->pcie_sge_cmd_rsp.eop[1] = sge_dbg_reg->debug_PC_Rsp_EOP0_cnt; 986 987 /* Get commands sent from SGE to CIM/uP*/ 988 wtp->sge_cim.sop[0] = sge_dbg_reg->debug_CIM_SOP0_cnt; 989 wtp->sge_cim.sop[1] = sge_dbg_reg->debug_CIM_SOP1_cnt; 990 991 wtp->sge_cim.eop[0] = sge_dbg_reg->debug_CIM_EOP0_cnt; 992 wtp->sge_cim.eop[1] = sge_dbg_reg->debug_CIM_EOP1_cnt; 993 994 /*Get SGE debug data high index 9*/ 995 value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_9); 996 wtp->sge_work_req_pkt.sop[0] = ((value >> 4) & 0x0F); 997 wtp->sge_work_req_pkt.eop[0] = ((value >> 0) & 0x0F); 998 999 for (i = 0; i < 2; i++) { 1000 value = t4_read_reg(padap, A_PCIE_T5_DMA_STAT2 + (i * 0x10)); 1001 wtp->pcie_dma1_stat2.sop[i] = ((value >> 8) & 0x0F); 1002 wtp->pcie_dma1_stat2.eop[i] = ((value >> 8) & 0x0F); 1003 wtp->pcie_dma1_stat2_core.sop[i] = value & 0x0F; 1004 wtp->pcie_dma1_stat2_core.eop[i] = value & 0x0F; 1005 } 1006 1007 /* Get DMA0 stats3*/ 1008 for (i = 0; i < 2; i++) { 1009 value = t4_read_reg(padap, A_PCIE_T5_DMA_STAT3 + (i * 0x10)); 1010 wtp->pcie_t5_dma_stat3.sop[i] = value & 0xFF; 1011 wtp->pcie_t5_dma_stat3.eop[i] = ((value >> 16) & 0xFF); 1012 } 1013 1014 /* Get ULP SE CNT CHx*/ 1015 for (i = 0; i < 4; i++) { 1016 value = t4_read_reg(padap, A_ULP_TX_SE_CNT_CH0 + (i * 4)); 1017 wtp->ulp_se_cnt_chx.sop[i] = ((value >> 28) & 0x0F); 1018 wtp->ulp_se_cnt_chx.eop[i] = ((value >> 24) & 0x0F); 1019 } 1020 1021 /* Get TP_DBG_CSIDE registers*/ 1022 for (i = 0; i < 4; i++) { 1023 t4_tp_pio_read(padap, &value, 1, (u32)(A_TP_DBG_CSIDE_RX0 + i), 1024 true); 1025 1026 wtp->utx_tpcside.sop[i] = ((value >> 28) & 0xF);/*bits 28:31*/ 1027 wtp->utx_tpcside.eop[i] = ((value >> 24) & 0xF);/*bits 24:27*/ 1028 wtp->tpcside_rxarb.sop[i] = ((value >> 12) & 0xF);/*bits 12:15*/ 1029 wtp->tpcside_rxarb.eop[i] = ((value >> 8) & 0xF); /*bits 8:11*/ 1030 } 1031 1032 for (i = 0; i < 4; i++) { 1033 t4_tp_pio_read(padap, &value, 1, (u32)(A_TP_DBG_ESIDE_PKT0 + i), 1034 true); 1035 1036 1037 wtp->tpeside_mps.sop[i] = ((value >> 28) & 0xF); /*bits 28:31*/ 1038 wtp->tpeside_mps.eop[i] = ((value >> 24) & 0xF); /*bits 24:27*/ 1039 } 1040 1041 for (i = 0; i < 2; i++) { 1042 value = t4_read_reg(padap, (A_MPS_TX_SE_CNT_TP01 + (i << 2))); 1043 wtp->tp_mps.sop[(i*2)] = ((value >> 8) & 0xFF); /*bit 8:15*/ 1044 wtp->tp_mps.eop[(i*2)] = ((value >> 0) & 0xFF); /*bit 0:7*/ 1045 wtp->tp_mps.sop[(i*2) + 1] = ((value >> 24) & 0xFF);/*bit 24:31 1046 */ 1047 wtp->tp_mps.eop[(i*2) + 1] = ((value >> 16) & 0xFF);/*bit 16:23 1048 */ 1049 } 1050 1051 for (i = 0; i < 2; i++) { 1052 value = t4_read_reg(padap, (A_MPS_TX_SE_CNT_MAC01 + (i << 2))); 1053 wtp->mps_xgm.sop[(i*2)] = ((value >> 8) & 0xFF);/*bit 8:15*/ 1054 wtp->mps_xgm.eop[(i*2)] = ((value >> 0) & 0xFF); /*bit 0:7*/ 1055 wtp->mps_xgm.sop[(i*2) + 1] = ((value >> 24) & 0xFF);/*bit 24:31 1056 */ 1057 wtp->mps_xgm.eop[(i*2) + 1] = ((value >> 16) & 0xFF);/*bit 16:23 1058 */ 1059 } 1060 1061 /* Get MAC PORTx PKT COUNT*/ 1062 for (i = 0; i < 2; i++) { 1063 value = t4_read_reg(padap, 0x3081c + ((i * 4) << 12)); 1064 wtp->mac_portx_pkt_count.sop[i] = ((value >> 24) & 0xFF); 1065 wtp->mac_portx_pkt_count.eop[i] = ((value >> 16) & 0xFF); 1066 wtp->mac_porrx_pkt_count.sop[i] = ((value >> 8) & 0xFF); 1067 wtp->mac_porrx_pkt_count.eop[i] = ((value >> 0) & 0xFF); 1068 } 1069 1070 for (i = 0; i < 2; i++) { 1071 value = t4_read_reg(padap, 0x30f20 + ((i * 4) << 12)); 1072 wtp->mac_portx_aframestra_ok.sop[i] = value & 0xff; 1073 wtp->mac_portx_aframestra_ok.eop[i] = value & 0xff; 1074 } 1075 1076 /*MAC_PORT_MTIP_1G10G_TX_etherStatsPkts*/ 1077 1078 for (i = 0; i < 2; i++) { 1079 value = t4_read_reg(padap, 0x30f60 + ((i * 4) << 12)); 1080 wtp->mac_portx_etherstatspkts.sop[i] = value & 0xff; 1081 wtp->mac_portx_etherstatspkts.eop[i] = value & 0xff; 1082 } 1083 1084 /*RX path*/ 1085 1086 value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_7); 1087 wtp->sge_debug_data_high_indx7.sop[0] = ((value >> 4) & 0x0F); 1088 wtp->sge_debug_data_high_indx7.eop[0] = ((value >> 0) & 0x0F); 1089 wtp->sge_debug_data_high_indx7.sop[1] = ((value >> 12) & 0x0F); 1090 wtp->sge_debug_data_high_indx7.eop[1] = ((value >> 8) & 0x0F); 1091 1092 /*Get SGE debug data high index 1*/ 1093 value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_1); 1094 wtp->sge_debug_data_high_indx1.sop[0] = ((value >> 20) & 0x0F); 1095 wtp->sge_debug_data_high_indx1.eop[0] = ((value >> 16) & 0x0F); 1096 wtp->sge_debug_data_high_indx1.sop[1] = ((value >> 28) & 0x0F); 1097 wtp->sge_debug_data_high_indx1.eop[1] = ((value >> 24) & 0x0F); 1098 1099 value = t4_read_reg(padap, A_SGE_DEBUG_DATA_HIGH_INDEX_9); 1100 wtp->sge_debug_data_high_indx9.sop[0] = ((value >> 20) & 0x0F); 1101 wtp->sge_debug_data_high_indx9.sop[1] = ((value >> 28) & 0x0F); 1102 1103 wtp->sge_debug_data_high_indx9.eop[0] = ((value >> 16) & 0x0F); 1104 wtp->sge_debug_data_high_indx9.eop[1] = ((value >> 24) & 0x0F); 1105 1106 for (i = 0; i < 2; i++) { 1107 t4_tp_pio_read(padap, &value, 1, (u32)(A_TP_DBG_CSIDE_TX0 + i), 1108 true); 1109 1110 wtp->utx_tpcside_tx.sop[i] = ((value >> 28) & 0xF);/*bits 28:31 1111 */ 1112 wtp->utx_tpcside_tx.eop[i] = ((value >> 24) & 0xF); 1113 } 1114 1115 /*ULP_RX input/output*/ 1116 for (i = 0; i < 2; i++) { 1117 value = t4_read_reg(padap, (A_ULP_RX_SE_CNT_CH0 + (i*4))); 1118 1119 wtp->pmrx_ulprx.sop[i] = ((value >> 4) & 0xF); /*bits 4:7*/ 1120 wtp->pmrx_ulprx.eop[i] = ((value >> 0) & 0xF); /*bits 0:3*/ 1121 wtp->ulprx_tpcside.sop[i] = ((value >> 28) & 0xF);/*bits 28:31*/ 1122 wtp->ulprx_tpcside.eop[i] = ((value >> 24) & 0xF);/*bits 24:27*/ 1123 } 1124 1125 /*Get LE DB response count*/ 1126 value = t4_read_reg(padap, A_LE_DB_REQ_RSP_CNT); 1127 wtp->le_db_rsp_cnt.sop = value & 0xF; 1128 wtp->le_db_rsp_cnt.eop = (value >> 16) & 0xF; 1129 1130 /*Get TP debug Eside PKTx*/ 1131 for (i = 0; i < 4; i++) { 1132 t4_tp_pio_read(padap, &value, 1, (u32)(A_TP_DBG_ESIDE_PKT0 + i), 1133 true); 1134 1135 wtp->tp_dbg_eside_pktx.sop[i] = ((value >> 12) & 0xF); 1136 wtp->tp_dbg_eside_pktx.eop[i] = ((value >> 8) & 0xF); 1137 } 1138 1139 drop = 0; 1140 /*MPS_RX_SE_CNT_OUT01*/ 1141 value = t4_read_reg(padap, (A_MPS_RX_SE_CNT_OUT01 + (i << 2))); 1142 wtp->mps_tp.sop[0] = ((value >> 8) & 0xFF); /*bit 8:15*/ 1143 wtp->mps_tp.eop[0] = ((value >> 0) & 0xFF); /*bit 0:7*/ 1144 wtp->mps_tp.sop[1] = ((value >> 24) & 0xFF); /*bit 24:31*/ 1145 wtp->mps_tp.eop[1] = ((value >> 16) & 0xFF); /*bit 16:23*/ 1146 1147 drop = ptp_mib->TP_MIB_TNL_CNG_DROP_0.value; 1148 drop += ptp_mib->TP_MIB_TNL_CNG_DROP_1.value; 1149 drop += ptp_mib->TP_MIB_OFD_CHN_DROP_0.value; 1150 drop += ptp_mib->TP_MIB_OFD_CHN_DROP_1.value; 1151 drop += ptp_mib->TP_MIB_FCOE_DROP_0.value; 1152 drop += ptp_mib->TP_MIB_FCOE_DROP_1.value; 1153 drop += ptp_mib->TP_MIB_OFD_VLN_DROP_0.value; 1154 drop += ptp_mib->TP_MIB_OFD_VLN_DROP_1.value; 1155 drop += ptp_mib->TP_MIB_USM_DROP.value; 1156 1157 wtp->mps_tp.drops = drop; 1158 1159 drop = 0; 1160 for (i = 0; i < 8; i++) { 1161 value = t4_read_reg(padap, (A_MPS_RX_SE_CNT_IN0 + (i << 2))); 1162 1163 wtp->xgm_mps.sop[i] = ((value >> 8) & 0xFF); /*bits 8:15*/ 1164 wtp->xgm_mps.eop[i] = ((value >> 0) & 0xFF); /*bits 0:7*/ 1165 } 1166 for (i = 0; i < 2; i++) { 1167 value = t4_read_reg(padap, (A_MPS_RX_CLS_DROP_CNT0 + (i << 2))); 1168 drop += (value & 0xFFFF) + ((value >> 16) & 0xFFFF); 1169 } 1170 wtp->xgm_mps.cls_drop = drop & 0xFF; 1171 1172 for (i = 0; i < 2; i++) { 1173 value = t4_read_reg(padap, 0x30e20 + ((i * 4) << 12)); 1174 wtp->mac_porrx_aframestra_ok.sop[i] = value & 0xff; 1175 wtp->mac_porrx_aframestra_ok.eop[i] = value & 0xff; 1176 } 1177 1178 /*MAC_PORT_MTIP_1G10G_RX_etherStatsPkts*/ 1179 for (i = 0; i < 2; i++) { 1180 value = t4_read_reg(padap, 0x30e60 + ((i * 4) << 12)); 1181 wtp->mac_porrx_etherstatspkts.sop[i] = value & 0xff; 1182 wtp->mac_porrx_etherstatspkts.eop[i] = value & 0xff; 1183 } 1184 1185 wtp->sge_pcie_ints.sop[0] = sge_dbg_reg->debug_PD_Req_Int0_cnt; 1186 wtp->sge_pcie_ints.sop[1] = sge_dbg_reg->debug_PD_Req_Int1_cnt; 1187 wtp->sge_pcie_ints.sop[2] = sge_dbg_reg->debug_PD_Req_Int2_cnt; 1188 wtp->sge_pcie_ints.sop[3] = sge_dbg_reg->debug_PD_Req_Int3_cnt; 1189 1190 /* Add up the overflow drops on all 4 ports.*/ 1191 drop = 0; 1192 for (i = 0; i < 2; i++) { 1193 value = t4_read_reg(padap, 1194 (A_MPS_STAT_RX_BG_0_MAC_DROP_FRAME_L + 1195 (i << 3))); 1196 drop += value; 1197 value = t4_read_reg(padap, 1198 (A_MPS_STAT_RX_BG_0_MAC_DROP_FRAME_L + 1199 (i << 2))); 1200 value = t4_read_reg(padap, 1201 (A_MPS_STAT_RX_BG_0_LB_DROP_FRAME_L + 1202 (i << 3))); 1203 drop += value; 1204 value = t4_read_reg(padap, 1205 (A_MPS_STAT_RX_BG_0_MAC_DROP_FRAME_L + 1206 (i << 2))); 1207 1208 value = t4_read_reg(padap, 1209 (A_MPS_STAT_RX_BG_0_MAC_TRUNC_FRAME_L + 1210 (i << 3))); 1211 drop += value; 1212 value = t4_read_reg(padap, 1213 (A_MPS_STAT_RX_BG_0_MAC_TRUNC_FRAME_L + 1214 (i << 3))); 1215 value = t4_read_reg(padap, 1216 (A_MPS_STAT_RX_BG_0_LB_TRUNC_FRAME_L + 1217 (i << 3))); 1218 drop += value; 1219 value = t4_read_reg(padap, 1220 (A_MPS_STAT_RX_BG_0_MAC_TRUNC_FRAME_L + 1221 (i << 3))); 1222 1223 value = t4_read_reg(padap, 1224 (T5_PORT0_REG(A_MPS_PORT_STAT_LB_PORT_DROP_FRAMES) + 1225 (i * T5_PORT_STRIDE))); 1226 drop += value; 1227 } 1228 wtp->xgm_mps.drop = (drop & 0xFF); 1229 1230 /* Add up the MPS errors that should result in dropped packets*/ 1231 err = 0; 1232 for (i = 0; i < 2; i++) { 1233 1234 value = t4_read_reg(padap, 1235 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_MTU_ERROR_L) + 1236 (i * T5_PORT_STRIDE))); 1237 err += value; 1238 value = t4_read_reg(padap, 1239 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_MTU_ERROR_L) + 1240 (i * T5_PORT_STRIDE) + 4)); 1241 1242 value = t4_read_reg(padap, 1243 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_MTU_CRC_ERROR_L) + 1244 (i * T5_PORT_STRIDE))); 1245 err += value; 1246 value = t4_read_reg(padap, 1247 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_MTU_CRC_ERROR_L) + 1248 (i * T5_PORT_STRIDE) + 4)); 1249 1250 value = t4_read_reg(padap, 1251 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_CRC_ERROR_L) + 1252 (i * T5_PORT_STRIDE))); 1253 err += value; 1254 value = t4_read_reg(padap, 1255 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_CRC_ERROR_L) + 1256 (i * T5_PORT_STRIDE) + 4)); 1257 1258 value = t4_read_reg(padap, 1259 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_LEN_ERROR_L) + 1260 (i * T5_PORT_STRIDE))); 1261 err += value; 1262 value = t4_read_reg(padap, 1263 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_LEN_ERROR_L) + 1264 (i * T5_PORT_STRIDE) + 4)); 1265 1266 value = t4_read_reg(padap, 1267 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_SYM_ERROR_L) + 1268 (i * T5_PORT_STRIDE))); 1269 err += value; 1270 value = t4_read_reg(padap, 1271 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_SYM_ERROR_L) + 1272 (i * T5_PORT_STRIDE) + 4)); 1273 1274 value = t4_read_reg(padap, 1275 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_LESS_64B_L) + 1276 (i * T5_PORT_STRIDE))); 1277 err += value; 1278 value = t4_read_reg(padap, 1279 (T5_PORT0_REG(A_MPS_PORT_STAT_RX_PORT_LESS_64B_L) + 1280 (i * T5_PORT_STRIDE) + 4)); 1281 } 1282 wtp->xgm_mps.err = (err & 0xFF); 1283 1284 rc = write_compression_hdr(&scratch_buff, dbg_buff); 1285 1286 if (rc) 1287 goto err1; 1288 1289 rc = compress_buff(&scratch_buff, dbg_buff); 1290 1291 err1: 1292 release_scratch_buff(&scratch_buff, dbg_buff); 1293 err: 1294 return rc; 1295 } 1296 1297 int collect_wtp_data(struct cudbg_init *pdbg_init, 1298 struct cudbg_buffer *dbg_buff, 1299 struct cudbg_error *cudbg_err) 1300 { 1301 struct adapter *padap = pdbg_init->adap; 1302 int rc = -1; 1303 1304 if (is_t5(padap)) 1305 rc = t5_wtp_data(pdbg_init, dbg_buff, cudbg_err); 1306 else if (is_t6(padap)) 1307 rc = t6_wtp_data(pdbg_init, dbg_buff, cudbg_err); 1308 1309 return rc; 1310 } 1311