xref: /linux/drivers/net/ethernet/chelsio/cxgb4/cxgb4_cudbg.c (revision 26fbb4c8c7c3ee9a4c3b4de555a8587b5a19154e)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  *  Copyright (C) 2017 Chelsio Communications.  All rights reserved.
4  */
5 
6 #include "t4_regs.h"
7 #include "cxgb4.h"
8 #include "cxgb4_cudbg.h"
9 #include "cudbg_zlib.h"
10 
11 static const struct cxgb4_collect_entity cxgb4_collect_mem_dump[] = {
12 	{ CUDBG_EDC0, cudbg_collect_edc0_meminfo },
13 	{ CUDBG_EDC1, cudbg_collect_edc1_meminfo },
14 	{ CUDBG_MC0, cudbg_collect_mc0_meminfo },
15 	{ CUDBG_MC1, cudbg_collect_mc1_meminfo },
16 	{ CUDBG_HMA, cudbg_collect_hma_meminfo },
17 };
18 
19 static const struct cxgb4_collect_entity cxgb4_collect_hw_dump[] = {
20 	{ CUDBG_MBOX_LOG, cudbg_collect_mbox_log },
21 	{ CUDBG_QDESC, cudbg_collect_qdesc },
22 	{ CUDBG_DEV_LOG, cudbg_collect_fw_devlog },
23 	{ CUDBG_REG_DUMP, cudbg_collect_reg_dump },
24 	{ CUDBG_CIM_LA, cudbg_collect_cim_la },
25 	{ CUDBG_CIM_MA_LA, cudbg_collect_cim_ma_la },
26 	{ CUDBG_CIM_QCFG, cudbg_collect_cim_qcfg },
27 	{ CUDBG_CIM_IBQ_TP0, cudbg_collect_cim_ibq_tp0 },
28 	{ CUDBG_CIM_IBQ_TP1, cudbg_collect_cim_ibq_tp1 },
29 	{ CUDBG_CIM_IBQ_ULP, cudbg_collect_cim_ibq_ulp },
30 	{ CUDBG_CIM_IBQ_SGE0, cudbg_collect_cim_ibq_sge0 },
31 	{ CUDBG_CIM_IBQ_SGE1, cudbg_collect_cim_ibq_sge1 },
32 	{ CUDBG_CIM_IBQ_NCSI, cudbg_collect_cim_ibq_ncsi },
33 	{ CUDBG_CIM_OBQ_ULP0, cudbg_collect_cim_obq_ulp0 },
34 	{ CUDBG_CIM_OBQ_ULP1, cudbg_collect_cim_obq_ulp1 },
35 	{ CUDBG_CIM_OBQ_ULP2, cudbg_collect_cim_obq_ulp2 },
36 	{ CUDBG_CIM_OBQ_ULP3, cudbg_collect_cim_obq_ulp3 },
37 	{ CUDBG_CIM_OBQ_SGE, cudbg_collect_cim_obq_sge },
38 	{ CUDBG_CIM_OBQ_NCSI, cudbg_collect_cim_obq_ncsi },
39 	{ CUDBG_RSS, cudbg_collect_rss },
40 	{ CUDBG_RSS_VF_CONF, cudbg_collect_rss_vf_config },
41 	{ CUDBG_PATH_MTU, cudbg_collect_path_mtu },
42 	{ CUDBG_PM_STATS, cudbg_collect_pm_stats },
43 	{ CUDBG_HW_SCHED, cudbg_collect_hw_sched },
44 	{ CUDBG_TP_INDIRECT, cudbg_collect_tp_indirect },
45 	{ CUDBG_SGE_INDIRECT, cudbg_collect_sge_indirect },
46 	{ CUDBG_ULPRX_LA, cudbg_collect_ulprx_la },
47 	{ CUDBG_TP_LA, cudbg_collect_tp_la },
48 	{ CUDBG_MEMINFO, cudbg_collect_meminfo },
49 	{ CUDBG_CIM_PIF_LA, cudbg_collect_cim_pif_la },
50 	{ CUDBG_CLK, cudbg_collect_clk_info },
51 	{ CUDBG_CIM_OBQ_RXQ0, cudbg_collect_obq_sge_rx_q0 },
52 	{ CUDBG_CIM_OBQ_RXQ1, cudbg_collect_obq_sge_rx_q1 },
53 	{ CUDBG_PCIE_INDIRECT, cudbg_collect_pcie_indirect },
54 	{ CUDBG_PM_INDIRECT, cudbg_collect_pm_indirect },
55 	{ CUDBG_TID_INFO, cudbg_collect_tid },
56 	{ CUDBG_PCIE_CONFIG, cudbg_collect_pcie_config },
57 	{ CUDBG_DUMP_CONTEXT, cudbg_collect_dump_context },
58 	{ CUDBG_MPS_TCAM, cudbg_collect_mps_tcam },
59 	{ CUDBG_VPD_DATA, cudbg_collect_vpd_data },
60 	{ CUDBG_LE_TCAM, cudbg_collect_le_tcam },
61 	{ CUDBG_CCTRL, cudbg_collect_cctrl },
62 	{ CUDBG_MA_INDIRECT, cudbg_collect_ma_indirect },
63 	{ CUDBG_ULPTX_LA, cudbg_collect_ulptx_la },
64 	{ CUDBG_UP_CIM_INDIRECT, cudbg_collect_up_cim_indirect },
65 	{ CUDBG_PBT_TABLE, cudbg_collect_pbt_tables },
66 	{ CUDBG_HMA_INDIRECT, cudbg_collect_hma_indirect },
67 };
68 
69 static const struct cxgb4_collect_entity cxgb4_collect_flash_dump[] = {
70 	{ CUDBG_FLASH, cudbg_collect_flash },
71 };
72 
73 u32 cxgb4_get_dump_length(struct adapter *adap, u32 flag)
74 {
75 	u32 i, entity;
76 	u32 len = 0;
77 	u32 wsize;
78 
79 	if (flag & CXGB4_ETH_DUMP_HW) {
80 		for (i = 0; i < ARRAY_SIZE(cxgb4_collect_hw_dump); i++) {
81 			entity = cxgb4_collect_hw_dump[i].entity;
82 			len += cudbg_get_entity_length(adap, entity);
83 		}
84 	}
85 
86 	if (flag & CXGB4_ETH_DUMP_MEM) {
87 		for (i = 0; i < ARRAY_SIZE(cxgb4_collect_mem_dump); i++) {
88 			entity = cxgb4_collect_mem_dump[i].entity;
89 			len += cudbg_get_entity_length(adap, entity);
90 		}
91 	}
92 
93 	if (flag & CXGB4_ETH_DUMP_FLASH)
94 		len += adap->params.sf_size;
95 
96 	/* If compression is enabled, a smaller destination buffer is enough */
97 	wsize = cudbg_get_workspace_size();
98 	if (wsize && len > CUDBG_DUMP_BUFF_SIZE)
99 		len = CUDBG_DUMP_BUFF_SIZE;
100 
101 	return len;
102 }
103 
104 static void cxgb4_cudbg_collect_entity(struct cudbg_init *pdbg_init,
105 				       struct cudbg_buffer *dbg_buff,
106 				       const struct cxgb4_collect_entity *e_arr,
107 				       u32 arr_size, void *buf, u32 *tot_size)
108 {
109 	struct cudbg_error cudbg_err = { 0 };
110 	struct cudbg_entity_hdr *entity_hdr;
111 	u32 i, total_size = 0;
112 	int ret;
113 
114 	for (i = 0; i < arr_size; i++) {
115 		const struct cxgb4_collect_entity *e = &e_arr[i];
116 
117 		entity_hdr = cudbg_get_entity_hdr(buf, e->entity);
118 		entity_hdr->entity_type = e->entity;
119 		entity_hdr->start_offset = dbg_buff->offset;
120 		memset(&cudbg_err, 0, sizeof(struct cudbg_error));
121 		ret = e->collect_cb(pdbg_init, dbg_buff, &cudbg_err);
122 		if (ret) {
123 			entity_hdr->size = 0;
124 			dbg_buff->offset = entity_hdr->start_offset;
125 		} else {
126 			cudbg_align_debug_buffer(dbg_buff, entity_hdr);
127 		}
128 
129 		/* Log error and continue with next entity */
130 		if (cudbg_err.sys_err)
131 			ret = CUDBG_SYSTEM_ERROR;
132 
133 		entity_hdr->hdr_flags = ret;
134 		entity_hdr->sys_err = cudbg_err.sys_err;
135 		entity_hdr->sys_warn = cudbg_err.sys_warn;
136 		total_size += entity_hdr->size;
137 	}
138 
139 	*tot_size += total_size;
140 }
141 
142 static int cudbg_alloc_compress_buff(struct cudbg_init *pdbg_init)
143 {
144 	u32 workspace_size;
145 
146 	workspace_size = cudbg_get_workspace_size();
147 	pdbg_init->compress_buff = vzalloc(CUDBG_COMPRESS_BUFF_SIZE +
148 					   workspace_size);
149 	if (!pdbg_init->compress_buff)
150 		return -ENOMEM;
151 
152 	pdbg_init->compress_buff_size = CUDBG_COMPRESS_BUFF_SIZE;
153 	pdbg_init->workspace = (u8 *)pdbg_init->compress_buff +
154 			       CUDBG_COMPRESS_BUFF_SIZE - workspace_size;
155 	return 0;
156 }
157 
158 static void cudbg_free_compress_buff(struct cudbg_init *pdbg_init)
159 {
160 	if (pdbg_init->compress_buff)
161 		vfree(pdbg_init->compress_buff);
162 }
163 
164 int cxgb4_cudbg_collect(struct adapter *adap, void *buf, u32 *buf_size,
165 			u32 flag)
166 {
167 	struct cudbg_buffer dbg_buff = { 0 };
168 	u32 size, min_size, total_size = 0;
169 	struct cudbg_init cudbg_init;
170 	struct cudbg_hdr *cudbg_hdr;
171 	int rc;
172 
173 	size = *buf_size;
174 
175 	memset(&cudbg_init, 0, sizeof(struct cudbg_init));
176 	cudbg_init.adap = adap;
177 	cudbg_init.outbuf = buf;
178 	cudbg_init.outbuf_size = size;
179 
180 	dbg_buff.data = buf;
181 	dbg_buff.size = size;
182 	dbg_buff.offset = 0;
183 
184 	cudbg_hdr = (struct cudbg_hdr *)buf;
185 	cudbg_hdr->signature = CUDBG_SIGNATURE;
186 	cudbg_hdr->hdr_len = sizeof(struct cudbg_hdr);
187 	cudbg_hdr->major_ver = CUDBG_MAJOR_VERSION;
188 	cudbg_hdr->minor_ver = CUDBG_MINOR_VERSION;
189 	cudbg_hdr->max_entities = CUDBG_MAX_ENTITY;
190 	cudbg_hdr->chip_ver = adap->params.chip;
191 	cudbg_hdr->dump_type = CUDBG_DUMP_TYPE_MINI;
192 
193 	min_size = sizeof(struct cudbg_hdr) +
194 		   sizeof(struct cudbg_entity_hdr) *
195 		   cudbg_hdr->max_entities;
196 	if (size < min_size)
197 		return -ENOMEM;
198 
199 	rc = cudbg_get_workspace_size();
200 	if (rc) {
201 		/* Zlib available.  So, use zlib deflate */
202 		cudbg_init.compress_type = CUDBG_COMPRESSION_ZLIB;
203 		rc = cudbg_alloc_compress_buff(&cudbg_init);
204 		if (rc) {
205 			/* Ignore error and continue without compression. */
206 			dev_warn(adap->pdev_dev,
207 				 "Fail allocating compression buffer ret: %d.  Continuing without compression.\n",
208 				 rc);
209 			cudbg_init.compress_type = CUDBG_COMPRESSION_NONE;
210 			rc = 0;
211 		}
212 	} else {
213 		cudbg_init.compress_type = CUDBG_COMPRESSION_NONE;
214 	}
215 
216 	cudbg_hdr->compress_type = cudbg_init.compress_type;
217 	dbg_buff.offset += min_size;
218 	total_size = dbg_buff.offset;
219 
220 	if (flag & CXGB4_ETH_DUMP_HW)
221 		cxgb4_cudbg_collect_entity(&cudbg_init, &dbg_buff,
222 					   cxgb4_collect_hw_dump,
223 					   ARRAY_SIZE(cxgb4_collect_hw_dump),
224 					   buf,
225 					   &total_size);
226 
227 	if (flag & CXGB4_ETH_DUMP_MEM)
228 		cxgb4_cudbg_collect_entity(&cudbg_init, &dbg_buff,
229 					   cxgb4_collect_mem_dump,
230 					   ARRAY_SIZE(cxgb4_collect_mem_dump),
231 					   buf,
232 					   &total_size);
233 
234 	if (flag & CXGB4_ETH_DUMP_FLASH)
235 		cxgb4_cudbg_collect_entity(&cudbg_init, &dbg_buff,
236 					   cxgb4_collect_flash_dump,
237 					   ARRAY_SIZE(cxgb4_collect_flash_dump),
238 					   buf,
239 					   &total_size);
240 
241 	cudbg_free_compress_buff(&cudbg_init);
242 	cudbg_hdr->data_len = total_size;
243 	if (cudbg_init.compress_type != CUDBG_COMPRESSION_NONE)
244 		*buf_size = size;
245 	else
246 		*buf_size = total_size;
247 	return 0;
248 }
249 
250 void cxgb4_init_ethtool_dump(struct adapter *adapter)
251 {
252 	adapter->eth_dump.flag = CXGB4_ETH_DUMP_NONE;
253 	adapter->eth_dump.version = adapter->params.fw_vers;
254 	adapter->eth_dump.len = 0;
255 }
256 
257 static int cxgb4_cudbg_vmcoredd_collect(struct vmcoredd_data *data, void *buf)
258 {
259 	struct adapter *adap = container_of(data, struct adapter, vmcoredd);
260 	u32 len = data->size;
261 
262 	return cxgb4_cudbg_collect(adap, buf, &len, CXGB4_ETH_DUMP_ALL);
263 }
264 
265 int cxgb4_cudbg_vmcore_add_dump(struct adapter *adap)
266 {
267 	struct vmcoredd_data *data = &adap->vmcoredd;
268 	u32 len;
269 
270 	len = sizeof(struct cudbg_hdr) +
271 	      sizeof(struct cudbg_entity_hdr) * CUDBG_MAX_ENTITY;
272 	len += CUDBG_DUMP_BUFF_SIZE;
273 
274 	data->size = len;
275 	snprintf(data->dump_name, sizeof(data->dump_name), "%s_%s",
276 		 cxgb4_driver_name, adap->name);
277 	data->vmcoredd_callback = cxgb4_cudbg_vmcoredd_collect;
278 
279 	return vmcore_add_device_dump(data);
280 }
281