xref: /linux/drivers/gpu/drm/xe/xe_debugfs.c (revision 44343e8b250abb2f6bfd615493ca07a7f11f3cc2)
1 // SPDX-License-Identifier: MIT
2 /*
3  * Copyright © 2022 Intel Corporation
4  */
5 
6 #include "xe_debugfs.h"
7 
8 #include <linux/debugfs.h>
9 #include <linux/fault-inject.h>
10 #include <linux/string_helpers.h>
11 
12 #include <drm/drm_debugfs.h>
13 
14 #include "regs/xe_pmt.h"
15 #include "xe_bo.h"
16 #include "xe_device.h"
17 #include "xe_force_wake.h"
18 #include "xe_gt_debugfs.h"
19 #include "xe_gt_printk.h"
20 #include "xe_guc_ads.h"
21 #include "xe_mmio.h"
22 #include "xe_pm.h"
23 #include "xe_psmi.h"
24 #include "xe_pxp_debugfs.h"
25 #include "xe_sriov.h"
26 #include "xe_sriov_pf.h"
27 #include "xe_step.h"
28 #include "xe_wa.h"
29 #include "xe_vsec.h"
30 
31 #ifdef CONFIG_DRM_XE_DEBUG
32 #include "xe_bo_evict.h"
33 #include "xe_migrate.h"
34 #include "xe_vm.h"
35 #endif
36 
37 DECLARE_FAULT_ATTR(gt_reset_failure);
38 DECLARE_FAULT_ATTR(inject_csc_hw_error);
39 
40 static void read_residency_counter(struct xe_device *xe, struct xe_mmio *mmio,
41 				   u32 offset, char *name, struct drm_printer *p)
42 {
43 	u64 residency = 0;
44 	int ret;
45 
46 	ret = xe_pmt_telem_read(to_pci_dev(xe->drm.dev),
47 				xe_mmio_read32(mmio, PUNIT_TELEMETRY_GUID),
48 				&residency, offset, sizeof(residency));
49 	if (ret != sizeof(residency)) {
50 		drm_warn(&xe->drm, "%s counter failed to read, ret %d\n", name, ret);
51 		return;
52 	}
53 
54 	drm_printf(p, "%s : %llu\n", name, residency);
55 }
56 
57 static struct xe_device *node_to_xe(struct drm_info_node *node)
58 {
59 	return to_xe_device(node->minor->dev);
60 }
61 
62 static int info(struct seq_file *m, void *data)
63 {
64 	struct xe_device *xe = node_to_xe(m->private);
65 	struct drm_printer p = drm_seq_file_printer(m);
66 	struct xe_gt *gt;
67 	u8 id;
68 
69 	xe_pm_runtime_get(xe);
70 
71 	drm_printf(&p, "graphics_verx100 %d\n", xe->info.graphics_verx100);
72 	drm_printf(&p, "media_verx100 %d\n", xe->info.media_verx100);
73 	drm_printf(&p, "stepping G:%s M:%s B:%s\n",
74 		   xe_step_name(xe->info.step.graphics),
75 		   xe_step_name(xe->info.step.media),
76 		   xe_step_name(xe->info.step.basedie));
77 	drm_printf(&p, "is_dgfx %s\n", str_yes_no(xe->info.is_dgfx));
78 	drm_printf(&p, "platform %d\n", xe->info.platform);
79 	drm_printf(&p, "subplatform %d\n",
80 		   xe->info.subplatform > XE_SUBPLATFORM_NONE ? xe->info.subplatform : 0);
81 	drm_printf(&p, "devid 0x%x\n", xe->info.devid);
82 	drm_printf(&p, "revid %d\n", xe->info.revid);
83 	drm_printf(&p, "tile_count %d\n", xe->info.tile_count);
84 	drm_printf(&p, "vm_max_level %d\n", xe->info.vm_max_level);
85 	drm_printf(&p, "force_execlist %s\n", str_yes_no(xe->info.force_execlist));
86 	drm_printf(&p, "has_flat_ccs %s\n", str_yes_no(xe->info.has_flat_ccs));
87 	drm_printf(&p, "has_usm %s\n", str_yes_no(xe->info.has_usm));
88 	drm_printf(&p, "skip_guc_pc %s\n", str_yes_no(xe->info.skip_guc_pc));
89 	for_each_gt(gt, xe, id) {
90 		drm_printf(&p, "gt%d force wake %d\n", id,
91 			   xe_force_wake_ref(gt_to_fw(gt), XE_FW_GT));
92 		drm_printf(&p, "gt%d engine_mask 0x%llx\n", id,
93 			   gt->info.engine_mask);
94 	}
95 
96 	xe_pm_runtime_put(xe);
97 	return 0;
98 }
99 
100 static int sriov_info(struct seq_file *m, void *data)
101 {
102 	struct xe_device *xe = node_to_xe(m->private);
103 	struct drm_printer p = drm_seq_file_printer(m);
104 
105 	xe_sriov_print_info(xe, &p);
106 	return 0;
107 }
108 
109 static int workarounds(struct xe_device *xe, struct drm_printer *p)
110 {
111 	xe_pm_runtime_get(xe);
112 	xe_wa_device_dump(xe, p);
113 	xe_pm_runtime_put(xe);
114 
115 	return 0;
116 }
117 
118 static int workaround_info(struct seq_file *m, void *data)
119 {
120 	struct xe_device *xe = node_to_xe(m->private);
121 	struct drm_printer p = drm_seq_file_printer(m);
122 
123 	workarounds(xe, &p);
124 	return 0;
125 }
126 
127 static int dgfx_pkg_residencies_show(struct seq_file *m, void *data)
128 {
129 	struct xe_device *xe;
130 	struct xe_mmio *mmio;
131 	struct drm_printer p;
132 
133 	xe = node_to_xe(m->private);
134 	p = drm_seq_file_printer(m);
135 	xe_pm_runtime_get(xe);
136 	mmio = xe_root_tile_mmio(xe);
137 	struct {
138 		u32 offset;
139 		char *name;
140 	} residencies[] = {
141 		{BMG_G2_RESIDENCY_OFFSET, "Package G2"},
142 		{BMG_G6_RESIDENCY_OFFSET, "Package G6"},
143 		{BMG_G8_RESIDENCY_OFFSET, "Package G8"},
144 		{BMG_G10_RESIDENCY_OFFSET, "Package G10"},
145 		{BMG_MODS_RESIDENCY_OFFSET, "Package ModS"}
146 	};
147 
148 	for (int i = 0; i < ARRAY_SIZE(residencies); i++)
149 		read_residency_counter(xe, mmio, residencies[i].offset, residencies[i].name, &p);
150 
151 	xe_pm_runtime_put(xe);
152 	return 0;
153 }
154 
155 static int dgfx_pcie_link_residencies_show(struct seq_file *m, void *data)
156 {
157 	struct xe_device *xe;
158 	struct xe_mmio *mmio;
159 	struct drm_printer p;
160 
161 	xe = node_to_xe(m->private);
162 	p = drm_seq_file_printer(m);
163 	xe_pm_runtime_get(xe);
164 	mmio = xe_root_tile_mmio(xe);
165 
166 	struct {
167 		u32 offset;
168 		char *name;
169 	} residencies[] = {
170 		{BMG_PCIE_LINK_L0_RESIDENCY_OFFSET, "PCIE LINK L0 RESIDENCY"},
171 		{BMG_PCIE_LINK_L1_RESIDENCY_OFFSET, "PCIE LINK L1 RESIDENCY"},
172 		{BMG_PCIE_LINK_L1_2_RESIDENCY_OFFSET, "PCIE LINK L1.2 RESIDENCY"}
173 	};
174 
175 	for (int i = 0; i < ARRAY_SIZE(residencies); i++)
176 		read_residency_counter(xe, mmio, residencies[i].offset, residencies[i].name, &p);
177 
178 	xe_pm_runtime_put(xe);
179 	return 0;
180 }
181 
182 static const struct drm_info_list debugfs_list[] = {
183 	{"info", info, 0},
184 	{ .name = "sriov_info", .show = sriov_info, },
185 	{ .name = "workarounds", .show = workaround_info, },
186 };
187 
188 static const struct drm_info_list debugfs_residencies[] = {
189 	{ .name = "dgfx_pkg_residencies", .show = dgfx_pkg_residencies_show, },
190 	{ .name = "dgfx_pcie_link_residencies", .show = dgfx_pcie_link_residencies_show, },
191 };
192 
193 static int forcewake_open(struct inode *inode, struct file *file)
194 {
195 	struct xe_device *xe = inode->i_private;
196 	struct xe_gt *gt;
197 	u8 id, last_gt;
198 	unsigned int fw_ref;
199 
200 	xe_pm_runtime_get(xe);
201 	for_each_gt(gt, xe, id) {
202 		last_gt = id;
203 
204 		fw_ref = xe_force_wake_get(gt_to_fw(gt), XE_FORCEWAKE_ALL);
205 		if (!xe_force_wake_ref_has_domain(fw_ref, XE_FORCEWAKE_ALL))
206 			goto err_fw_get;
207 	}
208 
209 	return 0;
210 
211 err_fw_get:
212 	for_each_gt(gt, xe, id) {
213 		if (id < last_gt)
214 			xe_force_wake_put(gt_to_fw(gt), XE_FORCEWAKE_ALL);
215 		else if (id == last_gt)
216 			xe_force_wake_put(gt_to_fw(gt), fw_ref);
217 		else
218 			break;
219 	}
220 
221 	xe_pm_runtime_put(xe);
222 	return -ETIMEDOUT;
223 }
224 
225 static int forcewake_release(struct inode *inode, struct file *file)
226 {
227 	struct xe_device *xe = inode->i_private;
228 	struct xe_gt *gt;
229 	u8 id;
230 
231 	for_each_gt(gt, xe, id)
232 		xe_force_wake_put(gt_to_fw(gt), XE_FORCEWAKE_ALL);
233 	xe_pm_runtime_put(xe);
234 
235 	return 0;
236 }
237 
238 static const struct file_operations forcewake_all_fops = {
239 	.owner = THIS_MODULE,
240 	.open = forcewake_open,
241 	.release = forcewake_release,
242 };
243 
244 static ssize_t wedged_mode_show(struct file *f, char __user *ubuf,
245 				size_t size, loff_t *pos)
246 {
247 	struct xe_device *xe = file_inode(f)->i_private;
248 	char buf[32];
249 	int len = 0;
250 
251 	len = scnprintf(buf, sizeof(buf), "%d\n", xe->wedged.mode);
252 
253 	return simple_read_from_buffer(ubuf, size, pos, buf, len);
254 }
255 
256 static ssize_t wedged_mode_set(struct file *f, const char __user *ubuf,
257 			       size_t size, loff_t *pos)
258 {
259 	struct xe_device *xe = file_inode(f)->i_private;
260 	struct xe_gt *gt;
261 	u32 wedged_mode;
262 	ssize_t ret;
263 	u8 id;
264 
265 	ret = kstrtouint_from_user(ubuf, size, 0, &wedged_mode);
266 	if (ret)
267 		return ret;
268 
269 	if (wedged_mode > 2)
270 		return -EINVAL;
271 
272 	if (xe->wedged.mode == wedged_mode)
273 		return size;
274 
275 	xe->wedged.mode = wedged_mode;
276 
277 	xe_pm_runtime_get(xe);
278 	for_each_gt(gt, xe, id) {
279 		ret = xe_guc_ads_scheduler_policy_toggle_reset(&gt->uc.guc.ads);
280 		if (ret) {
281 			xe_gt_err(gt, "Failed to update GuC ADS scheduler policy. GuC may still cause engine reset even with wedged_mode=2\n");
282 			xe_pm_runtime_put(xe);
283 			return -EIO;
284 		}
285 	}
286 	xe_pm_runtime_put(xe);
287 
288 	return size;
289 }
290 
291 static const struct file_operations wedged_mode_fops = {
292 	.owner = THIS_MODULE,
293 	.read = wedged_mode_show,
294 	.write = wedged_mode_set,
295 };
296 
297 static ssize_t atomic_svm_timeslice_ms_show(struct file *f, char __user *ubuf,
298 					    size_t size, loff_t *pos)
299 {
300 	struct xe_device *xe = file_inode(f)->i_private;
301 	char buf[32];
302 	int len = 0;
303 
304 	len = scnprintf(buf, sizeof(buf), "%d\n", xe->atomic_svm_timeslice_ms);
305 
306 	return simple_read_from_buffer(ubuf, size, pos, buf, len);
307 }
308 
309 static ssize_t atomic_svm_timeslice_ms_set(struct file *f,
310 					   const char __user *ubuf,
311 					   size_t size, loff_t *pos)
312 {
313 	struct xe_device *xe = file_inode(f)->i_private;
314 	u32 atomic_svm_timeslice_ms;
315 	ssize_t ret;
316 
317 	ret = kstrtouint_from_user(ubuf, size, 0, &atomic_svm_timeslice_ms);
318 	if (ret)
319 		return ret;
320 
321 	xe->atomic_svm_timeslice_ms = atomic_svm_timeslice_ms;
322 
323 	return size;
324 }
325 
326 static const struct file_operations atomic_svm_timeslice_ms_fops = {
327 	.owner = THIS_MODULE,
328 	.read = atomic_svm_timeslice_ms_show,
329 	.write = atomic_svm_timeslice_ms_set,
330 };
331 
332 static void create_tile_debugfs(struct xe_tile *tile, struct dentry *root)
333 {
334 	char name[8];
335 
336 	snprintf(name, sizeof(name), "tile%u", tile->id);
337 	tile->debugfs = debugfs_create_dir(name, root);
338 	if (IS_ERR(tile->debugfs))
339 		return;
340 
341 	/*
342 	 * Store the xe_tile pointer as private data of the tile/ directory
343 	 * node so other tile specific attributes under that directory may
344 	 * refer to it by looking at its parent node private data.
345 	 */
346 	tile->debugfs->d_inode->i_private = tile;
347 }
348 
349 void xe_debugfs_register(struct xe_device *xe)
350 {
351 	struct ttm_device *bdev = &xe->ttm;
352 	struct drm_minor *minor = xe->drm.primary;
353 	struct dentry *root = minor->debugfs_root;
354 	struct ttm_resource_manager *man;
355 	struct xe_tile *tile;
356 	struct xe_gt *gt;
357 	u32 mem_type;
358 	u8 tile_id;
359 	u8 id;
360 
361 	drm_debugfs_create_files(debugfs_list,
362 				 ARRAY_SIZE(debugfs_list),
363 				 root, minor);
364 
365 	if (xe->info.platform == XE_BATTLEMAGE) {
366 		drm_debugfs_create_files(debugfs_residencies,
367 					 ARRAY_SIZE(debugfs_residencies),
368 					 root, minor);
369 		fault_create_debugfs_attr("inject_csc_hw_error", root,
370 					  &inject_csc_hw_error);
371 	}
372 
373 	debugfs_create_file("forcewake_all", 0400, root, xe,
374 			    &forcewake_all_fops);
375 
376 	debugfs_create_file("wedged_mode", 0600, root, xe,
377 			    &wedged_mode_fops);
378 
379 	debugfs_create_file("atomic_svm_timeslice_ms", 0600, root, xe,
380 			    &atomic_svm_timeslice_ms_fops);
381 
382 	for (mem_type = XE_PL_VRAM0; mem_type <= XE_PL_VRAM1; ++mem_type) {
383 		man = ttm_manager_type(bdev, mem_type);
384 
385 		if (man) {
386 			char name[16];
387 
388 			snprintf(name, sizeof(name), "vram%d_mm", mem_type - XE_PL_VRAM0);
389 			ttm_resource_manager_create_debugfs(man, root, name);
390 		}
391 	}
392 
393 	man = ttm_manager_type(bdev, XE_PL_TT);
394 	ttm_resource_manager_create_debugfs(man, root, "gtt_mm");
395 
396 	man = ttm_manager_type(bdev, XE_PL_STOLEN);
397 	if (man)
398 		ttm_resource_manager_create_debugfs(man, root, "stolen_mm");
399 
400 	for_each_tile(tile, xe, tile_id)
401 		create_tile_debugfs(tile, root);
402 
403 	for_each_gt(gt, xe, id)
404 		xe_gt_debugfs_register(gt);
405 
406 	xe_pxp_debugfs_register(xe->pxp);
407 
408 	xe_psmi_debugfs_register(xe);
409 
410 	fault_create_debugfs_attr("fail_gt_reset", root, &gt_reset_failure);
411 
412 	if (IS_SRIOV_PF(xe))
413 		xe_sriov_pf_debugfs_register(xe, root);
414 }
415