xref: /linux/drivers/gpu/drm/xe/xe_debugfs.c (revision 68a052239fc4b351e961f698b824f7654a346091)
1 // SPDX-License-Identifier: MIT
2 /*
3  * Copyright © 2022 Intel Corporation
4  */
5 
6 #include "xe_debugfs.h"
7 
8 #include <linux/debugfs.h>
9 #include <linux/fault-inject.h>
10 #include <linux/string_helpers.h>
11 
12 #include <drm/drm_debugfs.h>
13 
14 #include "regs/xe_pmt.h"
15 #include "xe_bo.h"
16 #include "xe_device.h"
17 #include "xe_force_wake.h"
18 #include "xe_gt_debugfs.h"
19 #include "xe_gt_printk.h"
20 #include "xe_guc_ads.h"
21 #include "xe_mmio.h"
22 #include "xe_pm.h"
23 #include "xe_psmi.h"
24 #include "xe_pxp_debugfs.h"
25 #include "xe_sriov.h"
26 #include "xe_sriov_pf.h"
27 #include "xe_sriov_vf.h"
28 #include "xe_step.h"
29 #include "xe_tile_debugfs.h"
30 #include "xe_wa.h"
31 #include "xe_vsec.h"
32 
33 #ifdef CONFIG_DRM_XE_DEBUG
34 #include "xe_bo_evict.h"
35 #include "xe_migrate.h"
36 #include "xe_vm.h"
37 #endif
38 
39 DECLARE_FAULT_ATTR(gt_reset_failure);
40 DECLARE_FAULT_ATTR(inject_csc_hw_error);
41 
42 static void read_residency_counter(struct xe_device *xe, struct xe_mmio *mmio,
43 				   u32 offset, const char *name, struct drm_printer *p)
44 {
45 	u64 residency = 0;
46 	int ret;
47 
48 	ret = xe_pmt_telem_read(to_pci_dev(xe->drm.dev),
49 				xe_mmio_read32(mmio, PUNIT_TELEMETRY_GUID),
50 				&residency, offset, sizeof(residency));
51 	if (ret != sizeof(residency)) {
52 		drm_warn(&xe->drm, "%s counter failed to read, ret %d\n", name, ret);
53 		return;
54 	}
55 
56 	drm_printf(p, "%s : %llu\n", name, residency);
57 }
58 
59 static struct xe_device *node_to_xe(struct drm_info_node *node)
60 {
61 	return to_xe_device(node->minor->dev);
62 }
63 
64 static int info(struct seq_file *m, void *data)
65 {
66 	struct xe_device *xe = node_to_xe(m->private);
67 	struct drm_printer p = drm_seq_file_printer(m);
68 	struct xe_gt *gt;
69 	u8 id;
70 
71 	xe_pm_runtime_get(xe);
72 
73 	drm_printf(&p, "graphics_verx100 %d\n", xe->info.graphics_verx100);
74 	drm_printf(&p, "media_verx100 %d\n", xe->info.media_verx100);
75 	drm_printf(&p, "stepping G:%s M:%s B:%s\n",
76 		   xe_step_name(xe->info.step.graphics),
77 		   xe_step_name(xe->info.step.media),
78 		   xe_step_name(xe->info.step.basedie));
79 	drm_printf(&p, "is_dgfx %s\n", str_yes_no(xe->info.is_dgfx));
80 	drm_printf(&p, "platform %d\n", xe->info.platform);
81 	drm_printf(&p, "subplatform %d\n",
82 		   xe->info.subplatform > XE_SUBPLATFORM_NONE ? xe->info.subplatform : 0);
83 	drm_printf(&p, "devid 0x%x\n", xe->info.devid);
84 	drm_printf(&p, "revid %d\n", xe->info.revid);
85 	drm_printf(&p, "tile_count %d\n", xe->info.tile_count);
86 	drm_printf(&p, "vm_max_level %d\n", xe->info.vm_max_level);
87 	drm_printf(&p, "force_execlist %s\n", str_yes_no(xe->info.force_execlist));
88 	drm_printf(&p, "has_flat_ccs %s\n", str_yes_no(xe->info.has_flat_ccs));
89 	drm_printf(&p, "has_usm %s\n", str_yes_no(xe->info.has_usm));
90 	drm_printf(&p, "skip_guc_pc %s\n", str_yes_no(xe->info.skip_guc_pc));
91 	for_each_gt(gt, xe, id) {
92 		drm_printf(&p, "gt%d force wake %d\n", id,
93 			   xe_force_wake_ref(gt_to_fw(gt), XE_FW_GT));
94 		drm_printf(&p, "gt%d engine_mask 0x%llx\n", id,
95 			   gt->info.engine_mask);
96 	}
97 
98 	xe_pm_runtime_put(xe);
99 	return 0;
100 }
101 
102 static int sriov_info(struct seq_file *m, void *data)
103 {
104 	struct xe_device *xe = node_to_xe(m->private);
105 	struct drm_printer p = drm_seq_file_printer(m);
106 
107 	xe_sriov_print_info(xe, &p);
108 	return 0;
109 }
110 
111 static int workarounds(struct xe_device *xe, struct drm_printer *p)
112 {
113 	xe_pm_runtime_get(xe);
114 	xe_wa_device_dump(xe, p);
115 	xe_pm_runtime_put(xe);
116 
117 	return 0;
118 }
119 
120 static int workaround_info(struct seq_file *m, void *data)
121 {
122 	struct xe_device *xe = node_to_xe(m->private);
123 	struct drm_printer p = drm_seq_file_printer(m);
124 
125 	workarounds(xe, &p);
126 	return 0;
127 }
128 
129 static int dgfx_pkg_residencies_show(struct seq_file *m, void *data)
130 {
131 	struct xe_device *xe;
132 	struct xe_mmio *mmio;
133 	struct drm_printer p;
134 
135 	xe = node_to_xe(m->private);
136 	p = drm_seq_file_printer(m);
137 	xe_pm_runtime_get(xe);
138 	mmio = xe_root_tile_mmio(xe);
139 	static const struct {
140 		u32 offset;
141 		const char *name;
142 	} residencies[] = {
143 		{BMG_G2_RESIDENCY_OFFSET, "Package G2"},
144 		{BMG_G6_RESIDENCY_OFFSET, "Package G6"},
145 		{BMG_G8_RESIDENCY_OFFSET, "Package G8"},
146 		{BMG_G10_RESIDENCY_OFFSET, "Package G10"},
147 		{BMG_MODS_RESIDENCY_OFFSET, "Package ModS"}
148 	};
149 
150 	for (int i = 0; i < ARRAY_SIZE(residencies); i++)
151 		read_residency_counter(xe, mmio, residencies[i].offset, residencies[i].name, &p);
152 
153 	xe_pm_runtime_put(xe);
154 	return 0;
155 }
156 
157 static int dgfx_pcie_link_residencies_show(struct seq_file *m, void *data)
158 {
159 	struct xe_device *xe;
160 	struct xe_mmio *mmio;
161 	struct drm_printer p;
162 
163 	xe = node_to_xe(m->private);
164 	p = drm_seq_file_printer(m);
165 	xe_pm_runtime_get(xe);
166 	mmio = xe_root_tile_mmio(xe);
167 
168 	static const struct {
169 		u32 offset;
170 		const char *name;
171 	} residencies[] = {
172 		{BMG_PCIE_LINK_L0_RESIDENCY_OFFSET, "PCIE LINK L0 RESIDENCY"},
173 		{BMG_PCIE_LINK_L1_RESIDENCY_OFFSET, "PCIE LINK L1 RESIDENCY"},
174 		{BMG_PCIE_LINK_L1_2_RESIDENCY_OFFSET, "PCIE LINK L1.2 RESIDENCY"}
175 	};
176 
177 	for (int i = 0; i < ARRAY_SIZE(residencies); i++)
178 		read_residency_counter(xe, mmio, residencies[i].offset, residencies[i].name, &p);
179 
180 	xe_pm_runtime_put(xe);
181 	return 0;
182 }
183 
184 static const struct drm_info_list debugfs_list[] = {
185 	{"info", info, 0},
186 	{ .name = "sriov_info", .show = sriov_info, },
187 	{ .name = "workarounds", .show = workaround_info, },
188 };
189 
190 static const struct drm_info_list debugfs_residencies[] = {
191 	{ .name = "dgfx_pkg_residencies", .show = dgfx_pkg_residencies_show, },
192 	{ .name = "dgfx_pcie_link_residencies", .show = dgfx_pcie_link_residencies_show, },
193 };
194 
195 static int forcewake_open(struct inode *inode, struct file *file)
196 {
197 	struct xe_device *xe = inode->i_private;
198 	struct xe_gt *gt;
199 	u8 id, last_gt;
200 	unsigned int fw_ref;
201 
202 	xe_pm_runtime_get(xe);
203 	for_each_gt(gt, xe, id) {
204 		last_gt = id;
205 
206 		fw_ref = xe_force_wake_get(gt_to_fw(gt), XE_FORCEWAKE_ALL);
207 		if (!xe_force_wake_ref_has_domain(fw_ref, XE_FORCEWAKE_ALL))
208 			goto err_fw_get;
209 	}
210 
211 	return 0;
212 
213 err_fw_get:
214 	for_each_gt(gt, xe, id) {
215 		if (id < last_gt)
216 			xe_force_wake_put(gt_to_fw(gt), XE_FORCEWAKE_ALL);
217 		else if (id == last_gt)
218 			xe_force_wake_put(gt_to_fw(gt), fw_ref);
219 		else
220 			break;
221 	}
222 
223 	xe_pm_runtime_put(xe);
224 	return -ETIMEDOUT;
225 }
226 
227 static int forcewake_release(struct inode *inode, struct file *file)
228 {
229 	struct xe_device *xe = inode->i_private;
230 	struct xe_gt *gt;
231 	u8 id;
232 
233 	for_each_gt(gt, xe, id)
234 		xe_force_wake_put(gt_to_fw(gt), XE_FORCEWAKE_ALL);
235 	xe_pm_runtime_put(xe);
236 
237 	return 0;
238 }
239 
240 static const struct file_operations forcewake_all_fops = {
241 	.owner = THIS_MODULE,
242 	.open = forcewake_open,
243 	.release = forcewake_release,
244 };
245 
246 static ssize_t wedged_mode_show(struct file *f, char __user *ubuf,
247 				size_t size, loff_t *pos)
248 {
249 	struct xe_device *xe = file_inode(f)->i_private;
250 	char buf[32];
251 	int len = 0;
252 
253 	len = scnprintf(buf, sizeof(buf), "%d\n", xe->wedged.mode);
254 
255 	return simple_read_from_buffer(ubuf, size, pos, buf, len);
256 }
257 
258 static ssize_t wedged_mode_set(struct file *f, const char __user *ubuf,
259 			       size_t size, loff_t *pos)
260 {
261 	struct xe_device *xe = file_inode(f)->i_private;
262 	struct xe_gt *gt;
263 	u32 wedged_mode;
264 	ssize_t ret;
265 	u8 id;
266 
267 	ret = kstrtouint_from_user(ubuf, size, 0, &wedged_mode);
268 	if (ret)
269 		return ret;
270 
271 	if (wedged_mode > 2)
272 		return -EINVAL;
273 
274 	if (xe->wedged.mode == wedged_mode)
275 		return size;
276 
277 	xe->wedged.mode = wedged_mode;
278 
279 	xe_pm_runtime_get(xe);
280 	for_each_gt(gt, xe, id) {
281 		ret = xe_guc_ads_scheduler_policy_toggle_reset(&gt->uc.guc.ads);
282 		if (ret) {
283 			xe_gt_err(gt, "Failed to update GuC ADS scheduler policy. GuC may still cause engine reset even with wedged_mode=2\n");
284 			xe_pm_runtime_put(xe);
285 			return -EIO;
286 		}
287 	}
288 	xe_pm_runtime_put(xe);
289 
290 	return size;
291 }
292 
293 static const struct file_operations wedged_mode_fops = {
294 	.owner = THIS_MODULE,
295 	.read = wedged_mode_show,
296 	.write = wedged_mode_set,
297 };
298 
299 static ssize_t atomic_svm_timeslice_ms_show(struct file *f, char __user *ubuf,
300 					    size_t size, loff_t *pos)
301 {
302 	struct xe_device *xe = file_inode(f)->i_private;
303 	char buf[32];
304 	int len = 0;
305 
306 	len = scnprintf(buf, sizeof(buf), "%d\n", xe->atomic_svm_timeslice_ms);
307 
308 	return simple_read_from_buffer(ubuf, size, pos, buf, len);
309 }
310 
311 static ssize_t atomic_svm_timeslice_ms_set(struct file *f,
312 					   const char __user *ubuf,
313 					   size_t size, loff_t *pos)
314 {
315 	struct xe_device *xe = file_inode(f)->i_private;
316 	u32 atomic_svm_timeslice_ms;
317 	ssize_t ret;
318 
319 	ret = kstrtouint_from_user(ubuf, size, 0, &atomic_svm_timeslice_ms);
320 	if (ret)
321 		return ret;
322 
323 	xe->atomic_svm_timeslice_ms = atomic_svm_timeslice_ms;
324 
325 	return size;
326 }
327 
328 static const struct file_operations atomic_svm_timeslice_ms_fops = {
329 	.owner = THIS_MODULE,
330 	.read = atomic_svm_timeslice_ms_show,
331 	.write = atomic_svm_timeslice_ms_set,
332 };
333 
334 static ssize_t disable_late_binding_show(struct file *f, char __user *ubuf,
335 					 size_t size, loff_t *pos)
336 {
337 	struct xe_device *xe = file_inode(f)->i_private;
338 	struct xe_late_bind *late_bind = &xe->late_bind;
339 	char buf[32];
340 	int len;
341 
342 	len = scnprintf(buf, sizeof(buf), "%d\n", late_bind->disable);
343 
344 	return simple_read_from_buffer(ubuf, size, pos, buf, len);
345 }
346 
347 static ssize_t disable_late_binding_set(struct file *f, const char __user *ubuf,
348 					size_t size, loff_t *pos)
349 {
350 	struct xe_device *xe = file_inode(f)->i_private;
351 	struct xe_late_bind *late_bind = &xe->late_bind;
352 	u32 uval;
353 	ssize_t ret;
354 
355 	ret = kstrtouint_from_user(ubuf, size, sizeof(uval), &uval);
356 	if (ret)
357 		return ret;
358 
359 	if (uval > 1)
360 		return -EINVAL;
361 
362 	late_bind->disable = !!uval;
363 	return size;
364 }
365 
366 static const struct file_operations disable_late_binding_fops = {
367 	.owner = THIS_MODULE,
368 	.read = disable_late_binding_show,
369 	.write = disable_late_binding_set,
370 };
371 
372 void xe_debugfs_register(struct xe_device *xe)
373 {
374 	struct ttm_device *bdev = &xe->ttm;
375 	struct drm_minor *minor = xe->drm.primary;
376 	struct dentry *root = minor->debugfs_root;
377 	struct ttm_resource_manager *man;
378 	struct xe_tile *tile;
379 	struct xe_gt *gt;
380 	u32 mem_type;
381 	u8 tile_id;
382 	u8 id;
383 
384 	drm_debugfs_create_files(debugfs_list,
385 				 ARRAY_SIZE(debugfs_list),
386 				 root, minor);
387 
388 	if (xe->info.platform == XE_BATTLEMAGE && !IS_SRIOV_VF(xe)) {
389 		drm_debugfs_create_files(debugfs_residencies,
390 					 ARRAY_SIZE(debugfs_residencies),
391 					 root, minor);
392 		fault_create_debugfs_attr("inject_csc_hw_error", root,
393 					  &inject_csc_hw_error);
394 	}
395 
396 	debugfs_create_file("forcewake_all", 0400, root, xe,
397 			    &forcewake_all_fops);
398 
399 	debugfs_create_file("wedged_mode", 0600, root, xe,
400 			    &wedged_mode_fops);
401 
402 	debugfs_create_file("atomic_svm_timeslice_ms", 0600, root, xe,
403 			    &atomic_svm_timeslice_ms_fops);
404 
405 	debugfs_create_file("disable_late_binding", 0600, root, xe,
406 			    &disable_late_binding_fops);
407 
408 	for (mem_type = XE_PL_VRAM0; mem_type <= XE_PL_VRAM1; ++mem_type) {
409 		man = ttm_manager_type(bdev, mem_type);
410 
411 		if (man) {
412 			char name[16];
413 
414 			snprintf(name, sizeof(name), "vram%d_mm", mem_type - XE_PL_VRAM0);
415 			ttm_resource_manager_create_debugfs(man, root, name);
416 		}
417 	}
418 
419 	man = ttm_manager_type(bdev, XE_PL_TT);
420 	ttm_resource_manager_create_debugfs(man, root, "gtt_mm");
421 
422 	man = ttm_manager_type(bdev, XE_PL_STOLEN);
423 	if (man)
424 		ttm_resource_manager_create_debugfs(man, root, "stolen_mm");
425 
426 	for_each_tile(tile, xe, tile_id)
427 		xe_tile_debugfs_register(tile);
428 
429 	for_each_gt(gt, xe, id)
430 		xe_gt_debugfs_register(gt);
431 
432 	xe_pxp_debugfs_register(xe->pxp);
433 
434 	xe_psmi_debugfs_register(xe);
435 
436 	fault_create_debugfs_attr("fail_gt_reset", root, &gt_reset_failure);
437 
438 	if (IS_SRIOV_PF(xe))
439 		xe_sriov_pf_debugfs_register(xe, root);
440 	else if (IS_SRIOV_VF(xe))
441 		xe_sriov_vf_debugfs_register(xe, root);
442 }
443