1 // SPDX-License-Identifier: MIT
2 /*
3 * Copyright © 2022 Intel Corporation
4 */
5
6 #include "xe_debugfs.h"
7
8 #include <linux/debugfs.h>
9 #include <linux/fault-inject.h>
10 #include <linux/string_helpers.h>
11
12 #include <drm/drm_debugfs.h>
13
14 #include "regs/xe_pmt.h"
15 #include "xe_bo.h"
16 #include "xe_device.h"
17 #include "xe_force_wake.h"
18 #include "xe_gt_debugfs.h"
19 #include "xe_gt_printk.h"
20 #include "xe_guc_ads.h"
21 #include "xe_mmio.h"
22 #include "xe_pm.h"
23 #include "xe_psmi.h"
24 #include "xe_pxp_debugfs.h"
25 #include "xe_sriov.h"
26 #include "xe_sriov_pf_debugfs.h"
27 #include "xe_sriov_vf.h"
28 #include "xe_step.h"
29 #include "xe_tile_debugfs.h"
30 #include "xe_vsec.h"
31 #include "xe_wa.h"
32
33 #ifdef CONFIG_DRM_XE_DEBUG
34 #include "xe_bo_evict.h"
35 #include "xe_migrate.h"
36 #include "xe_vm.h"
37 #endif
38
39 DECLARE_FAULT_ATTR(gt_reset_failure);
40 DECLARE_FAULT_ATTR(inject_csc_hw_error);
41
read_residency_counter(struct xe_device * xe,struct xe_mmio * mmio,u32 offset,const char * name,struct drm_printer * p)42 static void read_residency_counter(struct xe_device *xe, struct xe_mmio *mmio,
43 u32 offset, const char *name, struct drm_printer *p)
44 {
45 u64 residency = 0;
46 int ret;
47
48 ret = xe_pmt_telem_read(to_pci_dev(xe->drm.dev),
49 xe_mmio_read32(mmio, PUNIT_TELEMETRY_GUID),
50 &residency, offset, sizeof(residency));
51 if (ret != sizeof(residency)) {
52 drm_warn(&xe->drm, "%s counter failed to read, ret %d\n", name, ret);
53 return;
54 }
55
56 drm_printf(p, "%s : %llu\n", name, residency);
57 }
58
node_to_xe(struct drm_info_node * node)59 static struct xe_device *node_to_xe(struct drm_info_node *node)
60 {
61 return to_xe_device(node->minor->dev);
62 }
63
info(struct seq_file * m,void * data)64 static int info(struct seq_file *m, void *data)
65 {
66 struct xe_device *xe = node_to_xe(m->private);
67 struct drm_printer p = drm_seq_file_printer(m);
68 struct xe_gt *gt;
69 u8 id;
70
71 guard(xe_pm_runtime)(xe);
72
73 drm_printf(&p, "graphics_verx100 %d\n", xe->info.graphics_verx100);
74 drm_printf(&p, "media_verx100 %d\n", xe->info.media_verx100);
75 drm_printf(&p, "stepping G:%s M:%s B:%s\n",
76 xe_step_name(xe->info.step.graphics),
77 xe_step_name(xe->info.step.media),
78 xe_step_name(xe->info.step.basedie));
79 drm_printf(&p, "is_dgfx %s\n", str_yes_no(xe->info.is_dgfx));
80 drm_printf(&p, "platform %d\n", xe->info.platform);
81 drm_printf(&p, "subplatform %d\n",
82 xe->info.subplatform > XE_SUBPLATFORM_NONE ? xe->info.subplatform : 0);
83 drm_printf(&p, "devid 0x%x\n", xe->info.devid);
84 drm_printf(&p, "revid %d\n", xe->info.revid);
85 drm_printf(&p, "tile_count %d\n", xe->info.tile_count);
86 drm_printf(&p, "vm_max_level %d\n", xe->info.vm_max_level);
87 drm_printf(&p, "force_execlist %s\n", str_yes_no(xe->info.force_execlist));
88 drm_printf(&p, "has_flat_ccs %s\n", str_yes_no(xe->info.has_flat_ccs));
89 drm_printf(&p, "has_usm %s\n", str_yes_no(xe->info.has_usm));
90 drm_printf(&p, "skip_guc_pc %s\n", str_yes_no(xe->info.skip_guc_pc));
91 for_each_gt(gt, xe, id) {
92 drm_printf(&p, "gt%d force wake %d\n", id,
93 xe_force_wake_ref(gt_to_fw(gt), XE_FW_GT));
94 drm_printf(&p, "gt%d engine_mask 0x%llx\n", id,
95 gt->info.engine_mask);
96 drm_printf(&p, "gt%d multi_queue_engine_class_mask 0x%x\n", id,
97 gt->info.multi_queue_engine_class_mask);
98 }
99
100 return 0;
101 }
102
sriov_info(struct seq_file * m,void * data)103 static int sriov_info(struct seq_file *m, void *data)
104 {
105 struct xe_device *xe = node_to_xe(m->private);
106 struct drm_printer p = drm_seq_file_printer(m);
107
108 xe_sriov_print_info(xe, &p);
109 return 0;
110 }
111
workarounds(struct xe_device * xe,struct drm_printer * p)112 static int workarounds(struct xe_device *xe, struct drm_printer *p)
113 {
114 guard(xe_pm_runtime)(xe);
115 xe_wa_device_dump(xe, p);
116
117 return 0;
118 }
119
workaround_info(struct seq_file * m,void * data)120 static int workaround_info(struct seq_file *m, void *data)
121 {
122 struct xe_device *xe = node_to_xe(m->private);
123 struct drm_printer p = drm_seq_file_printer(m);
124
125 workarounds(xe, &p);
126 return 0;
127 }
128
dgfx_pkg_residencies_show(struct seq_file * m,void * data)129 static int dgfx_pkg_residencies_show(struct seq_file *m, void *data)
130 {
131 struct xe_device *xe;
132 struct xe_mmio *mmio;
133 struct drm_printer p;
134
135 xe = node_to_xe(m->private);
136 p = drm_seq_file_printer(m);
137 guard(xe_pm_runtime)(xe);
138 mmio = xe_root_tile_mmio(xe);
139 static const struct {
140 u32 offset;
141 const char *name;
142 } residencies[] = {
143 {BMG_G2_RESIDENCY_OFFSET, "Package G2"},
144 {BMG_G6_RESIDENCY_OFFSET, "Package G6"},
145 {BMG_G7_RESIDENCY_OFFSET, "Package G7"},
146 {BMG_G8_RESIDENCY_OFFSET, "Package G8"},
147 {BMG_G10_RESIDENCY_OFFSET, "Package G10"},
148 {BMG_MODS_RESIDENCY_OFFSET, "Package ModS"}
149 };
150
151 for (int i = 0; i < ARRAY_SIZE(residencies); i++)
152 read_residency_counter(xe, mmio, residencies[i].offset, residencies[i].name, &p);
153
154 return 0;
155 }
156
dgfx_pcie_link_residencies_show(struct seq_file * m,void * data)157 static int dgfx_pcie_link_residencies_show(struct seq_file *m, void *data)
158 {
159 struct xe_device *xe;
160 struct xe_mmio *mmio;
161 struct drm_printer p;
162
163 xe = node_to_xe(m->private);
164 p = drm_seq_file_printer(m);
165 guard(xe_pm_runtime)(xe);
166 mmio = xe_root_tile_mmio(xe);
167
168 static const struct {
169 u32 offset;
170 const char *name;
171 } residencies[] = {
172 {BMG_PCIE_LINK_L0_RESIDENCY_OFFSET, "PCIE LINK L0 RESIDENCY"},
173 {BMG_PCIE_LINK_L1_RESIDENCY_OFFSET, "PCIE LINK L1 RESIDENCY"},
174 {BMG_PCIE_LINK_L1_2_RESIDENCY_OFFSET, "PCIE LINK L1.2 RESIDENCY"}
175 };
176
177 for (int i = 0; i < ARRAY_SIZE(residencies); i++)
178 read_residency_counter(xe, mmio, residencies[i].offset, residencies[i].name, &p);
179
180 return 0;
181 }
182
183 static const struct drm_info_list debugfs_list[] = {
184 {"info", info, 0},
185 { .name = "sriov_info", .show = sriov_info, },
186 { .name = "workarounds", .show = workaround_info, },
187 };
188
189 static const struct drm_info_list debugfs_residencies[] = {
190 { .name = "dgfx_pkg_residencies", .show = dgfx_pkg_residencies_show, },
191 { .name = "dgfx_pcie_link_residencies", .show = dgfx_pcie_link_residencies_show, },
192 };
193
forcewake_open(struct inode * inode,struct file * file)194 static int forcewake_open(struct inode *inode, struct file *file)
195 {
196 struct xe_device *xe = inode->i_private;
197 struct xe_gt *gt;
198 u8 id, last_gt;
199 unsigned int fw_ref;
200
201 xe_pm_runtime_get(xe);
202 for_each_gt(gt, xe, id) {
203 last_gt = id;
204
205 fw_ref = xe_force_wake_get(gt_to_fw(gt), XE_FORCEWAKE_ALL);
206 if (!xe_force_wake_ref_has_domain(fw_ref, XE_FORCEWAKE_ALL))
207 goto err_fw_get;
208 }
209
210 return 0;
211
212 err_fw_get:
213 for_each_gt(gt, xe, id) {
214 if (id < last_gt)
215 xe_force_wake_put(gt_to_fw(gt), XE_FORCEWAKE_ALL);
216 else if (id == last_gt)
217 xe_force_wake_put(gt_to_fw(gt), fw_ref);
218 else
219 break;
220 }
221
222 xe_pm_runtime_put(xe);
223 return -ETIMEDOUT;
224 }
225
forcewake_release(struct inode * inode,struct file * file)226 static int forcewake_release(struct inode *inode, struct file *file)
227 {
228 struct xe_device *xe = inode->i_private;
229 struct xe_gt *gt;
230 u8 id;
231
232 for_each_gt(gt, xe, id)
233 xe_force_wake_put(gt_to_fw(gt), XE_FORCEWAKE_ALL);
234 xe_pm_runtime_put(xe);
235
236 return 0;
237 }
238
239 static const struct file_operations forcewake_all_fops = {
240 .owner = THIS_MODULE,
241 .open = forcewake_open,
242 .release = forcewake_release,
243 };
244
wedged_mode_show(struct file * f,char __user * ubuf,size_t size,loff_t * pos)245 static ssize_t wedged_mode_show(struct file *f, char __user *ubuf,
246 size_t size, loff_t *pos)
247 {
248 struct xe_device *xe = file_inode(f)->i_private;
249 char buf[32];
250 int len = 0;
251
252 len = scnprintf(buf, sizeof(buf), "%d\n", xe->wedged.mode);
253
254 return simple_read_from_buffer(ubuf, size, pos, buf, len);
255 }
256
__wedged_mode_set_reset_policy(struct xe_gt * gt,enum xe_wedged_mode mode)257 static int __wedged_mode_set_reset_policy(struct xe_gt *gt, enum xe_wedged_mode mode)
258 {
259 bool enable_engine_reset;
260 int ret;
261
262 enable_engine_reset = (mode != XE_WEDGED_MODE_UPON_ANY_HANG_NO_RESET);
263 ret = xe_guc_ads_scheduler_policy_toggle_reset(>->uc.guc.ads,
264 enable_engine_reset);
265 if (ret)
266 xe_gt_err(gt, "Failed to update GuC ADS scheduler policy (%pe)\n", ERR_PTR(ret));
267
268 return ret;
269 }
270
wedged_mode_set_reset_policy(struct xe_device * xe,enum xe_wedged_mode mode)271 static int wedged_mode_set_reset_policy(struct xe_device *xe, enum xe_wedged_mode mode)
272 {
273 struct xe_gt *gt;
274 int ret;
275 u8 id;
276
277 guard(xe_pm_runtime)(xe);
278 for_each_gt(gt, xe, id) {
279 ret = __wedged_mode_set_reset_policy(gt, mode);
280 if (ret) {
281 if (id > 0) {
282 xe->wedged.inconsistent_reset = true;
283 drm_err(&xe->drm, "Inconsistent reset policy state between GTs\n");
284 }
285 return ret;
286 }
287 }
288
289 xe->wedged.inconsistent_reset = false;
290
291 return 0;
292 }
293
wedged_mode_needs_policy_update(struct xe_device * xe,enum xe_wedged_mode mode)294 static bool wedged_mode_needs_policy_update(struct xe_device *xe, enum xe_wedged_mode mode)
295 {
296 if (xe->wedged.inconsistent_reset)
297 return true;
298
299 if (xe->wedged.mode == mode)
300 return false;
301
302 if (xe->wedged.mode == XE_WEDGED_MODE_UPON_ANY_HANG_NO_RESET ||
303 mode == XE_WEDGED_MODE_UPON_ANY_HANG_NO_RESET)
304 return true;
305
306 return false;
307 }
308
wedged_mode_set(struct file * f,const char __user * ubuf,size_t size,loff_t * pos)309 static ssize_t wedged_mode_set(struct file *f, const char __user *ubuf,
310 size_t size, loff_t *pos)
311 {
312 struct xe_device *xe = file_inode(f)->i_private;
313 u32 wedged_mode;
314 ssize_t ret;
315
316 ret = kstrtouint_from_user(ubuf, size, 0, &wedged_mode);
317 if (ret)
318 return ret;
319
320 ret = xe_device_validate_wedged_mode(xe, wedged_mode);
321 if (ret)
322 return ret;
323
324 if (wedged_mode_needs_policy_update(xe, wedged_mode)) {
325 ret = wedged_mode_set_reset_policy(xe, wedged_mode);
326 if (ret)
327 return ret;
328 }
329
330 xe->wedged.mode = wedged_mode;
331
332 return size;
333 }
334
335 static const struct file_operations wedged_mode_fops = {
336 .owner = THIS_MODULE,
337 .read = wedged_mode_show,
338 .write = wedged_mode_set,
339 };
340
page_reclaim_hw_assist_show(struct file * f,char __user * ubuf,size_t size,loff_t * pos)341 static ssize_t page_reclaim_hw_assist_show(struct file *f, char __user *ubuf,
342 size_t size, loff_t *pos)
343 {
344 struct xe_device *xe = file_inode(f)->i_private;
345 char buf[8];
346 int len;
347
348 len = scnprintf(buf, sizeof(buf), "%d\n", xe->info.has_page_reclaim_hw_assist);
349 return simple_read_from_buffer(ubuf, size, pos, buf, len);
350 }
351
page_reclaim_hw_assist_set(struct file * f,const char __user * ubuf,size_t size,loff_t * pos)352 static ssize_t page_reclaim_hw_assist_set(struct file *f, const char __user *ubuf,
353 size_t size, loff_t *pos)
354 {
355 struct xe_device *xe = file_inode(f)->i_private;
356 bool val;
357 ssize_t ret;
358
359 ret = kstrtobool_from_user(ubuf, size, &val);
360 if (ret)
361 return ret;
362
363 xe->info.has_page_reclaim_hw_assist = val;
364
365 return size;
366 }
367
368 static const struct file_operations page_reclaim_hw_assist_fops = {
369 .owner = THIS_MODULE,
370 .read = page_reclaim_hw_assist_show,
371 .write = page_reclaim_hw_assist_set,
372 };
373
atomic_svm_timeslice_ms_show(struct file * f,char __user * ubuf,size_t size,loff_t * pos)374 static ssize_t atomic_svm_timeslice_ms_show(struct file *f, char __user *ubuf,
375 size_t size, loff_t *pos)
376 {
377 struct xe_device *xe = file_inode(f)->i_private;
378 char buf[32];
379 int len = 0;
380
381 len = scnprintf(buf, sizeof(buf), "%d\n", xe->atomic_svm_timeslice_ms);
382
383 return simple_read_from_buffer(ubuf, size, pos, buf, len);
384 }
385
atomic_svm_timeslice_ms_set(struct file * f,const char __user * ubuf,size_t size,loff_t * pos)386 static ssize_t atomic_svm_timeslice_ms_set(struct file *f,
387 const char __user *ubuf,
388 size_t size, loff_t *pos)
389 {
390 struct xe_device *xe = file_inode(f)->i_private;
391 u32 atomic_svm_timeslice_ms;
392 ssize_t ret;
393
394 ret = kstrtouint_from_user(ubuf, size, 0, &atomic_svm_timeslice_ms);
395 if (ret)
396 return ret;
397
398 xe->atomic_svm_timeslice_ms = atomic_svm_timeslice_ms;
399
400 return size;
401 }
402
403 static const struct file_operations atomic_svm_timeslice_ms_fops = {
404 .owner = THIS_MODULE,
405 .read = atomic_svm_timeslice_ms_show,
406 .write = atomic_svm_timeslice_ms_set,
407 };
408
min_run_period_lr_ms_show(struct file * f,char __user * ubuf,size_t size,loff_t * pos)409 static ssize_t min_run_period_lr_ms_show(struct file *f, char __user *ubuf,
410 size_t size, loff_t *pos)
411 {
412 struct xe_device *xe = file_inode(f)->i_private;
413 char buf[32];
414 int len = 0;
415
416 len = scnprintf(buf, sizeof(buf), "%d\n", xe->min_run_period_lr_ms);
417
418 return simple_read_from_buffer(ubuf, size, pos, buf, len);
419 }
420
min_run_period_lr_ms_set(struct file * f,const char __user * ubuf,size_t size,loff_t * pos)421 static ssize_t min_run_period_lr_ms_set(struct file *f, const char __user *ubuf,
422 size_t size, loff_t *pos)
423 {
424 struct xe_device *xe = file_inode(f)->i_private;
425 u32 min_run_period_lr_ms;
426 ssize_t ret;
427
428 ret = kstrtouint_from_user(ubuf, size, 0, &min_run_period_lr_ms);
429 if (ret)
430 return ret;
431
432 xe->min_run_period_lr_ms = min_run_period_lr_ms;
433
434 return size;
435 }
436
437 static const struct file_operations min_run_period_lr_ms_fops = {
438 .owner = THIS_MODULE,
439 .read = min_run_period_lr_ms_show,
440 .write = min_run_period_lr_ms_set,
441 };
442
min_run_period_pf_ms_show(struct file * f,char __user * ubuf,size_t size,loff_t * pos)443 static ssize_t min_run_period_pf_ms_show(struct file *f, char __user *ubuf,
444 size_t size, loff_t *pos)
445 {
446 struct xe_device *xe = file_inode(f)->i_private;
447 char buf[32];
448 int len = 0;
449
450 len = scnprintf(buf, sizeof(buf), "%d\n", xe->min_run_period_pf_ms);
451
452 return simple_read_from_buffer(ubuf, size, pos, buf, len);
453 }
454
min_run_period_pf_ms_set(struct file * f,const char __user * ubuf,size_t size,loff_t * pos)455 static ssize_t min_run_period_pf_ms_set(struct file *f, const char __user *ubuf,
456 size_t size, loff_t *pos)
457 {
458 struct xe_device *xe = file_inode(f)->i_private;
459 u32 min_run_period_pf_ms;
460 ssize_t ret;
461
462 ret = kstrtouint_from_user(ubuf, size, 0, &min_run_period_pf_ms);
463 if (ret)
464 return ret;
465
466 xe->min_run_period_pf_ms = min_run_period_pf_ms;
467
468 return size;
469 }
470
471 static const struct file_operations min_run_period_pf_ms_fops = {
472 .owner = THIS_MODULE,
473 .read = min_run_period_pf_ms_show,
474 .write = min_run_period_pf_ms_set,
475 };
476
disable_late_binding_show(struct file * f,char __user * ubuf,size_t size,loff_t * pos)477 static ssize_t disable_late_binding_show(struct file *f, char __user *ubuf,
478 size_t size, loff_t *pos)
479 {
480 struct xe_device *xe = file_inode(f)->i_private;
481 struct xe_late_bind *late_bind = &xe->late_bind;
482 char buf[32];
483 int len;
484
485 len = scnprintf(buf, sizeof(buf), "%d\n", late_bind->disable);
486
487 return simple_read_from_buffer(ubuf, size, pos, buf, len);
488 }
489
disable_late_binding_set(struct file * f,const char __user * ubuf,size_t size,loff_t * pos)490 static ssize_t disable_late_binding_set(struct file *f, const char __user *ubuf,
491 size_t size, loff_t *pos)
492 {
493 struct xe_device *xe = file_inode(f)->i_private;
494 struct xe_late_bind *late_bind = &xe->late_bind;
495 bool val;
496 int ret;
497
498 ret = kstrtobool_from_user(ubuf, size, &val);
499 if (ret)
500 return ret;
501
502 late_bind->disable = val;
503 return size;
504 }
505
506 static const struct file_operations disable_late_binding_fops = {
507 .owner = THIS_MODULE,
508 .read = disable_late_binding_show,
509 .write = disable_late_binding_set,
510 };
511
xe_debugfs_register(struct xe_device * xe)512 void xe_debugfs_register(struct xe_device *xe)
513 {
514 struct ttm_device *bdev = &xe->ttm;
515 struct drm_minor *minor = xe->drm.primary;
516 struct dentry *root = minor->debugfs_root;
517 struct ttm_resource_manager *man;
518 struct xe_tile *tile;
519 struct xe_gt *gt;
520 u8 tile_id;
521 u8 id;
522
523 drm_debugfs_create_files(debugfs_list,
524 ARRAY_SIZE(debugfs_list),
525 root, minor);
526
527 if (xe->info.platform == XE_BATTLEMAGE && !IS_SRIOV_VF(xe)) {
528 drm_debugfs_create_files(debugfs_residencies,
529 ARRAY_SIZE(debugfs_residencies),
530 root, minor);
531 fault_create_debugfs_attr("inject_csc_hw_error", root,
532 &inject_csc_hw_error);
533 }
534
535 debugfs_create_file("forcewake_all", 0400, root, xe,
536 &forcewake_all_fops);
537
538 debugfs_create_file("wedged_mode", 0600, root, xe,
539 &wedged_mode_fops);
540
541 debugfs_create_file("atomic_svm_timeslice_ms", 0600, root, xe,
542 &atomic_svm_timeslice_ms_fops);
543
544 debugfs_create_file("min_run_period_lr_ms", 0600, root, xe,
545 &min_run_period_lr_ms_fops);
546
547 debugfs_create_file("min_run_period_pf_ms", 0600, root, xe,
548 &min_run_period_pf_ms_fops);
549
550 debugfs_create_file("disable_late_binding", 0600, root, xe,
551 &disable_late_binding_fops);
552
553 /*
554 * Don't expose page reclaim configuration file if not supported by the
555 * hardware initially.
556 */
557 if (xe->info.has_page_reclaim_hw_assist)
558 debugfs_create_file("page_reclaim_hw_assist", 0600, root, xe,
559 &page_reclaim_hw_assist_fops);
560
561 man = ttm_manager_type(bdev, XE_PL_TT);
562 ttm_resource_manager_create_debugfs(man, root, "gtt_mm");
563
564 man = ttm_manager_type(bdev, XE_PL_STOLEN);
565 if (man)
566 ttm_resource_manager_create_debugfs(man, root, "stolen_mm");
567
568 for_each_tile(tile, xe, tile_id)
569 xe_tile_debugfs_register(tile);
570
571 for_each_gt(gt, xe, id)
572 xe_gt_debugfs_register(gt);
573
574 xe_pxp_debugfs_register(xe->pxp);
575
576 xe_psmi_debugfs_register(xe);
577
578 fault_create_debugfs_attr("fail_gt_reset", root, >_reset_failure);
579
580 if (IS_SRIOV_PF(xe))
581 xe_sriov_pf_debugfs_register(xe, root);
582 else if (IS_SRIOV_VF(xe))
583 xe_sriov_vf_debugfs_register(xe, root);
584 }
585