1 // SPDX-License-Identifier: MIT 2 /* 3 * Copyright © 2022 Intel Corporation 4 */ 5 6 #include "xe_debugfs.h" 7 8 #include <linux/debugfs.h> 9 #include <linux/fault-inject.h> 10 #include <linux/string_helpers.h> 11 12 #include <drm/drm_debugfs.h> 13 14 #include "regs/xe_pmt.h" 15 #include "xe_bo.h" 16 #include "xe_device.h" 17 #include "xe_force_wake.h" 18 #include "xe_gt_debugfs.h" 19 #include "xe_gt_printk.h" 20 #include "xe_guc_ads.h" 21 #include "xe_mmio.h" 22 #include "xe_pm.h" 23 #include "xe_pxp_debugfs.h" 24 #include "xe_sriov.h" 25 #include "xe_sriov_pf.h" 26 #include "xe_step.h" 27 #include "xe_wa.h" 28 #include "xe_vsec.h" 29 30 #ifdef CONFIG_DRM_XE_DEBUG 31 #include "xe_bo_evict.h" 32 #include "xe_migrate.h" 33 #include "xe_vm.h" 34 #endif 35 36 DECLARE_FAULT_ATTR(gt_reset_failure); 37 38 static void read_residency_counter(struct xe_device *xe, struct xe_mmio *mmio, 39 u32 offset, char *name, struct drm_printer *p) 40 { 41 u64 residency = 0; 42 int ret; 43 44 ret = xe_pmt_telem_read(to_pci_dev(xe->drm.dev), 45 xe_mmio_read32(mmio, PUNIT_TELEMETRY_GUID), 46 &residency, offset, sizeof(residency)); 47 if (ret != sizeof(residency)) { 48 drm_warn(&xe->drm, "%s counter failed to read, ret %d\n", name, ret); 49 return; 50 } 51 52 drm_printf(p, "%s : %llu\n", name, residency); 53 } 54 55 static struct xe_device *node_to_xe(struct drm_info_node *node) 56 { 57 return to_xe_device(node->minor->dev); 58 } 59 60 static int info(struct seq_file *m, void *data) 61 { 62 struct xe_device *xe = node_to_xe(m->private); 63 struct drm_printer p = drm_seq_file_printer(m); 64 struct xe_gt *gt; 65 u8 id; 66 67 xe_pm_runtime_get(xe); 68 69 drm_printf(&p, "graphics_verx100 %d\n", xe->info.graphics_verx100); 70 drm_printf(&p, "media_verx100 %d\n", xe->info.media_verx100); 71 drm_printf(&p, "stepping G:%s M:%s B:%s\n", 72 xe_step_name(xe->info.step.graphics), 73 xe_step_name(xe->info.step.media), 74 xe_step_name(xe->info.step.basedie)); 75 drm_printf(&p, "is_dgfx %s\n", str_yes_no(xe->info.is_dgfx)); 76 drm_printf(&p, "platform %d\n", xe->info.platform); 77 drm_printf(&p, "subplatform %d\n", 78 xe->info.subplatform > XE_SUBPLATFORM_NONE ? xe->info.subplatform : 0); 79 drm_printf(&p, "devid 0x%x\n", xe->info.devid); 80 drm_printf(&p, "revid %d\n", xe->info.revid); 81 drm_printf(&p, "tile_count %d\n", xe->info.tile_count); 82 drm_printf(&p, "vm_max_level %d\n", xe->info.vm_max_level); 83 drm_printf(&p, "force_execlist %s\n", str_yes_no(xe->info.force_execlist)); 84 drm_printf(&p, "has_flat_ccs %s\n", str_yes_no(xe->info.has_flat_ccs)); 85 drm_printf(&p, "has_usm %s\n", str_yes_no(xe->info.has_usm)); 86 drm_printf(&p, "skip_guc_pc %s\n", str_yes_no(xe->info.skip_guc_pc)); 87 for_each_gt(gt, xe, id) { 88 drm_printf(&p, "gt%d force wake %d\n", id, 89 xe_force_wake_ref(gt_to_fw(gt), XE_FW_GT)); 90 drm_printf(&p, "gt%d engine_mask 0x%llx\n", id, 91 gt->info.engine_mask); 92 } 93 94 xe_pm_runtime_put(xe); 95 return 0; 96 } 97 98 static int sriov_info(struct seq_file *m, void *data) 99 { 100 struct xe_device *xe = node_to_xe(m->private); 101 struct drm_printer p = drm_seq_file_printer(m); 102 103 xe_sriov_print_info(xe, &p); 104 return 0; 105 } 106 107 static int workarounds(struct xe_device *xe, struct drm_printer *p) 108 { 109 xe_pm_runtime_get(xe); 110 xe_wa_device_dump(xe, p); 111 xe_pm_runtime_put(xe); 112 113 return 0; 114 } 115 116 static int workaround_info(struct seq_file *m, void *data) 117 { 118 struct xe_device *xe = node_to_xe(m->private); 119 struct drm_printer p = drm_seq_file_printer(m); 120 121 workarounds(xe, &p); 122 return 0; 123 } 124 125 static int dgfx_pkg_residencies_show(struct seq_file *m, void *data) 126 { 127 struct xe_device *xe; 128 struct xe_mmio *mmio; 129 struct drm_printer p; 130 131 xe = node_to_xe(m->private); 132 p = drm_seq_file_printer(m); 133 xe_pm_runtime_get(xe); 134 mmio = xe_root_tile_mmio(xe); 135 struct { 136 u32 offset; 137 char *name; 138 } residencies[] = { 139 {BMG_G2_RESIDENCY_OFFSET, "Package G2"}, 140 {BMG_G6_RESIDENCY_OFFSET, "Package G6"}, 141 {BMG_G8_RESIDENCY_OFFSET, "Package G8"}, 142 {BMG_G10_RESIDENCY_OFFSET, "Package G10"}, 143 {BMG_MODS_RESIDENCY_OFFSET, "Package ModS"} 144 }; 145 146 for (int i = 0; i < ARRAY_SIZE(residencies); i++) 147 read_residency_counter(xe, mmio, residencies[i].offset, residencies[i].name, &p); 148 149 xe_pm_runtime_put(xe); 150 return 0; 151 } 152 153 static int dgfx_pcie_link_residencies_show(struct seq_file *m, void *data) 154 { 155 struct xe_device *xe; 156 struct xe_mmio *mmio; 157 struct drm_printer p; 158 159 xe = node_to_xe(m->private); 160 p = drm_seq_file_printer(m); 161 xe_pm_runtime_get(xe); 162 mmio = xe_root_tile_mmio(xe); 163 164 struct { 165 u32 offset; 166 char *name; 167 } residencies[] = { 168 {BMG_PCIE_LINK_L0_RESIDENCY_OFFSET, "PCIE LINK L0 RESIDENCY"}, 169 {BMG_PCIE_LINK_L1_RESIDENCY_OFFSET, "PCIE LINK L1 RESIDENCY"}, 170 {BMG_PCIE_LINK_L1_2_RESIDENCY_OFFSET, "PCIE LINK L1.2 RESIDENCY"} 171 }; 172 173 for (int i = 0; i < ARRAY_SIZE(residencies); i++) 174 read_residency_counter(xe, mmio, residencies[i].offset, residencies[i].name, &p); 175 176 xe_pm_runtime_put(xe); 177 return 0; 178 } 179 180 static const struct drm_info_list debugfs_list[] = { 181 {"info", info, 0}, 182 { .name = "sriov_info", .show = sriov_info, }, 183 { .name = "workarounds", .show = workaround_info, }, 184 }; 185 186 static const struct drm_info_list debugfs_residencies[] = { 187 { .name = "dgfx_pkg_residencies", .show = dgfx_pkg_residencies_show, }, 188 { .name = "dgfx_pcie_link_residencies", .show = dgfx_pcie_link_residencies_show, }, 189 }; 190 191 static int forcewake_open(struct inode *inode, struct file *file) 192 { 193 struct xe_device *xe = inode->i_private; 194 struct xe_gt *gt; 195 u8 id, last_gt; 196 unsigned int fw_ref; 197 198 xe_pm_runtime_get(xe); 199 for_each_gt(gt, xe, id) { 200 last_gt = id; 201 202 fw_ref = xe_force_wake_get(gt_to_fw(gt), XE_FORCEWAKE_ALL); 203 if (!xe_force_wake_ref_has_domain(fw_ref, XE_FORCEWAKE_ALL)) 204 goto err_fw_get; 205 } 206 207 return 0; 208 209 err_fw_get: 210 for_each_gt(gt, xe, id) { 211 if (id < last_gt) 212 xe_force_wake_put(gt_to_fw(gt), XE_FORCEWAKE_ALL); 213 else if (id == last_gt) 214 xe_force_wake_put(gt_to_fw(gt), fw_ref); 215 else 216 break; 217 } 218 219 xe_pm_runtime_put(xe); 220 return -ETIMEDOUT; 221 } 222 223 static int forcewake_release(struct inode *inode, struct file *file) 224 { 225 struct xe_device *xe = inode->i_private; 226 struct xe_gt *gt; 227 u8 id; 228 229 for_each_gt(gt, xe, id) 230 xe_force_wake_put(gt_to_fw(gt), XE_FORCEWAKE_ALL); 231 xe_pm_runtime_put(xe); 232 233 return 0; 234 } 235 236 static const struct file_operations forcewake_all_fops = { 237 .owner = THIS_MODULE, 238 .open = forcewake_open, 239 .release = forcewake_release, 240 }; 241 242 static ssize_t wedged_mode_show(struct file *f, char __user *ubuf, 243 size_t size, loff_t *pos) 244 { 245 struct xe_device *xe = file_inode(f)->i_private; 246 char buf[32]; 247 int len = 0; 248 249 len = scnprintf(buf, sizeof(buf), "%d\n", xe->wedged.mode); 250 251 return simple_read_from_buffer(ubuf, size, pos, buf, len); 252 } 253 254 static ssize_t wedged_mode_set(struct file *f, const char __user *ubuf, 255 size_t size, loff_t *pos) 256 { 257 struct xe_device *xe = file_inode(f)->i_private; 258 struct xe_gt *gt; 259 u32 wedged_mode; 260 ssize_t ret; 261 u8 id; 262 263 ret = kstrtouint_from_user(ubuf, size, 0, &wedged_mode); 264 if (ret) 265 return ret; 266 267 if (wedged_mode > 2) 268 return -EINVAL; 269 270 if (xe->wedged.mode == wedged_mode) 271 return size; 272 273 xe->wedged.mode = wedged_mode; 274 275 xe_pm_runtime_get(xe); 276 for_each_gt(gt, xe, id) { 277 ret = xe_guc_ads_scheduler_policy_toggle_reset(>->uc.guc.ads); 278 if (ret) { 279 xe_gt_err(gt, "Failed to update GuC ADS scheduler policy. GuC may still cause engine reset even with wedged_mode=2\n"); 280 xe_pm_runtime_put(xe); 281 return -EIO; 282 } 283 } 284 xe_pm_runtime_put(xe); 285 286 return size; 287 } 288 289 static const struct file_operations wedged_mode_fops = { 290 .owner = THIS_MODULE, 291 .read = wedged_mode_show, 292 .write = wedged_mode_set, 293 }; 294 295 static ssize_t atomic_svm_timeslice_ms_show(struct file *f, char __user *ubuf, 296 size_t size, loff_t *pos) 297 { 298 struct xe_device *xe = file_inode(f)->i_private; 299 char buf[32]; 300 int len = 0; 301 302 len = scnprintf(buf, sizeof(buf), "%d\n", xe->atomic_svm_timeslice_ms); 303 304 return simple_read_from_buffer(ubuf, size, pos, buf, len); 305 } 306 307 static ssize_t atomic_svm_timeslice_ms_set(struct file *f, 308 const char __user *ubuf, 309 size_t size, loff_t *pos) 310 { 311 struct xe_device *xe = file_inode(f)->i_private; 312 u32 atomic_svm_timeslice_ms; 313 ssize_t ret; 314 315 ret = kstrtouint_from_user(ubuf, size, 0, &atomic_svm_timeslice_ms); 316 if (ret) 317 return ret; 318 319 xe->atomic_svm_timeslice_ms = atomic_svm_timeslice_ms; 320 321 return size; 322 } 323 324 static const struct file_operations atomic_svm_timeslice_ms_fops = { 325 .owner = THIS_MODULE, 326 .read = atomic_svm_timeslice_ms_show, 327 .write = atomic_svm_timeslice_ms_set, 328 }; 329 330 static void create_tile_debugfs(struct xe_tile *tile, struct dentry *root) 331 { 332 char name[8]; 333 334 snprintf(name, sizeof(name), "tile%u", tile->id); 335 tile->debugfs = debugfs_create_dir(name, root); 336 if (IS_ERR(tile->debugfs)) 337 return; 338 339 /* 340 * Store the xe_tile pointer as private data of the tile/ directory 341 * node so other tile specific attributes under that directory may 342 * refer to it by looking at its parent node private data. 343 */ 344 tile->debugfs->d_inode->i_private = tile; 345 } 346 347 void xe_debugfs_register(struct xe_device *xe) 348 { 349 struct ttm_device *bdev = &xe->ttm; 350 struct drm_minor *minor = xe->drm.primary; 351 struct dentry *root = minor->debugfs_root; 352 struct ttm_resource_manager *man; 353 struct xe_tile *tile; 354 struct xe_gt *gt; 355 u32 mem_type; 356 u8 tile_id; 357 u8 id; 358 359 drm_debugfs_create_files(debugfs_list, 360 ARRAY_SIZE(debugfs_list), 361 root, minor); 362 363 if (xe->info.platform == XE_BATTLEMAGE) 364 drm_debugfs_create_files(debugfs_residencies, 365 ARRAY_SIZE(debugfs_residencies), 366 root, minor); 367 368 debugfs_create_file("forcewake_all", 0400, root, xe, 369 &forcewake_all_fops); 370 371 debugfs_create_file("wedged_mode", 0600, root, xe, 372 &wedged_mode_fops); 373 374 debugfs_create_file("atomic_svm_timeslice_ms", 0600, root, xe, 375 &atomic_svm_timeslice_ms_fops); 376 377 for (mem_type = XE_PL_VRAM0; mem_type <= XE_PL_VRAM1; ++mem_type) { 378 man = ttm_manager_type(bdev, mem_type); 379 380 if (man) { 381 char name[16]; 382 383 snprintf(name, sizeof(name), "vram%d_mm", mem_type - XE_PL_VRAM0); 384 ttm_resource_manager_create_debugfs(man, root, name); 385 } 386 } 387 388 man = ttm_manager_type(bdev, XE_PL_TT); 389 ttm_resource_manager_create_debugfs(man, root, "gtt_mm"); 390 391 man = ttm_manager_type(bdev, XE_PL_STOLEN); 392 if (man) 393 ttm_resource_manager_create_debugfs(man, root, "stolen_mm"); 394 395 for_each_tile(tile, xe, tile_id) 396 create_tile_debugfs(tile, root); 397 398 for_each_gt(gt, xe, id) 399 xe_gt_debugfs_register(gt); 400 401 xe_pxp_debugfs_register(xe->pxp); 402 403 fault_create_debugfs_attr("fail_gt_reset", root, >_reset_failure); 404 405 if (IS_SRIOV_PF(xe)) 406 xe_sriov_pf_debugfs_register(xe, root); 407 } 408