1 // SPDX-License-Identifier: MIT 2 /* 3 * Copyright © 2022 Intel Corporation 4 */ 5 6 #include "xe_debugfs.h" 7 8 #include <linux/debugfs.h> 9 #include <linux/fault-inject.h> 10 #include <linux/string_helpers.h> 11 12 #include <drm/drm_debugfs.h> 13 14 #include "regs/xe_pmt.h" 15 #include "xe_bo.h" 16 #include "xe_device.h" 17 #include "xe_force_wake.h" 18 #include "xe_gt_debugfs.h" 19 #include "xe_gt_printk.h" 20 #include "xe_guc_ads.h" 21 #include "xe_mmio.h" 22 #include "xe_pm.h" 23 #include "xe_psmi.h" 24 #include "xe_pxp_debugfs.h" 25 #include "xe_sriov.h" 26 #include "xe_sriov_pf.h" 27 #include "xe_step.h" 28 #include "xe_wa.h" 29 #include "xe_vsec.h" 30 31 #ifdef CONFIG_DRM_XE_DEBUG 32 #include "xe_bo_evict.h" 33 #include "xe_migrate.h" 34 #include "xe_vm.h" 35 #endif 36 37 DECLARE_FAULT_ATTR(gt_reset_failure); 38 39 static void read_residency_counter(struct xe_device *xe, struct xe_mmio *mmio, 40 u32 offset, char *name, struct drm_printer *p) 41 { 42 u64 residency = 0; 43 int ret; 44 45 ret = xe_pmt_telem_read(to_pci_dev(xe->drm.dev), 46 xe_mmio_read32(mmio, PUNIT_TELEMETRY_GUID), 47 &residency, offset, sizeof(residency)); 48 if (ret != sizeof(residency)) { 49 drm_warn(&xe->drm, "%s counter failed to read, ret %d\n", name, ret); 50 return; 51 } 52 53 drm_printf(p, "%s : %llu\n", name, residency); 54 } 55 56 static struct xe_device *node_to_xe(struct drm_info_node *node) 57 { 58 return to_xe_device(node->minor->dev); 59 } 60 61 static int info(struct seq_file *m, void *data) 62 { 63 struct xe_device *xe = node_to_xe(m->private); 64 struct drm_printer p = drm_seq_file_printer(m); 65 struct xe_gt *gt; 66 u8 id; 67 68 xe_pm_runtime_get(xe); 69 70 drm_printf(&p, "graphics_verx100 %d\n", xe->info.graphics_verx100); 71 drm_printf(&p, "media_verx100 %d\n", xe->info.media_verx100); 72 drm_printf(&p, "stepping G:%s M:%s B:%s\n", 73 xe_step_name(xe->info.step.graphics), 74 xe_step_name(xe->info.step.media), 75 xe_step_name(xe->info.step.basedie)); 76 drm_printf(&p, "is_dgfx %s\n", str_yes_no(xe->info.is_dgfx)); 77 drm_printf(&p, "platform %d\n", xe->info.platform); 78 drm_printf(&p, "subplatform %d\n", 79 xe->info.subplatform > XE_SUBPLATFORM_NONE ? xe->info.subplatform : 0); 80 drm_printf(&p, "devid 0x%x\n", xe->info.devid); 81 drm_printf(&p, "revid %d\n", xe->info.revid); 82 drm_printf(&p, "tile_count %d\n", xe->info.tile_count); 83 drm_printf(&p, "vm_max_level %d\n", xe->info.vm_max_level); 84 drm_printf(&p, "force_execlist %s\n", str_yes_no(xe->info.force_execlist)); 85 drm_printf(&p, "has_flat_ccs %s\n", str_yes_no(xe->info.has_flat_ccs)); 86 drm_printf(&p, "has_usm %s\n", str_yes_no(xe->info.has_usm)); 87 drm_printf(&p, "skip_guc_pc %s\n", str_yes_no(xe->info.skip_guc_pc)); 88 for_each_gt(gt, xe, id) { 89 drm_printf(&p, "gt%d force wake %d\n", id, 90 xe_force_wake_ref(gt_to_fw(gt), XE_FW_GT)); 91 drm_printf(&p, "gt%d engine_mask 0x%llx\n", id, 92 gt->info.engine_mask); 93 } 94 95 xe_pm_runtime_put(xe); 96 return 0; 97 } 98 99 static int sriov_info(struct seq_file *m, void *data) 100 { 101 struct xe_device *xe = node_to_xe(m->private); 102 struct drm_printer p = drm_seq_file_printer(m); 103 104 xe_sriov_print_info(xe, &p); 105 return 0; 106 } 107 108 static int workarounds(struct xe_device *xe, struct drm_printer *p) 109 { 110 xe_pm_runtime_get(xe); 111 xe_wa_device_dump(xe, p); 112 xe_pm_runtime_put(xe); 113 114 return 0; 115 } 116 117 static int workaround_info(struct seq_file *m, void *data) 118 { 119 struct xe_device *xe = node_to_xe(m->private); 120 struct drm_printer p = drm_seq_file_printer(m); 121 122 workarounds(xe, &p); 123 return 0; 124 } 125 126 static int dgfx_pkg_residencies_show(struct seq_file *m, void *data) 127 { 128 struct xe_device *xe; 129 struct xe_mmio *mmio; 130 struct drm_printer p; 131 132 xe = node_to_xe(m->private); 133 p = drm_seq_file_printer(m); 134 xe_pm_runtime_get(xe); 135 mmio = xe_root_tile_mmio(xe); 136 struct { 137 u32 offset; 138 char *name; 139 } residencies[] = { 140 {BMG_G2_RESIDENCY_OFFSET, "Package G2"}, 141 {BMG_G6_RESIDENCY_OFFSET, "Package G6"}, 142 {BMG_G8_RESIDENCY_OFFSET, "Package G8"}, 143 {BMG_G10_RESIDENCY_OFFSET, "Package G10"}, 144 {BMG_MODS_RESIDENCY_OFFSET, "Package ModS"} 145 }; 146 147 for (int i = 0; i < ARRAY_SIZE(residencies); i++) 148 read_residency_counter(xe, mmio, residencies[i].offset, residencies[i].name, &p); 149 150 xe_pm_runtime_put(xe); 151 return 0; 152 } 153 154 static int dgfx_pcie_link_residencies_show(struct seq_file *m, void *data) 155 { 156 struct xe_device *xe; 157 struct xe_mmio *mmio; 158 struct drm_printer p; 159 160 xe = node_to_xe(m->private); 161 p = drm_seq_file_printer(m); 162 xe_pm_runtime_get(xe); 163 mmio = xe_root_tile_mmio(xe); 164 165 struct { 166 u32 offset; 167 char *name; 168 } residencies[] = { 169 {BMG_PCIE_LINK_L0_RESIDENCY_OFFSET, "PCIE LINK L0 RESIDENCY"}, 170 {BMG_PCIE_LINK_L1_RESIDENCY_OFFSET, "PCIE LINK L1 RESIDENCY"}, 171 {BMG_PCIE_LINK_L1_2_RESIDENCY_OFFSET, "PCIE LINK L1.2 RESIDENCY"} 172 }; 173 174 for (int i = 0; i < ARRAY_SIZE(residencies); i++) 175 read_residency_counter(xe, mmio, residencies[i].offset, residencies[i].name, &p); 176 177 xe_pm_runtime_put(xe); 178 return 0; 179 } 180 181 static const struct drm_info_list debugfs_list[] = { 182 {"info", info, 0}, 183 { .name = "sriov_info", .show = sriov_info, }, 184 { .name = "workarounds", .show = workaround_info, }, 185 }; 186 187 static const struct drm_info_list debugfs_residencies[] = { 188 { .name = "dgfx_pkg_residencies", .show = dgfx_pkg_residencies_show, }, 189 { .name = "dgfx_pcie_link_residencies", .show = dgfx_pcie_link_residencies_show, }, 190 }; 191 192 static int forcewake_open(struct inode *inode, struct file *file) 193 { 194 struct xe_device *xe = inode->i_private; 195 struct xe_gt *gt; 196 u8 id, last_gt; 197 unsigned int fw_ref; 198 199 xe_pm_runtime_get(xe); 200 for_each_gt(gt, xe, id) { 201 last_gt = id; 202 203 fw_ref = xe_force_wake_get(gt_to_fw(gt), XE_FORCEWAKE_ALL); 204 if (!xe_force_wake_ref_has_domain(fw_ref, XE_FORCEWAKE_ALL)) 205 goto err_fw_get; 206 } 207 208 return 0; 209 210 err_fw_get: 211 for_each_gt(gt, xe, id) { 212 if (id < last_gt) 213 xe_force_wake_put(gt_to_fw(gt), XE_FORCEWAKE_ALL); 214 else if (id == last_gt) 215 xe_force_wake_put(gt_to_fw(gt), fw_ref); 216 else 217 break; 218 } 219 220 xe_pm_runtime_put(xe); 221 return -ETIMEDOUT; 222 } 223 224 static int forcewake_release(struct inode *inode, struct file *file) 225 { 226 struct xe_device *xe = inode->i_private; 227 struct xe_gt *gt; 228 u8 id; 229 230 for_each_gt(gt, xe, id) 231 xe_force_wake_put(gt_to_fw(gt), XE_FORCEWAKE_ALL); 232 xe_pm_runtime_put(xe); 233 234 return 0; 235 } 236 237 static const struct file_operations forcewake_all_fops = { 238 .owner = THIS_MODULE, 239 .open = forcewake_open, 240 .release = forcewake_release, 241 }; 242 243 static ssize_t wedged_mode_show(struct file *f, char __user *ubuf, 244 size_t size, loff_t *pos) 245 { 246 struct xe_device *xe = file_inode(f)->i_private; 247 char buf[32]; 248 int len = 0; 249 250 len = scnprintf(buf, sizeof(buf), "%d\n", xe->wedged.mode); 251 252 return simple_read_from_buffer(ubuf, size, pos, buf, len); 253 } 254 255 static ssize_t wedged_mode_set(struct file *f, const char __user *ubuf, 256 size_t size, loff_t *pos) 257 { 258 struct xe_device *xe = file_inode(f)->i_private; 259 struct xe_gt *gt; 260 u32 wedged_mode; 261 ssize_t ret; 262 u8 id; 263 264 ret = kstrtouint_from_user(ubuf, size, 0, &wedged_mode); 265 if (ret) 266 return ret; 267 268 if (wedged_mode > 2) 269 return -EINVAL; 270 271 if (xe->wedged.mode == wedged_mode) 272 return size; 273 274 xe->wedged.mode = wedged_mode; 275 276 xe_pm_runtime_get(xe); 277 for_each_gt(gt, xe, id) { 278 ret = xe_guc_ads_scheduler_policy_toggle_reset(>->uc.guc.ads); 279 if (ret) { 280 xe_gt_err(gt, "Failed to update GuC ADS scheduler policy. GuC may still cause engine reset even with wedged_mode=2\n"); 281 xe_pm_runtime_put(xe); 282 return -EIO; 283 } 284 } 285 xe_pm_runtime_put(xe); 286 287 return size; 288 } 289 290 static const struct file_operations wedged_mode_fops = { 291 .owner = THIS_MODULE, 292 .read = wedged_mode_show, 293 .write = wedged_mode_set, 294 }; 295 296 static ssize_t atomic_svm_timeslice_ms_show(struct file *f, char __user *ubuf, 297 size_t size, loff_t *pos) 298 { 299 struct xe_device *xe = file_inode(f)->i_private; 300 char buf[32]; 301 int len = 0; 302 303 len = scnprintf(buf, sizeof(buf), "%d\n", xe->atomic_svm_timeslice_ms); 304 305 return simple_read_from_buffer(ubuf, size, pos, buf, len); 306 } 307 308 static ssize_t atomic_svm_timeslice_ms_set(struct file *f, 309 const char __user *ubuf, 310 size_t size, loff_t *pos) 311 { 312 struct xe_device *xe = file_inode(f)->i_private; 313 u32 atomic_svm_timeslice_ms; 314 ssize_t ret; 315 316 ret = kstrtouint_from_user(ubuf, size, 0, &atomic_svm_timeslice_ms); 317 if (ret) 318 return ret; 319 320 xe->atomic_svm_timeslice_ms = atomic_svm_timeslice_ms; 321 322 return size; 323 } 324 325 static const struct file_operations atomic_svm_timeslice_ms_fops = { 326 .owner = THIS_MODULE, 327 .read = atomic_svm_timeslice_ms_show, 328 .write = atomic_svm_timeslice_ms_set, 329 }; 330 331 static void create_tile_debugfs(struct xe_tile *tile, struct dentry *root) 332 { 333 char name[8]; 334 335 snprintf(name, sizeof(name), "tile%u", tile->id); 336 tile->debugfs = debugfs_create_dir(name, root); 337 if (IS_ERR(tile->debugfs)) 338 return; 339 340 /* 341 * Store the xe_tile pointer as private data of the tile/ directory 342 * node so other tile specific attributes under that directory may 343 * refer to it by looking at its parent node private data. 344 */ 345 tile->debugfs->d_inode->i_private = tile; 346 } 347 348 void xe_debugfs_register(struct xe_device *xe) 349 { 350 struct ttm_device *bdev = &xe->ttm; 351 struct drm_minor *minor = xe->drm.primary; 352 struct dentry *root = minor->debugfs_root; 353 struct ttm_resource_manager *man; 354 struct xe_tile *tile; 355 struct xe_gt *gt; 356 u32 mem_type; 357 u8 tile_id; 358 u8 id; 359 360 drm_debugfs_create_files(debugfs_list, 361 ARRAY_SIZE(debugfs_list), 362 root, minor); 363 364 if (xe->info.platform == XE_BATTLEMAGE) 365 drm_debugfs_create_files(debugfs_residencies, 366 ARRAY_SIZE(debugfs_residencies), 367 root, minor); 368 369 debugfs_create_file("forcewake_all", 0400, root, xe, 370 &forcewake_all_fops); 371 372 debugfs_create_file("wedged_mode", 0600, root, xe, 373 &wedged_mode_fops); 374 375 debugfs_create_file("atomic_svm_timeslice_ms", 0600, root, xe, 376 &atomic_svm_timeslice_ms_fops); 377 378 for (mem_type = XE_PL_VRAM0; mem_type <= XE_PL_VRAM1; ++mem_type) { 379 man = ttm_manager_type(bdev, mem_type); 380 381 if (man) { 382 char name[16]; 383 384 snprintf(name, sizeof(name), "vram%d_mm", mem_type - XE_PL_VRAM0); 385 ttm_resource_manager_create_debugfs(man, root, name); 386 } 387 } 388 389 man = ttm_manager_type(bdev, XE_PL_TT); 390 ttm_resource_manager_create_debugfs(man, root, "gtt_mm"); 391 392 man = ttm_manager_type(bdev, XE_PL_STOLEN); 393 if (man) 394 ttm_resource_manager_create_debugfs(man, root, "stolen_mm"); 395 396 for_each_tile(tile, xe, tile_id) 397 create_tile_debugfs(tile, root); 398 399 for_each_gt(gt, xe, id) 400 xe_gt_debugfs_register(gt); 401 402 xe_pxp_debugfs_register(xe->pxp); 403 404 xe_psmi_debugfs_register(xe); 405 406 fault_create_debugfs_attr("fail_gt_reset", root, >_reset_failure); 407 408 if (IS_SRIOV_PF(xe)) 409 xe_sriov_pf_debugfs_register(xe, root); 410 } 411