1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * PCIe AER software error injection support. 4 * 5 * Debugging PCIe AER code is quite difficult because it is hard to 6 * trigger various real hardware errors. Software based error 7 * injection can fake almost all kinds of errors with the help of a 8 * user space helper tool aer-inject, which can be gotten from: 9 * https://github.com/intel/aer-inject.git 10 * 11 * Copyright 2009 Intel Corporation. 12 * Huang Ying <ying.huang@intel.com> 13 */ 14 15 #define dev_fmt(fmt) "aer_inject: " fmt 16 17 #include <linux/module.h> 18 #include <linux/init.h> 19 #include <linux/interrupt.h> 20 #include <linux/miscdevice.h> 21 #include <linux/pci.h> 22 #include <linux/slab.h> 23 #include <linux/fs.h> 24 #include <linux/uaccess.h> 25 #include <linux/stddef.h> 26 #include <linux/device.h> 27 28 #include "portdrv.h" 29 30 /* Override the existing corrected and uncorrected error masks */ 31 static bool aer_mask_override; 32 module_param(aer_mask_override, bool, 0); 33 34 struct aer_error_inj { 35 u8 bus; 36 u8 dev; 37 u8 fn; 38 u32 uncor_status; 39 u32 cor_status; 40 u32 header_log0; 41 u32 header_log1; 42 u32 header_log2; 43 u32 header_log3; 44 u32 domain; 45 }; 46 47 struct aer_error { 48 struct list_head list; 49 u32 domain; 50 unsigned int bus; 51 unsigned int devfn; 52 int pos_cap_err; 53 54 u32 uncor_status; 55 u32 cor_status; 56 u32 header_log0; 57 u32 header_log1; 58 u32 header_log2; 59 u32 header_log3; 60 u32 root_status; 61 u32 source_id; 62 }; 63 64 struct pci_bus_ops { 65 struct list_head list; 66 struct pci_bus *bus; 67 struct pci_ops *ops; 68 }; 69 70 static LIST_HEAD(einjected); 71 72 static LIST_HEAD(pci_bus_ops_list); 73 74 /* Protect einjected and pci_bus_ops_list */ 75 static DEFINE_SPINLOCK(inject_lock); 76 77 static void aer_error_init(struct aer_error *err, u32 domain, 78 unsigned int bus, unsigned int devfn, 79 int pos_cap_err) 80 { 81 INIT_LIST_HEAD(&err->list); 82 err->domain = domain; 83 err->bus = bus; 84 err->devfn = devfn; 85 err->pos_cap_err = pos_cap_err; 86 } 87 88 /* inject_lock must be held before calling */ 89 static struct aer_error *__find_aer_error(u32 domain, unsigned int bus, 90 unsigned int devfn) 91 { 92 struct aer_error *err; 93 94 list_for_each_entry(err, &einjected, list) { 95 if (domain == err->domain && 96 bus == err->bus && 97 devfn == err->devfn) 98 return err; 99 } 100 return NULL; 101 } 102 103 /* inject_lock must be held before calling */ 104 static struct aer_error *__find_aer_error_by_dev(struct pci_dev *dev) 105 { 106 int domain = pci_domain_nr(dev->bus); 107 if (domain < 0) 108 return NULL; 109 return __find_aer_error(domain, dev->bus->number, dev->devfn); 110 } 111 112 /* inject_lock must be held before calling */ 113 static struct pci_ops *__find_pci_bus_ops(struct pci_bus *bus) 114 { 115 struct pci_bus_ops *bus_ops; 116 117 list_for_each_entry(bus_ops, &pci_bus_ops_list, list) { 118 if (bus_ops->bus == bus) 119 return bus_ops->ops; 120 } 121 return NULL; 122 } 123 124 static struct pci_bus_ops *pci_bus_ops_pop(void) 125 { 126 unsigned long flags; 127 struct pci_bus_ops *bus_ops; 128 129 spin_lock_irqsave(&inject_lock, flags); 130 bus_ops = list_first_entry_or_null(&pci_bus_ops_list, 131 struct pci_bus_ops, list); 132 if (bus_ops) 133 list_del(&bus_ops->list); 134 spin_unlock_irqrestore(&inject_lock, flags); 135 return bus_ops; 136 } 137 138 static u32 *find_pci_config_dword(struct aer_error *err, int where, 139 int *prw1cs) 140 { 141 int rw1cs = 0; 142 u32 *target = NULL; 143 144 if (err->pos_cap_err == -1) 145 return NULL; 146 147 switch (where - err->pos_cap_err) { 148 case PCI_ERR_UNCOR_STATUS: 149 target = &err->uncor_status; 150 rw1cs = 1; 151 break; 152 case PCI_ERR_COR_STATUS: 153 target = &err->cor_status; 154 rw1cs = 1; 155 break; 156 case PCI_ERR_HEADER_LOG: 157 target = &err->header_log0; 158 break; 159 case PCI_ERR_HEADER_LOG+4: 160 target = &err->header_log1; 161 break; 162 case PCI_ERR_HEADER_LOG+8: 163 target = &err->header_log2; 164 break; 165 case PCI_ERR_HEADER_LOG+12: 166 target = &err->header_log3; 167 break; 168 case PCI_ERR_ROOT_STATUS: 169 target = &err->root_status; 170 rw1cs = 1; 171 break; 172 case PCI_ERR_ROOT_ERR_SRC: 173 target = &err->source_id; 174 break; 175 } 176 if (prw1cs) 177 *prw1cs = rw1cs; 178 return target; 179 } 180 181 static int aer_inj_read(struct pci_bus *bus, unsigned int devfn, int where, 182 int size, u32 *val) 183 { 184 struct pci_ops *ops, *my_ops; 185 int rv; 186 187 ops = __find_pci_bus_ops(bus); 188 if (!ops) 189 return -1; 190 191 my_ops = bus->ops; 192 bus->ops = ops; 193 rv = ops->read(bus, devfn, where, size, val); 194 bus->ops = my_ops; 195 196 return rv; 197 } 198 199 static int aer_inj_write(struct pci_bus *bus, unsigned int devfn, int where, 200 int size, u32 val) 201 { 202 struct pci_ops *ops, *my_ops; 203 int rv; 204 205 ops = __find_pci_bus_ops(bus); 206 if (!ops) 207 return -1; 208 209 my_ops = bus->ops; 210 bus->ops = ops; 211 rv = ops->write(bus, devfn, where, size, val); 212 bus->ops = my_ops; 213 214 return rv; 215 } 216 217 static int aer_inj_read_config(struct pci_bus *bus, unsigned int devfn, 218 int where, int size, u32 *val) 219 { 220 u32 *sim; 221 struct aer_error *err; 222 unsigned long flags; 223 int domain; 224 int rv; 225 226 spin_lock_irqsave(&inject_lock, flags); 227 if (size != sizeof(u32)) 228 goto out; 229 domain = pci_domain_nr(bus); 230 if (domain < 0) 231 goto out; 232 err = __find_aer_error(domain, bus->number, devfn); 233 if (!err) 234 goto out; 235 236 sim = find_pci_config_dword(err, where, NULL); 237 if (sim) { 238 *val = *sim; 239 spin_unlock_irqrestore(&inject_lock, flags); 240 return 0; 241 } 242 out: 243 rv = aer_inj_read(bus, devfn, where, size, val); 244 spin_unlock_irqrestore(&inject_lock, flags); 245 return rv; 246 } 247 248 static int aer_inj_write_config(struct pci_bus *bus, unsigned int devfn, 249 int where, int size, u32 val) 250 { 251 u32 *sim; 252 struct aer_error *err; 253 unsigned long flags; 254 int rw1cs; 255 int domain; 256 int rv; 257 258 spin_lock_irqsave(&inject_lock, flags); 259 if (size != sizeof(u32)) 260 goto out; 261 domain = pci_domain_nr(bus); 262 if (domain < 0) 263 goto out; 264 err = __find_aer_error(domain, bus->number, devfn); 265 if (!err) 266 goto out; 267 268 sim = find_pci_config_dword(err, where, &rw1cs); 269 if (sim) { 270 if (rw1cs) 271 *sim ^= val; 272 else 273 *sim = val; 274 spin_unlock_irqrestore(&inject_lock, flags); 275 return 0; 276 } 277 out: 278 rv = aer_inj_write(bus, devfn, where, size, val); 279 spin_unlock_irqrestore(&inject_lock, flags); 280 return rv; 281 } 282 283 static struct pci_ops aer_inj_pci_ops = { 284 .read = aer_inj_read_config, 285 .write = aer_inj_write_config, 286 }; 287 288 static void pci_bus_ops_init(struct pci_bus_ops *bus_ops, 289 struct pci_bus *bus, 290 struct pci_ops *ops) 291 { 292 INIT_LIST_HEAD(&bus_ops->list); 293 bus_ops->bus = bus; 294 bus_ops->ops = ops; 295 } 296 297 static int pci_bus_set_aer_ops(struct pci_bus *bus) 298 { 299 struct pci_ops *ops; 300 struct pci_bus_ops *bus_ops; 301 unsigned long flags; 302 303 bus_ops = kmalloc(sizeof(*bus_ops), GFP_KERNEL); 304 if (!bus_ops) 305 return -ENOMEM; 306 ops = pci_bus_set_ops(bus, &aer_inj_pci_ops); 307 spin_lock_irqsave(&inject_lock, flags); 308 if (ops == &aer_inj_pci_ops) 309 goto out; 310 pci_bus_ops_init(bus_ops, bus, ops); 311 list_add(&bus_ops->list, &pci_bus_ops_list); 312 bus_ops = NULL; 313 out: 314 spin_unlock_irqrestore(&inject_lock, flags); 315 kfree(bus_ops); 316 return 0; 317 } 318 319 static int aer_inject(struct aer_error_inj *einj) 320 { 321 struct aer_error *err, *rperr; 322 struct aer_error *err_alloc = NULL, *rperr_alloc = NULL; 323 struct pci_dev *dev, *rpdev; 324 struct pcie_device *edev; 325 struct device *device; 326 unsigned long flags; 327 unsigned int devfn = PCI_DEVFN(einj->dev, einj->fn); 328 int pos_cap_err, rp_pos_cap_err; 329 u32 sever, cor_mask, uncor_mask, cor_mask_orig = 0, uncor_mask_orig = 0; 330 int ret = 0; 331 332 dev = pci_get_domain_bus_and_slot(einj->domain, einj->bus, devfn); 333 if (!dev) 334 return -ENODEV; 335 rpdev = pcie_find_root_port(dev); 336 /* If Root Port not found, try to find an RCEC */ 337 if (!rpdev) 338 rpdev = dev->rcec; 339 if (!rpdev) { 340 pci_err(dev, "Neither Root Port nor RCEC found\n"); 341 ret = -ENODEV; 342 goto out_put; 343 } 344 345 pos_cap_err = dev->aer_cap; 346 if (!pos_cap_err) { 347 pci_err(dev, "Device doesn't support AER\n"); 348 ret = -EPROTONOSUPPORT; 349 goto out_put; 350 } 351 pci_read_config_dword(dev, pos_cap_err + PCI_ERR_UNCOR_SEVER, &sever); 352 pci_read_config_dword(dev, pos_cap_err + PCI_ERR_COR_MASK, &cor_mask); 353 pci_read_config_dword(dev, pos_cap_err + PCI_ERR_UNCOR_MASK, 354 &uncor_mask); 355 356 rp_pos_cap_err = rpdev->aer_cap; 357 if (!rp_pos_cap_err) { 358 pci_err(rpdev, "Root port doesn't support AER\n"); 359 ret = -EPROTONOSUPPORT; 360 goto out_put; 361 } 362 363 err_alloc = kzalloc(sizeof(struct aer_error), GFP_KERNEL); 364 if (!err_alloc) { 365 ret = -ENOMEM; 366 goto out_put; 367 } 368 rperr_alloc = kzalloc(sizeof(struct aer_error), GFP_KERNEL); 369 if (!rperr_alloc) { 370 ret = -ENOMEM; 371 goto out_put; 372 } 373 374 if (aer_mask_override) { 375 cor_mask_orig = cor_mask; 376 cor_mask &= !(einj->cor_status); 377 pci_write_config_dword(dev, pos_cap_err + PCI_ERR_COR_MASK, 378 cor_mask); 379 380 uncor_mask_orig = uncor_mask; 381 uncor_mask &= !(einj->uncor_status); 382 pci_write_config_dword(dev, pos_cap_err + PCI_ERR_UNCOR_MASK, 383 uncor_mask); 384 } 385 386 spin_lock_irqsave(&inject_lock, flags); 387 388 err = __find_aer_error_by_dev(dev); 389 if (!err) { 390 err = err_alloc; 391 err_alloc = NULL; 392 aer_error_init(err, einj->domain, einj->bus, devfn, 393 pos_cap_err); 394 list_add(&err->list, &einjected); 395 } 396 err->uncor_status |= einj->uncor_status; 397 err->cor_status |= einj->cor_status; 398 err->header_log0 = einj->header_log0; 399 err->header_log1 = einj->header_log1; 400 err->header_log2 = einj->header_log2; 401 err->header_log3 = einj->header_log3; 402 403 if (!aer_mask_override && einj->cor_status && 404 !(einj->cor_status & ~cor_mask)) { 405 ret = -EINVAL; 406 pci_warn(dev, "The correctable error(s) is masked by device\n"); 407 spin_unlock_irqrestore(&inject_lock, flags); 408 goto out_put; 409 } 410 if (!aer_mask_override && einj->uncor_status && 411 !(einj->uncor_status & ~uncor_mask)) { 412 ret = -EINVAL; 413 pci_warn(dev, "The uncorrectable error(s) is masked by device\n"); 414 spin_unlock_irqrestore(&inject_lock, flags); 415 goto out_put; 416 } 417 418 rperr = __find_aer_error_by_dev(rpdev); 419 if (!rperr) { 420 rperr = rperr_alloc; 421 rperr_alloc = NULL; 422 aer_error_init(rperr, pci_domain_nr(rpdev->bus), 423 rpdev->bus->number, rpdev->devfn, 424 rp_pos_cap_err); 425 list_add(&rperr->list, &einjected); 426 } 427 if (einj->cor_status) { 428 if (rperr->root_status & PCI_ERR_ROOT_COR_RCV) 429 rperr->root_status |= PCI_ERR_ROOT_MULTI_COR_RCV; 430 else 431 rperr->root_status |= PCI_ERR_ROOT_COR_RCV; 432 rperr->source_id &= 0xffff0000; 433 rperr->source_id |= PCI_DEVID(einj->bus, devfn); 434 } 435 if (einj->uncor_status) { 436 if (rperr->root_status & PCI_ERR_ROOT_UNCOR_RCV) 437 rperr->root_status |= PCI_ERR_ROOT_MULTI_UNCOR_RCV; 438 if (sever & einj->uncor_status) { 439 rperr->root_status |= PCI_ERR_ROOT_FATAL_RCV; 440 if (!(rperr->root_status & PCI_ERR_ROOT_UNCOR_RCV)) 441 rperr->root_status |= PCI_ERR_ROOT_FIRST_FATAL; 442 } else 443 rperr->root_status |= PCI_ERR_ROOT_NONFATAL_RCV; 444 rperr->root_status |= PCI_ERR_ROOT_UNCOR_RCV; 445 rperr->source_id &= 0x0000ffff; 446 rperr->source_id |= PCI_DEVID(einj->bus, devfn) << 16; 447 } 448 spin_unlock_irqrestore(&inject_lock, flags); 449 450 if (aer_mask_override) { 451 pci_write_config_dword(dev, pos_cap_err + PCI_ERR_COR_MASK, 452 cor_mask_orig); 453 pci_write_config_dword(dev, pos_cap_err + PCI_ERR_UNCOR_MASK, 454 uncor_mask_orig); 455 } 456 457 ret = pci_bus_set_aer_ops(dev->bus); 458 if (ret) 459 goto out_put; 460 ret = pci_bus_set_aer_ops(rpdev->bus); 461 if (ret) 462 goto out_put; 463 464 device = pcie_port_find_device(rpdev, PCIE_PORT_SERVICE_AER); 465 if (device) { 466 edev = to_pcie_device(device); 467 if (!get_service_data(edev)) { 468 pci_warn(edev->port, "AER service is not initialized\n"); 469 ret = -EPROTONOSUPPORT; 470 goto out_put; 471 } 472 pci_info(edev->port, "Injecting errors %08x/%08x into device %s\n", 473 einj->cor_status, einj->uncor_status, pci_name(dev)); 474 ret = irq_inject_interrupt(edev->irq); 475 } else { 476 pci_err(rpdev, "AER device not found\n"); 477 ret = -ENODEV; 478 } 479 out_put: 480 kfree(err_alloc); 481 kfree(rperr_alloc); 482 pci_dev_put(dev); 483 return ret; 484 } 485 486 static ssize_t aer_inject_write(struct file *filp, const char __user *ubuf, 487 size_t usize, loff_t *off) 488 { 489 struct aer_error_inj einj; 490 int ret; 491 492 if (!capable(CAP_SYS_ADMIN)) 493 return -EPERM; 494 if (usize < offsetof(struct aer_error_inj, domain) || 495 usize > sizeof(einj)) 496 return -EINVAL; 497 498 memset(&einj, 0, sizeof(einj)); 499 if (copy_from_user(&einj, ubuf, usize)) 500 return -EFAULT; 501 502 ret = aer_inject(&einj); 503 return ret ? ret : usize; 504 } 505 506 static const struct file_operations aer_inject_fops = { 507 .write = aer_inject_write, 508 .owner = THIS_MODULE, 509 .llseek = noop_llseek, 510 }; 511 512 static struct miscdevice aer_inject_device = { 513 .minor = MISC_DYNAMIC_MINOR, 514 .name = "aer_inject", 515 .fops = &aer_inject_fops, 516 }; 517 518 static int __init aer_inject_init(void) 519 { 520 return misc_register(&aer_inject_device); 521 } 522 523 static void __exit aer_inject_exit(void) 524 { 525 struct aer_error *err, *err_next; 526 unsigned long flags; 527 struct pci_bus_ops *bus_ops; 528 529 misc_deregister(&aer_inject_device); 530 531 while ((bus_ops = pci_bus_ops_pop())) { 532 pci_bus_set_ops(bus_ops->bus, bus_ops->ops); 533 kfree(bus_ops); 534 } 535 536 spin_lock_irqsave(&inject_lock, flags); 537 list_for_each_entry_safe(err, err_next, &einjected, list) { 538 list_del(&err->list); 539 kfree(err); 540 } 541 spin_unlock_irqrestore(&inject_lock, flags); 542 } 543 544 module_init(aer_inject_init); 545 module_exit(aer_inject_exit); 546 547 MODULE_DESCRIPTION("PCIe AER software error injector"); 548 MODULE_LICENSE("GPL"); 549