1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * Host side test driver to test endpoint functionality 4 * 5 * Copyright (C) 2017 Texas Instruments 6 * Author: Kishon Vijay Abraham I <kishon@ti.com> 7 */ 8 9 #include <linux/crc32.h> 10 #include <linux/delay.h> 11 #include <linux/fs.h> 12 #include <linux/io.h> 13 #include <linux/interrupt.h> 14 #include <linux/irq.h> 15 #include <linux/miscdevice.h> 16 #include <linux/module.h> 17 #include <linux/mutex.h> 18 #include <linux/random.h> 19 #include <linux/slab.h> 20 #include <linux/uaccess.h> 21 #include <linux/pci.h> 22 #include <linux/pci_ids.h> 23 24 #include <linux/pci_regs.h> 25 26 #include <uapi/linux/pcitest.h> 27 28 #define DRV_MODULE_NAME "pci-endpoint-test" 29 30 #define IRQ_TYPE_UNDEFINED -1 31 #define IRQ_TYPE_INTX 0 32 #define IRQ_TYPE_MSI 1 33 #define IRQ_TYPE_MSIX 2 34 35 #define PCI_ENDPOINT_TEST_MAGIC 0x0 36 37 #define PCI_ENDPOINT_TEST_COMMAND 0x4 38 #define COMMAND_RAISE_INTX_IRQ BIT(0) 39 #define COMMAND_RAISE_MSI_IRQ BIT(1) 40 #define COMMAND_RAISE_MSIX_IRQ BIT(2) 41 #define COMMAND_READ BIT(3) 42 #define COMMAND_WRITE BIT(4) 43 #define COMMAND_COPY BIT(5) 44 45 #define PCI_ENDPOINT_TEST_STATUS 0x8 46 #define STATUS_READ_SUCCESS BIT(0) 47 #define STATUS_READ_FAIL BIT(1) 48 #define STATUS_WRITE_SUCCESS BIT(2) 49 #define STATUS_WRITE_FAIL BIT(3) 50 #define STATUS_COPY_SUCCESS BIT(4) 51 #define STATUS_COPY_FAIL BIT(5) 52 #define STATUS_IRQ_RAISED BIT(6) 53 #define STATUS_SRC_ADDR_INVALID BIT(7) 54 #define STATUS_DST_ADDR_INVALID BIT(8) 55 56 #define PCI_ENDPOINT_TEST_LOWER_SRC_ADDR 0x0c 57 #define PCI_ENDPOINT_TEST_UPPER_SRC_ADDR 0x10 58 59 #define PCI_ENDPOINT_TEST_LOWER_DST_ADDR 0x14 60 #define PCI_ENDPOINT_TEST_UPPER_DST_ADDR 0x18 61 62 #define PCI_ENDPOINT_TEST_SIZE 0x1c 63 #define PCI_ENDPOINT_TEST_CHECKSUM 0x20 64 65 #define PCI_ENDPOINT_TEST_IRQ_TYPE 0x24 66 #define PCI_ENDPOINT_TEST_IRQ_NUMBER 0x28 67 68 #define PCI_ENDPOINT_TEST_FLAGS 0x2c 69 #define FLAG_USE_DMA BIT(0) 70 71 #define PCI_DEVICE_ID_TI_AM654 0xb00c 72 #define PCI_DEVICE_ID_TI_J7200 0xb00f 73 #define PCI_DEVICE_ID_TI_AM64 0xb010 74 #define PCI_DEVICE_ID_TI_J721S2 0xb013 75 #define PCI_DEVICE_ID_LS1088A 0x80c0 76 #define PCI_DEVICE_ID_IMX8 0x0808 77 78 #define is_am654_pci_dev(pdev) \ 79 ((pdev)->device == PCI_DEVICE_ID_TI_AM654) 80 81 #define PCI_DEVICE_ID_RENESAS_R8A774A1 0x0028 82 #define PCI_DEVICE_ID_RENESAS_R8A774B1 0x002b 83 #define PCI_DEVICE_ID_RENESAS_R8A774C0 0x002d 84 #define PCI_DEVICE_ID_RENESAS_R8A774E1 0x0025 85 #define PCI_DEVICE_ID_RENESAS_R8A779F0 0x0031 86 87 static DEFINE_IDA(pci_endpoint_test_ida); 88 89 #define to_endpoint_test(priv) container_of((priv), struct pci_endpoint_test, \ 90 miscdev) 91 92 static bool no_msi; 93 module_param(no_msi, bool, 0444); 94 MODULE_PARM_DESC(no_msi, "Disable MSI interrupt in pci_endpoint_test"); 95 96 static int irq_type = IRQ_TYPE_MSI; 97 module_param(irq_type, int, 0444); 98 MODULE_PARM_DESC(irq_type, "IRQ mode selection in pci_endpoint_test (0 - Legacy, 1 - MSI, 2 - MSI-X)"); 99 100 enum pci_barno { 101 BAR_0, 102 BAR_1, 103 BAR_2, 104 BAR_3, 105 BAR_4, 106 BAR_5, 107 }; 108 109 struct pci_endpoint_test { 110 struct pci_dev *pdev; 111 void __iomem *base; 112 void __iomem *bar[PCI_STD_NUM_BARS]; 113 struct completion irq_raised; 114 int last_irq; 115 int num_irqs; 116 int irq_type; 117 /* mutex to protect the ioctls */ 118 struct mutex mutex; 119 struct miscdevice miscdev; 120 enum pci_barno test_reg_bar; 121 size_t alignment; 122 const char *name; 123 }; 124 125 struct pci_endpoint_test_data { 126 enum pci_barno test_reg_bar; 127 size_t alignment; 128 int irq_type; 129 }; 130 131 static inline u32 pci_endpoint_test_readl(struct pci_endpoint_test *test, 132 u32 offset) 133 { 134 return readl(test->base + offset); 135 } 136 137 static inline void pci_endpoint_test_writel(struct pci_endpoint_test *test, 138 u32 offset, u32 value) 139 { 140 writel(value, test->base + offset); 141 } 142 143 static inline u32 pci_endpoint_test_bar_readl(struct pci_endpoint_test *test, 144 int bar, int offset) 145 { 146 return readl(test->bar[bar] + offset); 147 } 148 149 static inline void pci_endpoint_test_bar_writel(struct pci_endpoint_test *test, 150 int bar, u32 offset, u32 value) 151 { 152 writel(value, test->bar[bar] + offset); 153 } 154 155 static irqreturn_t pci_endpoint_test_irqhandler(int irq, void *dev_id) 156 { 157 struct pci_endpoint_test *test = dev_id; 158 u32 reg; 159 160 reg = pci_endpoint_test_readl(test, PCI_ENDPOINT_TEST_STATUS); 161 if (reg & STATUS_IRQ_RAISED) { 162 test->last_irq = irq; 163 complete(&test->irq_raised); 164 } 165 166 return IRQ_HANDLED; 167 } 168 169 static void pci_endpoint_test_free_irq_vectors(struct pci_endpoint_test *test) 170 { 171 struct pci_dev *pdev = test->pdev; 172 173 pci_free_irq_vectors(pdev); 174 test->irq_type = IRQ_TYPE_UNDEFINED; 175 } 176 177 static bool pci_endpoint_test_alloc_irq_vectors(struct pci_endpoint_test *test, 178 int type) 179 { 180 int irq = -1; 181 struct pci_dev *pdev = test->pdev; 182 struct device *dev = &pdev->dev; 183 bool res = true; 184 185 switch (type) { 186 case IRQ_TYPE_INTX: 187 irq = pci_alloc_irq_vectors(pdev, 1, 1, PCI_IRQ_INTX); 188 if (irq < 0) 189 dev_err(dev, "Failed to get Legacy interrupt\n"); 190 break; 191 case IRQ_TYPE_MSI: 192 irq = pci_alloc_irq_vectors(pdev, 1, 32, PCI_IRQ_MSI); 193 if (irq < 0) 194 dev_err(dev, "Failed to get MSI interrupts\n"); 195 break; 196 case IRQ_TYPE_MSIX: 197 irq = pci_alloc_irq_vectors(pdev, 1, 2048, PCI_IRQ_MSIX); 198 if (irq < 0) 199 dev_err(dev, "Failed to get MSI-X interrupts\n"); 200 break; 201 default: 202 dev_err(dev, "Invalid IRQ type selected\n"); 203 } 204 205 if (irq < 0) { 206 irq = 0; 207 res = false; 208 } 209 210 test->irq_type = type; 211 test->num_irqs = irq; 212 213 return res; 214 } 215 216 static void pci_endpoint_test_release_irq(struct pci_endpoint_test *test) 217 { 218 int i; 219 struct pci_dev *pdev = test->pdev; 220 struct device *dev = &pdev->dev; 221 222 for (i = 0; i < test->num_irqs; i++) 223 devm_free_irq(dev, pci_irq_vector(pdev, i), test); 224 225 test->num_irqs = 0; 226 } 227 228 static bool pci_endpoint_test_request_irq(struct pci_endpoint_test *test) 229 { 230 int i; 231 int err; 232 struct pci_dev *pdev = test->pdev; 233 struct device *dev = &pdev->dev; 234 235 for (i = 0; i < test->num_irqs; i++) { 236 err = devm_request_irq(dev, pci_irq_vector(pdev, i), 237 pci_endpoint_test_irqhandler, 238 IRQF_SHARED, test->name, test); 239 if (err) 240 goto fail; 241 } 242 243 return true; 244 245 fail: 246 switch (irq_type) { 247 case IRQ_TYPE_INTX: 248 dev_err(dev, "Failed to request IRQ %d for Legacy\n", 249 pci_irq_vector(pdev, i)); 250 break; 251 case IRQ_TYPE_MSI: 252 dev_err(dev, "Failed to request IRQ %d for MSI %d\n", 253 pci_irq_vector(pdev, i), 254 i + 1); 255 break; 256 case IRQ_TYPE_MSIX: 257 dev_err(dev, "Failed to request IRQ %d for MSI-X %d\n", 258 pci_irq_vector(pdev, i), 259 i + 1); 260 break; 261 } 262 263 return false; 264 } 265 266 static const u32 bar_test_pattern[] = { 267 0xA0A0A0A0, 268 0xA1A1A1A1, 269 0xA2A2A2A2, 270 0xA3A3A3A3, 271 0xA4A4A4A4, 272 0xA5A5A5A5, 273 }; 274 275 static bool pci_endpoint_test_bar(struct pci_endpoint_test *test, 276 enum pci_barno barno) 277 { 278 int j; 279 u32 val; 280 int size; 281 struct pci_dev *pdev = test->pdev; 282 283 if (!test->bar[barno]) 284 return false; 285 286 size = pci_resource_len(pdev, barno); 287 288 if (barno == test->test_reg_bar) 289 size = 0x4; 290 291 for (j = 0; j < size; j += 4) 292 pci_endpoint_test_bar_writel(test, barno, j, 293 bar_test_pattern[barno]); 294 295 for (j = 0; j < size; j += 4) { 296 val = pci_endpoint_test_bar_readl(test, barno, j); 297 if (val != bar_test_pattern[barno]) 298 return false; 299 } 300 301 return true; 302 } 303 304 static bool pci_endpoint_test_intx_irq(struct pci_endpoint_test *test) 305 { 306 u32 val; 307 308 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_TYPE, 309 IRQ_TYPE_INTX); 310 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_NUMBER, 0); 311 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_COMMAND, 312 COMMAND_RAISE_INTX_IRQ); 313 val = wait_for_completion_timeout(&test->irq_raised, 314 msecs_to_jiffies(1000)); 315 if (!val) 316 return false; 317 318 return true; 319 } 320 321 static bool pci_endpoint_test_msi_irq(struct pci_endpoint_test *test, 322 u16 msi_num, bool msix) 323 { 324 u32 val; 325 struct pci_dev *pdev = test->pdev; 326 327 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_TYPE, 328 msix ? IRQ_TYPE_MSIX : IRQ_TYPE_MSI); 329 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_NUMBER, msi_num); 330 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_COMMAND, 331 msix ? COMMAND_RAISE_MSIX_IRQ : 332 COMMAND_RAISE_MSI_IRQ); 333 val = wait_for_completion_timeout(&test->irq_raised, 334 msecs_to_jiffies(1000)); 335 if (!val) 336 return false; 337 338 return pci_irq_vector(pdev, msi_num - 1) == test->last_irq; 339 } 340 341 static int pci_endpoint_test_validate_xfer_params(struct device *dev, 342 struct pci_endpoint_test_xfer_param *param, size_t alignment) 343 { 344 if (!param->size) { 345 dev_dbg(dev, "Data size is zero\n"); 346 return -EINVAL; 347 } 348 349 if (param->size > SIZE_MAX - alignment) { 350 dev_dbg(dev, "Maximum transfer data size exceeded\n"); 351 return -EINVAL; 352 } 353 354 return 0; 355 } 356 357 static bool pci_endpoint_test_copy(struct pci_endpoint_test *test, 358 unsigned long arg) 359 { 360 struct pci_endpoint_test_xfer_param param; 361 bool ret = false; 362 void *src_addr; 363 void *dst_addr; 364 u32 flags = 0; 365 bool use_dma; 366 size_t size; 367 dma_addr_t src_phys_addr; 368 dma_addr_t dst_phys_addr; 369 struct pci_dev *pdev = test->pdev; 370 struct device *dev = &pdev->dev; 371 void *orig_src_addr; 372 dma_addr_t orig_src_phys_addr; 373 void *orig_dst_addr; 374 dma_addr_t orig_dst_phys_addr; 375 size_t offset; 376 size_t alignment = test->alignment; 377 int irq_type = test->irq_type; 378 u32 src_crc32; 379 u32 dst_crc32; 380 int err; 381 382 err = copy_from_user(¶m, (void __user *)arg, sizeof(param)); 383 if (err) { 384 dev_err(dev, "Failed to get transfer param\n"); 385 return false; 386 } 387 388 err = pci_endpoint_test_validate_xfer_params(dev, ¶m, alignment); 389 if (err) 390 return false; 391 392 size = param.size; 393 394 use_dma = !!(param.flags & PCITEST_FLAGS_USE_DMA); 395 if (use_dma) 396 flags |= FLAG_USE_DMA; 397 398 if (irq_type < IRQ_TYPE_INTX || irq_type > IRQ_TYPE_MSIX) { 399 dev_err(dev, "Invalid IRQ type option\n"); 400 goto err; 401 } 402 403 orig_src_addr = kzalloc(size + alignment, GFP_KERNEL); 404 if (!orig_src_addr) { 405 dev_err(dev, "Failed to allocate source buffer\n"); 406 ret = false; 407 goto err; 408 } 409 410 get_random_bytes(orig_src_addr, size + alignment); 411 orig_src_phys_addr = dma_map_single(dev, orig_src_addr, 412 size + alignment, DMA_TO_DEVICE); 413 if (dma_mapping_error(dev, orig_src_phys_addr)) { 414 dev_err(dev, "failed to map source buffer address\n"); 415 ret = false; 416 goto err_src_phys_addr; 417 } 418 419 if (alignment && !IS_ALIGNED(orig_src_phys_addr, alignment)) { 420 src_phys_addr = PTR_ALIGN(orig_src_phys_addr, alignment); 421 offset = src_phys_addr - orig_src_phys_addr; 422 src_addr = orig_src_addr + offset; 423 } else { 424 src_phys_addr = orig_src_phys_addr; 425 src_addr = orig_src_addr; 426 } 427 428 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_LOWER_SRC_ADDR, 429 lower_32_bits(src_phys_addr)); 430 431 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_UPPER_SRC_ADDR, 432 upper_32_bits(src_phys_addr)); 433 434 src_crc32 = crc32_le(~0, src_addr, size); 435 436 orig_dst_addr = kzalloc(size + alignment, GFP_KERNEL); 437 if (!orig_dst_addr) { 438 dev_err(dev, "Failed to allocate destination address\n"); 439 ret = false; 440 goto err_dst_addr; 441 } 442 443 orig_dst_phys_addr = dma_map_single(dev, orig_dst_addr, 444 size + alignment, DMA_FROM_DEVICE); 445 if (dma_mapping_error(dev, orig_dst_phys_addr)) { 446 dev_err(dev, "failed to map destination buffer address\n"); 447 ret = false; 448 goto err_dst_phys_addr; 449 } 450 451 if (alignment && !IS_ALIGNED(orig_dst_phys_addr, alignment)) { 452 dst_phys_addr = PTR_ALIGN(orig_dst_phys_addr, alignment); 453 offset = dst_phys_addr - orig_dst_phys_addr; 454 dst_addr = orig_dst_addr + offset; 455 } else { 456 dst_phys_addr = orig_dst_phys_addr; 457 dst_addr = orig_dst_addr; 458 } 459 460 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_LOWER_DST_ADDR, 461 lower_32_bits(dst_phys_addr)); 462 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_UPPER_DST_ADDR, 463 upper_32_bits(dst_phys_addr)); 464 465 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_SIZE, 466 size); 467 468 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_FLAGS, flags); 469 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_TYPE, irq_type); 470 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_NUMBER, 1); 471 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_COMMAND, 472 COMMAND_COPY); 473 474 wait_for_completion(&test->irq_raised); 475 476 dma_unmap_single(dev, orig_dst_phys_addr, size + alignment, 477 DMA_FROM_DEVICE); 478 479 dst_crc32 = crc32_le(~0, dst_addr, size); 480 if (dst_crc32 == src_crc32) 481 ret = true; 482 483 err_dst_phys_addr: 484 kfree(orig_dst_addr); 485 486 err_dst_addr: 487 dma_unmap_single(dev, orig_src_phys_addr, size + alignment, 488 DMA_TO_DEVICE); 489 490 err_src_phys_addr: 491 kfree(orig_src_addr); 492 493 err: 494 return ret; 495 } 496 497 static bool pci_endpoint_test_write(struct pci_endpoint_test *test, 498 unsigned long arg) 499 { 500 struct pci_endpoint_test_xfer_param param; 501 bool ret = false; 502 u32 flags = 0; 503 bool use_dma; 504 u32 reg; 505 void *addr; 506 dma_addr_t phys_addr; 507 struct pci_dev *pdev = test->pdev; 508 struct device *dev = &pdev->dev; 509 void *orig_addr; 510 dma_addr_t orig_phys_addr; 511 size_t offset; 512 size_t alignment = test->alignment; 513 int irq_type = test->irq_type; 514 size_t size; 515 u32 crc32; 516 int err; 517 518 err = copy_from_user(¶m, (void __user *)arg, sizeof(param)); 519 if (err != 0) { 520 dev_err(dev, "Failed to get transfer param\n"); 521 return false; 522 } 523 524 err = pci_endpoint_test_validate_xfer_params(dev, ¶m, alignment); 525 if (err) 526 return false; 527 528 size = param.size; 529 530 use_dma = !!(param.flags & PCITEST_FLAGS_USE_DMA); 531 if (use_dma) 532 flags |= FLAG_USE_DMA; 533 534 if (irq_type < IRQ_TYPE_INTX || irq_type > IRQ_TYPE_MSIX) { 535 dev_err(dev, "Invalid IRQ type option\n"); 536 goto err; 537 } 538 539 orig_addr = kzalloc(size + alignment, GFP_KERNEL); 540 if (!orig_addr) { 541 dev_err(dev, "Failed to allocate address\n"); 542 ret = false; 543 goto err; 544 } 545 546 get_random_bytes(orig_addr, size + alignment); 547 548 orig_phys_addr = dma_map_single(dev, orig_addr, size + alignment, 549 DMA_TO_DEVICE); 550 if (dma_mapping_error(dev, orig_phys_addr)) { 551 dev_err(dev, "failed to map source buffer address\n"); 552 ret = false; 553 goto err_phys_addr; 554 } 555 556 if (alignment && !IS_ALIGNED(orig_phys_addr, alignment)) { 557 phys_addr = PTR_ALIGN(orig_phys_addr, alignment); 558 offset = phys_addr - orig_phys_addr; 559 addr = orig_addr + offset; 560 } else { 561 phys_addr = orig_phys_addr; 562 addr = orig_addr; 563 } 564 565 crc32 = crc32_le(~0, addr, size); 566 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_CHECKSUM, 567 crc32); 568 569 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_LOWER_SRC_ADDR, 570 lower_32_bits(phys_addr)); 571 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_UPPER_SRC_ADDR, 572 upper_32_bits(phys_addr)); 573 574 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_SIZE, size); 575 576 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_FLAGS, flags); 577 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_TYPE, irq_type); 578 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_NUMBER, 1); 579 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_COMMAND, 580 COMMAND_READ); 581 582 wait_for_completion(&test->irq_raised); 583 584 reg = pci_endpoint_test_readl(test, PCI_ENDPOINT_TEST_STATUS); 585 if (reg & STATUS_READ_SUCCESS) 586 ret = true; 587 588 dma_unmap_single(dev, orig_phys_addr, size + alignment, 589 DMA_TO_DEVICE); 590 591 err_phys_addr: 592 kfree(orig_addr); 593 594 err: 595 return ret; 596 } 597 598 static bool pci_endpoint_test_read(struct pci_endpoint_test *test, 599 unsigned long arg) 600 { 601 struct pci_endpoint_test_xfer_param param; 602 bool ret = false; 603 u32 flags = 0; 604 bool use_dma; 605 size_t size; 606 void *addr; 607 dma_addr_t phys_addr; 608 struct pci_dev *pdev = test->pdev; 609 struct device *dev = &pdev->dev; 610 void *orig_addr; 611 dma_addr_t orig_phys_addr; 612 size_t offset; 613 size_t alignment = test->alignment; 614 int irq_type = test->irq_type; 615 u32 crc32; 616 int err; 617 618 err = copy_from_user(¶m, (void __user *)arg, sizeof(param)); 619 if (err) { 620 dev_err(dev, "Failed to get transfer param\n"); 621 return false; 622 } 623 624 err = pci_endpoint_test_validate_xfer_params(dev, ¶m, alignment); 625 if (err) 626 return false; 627 628 size = param.size; 629 630 use_dma = !!(param.flags & PCITEST_FLAGS_USE_DMA); 631 if (use_dma) 632 flags |= FLAG_USE_DMA; 633 634 if (irq_type < IRQ_TYPE_INTX || irq_type > IRQ_TYPE_MSIX) { 635 dev_err(dev, "Invalid IRQ type option\n"); 636 goto err; 637 } 638 639 orig_addr = kzalloc(size + alignment, GFP_KERNEL); 640 if (!orig_addr) { 641 dev_err(dev, "Failed to allocate destination address\n"); 642 ret = false; 643 goto err; 644 } 645 646 orig_phys_addr = dma_map_single(dev, orig_addr, size + alignment, 647 DMA_FROM_DEVICE); 648 if (dma_mapping_error(dev, orig_phys_addr)) { 649 dev_err(dev, "failed to map source buffer address\n"); 650 ret = false; 651 goto err_phys_addr; 652 } 653 654 if (alignment && !IS_ALIGNED(orig_phys_addr, alignment)) { 655 phys_addr = PTR_ALIGN(orig_phys_addr, alignment); 656 offset = phys_addr - orig_phys_addr; 657 addr = orig_addr + offset; 658 } else { 659 phys_addr = orig_phys_addr; 660 addr = orig_addr; 661 } 662 663 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_LOWER_DST_ADDR, 664 lower_32_bits(phys_addr)); 665 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_UPPER_DST_ADDR, 666 upper_32_bits(phys_addr)); 667 668 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_SIZE, size); 669 670 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_FLAGS, flags); 671 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_TYPE, irq_type); 672 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_NUMBER, 1); 673 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_COMMAND, 674 COMMAND_WRITE); 675 676 wait_for_completion(&test->irq_raised); 677 678 dma_unmap_single(dev, orig_phys_addr, size + alignment, 679 DMA_FROM_DEVICE); 680 681 crc32 = crc32_le(~0, addr, size); 682 if (crc32 == pci_endpoint_test_readl(test, PCI_ENDPOINT_TEST_CHECKSUM)) 683 ret = true; 684 685 err_phys_addr: 686 kfree(orig_addr); 687 err: 688 return ret; 689 } 690 691 static bool pci_endpoint_test_clear_irq(struct pci_endpoint_test *test) 692 { 693 pci_endpoint_test_release_irq(test); 694 pci_endpoint_test_free_irq_vectors(test); 695 return true; 696 } 697 698 static bool pci_endpoint_test_set_irq(struct pci_endpoint_test *test, 699 int req_irq_type) 700 { 701 struct pci_dev *pdev = test->pdev; 702 struct device *dev = &pdev->dev; 703 704 if (req_irq_type < IRQ_TYPE_INTX || req_irq_type > IRQ_TYPE_MSIX) { 705 dev_err(dev, "Invalid IRQ type option\n"); 706 return false; 707 } 708 709 if (test->irq_type == req_irq_type) 710 return true; 711 712 pci_endpoint_test_release_irq(test); 713 pci_endpoint_test_free_irq_vectors(test); 714 715 if (!pci_endpoint_test_alloc_irq_vectors(test, req_irq_type)) 716 goto err; 717 718 if (!pci_endpoint_test_request_irq(test)) 719 goto err; 720 721 return true; 722 723 err: 724 pci_endpoint_test_free_irq_vectors(test); 725 return false; 726 } 727 728 static long pci_endpoint_test_ioctl(struct file *file, unsigned int cmd, 729 unsigned long arg) 730 { 731 int ret = -EINVAL; 732 enum pci_barno bar; 733 struct pci_endpoint_test *test = to_endpoint_test(file->private_data); 734 struct pci_dev *pdev = test->pdev; 735 736 mutex_lock(&test->mutex); 737 738 reinit_completion(&test->irq_raised); 739 test->last_irq = -ENODATA; 740 741 switch (cmd) { 742 case PCITEST_BAR: 743 bar = arg; 744 if (bar > BAR_5) 745 goto ret; 746 if (is_am654_pci_dev(pdev) && bar == BAR_0) 747 goto ret; 748 ret = pci_endpoint_test_bar(test, bar); 749 break; 750 case PCITEST_INTX_IRQ: 751 ret = pci_endpoint_test_intx_irq(test); 752 break; 753 case PCITEST_MSI: 754 case PCITEST_MSIX: 755 ret = pci_endpoint_test_msi_irq(test, arg, cmd == PCITEST_MSIX); 756 break; 757 case PCITEST_WRITE: 758 ret = pci_endpoint_test_write(test, arg); 759 break; 760 case PCITEST_READ: 761 ret = pci_endpoint_test_read(test, arg); 762 break; 763 case PCITEST_COPY: 764 ret = pci_endpoint_test_copy(test, arg); 765 break; 766 case PCITEST_SET_IRQTYPE: 767 ret = pci_endpoint_test_set_irq(test, arg); 768 break; 769 case PCITEST_GET_IRQTYPE: 770 ret = irq_type; 771 break; 772 case PCITEST_CLEAR_IRQ: 773 ret = pci_endpoint_test_clear_irq(test); 774 break; 775 } 776 777 ret: 778 mutex_unlock(&test->mutex); 779 return ret; 780 } 781 782 static const struct file_operations pci_endpoint_test_fops = { 783 .owner = THIS_MODULE, 784 .unlocked_ioctl = pci_endpoint_test_ioctl, 785 }; 786 787 static int pci_endpoint_test_probe(struct pci_dev *pdev, 788 const struct pci_device_id *ent) 789 { 790 int err; 791 int id; 792 char name[24]; 793 enum pci_barno bar; 794 void __iomem *base; 795 struct device *dev = &pdev->dev; 796 struct pci_endpoint_test *test; 797 struct pci_endpoint_test_data *data; 798 enum pci_barno test_reg_bar = BAR_0; 799 struct miscdevice *misc_device; 800 801 if (pci_is_bridge(pdev)) 802 return -ENODEV; 803 804 test = devm_kzalloc(dev, sizeof(*test), GFP_KERNEL); 805 if (!test) 806 return -ENOMEM; 807 808 test->test_reg_bar = 0; 809 test->alignment = 0; 810 test->pdev = pdev; 811 test->irq_type = IRQ_TYPE_UNDEFINED; 812 813 if (no_msi) 814 irq_type = IRQ_TYPE_INTX; 815 816 data = (struct pci_endpoint_test_data *)ent->driver_data; 817 if (data) { 818 test_reg_bar = data->test_reg_bar; 819 test->test_reg_bar = test_reg_bar; 820 test->alignment = data->alignment; 821 irq_type = data->irq_type; 822 } 823 824 init_completion(&test->irq_raised); 825 mutex_init(&test->mutex); 826 827 if ((dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(48)) != 0) && 828 dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(32)) != 0) { 829 dev_err(dev, "Cannot set DMA mask\n"); 830 return -EINVAL; 831 } 832 833 err = pci_enable_device(pdev); 834 if (err) { 835 dev_err(dev, "Cannot enable PCI device\n"); 836 return err; 837 } 838 839 err = pci_request_regions(pdev, DRV_MODULE_NAME); 840 if (err) { 841 dev_err(dev, "Cannot obtain PCI resources\n"); 842 goto err_disable_pdev; 843 } 844 845 pci_set_master(pdev); 846 847 if (!pci_endpoint_test_alloc_irq_vectors(test, irq_type)) { 848 err = -EINVAL; 849 goto err_disable_irq; 850 } 851 852 for (bar = 0; bar < PCI_STD_NUM_BARS; bar++) { 853 if (pci_resource_flags(pdev, bar) & IORESOURCE_MEM) { 854 base = pci_ioremap_bar(pdev, bar); 855 if (!base) { 856 dev_err(dev, "Failed to read BAR%d\n", bar); 857 WARN_ON(bar == test_reg_bar); 858 } 859 test->bar[bar] = base; 860 } 861 } 862 863 test->base = test->bar[test_reg_bar]; 864 if (!test->base) { 865 err = -ENOMEM; 866 dev_err(dev, "Cannot perform PCI test without BAR%d\n", 867 test_reg_bar); 868 goto err_iounmap; 869 } 870 871 pci_set_drvdata(pdev, test); 872 873 id = ida_alloc(&pci_endpoint_test_ida, GFP_KERNEL); 874 if (id < 0) { 875 err = id; 876 dev_err(dev, "Unable to get id\n"); 877 goto err_iounmap; 878 } 879 880 snprintf(name, sizeof(name), DRV_MODULE_NAME ".%d", id); 881 test->name = kstrdup(name, GFP_KERNEL); 882 if (!test->name) { 883 err = -ENOMEM; 884 goto err_ida_remove; 885 } 886 887 if (!pci_endpoint_test_request_irq(test)) { 888 err = -EINVAL; 889 goto err_kfree_test_name; 890 } 891 892 misc_device = &test->miscdev; 893 misc_device->minor = MISC_DYNAMIC_MINOR; 894 misc_device->name = kstrdup(name, GFP_KERNEL); 895 if (!misc_device->name) { 896 err = -ENOMEM; 897 goto err_release_irq; 898 } 899 misc_device->parent = &pdev->dev; 900 misc_device->fops = &pci_endpoint_test_fops; 901 902 err = misc_register(misc_device); 903 if (err) { 904 dev_err(dev, "Failed to register device\n"); 905 goto err_kfree_name; 906 } 907 908 return 0; 909 910 err_kfree_name: 911 kfree(misc_device->name); 912 913 err_release_irq: 914 pci_endpoint_test_release_irq(test); 915 916 err_kfree_test_name: 917 kfree(test->name); 918 919 err_ida_remove: 920 ida_free(&pci_endpoint_test_ida, id); 921 922 err_iounmap: 923 for (bar = 0; bar < PCI_STD_NUM_BARS; bar++) { 924 if (test->bar[bar]) 925 pci_iounmap(pdev, test->bar[bar]); 926 } 927 928 err_disable_irq: 929 pci_endpoint_test_free_irq_vectors(test); 930 pci_release_regions(pdev); 931 932 err_disable_pdev: 933 pci_disable_device(pdev); 934 935 return err; 936 } 937 938 static void pci_endpoint_test_remove(struct pci_dev *pdev) 939 { 940 int id; 941 enum pci_barno bar; 942 struct pci_endpoint_test *test = pci_get_drvdata(pdev); 943 struct miscdevice *misc_device = &test->miscdev; 944 945 if (sscanf(misc_device->name, DRV_MODULE_NAME ".%d", &id) != 1) 946 return; 947 if (id < 0) 948 return; 949 950 pci_endpoint_test_release_irq(test); 951 pci_endpoint_test_free_irq_vectors(test); 952 953 misc_deregister(&test->miscdev); 954 kfree(misc_device->name); 955 kfree(test->name); 956 ida_free(&pci_endpoint_test_ida, id); 957 for (bar = 0; bar < PCI_STD_NUM_BARS; bar++) { 958 if (test->bar[bar]) 959 pci_iounmap(pdev, test->bar[bar]); 960 } 961 962 pci_release_regions(pdev); 963 pci_disable_device(pdev); 964 } 965 966 static const struct pci_endpoint_test_data default_data = { 967 .test_reg_bar = BAR_0, 968 .alignment = SZ_4K, 969 .irq_type = IRQ_TYPE_MSI, 970 }; 971 972 static const struct pci_endpoint_test_data am654_data = { 973 .test_reg_bar = BAR_2, 974 .alignment = SZ_64K, 975 .irq_type = IRQ_TYPE_MSI, 976 }; 977 978 static const struct pci_endpoint_test_data j721e_data = { 979 .alignment = 256, 980 .irq_type = IRQ_TYPE_MSI, 981 }; 982 983 static const struct pci_device_id pci_endpoint_test_tbl[] = { 984 { PCI_DEVICE(PCI_VENDOR_ID_TI, PCI_DEVICE_ID_TI_DRA74x), 985 .driver_data = (kernel_ulong_t)&default_data, 986 }, 987 { PCI_DEVICE(PCI_VENDOR_ID_TI, PCI_DEVICE_ID_TI_DRA72x), 988 .driver_data = (kernel_ulong_t)&default_data, 989 }, 990 { PCI_DEVICE(PCI_VENDOR_ID_FREESCALE, 0x81c0), 991 .driver_data = (kernel_ulong_t)&default_data, 992 }, 993 { PCI_DEVICE(PCI_VENDOR_ID_FREESCALE, PCI_DEVICE_ID_IMX8),}, 994 { PCI_DEVICE(PCI_VENDOR_ID_FREESCALE, PCI_DEVICE_ID_LS1088A), 995 .driver_data = (kernel_ulong_t)&default_data, 996 }, 997 { PCI_DEVICE_DATA(SYNOPSYS, EDDA, NULL) }, 998 { PCI_DEVICE(PCI_VENDOR_ID_TI, PCI_DEVICE_ID_TI_AM654), 999 .driver_data = (kernel_ulong_t)&am654_data 1000 }, 1001 { PCI_DEVICE(PCI_VENDOR_ID_RENESAS, PCI_DEVICE_ID_RENESAS_R8A774A1),}, 1002 { PCI_DEVICE(PCI_VENDOR_ID_RENESAS, PCI_DEVICE_ID_RENESAS_R8A774B1),}, 1003 { PCI_DEVICE(PCI_VENDOR_ID_RENESAS, PCI_DEVICE_ID_RENESAS_R8A774C0),}, 1004 { PCI_DEVICE(PCI_VENDOR_ID_RENESAS, PCI_DEVICE_ID_RENESAS_R8A774E1),}, 1005 { PCI_DEVICE(PCI_VENDOR_ID_RENESAS, PCI_DEVICE_ID_RENESAS_R8A779F0), 1006 .driver_data = (kernel_ulong_t)&default_data, 1007 }, 1008 { PCI_DEVICE(PCI_VENDOR_ID_TI, PCI_DEVICE_ID_TI_J721E), 1009 .driver_data = (kernel_ulong_t)&j721e_data, 1010 }, 1011 { PCI_DEVICE(PCI_VENDOR_ID_TI, PCI_DEVICE_ID_TI_J7200), 1012 .driver_data = (kernel_ulong_t)&j721e_data, 1013 }, 1014 { PCI_DEVICE(PCI_VENDOR_ID_TI, PCI_DEVICE_ID_TI_AM64), 1015 .driver_data = (kernel_ulong_t)&j721e_data, 1016 }, 1017 { PCI_DEVICE(PCI_VENDOR_ID_TI, PCI_DEVICE_ID_TI_J721S2), 1018 .driver_data = (kernel_ulong_t)&j721e_data, 1019 }, 1020 { } 1021 }; 1022 MODULE_DEVICE_TABLE(pci, pci_endpoint_test_tbl); 1023 1024 static struct pci_driver pci_endpoint_test_driver = { 1025 .name = DRV_MODULE_NAME, 1026 .id_table = pci_endpoint_test_tbl, 1027 .probe = pci_endpoint_test_probe, 1028 .remove = pci_endpoint_test_remove, 1029 .sriov_configure = pci_sriov_configure_simple, 1030 }; 1031 module_pci_driver(pci_endpoint_test_driver); 1032 1033 MODULE_DESCRIPTION("PCI ENDPOINT TEST HOST DRIVER"); 1034 MODULE_AUTHOR("Kishon Vijay Abraham I <kishon@ti.com>"); 1035 MODULE_LICENSE("GPL v2"); 1036