Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
1
fork

Configure Feed

Select the types of activity you want to include in your feed.

at v5.8-rc6 963 lines 24 kB view raw
1// SPDX-License-Identifier: GPL-2.0-only 2/** 3 * Host side test driver to test endpoint functionality 4 * 5 * Copyright (C) 2017 Texas Instruments 6 * Author: Kishon Vijay Abraham I <kishon@ti.com> 7 */ 8 9#include <linux/crc32.h> 10#include <linux/delay.h> 11#include <linux/fs.h> 12#include <linux/io.h> 13#include <linux/interrupt.h> 14#include <linux/irq.h> 15#include <linux/miscdevice.h> 16#include <linux/module.h> 17#include <linux/mutex.h> 18#include <linux/random.h> 19#include <linux/slab.h> 20#include <linux/uaccess.h> 21#include <linux/pci.h> 22#include <linux/pci_ids.h> 23 24#include <linux/pci_regs.h> 25 26#include <uapi/linux/pcitest.h> 27 28#define DRV_MODULE_NAME "pci-endpoint-test" 29 30#define IRQ_TYPE_UNDEFINED -1 31#define IRQ_TYPE_LEGACY 0 32#define IRQ_TYPE_MSI 1 33#define IRQ_TYPE_MSIX 2 34 35#define PCI_ENDPOINT_TEST_MAGIC 0x0 36 37#define PCI_ENDPOINT_TEST_COMMAND 0x4 38#define COMMAND_RAISE_LEGACY_IRQ BIT(0) 39#define COMMAND_RAISE_MSI_IRQ BIT(1) 40#define COMMAND_RAISE_MSIX_IRQ BIT(2) 41#define COMMAND_READ BIT(3) 42#define COMMAND_WRITE BIT(4) 43#define COMMAND_COPY BIT(5) 44 45#define PCI_ENDPOINT_TEST_STATUS 0x8 46#define STATUS_READ_SUCCESS BIT(0) 47#define STATUS_READ_FAIL BIT(1) 48#define STATUS_WRITE_SUCCESS BIT(2) 49#define STATUS_WRITE_FAIL BIT(3) 50#define STATUS_COPY_SUCCESS BIT(4) 51#define STATUS_COPY_FAIL BIT(5) 52#define STATUS_IRQ_RAISED BIT(6) 53#define STATUS_SRC_ADDR_INVALID BIT(7) 54#define STATUS_DST_ADDR_INVALID BIT(8) 55 56#define PCI_ENDPOINT_TEST_LOWER_SRC_ADDR 0x0c 57#define PCI_ENDPOINT_TEST_UPPER_SRC_ADDR 0x10 58 59#define PCI_ENDPOINT_TEST_LOWER_DST_ADDR 0x14 60#define PCI_ENDPOINT_TEST_UPPER_DST_ADDR 0x18 61 62#define PCI_ENDPOINT_TEST_SIZE 0x1c 63#define PCI_ENDPOINT_TEST_CHECKSUM 0x20 64 65#define PCI_ENDPOINT_TEST_IRQ_TYPE 0x24 66#define PCI_ENDPOINT_TEST_IRQ_NUMBER 0x28 67 68#define PCI_ENDPOINT_TEST_FLAGS 0x2c 69#define FLAG_USE_DMA BIT(0) 70 71#define PCI_DEVICE_ID_TI_AM654 0xb00c 72 73#define is_am654_pci_dev(pdev) \ 74 ((pdev)->device == PCI_DEVICE_ID_TI_AM654) 75 76#define PCI_DEVICE_ID_RENESAS_R8A774C0 0x002d 77 78static DEFINE_IDA(pci_endpoint_test_ida); 79 80#define to_endpoint_test(priv) container_of((priv), struct pci_endpoint_test, \ 81 miscdev) 82 83static bool no_msi; 84module_param(no_msi, bool, 0444); 85MODULE_PARM_DESC(no_msi, "Disable MSI interrupt in pci_endpoint_test"); 86 87static int irq_type = IRQ_TYPE_MSI; 88module_param(irq_type, int, 0444); 89MODULE_PARM_DESC(irq_type, "IRQ mode selection in pci_endpoint_test (0 - Legacy, 1 - MSI, 2 - MSI-X)"); 90 91enum pci_barno { 92 BAR_0, 93 BAR_1, 94 BAR_2, 95 BAR_3, 96 BAR_4, 97 BAR_5, 98}; 99 100struct pci_endpoint_test { 101 struct pci_dev *pdev; 102 void __iomem *base; 103 void __iomem *bar[PCI_STD_NUM_BARS]; 104 struct completion irq_raised; 105 int last_irq; 106 int num_irqs; 107 int irq_type; 108 /* mutex to protect the ioctls */ 109 struct mutex mutex; 110 struct miscdevice miscdev; 111 enum pci_barno test_reg_bar; 112 size_t alignment; 113 const char *name; 114}; 115 116struct pci_endpoint_test_data { 117 enum pci_barno test_reg_bar; 118 size_t alignment; 119 int irq_type; 120}; 121 122static inline u32 pci_endpoint_test_readl(struct pci_endpoint_test *test, 123 u32 offset) 124{ 125 return readl(test->base + offset); 126} 127 128static inline void pci_endpoint_test_writel(struct pci_endpoint_test *test, 129 u32 offset, u32 value) 130{ 131 writel(value, test->base + offset); 132} 133 134static inline u32 pci_endpoint_test_bar_readl(struct pci_endpoint_test *test, 135 int bar, int offset) 136{ 137 return readl(test->bar[bar] + offset); 138} 139 140static inline void pci_endpoint_test_bar_writel(struct pci_endpoint_test *test, 141 int bar, u32 offset, u32 value) 142{ 143 writel(value, test->bar[bar] + offset); 144} 145 146static irqreturn_t pci_endpoint_test_irqhandler(int irq, void *dev_id) 147{ 148 struct pci_endpoint_test *test = dev_id; 149 u32 reg; 150 151 reg = pci_endpoint_test_readl(test, PCI_ENDPOINT_TEST_STATUS); 152 if (reg & STATUS_IRQ_RAISED) { 153 test->last_irq = irq; 154 complete(&test->irq_raised); 155 reg &= ~STATUS_IRQ_RAISED; 156 } 157 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_STATUS, 158 reg); 159 160 return IRQ_HANDLED; 161} 162 163static void pci_endpoint_test_free_irq_vectors(struct pci_endpoint_test *test) 164{ 165 struct pci_dev *pdev = test->pdev; 166 167 pci_free_irq_vectors(pdev); 168 test->irq_type = IRQ_TYPE_UNDEFINED; 169} 170 171static bool pci_endpoint_test_alloc_irq_vectors(struct pci_endpoint_test *test, 172 int type) 173{ 174 int irq = -1; 175 struct pci_dev *pdev = test->pdev; 176 struct device *dev = &pdev->dev; 177 bool res = true; 178 179 switch (type) { 180 case IRQ_TYPE_LEGACY: 181 irq = pci_alloc_irq_vectors(pdev, 1, 1, PCI_IRQ_LEGACY); 182 if (irq < 0) 183 dev_err(dev, "Failed to get Legacy interrupt\n"); 184 break; 185 case IRQ_TYPE_MSI: 186 irq = pci_alloc_irq_vectors(pdev, 1, 32, PCI_IRQ_MSI); 187 if (irq < 0) 188 dev_err(dev, "Failed to get MSI interrupts\n"); 189 break; 190 case IRQ_TYPE_MSIX: 191 irq = pci_alloc_irq_vectors(pdev, 1, 2048, PCI_IRQ_MSIX); 192 if (irq < 0) 193 dev_err(dev, "Failed to get MSI-X interrupts\n"); 194 break; 195 default: 196 dev_err(dev, "Invalid IRQ type selected\n"); 197 } 198 199 if (irq < 0) { 200 irq = 0; 201 res = false; 202 } 203 204 test->irq_type = type; 205 test->num_irqs = irq; 206 207 return res; 208} 209 210static void pci_endpoint_test_release_irq(struct pci_endpoint_test *test) 211{ 212 int i; 213 struct pci_dev *pdev = test->pdev; 214 struct device *dev = &pdev->dev; 215 216 for (i = 0; i < test->num_irqs; i++) 217 devm_free_irq(dev, pci_irq_vector(pdev, i), test); 218 219 test->num_irqs = 0; 220} 221 222static bool pci_endpoint_test_request_irq(struct pci_endpoint_test *test) 223{ 224 int i; 225 int err; 226 struct pci_dev *pdev = test->pdev; 227 struct device *dev = &pdev->dev; 228 229 for (i = 0; i < test->num_irqs; i++) { 230 err = devm_request_irq(dev, pci_irq_vector(pdev, i), 231 pci_endpoint_test_irqhandler, 232 IRQF_SHARED, test->name, test); 233 if (err) 234 goto fail; 235 } 236 237 return true; 238 239fail: 240 switch (irq_type) { 241 case IRQ_TYPE_LEGACY: 242 dev_err(dev, "Failed to request IRQ %d for Legacy\n", 243 pci_irq_vector(pdev, i)); 244 break; 245 case IRQ_TYPE_MSI: 246 dev_err(dev, "Failed to request IRQ %d for MSI %d\n", 247 pci_irq_vector(pdev, i), 248 i + 1); 249 break; 250 case IRQ_TYPE_MSIX: 251 dev_err(dev, "Failed to request IRQ %d for MSI-X %d\n", 252 pci_irq_vector(pdev, i), 253 i + 1); 254 break; 255 } 256 257 return false; 258} 259 260static bool pci_endpoint_test_bar(struct pci_endpoint_test *test, 261 enum pci_barno barno) 262{ 263 int j; 264 u32 val; 265 int size; 266 struct pci_dev *pdev = test->pdev; 267 268 if (!test->bar[barno]) 269 return false; 270 271 size = pci_resource_len(pdev, barno); 272 273 if (barno == test->test_reg_bar) 274 size = 0x4; 275 276 for (j = 0; j < size; j += 4) 277 pci_endpoint_test_bar_writel(test, barno, j, 0xA0A0A0A0); 278 279 for (j = 0; j < size; j += 4) { 280 val = pci_endpoint_test_bar_readl(test, barno, j); 281 if (val != 0xA0A0A0A0) 282 return false; 283 } 284 285 return true; 286} 287 288static bool pci_endpoint_test_legacy_irq(struct pci_endpoint_test *test) 289{ 290 u32 val; 291 292 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_TYPE, 293 IRQ_TYPE_LEGACY); 294 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_NUMBER, 0); 295 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_COMMAND, 296 COMMAND_RAISE_LEGACY_IRQ); 297 val = wait_for_completion_timeout(&test->irq_raised, 298 msecs_to_jiffies(1000)); 299 if (!val) 300 return false; 301 302 return true; 303} 304 305static bool pci_endpoint_test_msi_irq(struct pci_endpoint_test *test, 306 u16 msi_num, bool msix) 307{ 308 u32 val; 309 struct pci_dev *pdev = test->pdev; 310 311 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_TYPE, 312 msix == false ? IRQ_TYPE_MSI : 313 IRQ_TYPE_MSIX); 314 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_NUMBER, msi_num); 315 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_COMMAND, 316 msix == false ? COMMAND_RAISE_MSI_IRQ : 317 COMMAND_RAISE_MSIX_IRQ); 318 val = wait_for_completion_timeout(&test->irq_raised, 319 msecs_to_jiffies(1000)); 320 if (!val) 321 return false; 322 323 if (pci_irq_vector(pdev, msi_num - 1) == test->last_irq) 324 return true; 325 326 return false; 327} 328 329static bool pci_endpoint_test_copy(struct pci_endpoint_test *test, 330 unsigned long arg) 331{ 332 struct pci_endpoint_test_xfer_param param; 333 bool ret = false; 334 void *src_addr; 335 void *dst_addr; 336 u32 flags = 0; 337 bool use_dma; 338 size_t size; 339 dma_addr_t src_phys_addr; 340 dma_addr_t dst_phys_addr; 341 struct pci_dev *pdev = test->pdev; 342 struct device *dev = &pdev->dev; 343 void *orig_src_addr; 344 dma_addr_t orig_src_phys_addr; 345 void *orig_dst_addr; 346 dma_addr_t orig_dst_phys_addr; 347 size_t offset; 348 size_t alignment = test->alignment; 349 int irq_type = test->irq_type; 350 u32 src_crc32; 351 u32 dst_crc32; 352 int err; 353 354 err = copy_from_user(&param, (void __user *)arg, sizeof(param)); 355 if (err) { 356 dev_err(dev, "Failed to get transfer param\n"); 357 return false; 358 } 359 360 size = param.size; 361 if (size > SIZE_MAX - alignment) 362 goto err; 363 364 use_dma = !!(param.flags & PCITEST_FLAGS_USE_DMA); 365 if (use_dma) 366 flags |= FLAG_USE_DMA; 367 368 if (irq_type < IRQ_TYPE_LEGACY || irq_type > IRQ_TYPE_MSIX) { 369 dev_err(dev, "Invalid IRQ type option\n"); 370 goto err; 371 } 372 373 orig_src_addr = kzalloc(size + alignment, GFP_KERNEL); 374 if (!orig_src_addr) { 375 dev_err(dev, "Failed to allocate source buffer\n"); 376 ret = false; 377 goto err; 378 } 379 380 get_random_bytes(orig_src_addr, size + alignment); 381 orig_src_phys_addr = dma_map_single(dev, orig_src_addr, 382 size + alignment, DMA_TO_DEVICE); 383 if (dma_mapping_error(dev, orig_src_phys_addr)) { 384 dev_err(dev, "failed to map source buffer address\n"); 385 ret = false; 386 goto err_src_phys_addr; 387 } 388 389 if (alignment && !IS_ALIGNED(orig_src_phys_addr, alignment)) { 390 src_phys_addr = PTR_ALIGN(orig_src_phys_addr, alignment); 391 offset = src_phys_addr - orig_src_phys_addr; 392 src_addr = orig_src_addr + offset; 393 } else { 394 src_phys_addr = orig_src_phys_addr; 395 src_addr = orig_src_addr; 396 } 397 398 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_LOWER_SRC_ADDR, 399 lower_32_bits(src_phys_addr)); 400 401 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_UPPER_SRC_ADDR, 402 upper_32_bits(src_phys_addr)); 403 404 src_crc32 = crc32_le(~0, src_addr, size); 405 406 orig_dst_addr = kzalloc(size + alignment, GFP_KERNEL); 407 if (!orig_dst_addr) { 408 dev_err(dev, "Failed to allocate destination address\n"); 409 ret = false; 410 goto err_dst_addr; 411 } 412 413 orig_dst_phys_addr = dma_map_single(dev, orig_dst_addr, 414 size + alignment, DMA_FROM_DEVICE); 415 if (dma_mapping_error(dev, orig_dst_phys_addr)) { 416 dev_err(dev, "failed to map destination buffer address\n"); 417 ret = false; 418 goto err_dst_phys_addr; 419 } 420 421 if (alignment && !IS_ALIGNED(orig_dst_phys_addr, alignment)) { 422 dst_phys_addr = PTR_ALIGN(orig_dst_phys_addr, alignment); 423 offset = dst_phys_addr - orig_dst_phys_addr; 424 dst_addr = orig_dst_addr + offset; 425 } else { 426 dst_phys_addr = orig_dst_phys_addr; 427 dst_addr = orig_dst_addr; 428 } 429 430 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_LOWER_DST_ADDR, 431 lower_32_bits(dst_phys_addr)); 432 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_UPPER_DST_ADDR, 433 upper_32_bits(dst_phys_addr)); 434 435 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_SIZE, 436 size); 437 438 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_FLAGS, flags); 439 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_TYPE, irq_type); 440 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_NUMBER, 1); 441 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_COMMAND, 442 COMMAND_COPY); 443 444 wait_for_completion(&test->irq_raised); 445 446 dma_unmap_single(dev, orig_dst_phys_addr, size + alignment, 447 DMA_FROM_DEVICE); 448 449 dst_crc32 = crc32_le(~0, dst_addr, size); 450 if (dst_crc32 == src_crc32) 451 ret = true; 452 453err_dst_phys_addr: 454 kfree(orig_dst_addr); 455 456err_dst_addr: 457 dma_unmap_single(dev, orig_src_phys_addr, size + alignment, 458 DMA_TO_DEVICE); 459 460err_src_phys_addr: 461 kfree(orig_src_addr); 462 463err: 464 return ret; 465} 466 467static bool pci_endpoint_test_write(struct pci_endpoint_test *test, 468 unsigned long arg) 469{ 470 struct pci_endpoint_test_xfer_param param; 471 bool ret = false; 472 u32 flags = 0; 473 bool use_dma; 474 u32 reg; 475 void *addr; 476 dma_addr_t phys_addr; 477 struct pci_dev *pdev = test->pdev; 478 struct device *dev = &pdev->dev; 479 void *orig_addr; 480 dma_addr_t orig_phys_addr; 481 size_t offset; 482 size_t alignment = test->alignment; 483 int irq_type = test->irq_type; 484 size_t size; 485 u32 crc32; 486 int err; 487 488 err = copy_from_user(&param, (void __user *)arg, sizeof(param)); 489 if (err != 0) { 490 dev_err(dev, "Failed to get transfer param\n"); 491 return false; 492 } 493 494 size = param.size; 495 if (size > SIZE_MAX - alignment) 496 goto err; 497 498 use_dma = !!(param.flags & PCITEST_FLAGS_USE_DMA); 499 if (use_dma) 500 flags |= FLAG_USE_DMA; 501 502 if (irq_type < IRQ_TYPE_LEGACY || irq_type > IRQ_TYPE_MSIX) { 503 dev_err(dev, "Invalid IRQ type option\n"); 504 goto err; 505 } 506 507 orig_addr = kzalloc(size + alignment, GFP_KERNEL); 508 if (!orig_addr) { 509 dev_err(dev, "Failed to allocate address\n"); 510 ret = false; 511 goto err; 512 } 513 514 get_random_bytes(orig_addr, size + alignment); 515 516 orig_phys_addr = dma_map_single(dev, orig_addr, size + alignment, 517 DMA_TO_DEVICE); 518 if (dma_mapping_error(dev, orig_phys_addr)) { 519 dev_err(dev, "failed to map source buffer address\n"); 520 ret = false; 521 goto err_phys_addr; 522 } 523 524 if (alignment && !IS_ALIGNED(orig_phys_addr, alignment)) { 525 phys_addr = PTR_ALIGN(orig_phys_addr, alignment); 526 offset = phys_addr - orig_phys_addr; 527 addr = orig_addr + offset; 528 } else { 529 phys_addr = orig_phys_addr; 530 addr = orig_addr; 531 } 532 533 crc32 = crc32_le(~0, addr, size); 534 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_CHECKSUM, 535 crc32); 536 537 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_LOWER_SRC_ADDR, 538 lower_32_bits(phys_addr)); 539 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_UPPER_SRC_ADDR, 540 upper_32_bits(phys_addr)); 541 542 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_SIZE, size); 543 544 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_FLAGS, flags); 545 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_TYPE, irq_type); 546 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_NUMBER, 1); 547 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_COMMAND, 548 COMMAND_READ); 549 550 wait_for_completion(&test->irq_raised); 551 552 reg = pci_endpoint_test_readl(test, PCI_ENDPOINT_TEST_STATUS); 553 if (reg & STATUS_READ_SUCCESS) 554 ret = true; 555 556 dma_unmap_single(dev, orig_phys_addr, size + alignment, 557 DMA_TO_DEVICE); 558 559err_phys_addr: 560 kfree(orig_addr); 561 562err: 563 return ret; 564} 565 566static bool pci_endpoint_test_read(struct pci_endpoint_test *test, 567 unsigned long arg) 568{ 569 struct pci_endpoint_test_xfer_param param; 570 bool ret = false; 571 u32 flags = 0; 572 bool use_dma; 573 size_t size; 574 void *addr; 575 dma_addr_t phys_addr; 576 struct pci_dev *pdev = test->pdev; 577 struct device *dev = &pdev->dev; 578 void *orig_addr; 579 dma_addr_t orig_phys_addr; 580 size_t offset; 581 size_t alignment = test->alignment; 582 int irq_type = test->irq_type; 583 u32 crc32; 584 int err; 585 586 err = copy_from_user(&param, (void __user *)arg, sizeof(param)); 587 if (err) { 588 dev_err(dev, "Failed to get transfer param\n"); 589 return false; 590 } 591 592 size = param.size; 593 if (size > SIZE_MAX - alignment) 594 goto err; 595 596 use_dma = !!(param.flags & PCITEST_FLAGS_USE_DMA); 597 if (use_dma) 598 flags |= FLAG_USE_DMA; 599 600 if (irq_type < IRQ_TYPE_LEGACY || irq_type > IRQ_TYPE_MSIX) { 601 dev_err(dev, "Invalid IRQ type option\n"); 602 goto err; 603 } 604 605 orig_addr = kzalloc(size + alignment, GFP_KERNEL); 606 if (!orig_addr) { 607 dev_err(dev, "Failed to allocate destination address\n"); 608 ret = false; 609 goto err; 610 } 611 612 orig_phys_addr = dma_map_single(dev, orig_addr, size + alignment, 613 DMA_FROM_DEVICE); 614 if (dma_mapping_error(dev, orig_phys_addr)) { 615 dev_err(dev, "failed to map source buffer address\n"); 616 ret = false; 617 goto err_phys_addr; 618 } 619 620 if (alignment && !IS_ALIGNED(orig_phys_addr, alignment)) { 621 phys_addr = PTR_ALIGN(orig_phys_addr, alignment); 622 offset = phys_addr - orig_phys_addr; 623 addr = orig_addr + offset; 624 } else { 625 phys_addr = orig_phys_addr; 626 addr = orig_addr; 627 } 628 629 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_LOWER_DST_ADDR, 630 lower_32_bits(phys_addr)); 631 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_UPPER_DST_ADDR, 632 upper_32_bits(phys_addr)); 633 634 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_SIZE, size); 635 636 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_FLAGS, flags); 637 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_TYPE, irq_type); 638 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_NUMBER, 1); 639 pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_COMMAND, 640 COMMAND_WRITE); 641 642 wait_for_completion(&test->irq_raised); 643 644 dma_unmap_single(dev, orig_phys_addr, size + alignment, 645 DMA_FROM_DEVICE); 646 647 crc32 = crc32_le(~0, addr, size); 648 if (crc32 == pci_endpoint_test_readl(test, PCI_ENDPOINT_TEST_CHECKSUM)) 649 ret = true; 650 651err_phys_addr: 652 kfree(orig_addr); 653err: 654 return ret; 655} 656 657static bool pci_endpoint_test_clear_irq(struct pci_endpoint_test *test) 658{ 659 pci_endpoint_test_release_irq(test); 660 pci_endpoint_test_free_irq_vectors(test); 661 return true; 662} 663 664static bool pci_endpoint_test_set_irq(struct pci_endpoint_test *test, 665 int req_irq_type) 666{ 667 struct pci_dev *pdev = test->pdev; 668 struct device *dev = &pdev->dev; 669 670 if (req_irq_type < IRQ_TYPE_LEGACY || req_irq_type > IRQ_TYPE_MSIX) { 671 dev_err(dev, "Invalid IRQ type option\n"); 672 return false; 673 } 674 675 if (test->irq_type == req_irq_type) 676 return true; 677 678 pci_endpoint_test_release_irq(test); 679 pci_endpoint_test_free_irq_vectors(test); 680 681 if (!pci_endpoint_test_alloc_irq_vectors(test, req_irq_type)) 682 goto err; 683 684 if (!pci_endpoint_test_request_irq(test)) 685 goto err; 686 687 return true; 688 689err: 690 pci_endpoint_test_free_irq_vectors(test); 691 return false; 692} 693 694static long pci_endpoint_test_ioctl(struct file *file, unsigned int cmd, 695 unsigned long arg) 696{ 697 int ret = -EINVAL; 698 enum pci_barno bar; 699 struct pci_endpoint_test *test = to_endpoint_test(file->private_data); 700 struct pci_dev *pdev = test->pdev; 701 702 mutex_lock(&test->mutex); 703 switch (cmd) { 704 case PCITEST_BAR: 705 bar = arg; 706 if (bar < 0 || bar > 5) 707 goto ret; 708 if (is_am654_pci_dev(pdev) && bar == BAR_0) 709 goto ret; 710 ret = pci_endpoint_test_bar(test, bar); 711 break; 712 case PCITEST_LEGACY_IRQ: 713 ret = pci_endpoint_test_legacy_irq(test); 714 break; 715 case PCITEST_MSI: 716 case PCITEST_MSIX: 717 ret = pci_endpoint_test_msi_irq(test, arg, cmd == PCITEST_MSIX); 718 break; 719 case PCITEST_WRITE: 720 ret = pci_endpoint_test_write(test, arg); 721 break; 722 case PCITEST_READ: 723 ret = pci_endpoint_test_read(test, arg); 724 break; 725 case PCITEST_COPY: 726 ret = pci_endpoint_test_copy(test, arg); 727 break; 728 case PCITEST_SET_IRQTYPE: 729 ret = pci_endpoint_test_set_irq(test, arg); 730 break; 731 case PCITEST_GET_IRQTYPE: 732 ret = irq_type; 733 break; 734 case PCITEST_CLEAR_IRQ: 735 ret = pci_endpoint_test_clear_irq(test); 736 break; 737 } 738 739ret: 740 mutex_unlock(&test->mutex); 741 return ret; 742} 743 744static const struct file_operations pci_endpoint_test_fops = { 745 .owner = THIS_MODULE, 746 .unlocked_ioctl = pci_endpoint_test_ioctl, 747}; 748 749static int pci_endpoint_test_probe(struct pci_dev *pdev, 750 const struct pci_device_id *ent) 751{ 752 int err; 753 int id; 754 char name[24]; 755 enum pci_barno bar; 756 void __iomem *base; 757 struct device *dev = &pdev->dev; 758 struct pci_endpoint_test *test; 759 struct pci_endpoint_test_data *data; 760 enum pci_barno test_reg_bar = BAR_0; 761 struct miscdevice *misc_device; 762 763 if (pci_is_bridge(pdev)) 764 return -ENODEV; 765 766 test = devm_kzalloc(dev, sizeof(*test), GFP_KERNEL); 767 if (!test) 768 return -ENOMEM; 769 770 test->test_reg_bar = 0; 771 test->alignment = 0; 772 test->pdev = pdev; 773 test->irq_type = IRQ_TYPE_UNDEFINED; 774 775 if (no_msi) 776 irq_type = IRQ_TYPE_LEGACY; 777 778 data = (struct pci_endpoint_test_data *)ent->driver_data; 779 if (data) { 780 test_reg_bar = data->test_reg_bar; 781 test->test_reg_bar = test_reg_bar; 782 test->alignment = data->alignment; 783 irq_type = data->irq_type; 784 } 785 786 init_completion(&test->irq_raised); 787 mutex_init(&test->mutex); 788 789 if ((dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(48)) != 0) && 790 dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(32)) != 0) { 791 dev_err(dev, "Cannot set DMA mask\n"); 792 return -EINVAL; 793 } 794 795 err = pci_enable_device(pdev); 796 if (err) { 797 dev_err(dev, "Cannot enable PCI device\n"); 798 return err; 799 } 800 801 err = pci_request_regions(pdev, DRV_MODULE_NAME); 802 if (err) { 803 dev_err(dev, "Cannot obtain PCI resources\n"); 804 goto err_disable_pdev; 805 } 806 807 pci_set_master(pdev); 808 809 if (!pci_endpoint_test_alloc_irq_vectors(test, irq_type)) 810 goto err_disable_irq; 811 812 for (bar = 0; bar < PCI_STD_NUM_BARS; bar++) { 813 if (pci_resource_flags(pdev, bar) & IORESOURCE_MEM) { 814 base = pci_ioremap_bar(pdev, bar); 815 if (!base) { 816 dev_err(dev, "Failed to read BAR%d\n", bar); 817 WARN_ON(bar == test_reg_bar); 818 } 819 test->bar[bar] = base; 820 } 821 } 822 823 test->base = test->bar[test_reg_bar]; 824 if (!test->base) { 825 err = -ENOMEM; 826 dev_err(dev, "Cannot perform PCI test without BAR%d\n", 827 test_reg_bar); 828 goto err_iounmap; 829 } 830 831 pci_set_drvdata(pdev, test); 832 833 id = ida_simple_get(&pci_endpoint_test_ida, 0, 0, GFP_KERNEL); 834 if (id < 0) { 835 err = id; 836 dev_err(dev, "Unable to get id\n"); 837 goto err_iounmap; 838 } 839 840 snprintf(name, sizeof(name), DRV_MODULE_NAME ".%d", id); 841 test->name = kstrdup(name, GFP_KERNEL); 842 if (!test->name) { 843 err = -ENOMEM; 844 goto err_ida_remove; 845 } 846 847 if (!pci_endpoint_test_request_irq(test)) 848 goto err_kfree_test_name; 849 850 misc_device = &test->miscdev; 851 misc_device->minor = MISC_DYNAMIC_MINOR; 852 misc_device->name = kstrdup(name, GFP_KERNEL); 853 if (!misc_device->name) { 854 err = -ENOMEM; 855 goto err_release_irq; 856 } 857 misc_device->fops = &pci_endpoint_test_fops, 858 859 err = misc_register(misc_device); 860 if (err) { 861 dev_err(dev, "Failed to register device\n"); 862 goto err_kfree_name; 863 } 864 865 return 0; 866 867err_kfree_name: 868 kfree(misc_device->name); 869 870err_release_irq: 871 pci_endpoint_test_release_irq(test); 872 873err_kfree_test_name: 874 kfree(test->name); 875 876err_ida_remove: 877 ida_simple_remove(&pci_endpoint_test_ida, id); 878 879err_iounmap: 880 for (bar = 0; bar < PCI_STD_NUM_BARS; bar++) { 881 if (test->bar[bar]) 882 pci_iounmap(pdev, test->bar[bar]); 883 } 884 885err_disable_irq: 886 pci_endpoint_test_free_irq_vectors(test); 887 pci_release_regions(pdev); 888 889err_disable_pdev: 890 pci_disable_device(pdev); 891 892 return err; 893} 894 895static void pci_endpoint_test_remove(struct pci_dev *pdev) 896{ 897 int id; 898 enum pci_barno bar; 899 struct pci_endpoint_test *test = pci_get_drvdata(pdev); 900 struct miscdevice *misc_device = &test->miscdev; 901 902 if (sscanf(misc_device->name, DRV_MODULE_NAME ".%d", &id) != 1) 903 return; 904 if (id < 0) 905 return; 906 907 misc_deregister(&test->miscdev); 908 kfree(misc_device->name); 909 kfree(test->name); 910 ida_simple_remove(&pci_endpoint_test_ida, id); 911 for (bar = 0; bar < PCI_STD_NUM_BARS; bar++) { 912 if (test->bar[bar]) 913 pci_iounmap(pdev, test->bar[bar]); 914 } 915 916 pci_endpoint_test_release_irq(test); 917 pci_endpoint_test_free_irq_vectors(test); 918 919 pci_release_regions(pdev); 920 pci_disable_device(pdev); 921} 922 923static const struct pci_endpoint_test_data default_data = { 924 .test_reg_bar = BAR_0, 925 .alignment = SZ_4K, 926 .irq_type = IRQ_TYPE_MSI, 927}; 928 929static const struct pci_endpoint_test_data am654_data = { 930 .test_reg_bar = BAR_2, 931 .alignment = SZ_64K, 932 .irq_type = IRQ_TYPE_MSI, 933}; 934 935static const struct pci_device_id pci_endpoint_test_tbl[] = { 936 { PCI_DEVICE(PCI_VENDOR_ID_TI, PCI_DEVICE_ID_TI_DRA74x), 937 .driver_data = (kernel_ulong_t)&default_data, 938 }, 939 { PCI_DEVICE(PCI_VENDOR_ID_TI, PCI_DEVICE_ID_TI_DRA72x), 940 .driver_data = (kernel_ulong_t)&default_data, 941 }, 942 { PCI_DEVICE(PCI_VENDOR_ID_FREESCALE, 0x81c0) }, 943 { PCI_DEVICE_DATA(SYNOPSYS, EDDA, NULL) }, 944 { PCI_DEVICE(PCI_VENDOR_ID_TI, PCI_DEVICE_ID_TI_AM654), 945 .driver_data = (kernel_ulong_t)&am654_data 946 }, 947 { PCI_DEVICE(PCI_VENDOR_ID_RENESAS, PCI_DEVICE_ID_RENESAS_R8A774C0), 948 }, 949 { } 950}; 951MODULE_DEVICE_TABLE(pci, pci_endpoint_test_tbl); 952 953static struct pci_driver pci_endpoint_test_driver = { 954 .name = DRV_MODULE_NAME, 955 .id_table = pci_endpoint_test_tbl, 956 .probe = pci_endpoint_test_probe, 957 .remove = pci_endpoint_test_remove, 958}; 959module_pci_driver(pci_endpoint_test_driver); 960 961MODULE_DESCRIPTION("PCI ENDPOINT TEST HOST DRIVER"); 962MODULE_AUTHOR("Kishon Vijay Abraham I <kishon@ti.com>"); 963MODULE_LICENSE("GPL v2");