1 #define pr_fmt(fmt) "irq: " fmt 2 3 #include <linux/debugfs.h> 4 #include <linux/hardirq.h> 5 #include <linux/interrupt.h> 6 #include <linux/irq.h> 7 #include <linux/irqdesc.h> 8 #include <linux/irqdomain.h> 9 #include <linux/module.h> 10 #include <linux/mutex.h> 11 #include <linux/of.h> 12 #include <linux/of_address.h> 13 #include <linux/topology.h> 14 #include <linux/seq_file.h> 15 #include <linux/slab.h> 16 #include <linux/smp.h> 17 #include <linux/fs.h> 18 19 static LIST_HEAD(irq_domain_list); 20 static DEFINE_MUTEX(irq_domain_mutex); 21 22 static DEFINE_MUTEX(revmap_trees_mutex); 23 static struct irq_domain *irq_default_domain; 24 25 /** 26 * __irq_domain_add() - Allocate a new irq_domain data structure 27 * @of_node: optional device-tree node of the interrupt controller 28 * @size: Size of linear map; 0 for radix mapping only 29 * @direct_max: Maximum value of direct maps; Use ~0 for no limit; 0 for no 30 * direct mapping 31 * @ops: map/unmap domain callbacks 32 * @host_data: Controller private data pointer 33 * 34 * Allocates and initialize and irq_domain structure. Caller is expected to 35 * register allocated irq_domain with irq_domain_register(). Returns pointer 36 * to IRQ domain, or NULL on failure. 37 */ 38 struct irq_domain *__irq_domain_add(struct device_node *of_node, int size, 39 irq_hw_number_t hwirq_max, int direct_max, 40 const struct irq_domain_ops *ops, 41 void *host_data) 42 { 43 struct irq_domain *domain; 44 45 domain = kzalloc_node(sizeof(*domain) + (sizeof(unsigned int) * size), 46 GFP_KERNEL, of_node_to_nid(of_node)); 47 if (WARN_ON(!domain)) 48 return NULL; 49 50 /* Fill structure */ 51 INIT_RADIX_TREE(&domain->revmap_tree, GFP_KERNEL); 52 domain->ops = ops; 53 domain->host_data = host_data; 54 domain->of_node = of_node_get(of_node); 55 domain->hwirq_max = hwirq_max; 56 domain->revmap_size = size; 57 domain->revmap_direct_max_irq = direct_max; 58 59 mutex_lock(&irq_domain_mutex); 60 list_add(&domain->link, &irq_domain_list); 61 mutex_unlock(&irq_domain_mutex); 62 63 pr_debug("Added domain %s\n", domain->name); 64 return domain; 65 } 66 EXPORT_SYMBOL_GPL(__irq_domain_add); 67 68 /** 69 * irq_domain_remove() - Remove an irq domain. 70 * @domain: domain to remove 71 * 72 * This routine is used to remove an irq domain. The caller must ensure 73 * that all mappings within the domain have been disposed of prior to 74 * use, depending on the revmap type. 75 */ 76 void irq_domain_remove(struct irq_domain *domain) 77 { 78 mutex_lock(&irq_domain_mutex); 79 80 /* 81 * radix_tree_delete() takes care of destroying the root 82 * node when all entries are removed. Shout if there are 83 * any mappings left. 84 */ 85 WARN_ON(domain->revmap_tree.height); 86 87 list_del(&domain->link); 88 89 /* 90 * If the going away domain is the default one, reset it. 91 */ 92 if (unlikely(irq_default_domain == domain)) 93 irq_set_default_host(NULL); 94 95 mutex_unlock(&irq_domain_mutex); 96 97 pr_debug("Removed domain %s\n", domain->name); 98 99 of_node_put(domain->of_node); 100 kfree(domain); 101 } 102 EXPORT_SYMBOL_GPL(irq_domain_remove); 103 104 /** 105 * irq_domain_add_simple() - Register an irq_domain and optionally map a range of irqs 106 * @of_node: pointer to interrupt controller's device tree node. 107 * @size: total number of irqs in mapping 108 * @first_irq: first number of irq block assigned to the domain, 109 * pass zero to assign irqs on-the-fly. If first_irq is non-zero, then 110 * pre-map all of the irqs in the domain to virqs starting at first_irq. 111 * @ops: map/unmap domain callbacks 112 * @host_data: Controller private data pointer 113 * 114 * Allocates an irq_domain, and optionally if first_irq is positive then also 115 * allocate irq_descs and map all of the hwirqs to virqs starting at first_irq. 116 * 117 * This is intended to implement the expected behaviour for most 118 * interrupt controllers. If device tree is used, then first_irq will be 0 and 119 * irqs get mapped dynamically on the fly. However, if the controller requires 120 * static virq assignments (non-DT boot) then it will set that up correctly. 121 */ 122 struct irq_domain *irq_domain_add_simple(struct device_node *of_node, 123 unsigned int size, 124 unsigned int first_irq, 125 const struct irq_domain_ops *ops, 126 void *host_data) 127 { 128 struct irq_domain *domain; 129 130 domain = __irq_domain_add(of_node, size, size, 0, ops, host_data); 131 if (!domain) 132 return NULL; 133 134 if (first_irq > 0) { 135 if (IS_ENABLED(CONFIG_SPARSE_IRQ)) { 136 /* attempt to allocated irq_descs */ 137 int rc = irq_alloc_descs(first_irq, first_irq, size, 138 of_node_to_nid(of_node)); 139 if (rc < 0) 140 pr_info("Cannot allocate irq_descs @ IRQ%d, assuming pre-allocated\n", 141 first_irq); 142 } 143 irq_domain_associate_many(domain, first_irq, 0, size); 144 } 145 146 return domain; 147 } 148 EXPORT_SYMBOL_GPL(irq_domain_add_simple); 149 150 /** 151 * irq_domain_add_legacy() - Allocate and register a legacy revmap irq_domain. 152 * @of_node: pointer to interrupt controller's device tree node. 153 * @size: total number of irqs in legacy mapping 154 * @first_irq: first number of irq block assigned to the domain 155 * @first_hwirq: first hwirq number to use for the translation. Should normally 156 * be '0', but a positive integer can be used if the effective 157 * hwirqs numbering does not begin at zero. 158 * @ops: map/unmap domain callbacks 159 * @host_data: Controller private data pointer 160 * 161 * Note: the map() callback will be called before this function returns 162 * for all legacy interrupts except 0 (which is always the invalid irq for 163 * a legacy controller). 164 */ 165 struct irq_domain *irq_domain_add_legacy(struct device_node *of_node, 166 unsigned int size, 167 unsigned int first_irq, 168 irq_hw_number_t first_hwirq, 169 const struct irq_domain_ops *ops, 170 void *host_data) 171 { 172 struct irq_domain *domain; 173 174 domain = __irq_domain_add(of_node, first_hwirq + size, 175 first_hwirq + size, 0, ops, host_data); 176 if (!domain) 177 return NULL; 178 179 irq_domain_associate_many(domain, first_irq, first_hwirq, size); 180 181 return domain; 182 } 183 EXPORT_SYMBOL_GPL(irq_domain_add_legacy); 184 185 /** 186 * irq_find_host() - Locates a domain for a given device node 187 * @node: device-tree node of the interrupt controller 188 */ 189 struct irq_domain *irq_find_host(struct device_node *node) 190 { 191 struct irq_domain *h, *found = NULL; 192 int rc; 193 194 /* We might want to match the legacy controller last since 195 * it might potentially be set to match all interrupts in 196 * the absence of a device node. This isn't a problem so far 197 * yet though... 198 */ 199 mutex_lock(&irq_domain_mutex); 200 list_for_each_entry(h, &irq_domain_list, link) { 201 if (h->ops->match) 202 rc = h->ops->match(h, node); 203 else 204 rc = (h->of_node != NULL) && (h->of_node == node); 205 206 if (rc) { 207 found = h; 208 break; 209 } 210 } 211 mutex_unlock(&irq_domain_mutex); 212 return found; 213 } 214 EXPORT_SYMBOL_GPL(irq_find_host); 215 216 /** 217 * irq_set_default_host() - Set a "default" irq domain 218 * @domain: default domain pointer 219 * 220 * For convenience, it's possible to set a "default" domain that will be used 221 * whenever NULL is passed to irq_create_mapping(). It makes life easier for 222 * platforms that want to manipulate a few hard coded interrupt numbers that 223 * aren't properly represented in the device-tree. 224 */ 225 void irq_set_default_host(struct irq_domain *domain) 226 { 227 pr_debug("Default domain set to @0x%p\n", domain); 228 229 irq_default_domain = domain; 230 } 231 EXPORT_SYMBOL_GPL(irq_set_default_host); 232 233 static void irq_domain_disassociate(struct irq_domain *domain, unsigned int irq) 234 { 235 struct irq_data *irq_data = irq_get_irq_data(irq); 236 irq_hw_number_t hwirq; 237 238 if (WARN(!irq_data || irq_data->domain != domain, 239 "virq%i doesn't exist; cannot disassociate\n", irq)) 240 return; 241 242 hwirq = irq_data->hwirq; 243 irq_set_status_flags(irq, IRQ_NOREQUEST); 244 245 /* remove chip and handler */ 246 irq_set_chip_and_handler(irq, NULL, NULL); 247 248 /* Make sure it's completed */ 249 synchronize_irq(irq); 250 251 /* Tell the PIC about it */ 252 if (domain->ops->unmap) 253 domain->ops->unmap(domain, irq); 254 smp_mb(); 255 256 irq_data->domain = NULL; 257 irq_data->hwirq = 0; 258 259 /* Clear reverse map for this hwirq */ 260 if (hwirq < domain->revmap_size) { 261 domain->linear_revmap[hwirq] = 0; 262 } else { 263 mutex_lock(&revmap_trees_mutex); 264 radix_tree_delete(&domain->revmap_tree, hwirq); 265 mutex_unlock(&revmap_trees_mutex); 266 } 267 } 268 269 int irq_domain_associate(struct irq_domain *domain, unsigned int virq, 270 irq_hw_number_t hwirq) 271 { 272 struct irq_data *irq_data = irq_get_irq_data(virq); 273 int ret; 274 275 if (WARN(hwirq >= domain->hwirq_max, 276 "error: hwirq 0x%x is too large for %s\n", (int)hwirq, domain->name)) 277 return -EINVAL; 278 if (WARN(!irq_data, "error: virq%i is not allocated", virq)) 279 return -EINVAL; 280 if (WARN(irq_data->domain, "error: virq%i is already associated", virq)) 281 return -EINVAL; 282 283 mutex_lock(&irq_domain_mutex); 284 irq_data->hwirq = hwirq; 285 irq_data->domain = domain; 286 if (domain->ops->map) { 287 ret = domain->ops->map(domain, virq, hwirq); 288 if (ret != 0) { 289 /* 290 * If map() returns -EPERM, this interrupt is protected 291 * by the firmware or some other service and shall not 292 * be mapped. Don't bother telling the user about it. 293 */ 294 if (ret != -EPERM) { 295 pr_info("%s didn't like hwirq-0x%lx to VIRQ%i mapping (rc=%d)\n", 296 domain->name, hwirq, virq, ret); 297 } 298 irq_data->domain = NULL; 299 irq_data->hwirq = 0; 300 mutex_unlock(&irq_domain_mutex); 301 return ret; 302 } 303 304 /* If not already assigned, give the domain the chip's name */ 305 if (!domain->name && irq_data->chip) 306 domain->name = irq_data->chip->name; 307 } 308 309 if (hwirq < domain->revmap_size) { 310 domain->linear_revmap[hwirq] = virq; 311 } else { 312 mutex_lock(&revmap_trees_mutex); 313 radix_tree_insert(&domain->revmap_tree, hwirq, irq_data); 314 mutex_unlock(&revmap_trees_mutex); 315 } 316 mutex_unlock(&irq_domain_mutex); 317 318 irq_clear_status_flags(virq, IRQ_NOREQUEST); 319 320 return 0; 321 } 322 EXPORT_SYMBOL_GPL(irq_domain_associate); 323 324 void irq_domain_associate_many(struct irq_domain *domain, unsigned int irq_base, 325 irq_hw_number_t hwirq_base, int count) 326 { 327 int i; 328 329 pr_debug("%s(%s, irqbase=%i, hwbase=%i, count=%i)\n", __func__, 330 of_node_full_name(domain->of_node), irq_base, (int)hwirq_base, count); 331 332 for (i = 0; i < count; i++) { 333 irq_domain_associate(domain, irq_base + i, hwirq_base + i); 334 } 335 } 336 EXPORT_SYMBOL_GPL(irq_domain_associate_many); 337 338 /** 339 * irq_create_direct_mapping() - Allocate an irq for direct mapping 340 * @domain: domain to allocate the irq for or NULL for default domain 341 * 342 * This routine is used for irq controllers which can choose the hardware 343 * interrupt numbers they generate. In such a case it's simplest to use 344 * the linux irq as the hardware interrupt number. It still uses the linear 345 * or radix tree to store the mapping, but the irq controller can optimize 346 * the revmap path by using the hwirq directly. 347 */ 348 unsigned int irq_create_direct_mapping(struct irq_domain *domain) 349 { 350 unsigned int virq; 351 352 if (domain == NULL) 353 domain = irq_default_domain; 354 355 virq = irq_alloc_desc_from(1, of_node_to_nid(domain->of_node)); 356 if (!virq) { 357 pr_debug("create_direct virq allocation failed\n"); 358 return 0; 359 } 360 if (virq >= domain->revmap_direct_max_irq) { 361 pr_err("ERROR: no free irqs available below %i maximum\n", 362 domain->revmap_direct_max_irq); 363 irq_free_desc(virq); 364 return 0; 365 } 366 pr_debug("create_direct obtained virq %d\n", virq); 367 368 if (irq_domain_associate(domain, virq, virq)) { 369 irq_free_desc(virq); 370 return 0; 371 } 372 373 return virq; 374 } 375 EXPORT_SYMBOL_GPL(irq_create_direct_mapping); 376 377 /** 378 * irq_create_mapping() - Map a hardware interrupt into linux irq space 379 * @domain: domain owning this hardware interrupt or NULL for default domain 380 * @hwirq: hardware irq number in that domain space 381 * 382 * Only one mapping per hardware interrupt is permitted. Returns a linux 383 * irq number. 384 * If the sense/trigger is to be specified, set_irq_type() should be called 385 * on the number returned from that call. 386 */ 387 unsigned int irq_create_mapping(struct irq_domain *domain, 388 irq_hw_number_t hwirq) 389 { 390 unsigned int hint; 391 int virq; 392 393 pr_debug("irq_create_mapping(0x%p, 0x%lx)\n", domain, hwirq); 394 395 /* Look for default domain if nececssary */ 396 if (domain == NULL) 397 domain = irq_default_domain; 398 if (domain == NULL) { 399 WARN(1, "%s(, %lx) called with NULL domain\n", __func__, hwirq); 400 return 0; 401 } 402 pr_debug("-> using domain @%p\n", domain); 403 404 /* Check if mapping already exists */ 405 virq = irq_find_mapping(domain, hwirq); 406 if (virq) { 407 pr_debug("-> existing mapping on virq %d\n", virq); 408 return virq; 409 } 410 411 /* Allocate a virtual interrupt number */ 412 hint = hwirq % nr_irqs; 413 if (hint == 0) 414 hint++; 415 virq = irq_alloc_desc_from(hint, of_node_to_nid(domain->of_node)); 416 if (virq <= 0) 417 virq = irq_alloc_desc_from(1, of_node_to_nid(domain->of_node)); 418 if (virq <= 0) { 419 pr_debug("-> virq allocation failed\n"); 420 return 0; 421 } 422 423 if (irq_domain_associate(domain, virq, hwirq)) { 424 irq_free_desc(virq); 425 return 0; 426 } 427 428 pr_debug("irq %lu on domain %s mapped to virtual irq %u\n", 429 hwirq, of_node_full_name(domain->of_node), virq); 430 431 return virq; 432 } 433 EXPORT_SYMBOL_GPL(irq_create_mapping); 434 435 /** 436 * irq_create_strict_mappings() - Map a range of hw irqs to fixed linux irqs 437 * @domain: domain owning the interrupt range 438 * @irq_base: beginning of linux IRQ range 439 * @hwirq_base: beginning of hardware IRQ range 440 * @count: Number of interrupts to map 441 * 442 * This routine is used for allocating and mapping a range of hardware 443 * irqs to linux irqs where the linux irq numbers are at pre-defined 444 * locations. For use by controllers that already have static mappings 445 * to insert in to the domain. 446 * 447 * Non-linear users can use irq_create_identity_mapping() for IRQ-at-a-time 448 * domain insertion. 449 * 450 * 0 is returned upon success, while any failure to establish a static 451 * mapping is treated as an error. 452 */ 453 int irq_create_strict_mappings(struct irq_domain *domain, unsigned int irq_base, 454 irq_hw_number_t hwirq_base, int count) 455 { 456 int ret; 457 458 ret = irq_alloc_descs(irq_base, irq_base, count, 459 of_node_to_nid(domain->of_node)); 460 if (unlikely(ret < 0)) 461 return ret; 462 463 irq_domain_associate_many(domain, irq_base, hwirq_base, count); 464 return 0; 465 } 466 EXPORT_SYMBOL_GPL(irq_create_strict_mappings); 467 468 unsigned int irq_create_of_mapping(struct device_node *controller, 469 const u32 *intspec, unsigned int intsize) 470 { 471 struct irq_domain *domain; 472 irq_hw_number_t hwirq; 473 unsigned int type = IRQ_TYPE_NONE; 474 unsigned int virq; 475 476 domain = controller ? irq_find_host(controller) : irq_default_domain; 477 if (!domain) { 478 #ifdef CONFIG_MIPS 479 /* 480 * Workaround to avoid breaking interrupt controller drivers 481 * that don't yet register an irq_domain. This is temporary 482 * code. ~~~gcl, Feb 24, 2012 483 * 484 * Scheduled for removal in Linux v3.6. That should be enough 485 * time. 486 */ 487 if (intsize > 0) 488 return intspec[0]; 489 #endif 490 pr_warn("no irq domain found for %s !\n", 491 of_node_full_name(controller)); 492 return 0; 493 } 494 495 /* If domain has no translation, then we assume interrupt line */ 496 if (domain->ops->xlate == NULL) 497 hwirq = intspec[0]; 498 else { 499 if (domain->ops->xlate(domain, controller, intspec, intsize, 500 &hwirq, &type)) 501 return 0; 502 } 503 504 /* Create mapping */ 505 virq = irq_create_mapping(domain, hwirq); 506 if (!virq) 507 return virq; 508 509 /* Set type if specified and different than the current one */ 510 if (type != IRQ_TYPE_NONE && 511 type != irq_get_trigger_type(virq)) 512 irq_set_irq_type(virq, type); 513 return virq; 514 } 515 EXPORT_SYMBOL_GPL(irq_create_of_mapping); 516 517 /** 518 * irq_dispose_mapping() - Unmap an interrupt 519 * @virq: linux irq number of the interrupt to unmap 520 */ 521 void irq_dispose_mapping(unsigned int virq) 522 { 523 struct irq_data *irq_data = irq_get_irq_data(virq); 524 struct irq_domain *domain; 525 526 if (!virq || !irq_data) 527 return; 528 529 domain = irq_data->domain; 530 if (WARN_ON(domain == NULL)) 531 return; 532 533 irq_domain_disassociate(domain, virq); 534 irq_free_desc(virq); 535 } 536 EXPORT_SYMBOL_GPL(irq_dispose_mapping); 537 538 /** 539 * irq_find_mapping() - Find a linux irq from an hw irq number. 540 * @domain: domain owning this hardware interrupt 541 * @hwirq: hardware irq number in that domain space 542 */ 543 unsigned int irq_find_mapping(struct irq_domain *domain, 544 irq_hw_number_t hwirq) 545 { 546 struct irq_data *data; 547 548 /* Look for default domain if nececssary */ 549 if (domain == NULL) 550 domain = irq_default_domain; 551 if (domain == NULL) 552 return 0; 553 554 if (hwirq < domain->revmap_direct_max_irq) { 555 data = irq_get_irq_data(hwirq); 556 if (data && (data->domain == domain) && (data->hwirq == hwirq)) 557 return hwirq; 558 } 559 560 /* Check if the hwirq is in the linear revmap. */ 561 if (hwirq < domain->revmap_size) 562 return domain->linear_revmap[hwirq]; 563 564 rcu_read_lock(); 565 data = radix_tree_lookup(&domain->revmap_tree, hwirq); 566 rcu_read_unlock(); 567 return data ? data->irq : 0; 568 } 569 EXPORT_SYMBOL_GPL(irq_find_mapping); 570 571 #ifdef CONFIG_IRQ_DOMAIN_DEBUG 572 static int virq_debug_show(struct seq_file *m, void *private) 573 { 574 unsigned long flags; 575 struct irq_desc *desc; 576 struct irq_domain *domain; 577 struct radix_tree_iter iter; 578 void *data, **slot; 579 int i; 580 581 seq_printf(m, " %-16s %-6s %-10s %-10s %s\n", 582 "name", "mapped", "linear-max", "direct-max", "devtree-node"); 583 mutex_lock(&irq_domain_mutex); 584 list_for_each_entry(domain, &irq_domain_list, link) { 585 int count = 0; 586 radix_tree_for_each_slot(slot, &domain->revmap_tree, &iter, 0) 587 count++; 588 seq_printf(m, "%c%-16s %6u %10u %10u %s\n", 589 domain == irq_default_domain ? '*' : ' ', domain->name, 590 domain->revmap_size + count, domain->revmap_size, 591 domain->revmap_direct_max_irq, 592 domain->of_node ? of_node_full_name(domain->of_node) : ""); 593 } 594 mutex_unlock(&irq_domain_mutex); 595 596 seq_printf(m, "%-5s %-7s %-15s %-*s %6s %-14s %s\n", "irq", "hwirq", 597 "chip name", (int)(2 * sizeof(void *) + 2), "chip data", 598 "active", "type", "domain"); 599 600 for (i = 1; i < nr_irqs; i++) { 601 desc = irq_to_desc(i); 602 if (!desc) 603 continue; 604 605 raw_spin_lock_irqsave(&desc->lock, flags); 606 domain = desc->irq_data.domain; 607 608 if (domain) { 609 struct irq_chip *chip; 610 int hwirq = desc->irq_data.hwirq; 611 bool direct; 612 613 seq_printf(m, "%5d ", i); 614 seq_printf(m, "0x%05x ", hwirq); 615 616 chip = irq_desc_get_chip(desc); 617 seq_printf(m, "%-15s ", (chip && chip->name) ? chip->name : "none"); 618 619 data = irq_desc_get_chip_data(desc); 620 seq_printf(m, data ? "0x%p " : " %p ", data); 621 622 seq_printf(m, " %c ", (desc->action && desc->action->handler) ? '*' : ' '); 623 direct = (i == hwirq) && (i < domain->revmap_direct_max_irq); 624 seq_printf(m, "%6s%-8s ", 625 (hwirq < domain->revmap_size) ? "LINEAR" : "RADIX", 626 direct ? "(DIRECT)" : ""); 627 seq_printf(m, "%s\n", desc->irq_data.domain->name); 628 } 629 630 raw_spin_unlock_irqrestore(&desc->lock, flags); 631 } 632 633 return 0; 634 } 635 636 static int virq_debug_open(struct inode *inode, struct file *file) 637 { 638 return single_open(file, virq_debug_show, inode->i_private); 639 } 640 641 static const struct file_operations virq_debug_fops = { 642 .open = virq_debug_open, 643 .read = seq_read, 644 .llseek = seq_lseek, 645 .release = single_release, 646 }; 647 648 static int __init irq_debugfs_init(void) 649 { 650 if (debugfs_create_file("irq_domain_mapping", S_IRUGO, NULL, 651 NULL, &virq_debug_fops) == NULL) 652 return -ENOMEM; 653 654 return 0; 655 } 656 __initcall(irq_debugfs_init); 657 #endif /* CONFIG_IRQ_DOMAIN_DEBUG */ 658 659 /** 660 * irq_domain_xlate_onecell() - Generic xlate for direct one cell bindings 661 * 662 * Device Tree IRQ specifier translation function which works with one cell 663 * bindings where the cell value maps directly to the hwirq number. 664 */ 665 int irq_domain_xlate_onecell(struct irq_domain *d, struct device_node *ctrlr, 666 const u32 *intspec, unsigned int intsize, 667 unsigned long *out_hwirq, unsigned int *out_type) 668 { 669 if (WARN_ON(intsize < 1)) 670 return -EINVAL; 671 *out_hwirq = intspec[0]; 672 *out_type = IRQ_TYPE_NONE; 673 return 0; 674 } 675 EXPORT_SYMBOL_GPL(irq_domain_xlate_onecell); 676 677 /** 678 * irq_domain_xlate_twocell() - Generic xlate for direct two cell bindings 679 * 680 * Device Tree IRQ specifier translation function which works with two cell 681 * bindings where the cell values map directly to the hwirq number 682 * and linux irq flags. 683 */ 684 int irq_domain_xlate_twocell(struct irq_domain *d, struct device_node *ctrlr, 685 const u32 *intspec, unsigned int intsize, 686 irq_hw_number_t *out_hwirq, unsigned int *out_type) 687 { 688 if (WARN_ON(intsize < 2)) 689 return -EINVAL; 690 *out_hwirq = intspec[0]; 691 *out_type = intspec[1] & IRQ_TYPE_SENSE_MASK; 692 return 0; 693 } 694 EXPORT_SYMBOL_GPL(irq_domain_xlate_twocell); 695 696 /** 697 * irq_domain_xlate_onetwocell() - Generic xlate for one or two cell bindings 698 * 699 * Device Tree IRQ specifier translation function which works with either one 700 * or two cell bindings where the cell values map directly to the hwirq number 701 * and linux irq flags. 702 * 703 * Note: don't use this function unless your interrupt controller explicitly 704 * supports both one and two cell bindings. For the majority of controllers 705 * the _onecell() or _twocell() variants above should be used. 706 */ 707 int irq_domain_xlate_onetwocell(struct irq_domain *d, 708 struct device_node *ctrlr, 709 const u32 *intspec, unsigned int intsize, 710 unsigned long *out_hwirq, unsigned int *out_type) 711 { 712 if (WARN_ON(intsize < 1)) 713 return -EINVAL; 714 *out_hwirq = intspec[0]; 715 *out_type = (intsize > 1) ? intspec[1] : IRQ_TYPE_NONE; 716 return 0; 717 } 718 EXPORT_SYMBOL_GPL(irq_domain_xlate_onetwocell); 719 720 const struct irq_domain_ops irq_domain_simple_ops = { 721 .xlate = irq_domain_xlate_onetwocell, 722 }; 723 EXPORT_SYMBOL_GPL(irq_domain_simple_ops); 724