1 /* 2 * linux/kernel/irq/msi.c 3 * 4 * Copyright (C) 2014 Intel Corp. 5 * Author: Jiang Liu <jiang.liu@linux.intel.com> 6 * 7 * This file is licensed under GPLv2. 8 * 9 * This file contains common code to support Message Signalled Interrupt for 10 * PCI compatible and non PCI compatible devices. 11 */ 12 #include <linux/types.h> 13 #include <linux/device.h> 14 #include <linux/irq.h> 15 #include <linux/irqdomain.h> 16 #include <linux/msi.h> 17 #include <linux/slab.h> 18 19 #include "internals.h" 20 21 /** 22 * alloc_msi_entry - Allocate an initialize msi_entry 23 * @dev: Pointer to the device for which this is allocated 24 * @nvec: The number of vectors used in this entry 25 * @affinity: Optional pointer to an affinity mask array size of @nvec 26 * 27 * If @affinity is not NULL then a an affinity array[@nvec] is allocated 28 * and the affinity masks from @affinity are copied. 29 */ 30 struct msi_desc * 31 alloc_msi_entry(struct device *dev, int nvec, const struct cpumask *affinity) 32 { 33 struct msi_desc *desc; 34 35 desc = kzalloc(sizeof(*desc), GFP_KERNEL); 36 if (!desc) 37 return NULL; 38 39 INIT_LIST_HEAD(&desc->list); 40 desc->dev = dev; 41 desc->nvec_used = nvec; 42 if (affinity) { 43 desc->affinity = kmemdup(affinity, 44 nvec * sizeof(*desc->affinity), GFP_KERNEL); 45 if (!desc->affinity) { 46 kfree(desc); 47 return NULL; 48 } 49 } 50 51 return desc; 52 } 53 54 void free_msi_entry(struct msi_desc *entry) 55 { 56 kfree(entry->affinity); 57 kfree(entry); 58 } 59 60 void __get_cached_msi_msg(struct msi_desc *entry, struct msi_msg *msg) 61 { 62 *msg = entry->msg; 63 } 64 65 void get_cached_msi_msg(unsigned int irq, struct msi_msg *msg) 66 { 67 struct msi_desc *entry = irq_get_msi_desc(irq); 68 69 __get_cached_msi_msg(entry, msg); 70 } 71 EXPORT_SYMBOL_GPL(get_cached_msi_msg); 72 73 #ifdef CONFIG_GENERIC_MSI_IRQ_DOMAIN 74 static inline void irq_chip_write_msi_msg(struct irq_data *data, 75 struct msi_msg *msg) 76 { 77 data->chip->irq_write_msi_msg(data, msg); 78 } 79 80 /** 81 * msi_domain_set_affinity - Generic affinity setter function for MSI domains 82 * @irq_data: The irq data associated to the interrupt 83 * @mask: The affinity mask to set 84 * @force: Flag to enforce setting (disable online checks) 85 * 86 * Intended to be used by MSI interrupt controllers which are 87 * implemented with hierarchical domains. 88 */ 89 int msi_domain_set_affinity(struct irq_data *irq_data, 90 const struct cpumask *mask, bool force) 91 { 92 struct irq_data *parent = irq_data->parent_data; 93 struct msi_msg msg; 94 int ret; 95 96 ret = parent->chip->irq_set_affinity(parent, mask, force); 97 if (ret >= 0 && ret != IRQ_SET_MASK_OK_DONE) { 98 BUG_ON(irq_chip_compose_msi_msg(irq_data, &msg)); 99 irq_chip_write_msi_msg(irq_data, &msg); 100 } 101 102 return ret; 103 } 104 105 static int msi_domain_activate(struct irq_domain *domain, 106 struct irq_data *irq_data, bool early) 107 { 108 struct msi_msg msg; 109 110 BUG_ON(irq_chip_compose_msi_msg(irq_data, &msg)); 111 irq_chip_write_msi_msg(irq_data, &msg); 112 return 0; 113 } 114 115 static void msi_domain_deactivate(struct irq_domain *domain, 116 struct irq_data *irq_data) 117 { 118 struct msi_msg msg; 119 120 memset(&msg, 0, sizeof(msg)); 121 irq_chip_write_msi_msg(irq_data, &msg); 122 } 123 124 static int msi_domain_alloc(struct irq_domain *domain, unsigned int virq, 125 unsigned int nr_irqs, void *arg) 126 { 127 struct msi_domain_info *info = domain->host_data; 128 struct msi_domain_ops *ops = info->ops; 129 irq_hw_number_t hwirq = ops->get_hwirq(info, arg); 130 int i, ret; 131 132 if (irq_find_mapping(domain, hwirq) > 0) 133 return -EEXIST; 134 135 if (domain->parent) { 136 ret = irq_domain_alloc_irqs_parent(domain, virq, nr_irqs, arg); 137 if (ret < 0) 138 return ret; 139 } 140 141 for (i = 0; i < nr_irqs; i++) { 142 ret = ops->msi_init(domain, info, virq + i, hwirq + i, arg); 143 if (ret < 0) { 144 if (ops->msi_free) { 145 for (i--; i > 0; i--) 146 ops->msi_free(domain, info, virq + i); 147 } 148 irq_domain_free_irqs_top(domain, virq, nr_irqs); 149 return ret; 150 } 151 } 152 153 return 0; 154 } 155 156 static void msi_domain_free(struct irq_domain *domain, unsigned int virq, 157 unsigned int nr_irqs) 158 { 159 struct msi_domain_info *info = domain->host_data; 160 int i; 161 162 if (info->ops->msi_free) { 163 for (i = 0; i < nr_irqs; i++) 164 info->ops->msi_free(domain, info, virq + i); 165 } 166 irq_domain_free_irqs_top(domain, virq, nr_irqs); 167 } 168 169 static const struct irq_domain_ops msi_domain_ops = { 170 .alloc = msi_domain_alloc, 171 .free = msi_domain_free, 172 .activate = msi_domain_activate, 173 .deactivate = msi_domain_deactivate, 174 }; 175 176 #ifdef GENERIC_MSI_DOMAIN_OPS 177 static irq_hw_number_t msi_domain_ops_get_hwirq(struct msi_domain_info *info, 178 msi_alloc_info_t *arg) 179 { 180 return arg->hwirq; 181 } 182 183 static int msi_domain_ops_prepare(struct irq_domain *domain, struct device *dev, 184 int nvec, msi_alloc_info_t *arg) 185 { 186 memset(arg, 0, sizeof(*arg)); 187 return 0; 188 } 189 190 static void msi_domain_ops_set_desc(msi_alloc_info_t *arg, 191 struct msi_desc *desc) 192 { 193 arg->desc = desc; 194 } 195 #else 196 #define msi_domain_ops_get_hwirq NULL 197 #define msi_domain_ops_prepare NULL 198 #define msi_domain_ops_set_desc NULL 199 #endif /* !GENERIC_MSI_DOMAIN_OPS */ 200 201 static int msi_domain_ops_init(struct irq_domain *domain, 202 struct msi_domain_info *info, 203 unsigned int virq, irq_hw_number_t hwirq, 204 msi_alloc_info_t *arg) 205 { 206 irq_domain_set_hwirq_and_chip(domain, virq, hwirq, info->chip, 207 info->chip_data); 208 if (info->handler && info->handler_name) { 209 __irq_set_handler(virq, info->handler, 0, info->handler_name); 210 if (info->handler_data) 211 irq_set_handler_data(virq, info->handler_data); 212 } 213 return 0; 214 } 215 216 static int msi_domain_ops_check(struct irq_domain *domain, 217 struct msi_domain_info *info, 218 struct device *dev) 219 { 220 return 0; 221 } 222 223 static struct msi_domain_ops msi_domain_ops_default = { 224 .get_hwirq = msi_domain_ops_get_hwirq, 225 .msi_init = msi_domain_ops_init, 226 .msi_check = msi_domain_ops_check, 227 .msi_prepare = msi_domain_ops_prepare, 228 .set_desc = msi_domain_ops_set_desc, 229 }; 230 231 static void msi_domain_update_dom_ops(struct msi_domain_info *info) 232 { 233 struct msi_domain_ops *ops = info->ops; 234 235 if (ops == NULL) { 236 info->ops = &msi_domain_ops_default; 237 return; 238 } 239 240 if (ops->get_hwirq == NULL) 241 ops->get_hwirq = msi_domain_ops_default.get_hwirq; 242 if (ops->msi_init == NULL) 243 ops->msi_init = msi_domain_ops_default.msi_init; 244 if (ops->msi_check == NULL) 245 ops->msi_check = msi_domain_ops_default.msi_check; 246 if (ops->msi_prepare == NULL) 247 ops->msi_prepare = msi_domain_ops_default.msi_prepare; 248 if (ops->set_desc == NULL) 249 ops->set_desc = msi_domain_ops_default.set_desc; 250 } 251 252 static void msi_domain_update_chip_ops(struct msi_domain_info *info) 253 { 254 struct irq_chip *chip = info->chip; 255 256 BUG_ON(!chip || !chip->irq_mask || !chip->irq_unmask); 257 if (!chip->irq_set_affinity) 258 chip->irq_set_affinity = msi_domain_set_affinity; 259 } 260 261 /** 262 * msi_create_irq_domain - Create a MSI interrupt domain 263 * @fwnode: Optional fwnode of the interrupt controller 264 * @info: MSI domain info 265 * @parent: Parent irq domain 266 */ 267 struct irq_domain *msi_create_irq_domain(struct fwnode_handle *fwnode, 268 struct msi_domain_info *info, 269 struct irq_domain *parent) 270 { 271 struct irq_domain *domain; 272 273 if (info->flags & MSI_FLAG_USE_DEF_DOM_OPS) 274 msi_domain_update_dom_ops(info); 275 if (info->flags & MSI_FLAG_USE_DEF_CHIP_OPS) 276 msi_domain_update_chip_ops(info); 277 278 domain = irq_domain_create_hierarchy(parent, IRQ_DOMAIN_FLAG_MSI, 0, 279 fwnode, &msi_domain_ops, info); 280 281 if (domain && !domain->name && info->chip) 282 domain->name = info->chip->name; 283 284 return domain; 285 } 286 287 int msi_domain_prepare_irqs(struct irq_domain *domain, struct device *dev, 288 int nvec, msi_alloc_info_t *arg) 289 { 290 struct msi_domain_info *info = domain->host_data; 291 struct msi_domain_ops *ops = info->ops; 292 int ret; 293 294 ret = ops->msi_check(domain, info, dev); 295 if (ret == 0) 296 ret = ops->msi_prepare(domain, dev, nvec, arg); 297 298 return ret; 299 } 300 301 int msi_domain_populate_irqs(struct irq_domain *domain, struct device *dev, 302 int virq, int nvec, msi_alloc_info_t *arg) 303 { 304 struct msi_domain_info *info = domain->host_data; 305 struct msi_domain_ops *ops = info->ops; 306 struct msi_desc *desc; 307 int ret = 0; 308 309 for_each_msi_entry(desc, dev) { 310 /* Don't even try the multi-MSI brain damage. */ 311 if (WARN_ON(!desc->irq || desc->nvec_used != 1)) { 312 ret = -EINVAL; 313 break; 314 } 315 316 if (!(desc->irq >= virq && desc->irq < (virq + nvec))) 317 continue; 318 319 ops->set_desc(arg, desc); 320 /* Assumes the domain mutex is held! */ 321 ret = irq_domain_alloc_irqs_hierarchy(domain, desc->irq, 1, 322 arg); 323 if (ret) 324 break; 325 326 irq_set_msi_desc_off(desc->irq, 0, desc); 327 } 328 329 if (ret) { 330 /* Mop up the damage */ 331 for_each_msi_entry(desc, dev) { 332 if (!(desc->irq >= virq && desc->irq < (virq + nvec))) 333 continue; 334 335 irq_domain_free_irqs_common(domain, desc->irq, 1); 336 } 337 } 338 339 return ret; 340 } 341 342 /** 343 * msi_domain_alloc_irqs - Allocate interrupts from a MSI interrupt domain 344 * @domain: The domain to allocate from 345 * @dev: Pointer to device struct of the device for which the interrupts 346 * are allocated 347 * @nvec: The number of interrupts to allocate 348 * 349 * Returns 0 on success or an error code. 350 */ 351 int msi_domain_alloc_irqs(struct irq_domain *domain, struct device *dev, 352 int nvec) 353 { 354 struct msi_domain_info *info = domain->host_data; 355 struct msi_domain_ops *ops = info->ops; 356 msi_alloc_info_t arg; 357 struct msi_desc *desc; 358 int i, ret, virq; 359 360 ret = msi_domain_prepare_irqs(domain, dev, nvec, &arg); 361 if (ret) 362 return ret; 363 364 for_each_msi_entry(desc, dev) { 365 ops->set_desc(&arg, desc); 366 367 virq = __irq_domain_alloc_irqs(domain, -1, desc->nvec_used, 368 dev_to_node(dev), &arg, false, 369 desc->affinity); 370 if (virq < 0) { 371 ret = -ENOSPC; 372 if (ops->handle_error) 373 ret = ops->handle_error(domain, desc, ret); 374 if (ops->msi_finish) 375 ops->msi_finish(&arg, ret); 376 return ret; 377 } 378 379 for (i = 0; i < desc->nvec_used; i++) { 380 irq_set_msi_desc_off(virq, i, desc); 381 irq_debugfs_copy_devname(virq + i, dev); 382 } 383 } 384 385 if (ops->msi_finish) 386 ops->msi_finish(&arg, 0); 387 388 for_each_msi_entry(desc, dev) { 389 virq = desc->irq; 390 if (desc->nvec_used == 1) 391 dev_dbg(dev, "irq %d for MSI\n", virq); 392 else 393 dev_dbg(dev, "irq [%d-%d] for MSI\n", 394 virq, virq + desc->nvec_used - 1); 395 /* 396 * This flag is set by the PCI layer as we need to activate 397 * the MSI entries before the PCI layer enables MSI in the 398 * card. Otherwise the card latches a random msi message. 399 */ 400 if (info->flags & MSI_FLAG_ACTIVATE_EARLY) { 401 struct irq_data *irq_data; 402 403 irq_data = irq_domain_get_irq_data(domain, desc->irq); 404 ret = irq_domain_activate_irq(irq_data, true); 405 if (ret) 406 goto cleanup; 407 if (info->flags & MSI_FLAG_MUST_REACTIVATE) 408 irqd_clr_activated(irq_data); 409 } 410 } 411 return 0; 412 413 cleanup: 414 for_each_msi_entry(desc, dev) { 415 struct irq_data *irqd; 416 417 if (desc->irq == virq) 418 break; 419 420 irqd = irq_domain_get_irq_data(domain, desc->irq); 421 if (irqd_is_activated(irqd)) 422 irq_domain_deactivate_irq(irqd); 423 } 424 msi_domain_free_irqs(domain, dev); 425 return ret; 426 } 427 428 /** 429 * msi_domain_free_irqs - Free interrupts from a MSI interrupt @domain associated tp @dev 430 * @domain: The domain to managing the interrupts 431 * @dev: Pointer to device struct of the device for which the interrupts 432 * are free 433 */ 434 void msi_domain_free_irqs(struct irq_domain *domain, struct device *dev) 435 { 436 struct msi_desc *desc; 437 438 for_each_msi_entry(desc, dev) { 439 /* 440 * We might have failed to allocate an MSI early 441 * enough that there is no IRQ associated to this 442 * entry. If that's the case, don't do anything. 443 */ 444 if (desc->irq) { 445 irq_domain_free_irqs(desc->irq, desc->nvec_used); 446 desc->irq = 0; 447 } 448 } 449 } 450 451 /** 452 * msi_get_domain_info - Get the MSI interrupt domain info for @domain 453 * @domain: The interrupt domain to retrieve data from 454 * 455 * Returns the pointer to the msi_domain_info stored in 456 * @domain->host_data. 457 */ 458 struct msi_domain_info *msi_get_domain_info(struct irq_domain *domain) 459 { 460 return (struct msi_domain_info *)domain->host_data; 461 } 462 463 #endif /* CONFIG_GENERIC_MSI_IRQ_DOMAIN */ 464