xref: /linux/drivers/irqchip/irq-gic-v2m.c (revision 856e7c4b619af622d56b3b454f7bec32a170ac99)
1 /*
2  * ARM GIC v2m MSI(-X) support
3  * Support for Message Signaled Interrupts for systems that
4  * implement ARM Generic Interrupt Controller: GICv2m.
5  *
6  * Copyright (C) 2014 Advanced Micro Devices, Inc.
7  * Authors: Suravee Suthikulpanit <suravee.suthikulpanit@amd.com>
8  *	    Harish Kasiviswanathan <harish.kasiviswanathan@amd.com>
9  *	    Brandon Anderson <brandon.anderson@amd.com>
10  *
11  * This program is free software; you can redistribute it and/or modify it
12  * under the terms of the GNU General Public License version 2 as published
13  * by the Free Software Foundation.
14  */
15 
16 #define pr_fmt(fmt) "GICv2m: " fmt
17 
18 #include <linux/acpi.h>
19 #include <linux/dma-iommu.h>
20 #include <linux/irq.h>
21 #include <linux/irqdomain.h>
22 #include <linux/kernel.h>
23 #include <linux/msi.h>
24 #include <linux/of_address.h>
25 #include <linux/of_pci.h>
26 #include <linux/slab.h>
27 #include <linux/spinlock.h>
28 #include <linux/irqchip/arm-gic.h>
29 
30 /*
31 * MSI_TYPER:
32 *     [31:26] Reserved
33 *     [25:16] lowest SPI assigned to MSI
34 *     [15:10] Reserved
35 *     [9:0]   Numer of SPIs assigned to MSI
36 */
37 #define V2M_MSI_TYPER		       0x008
38 #define V2M_MSI_TYPER_BASE_SHIFT       16
39 #define V2M_MSI_TYPER_BASE_MASK	       0x3FF
40 #define V2M_MSI_TYPER_NUM_MASK	       0x3FF
41 #define V2M_MSI_SETSPI_NS	       0x040
42 #define V2M_MIN_SPI		       32
43 #define V2M_MAX_SPI		       1019
44 #define V2M_MSI_IIDR		       0xFCC
45 
46 #define V2M_MSI_TYPER_BASE_SPI(x)      \
47 	       (((x) >> V2M_MSI_TYPER_BASE_SHIFT) & V2M_MSI_TYPER_BASE_MASK)
48 
49 #define V2M_MSI_TYPER_NUM_SPI(x)       ((x) & V2M_MSI_TYPER_NUM_MASK)
50 
51 /* APM X-Gene with GICv2m MSI_IIDR register value */
52 #define XGENE_GICV2M_MSI_IIDR		0x06000170
53 
54 /* Broadcom NS2 GICv2m MSI_IIDR register value */
55 #define BCM_NS2_GICV2M_MSI_IIDR		0x0000013f
56 
57 /* List of flags for specific v2m implementation */
58 #define GICV2M_NEEDS_SPI_OFFSET		0x00000001
59 
60 static LIST_HEAD(v2m_nodes);
61 static DEFINE_SPINLOCK(v2m_lock);
62 
63 struct v2m_data {
64 	struct list_head entry;
65 	struct fwnode_handle *fwnode;
66 	struct resource res;	/* GICv2m resource */
67 	void __iomem *base;	/* GICv2m virt address */
68 	u32 spi_start;		/* The SPI number that MSIs start */
69 	u32 nr_spis;		/* The number of SPIs for MSIs */
70 	u32 spi_offset;		/* offset to be subtracted from SPI number */
71 	unsigned long *bm;	/* MSI vector bitmap */
72 	u32 flags;		/* v2m flags for specific implementation */
73 };
74 
75 static void gicv2m_mask_msi_irq(struct irq_data *d)
76 {
77 	pci_msi_mask_irq(d);
78 	irq_chip_mask_parent(d);
79 }
80 
81 static void gicv2m_unmask_msi_irq(struct irq_data *d)
82 {
83 	pci_msi_unmask_irq(d);
84 	irq_chip_unmask_parent(d);
85 }
86 
87 static struct irq_chip gicv2m_msi_irq_chip = {
88 	.name			= "MSI",
89 	.irq_mask		= gicv2m_mask_msi_irq,
90 	.irq_unmask		= gicv2m_unmask_msi_irq,
91 	.irq_eoi		= irq_chip_eoi_parent,
92 	.irq_write_msi_msg	= pci_msi_domain_write_msg,
93 };
94 
95 static struct msi_domain_info gicv2m_msi_domain_info = {
96 	.flags	= (MSI_FLAG_USE_DEF_DOM_OPS | MSI_FLAG_USE_DEF_CHIP_OPS |
97 		   MSI_FLAG_PCI_MSIX | MSI_FLAG_MULTI_PCI_MSI),
98 	.chip	= &gicv2m_msi_irq_chip,
99 };
100 
101 static void gicv2m_compose_msi_msg(struct irq_data *data, struct msi_msg *msg)
102 {
103 	struct v2m_data *v2m = irq_data_get_irq_chip_data(data);
104 	phys_addr_t addr = v2m->res.start + V2M_MSI_SETSPI_NS;
105 
106 	msg->address_hi = upper_32_bits(addr);
107 	msg->address_lo = lower_32_bits(addr);
108 	msg->data = data->hwirq;
109 
110 	if (v2m->flags & GICV2M_NEEDS_SPI_OFFSET)
111 		msg->data -= v2m->spi_offset;
112 
113 	iommu_dma_map_msi_msg(data->irq, msg);
114 }
115 
116 static struct irq_chip gicv2m_irq_chip = {
117 	.name			= "GICv2m",
118 	.irq_mask		= irq_chip_mask_parent,
119 	.irq_unmask		= irq_chip_unmask_parent,
120 	.irq_eoi		= irq_chip_eoi_parent,
121 	.irq_set_affinity	= irq_chip_set_affinity_parent,
122 	.irq_compose_msi_msg	= gicv2m_compose_msi_msg,
123 };
124 
125 static int gicv2m_irq_gic_domain_alloc(struct irq_domain *domain,
126 				       unsigned int virq,
127 				       irq_hw_number_t hwirq)
128 {
129 	struct irq_fwspec fwspec;
130 	struct irq_data *d;
131 	int err;
132 
133 	if (is_of_node(domain->parent->fwnode)) {
134 		fwspec.fwnode = domain->parent->fwnode;
135 		fwspec.param_count = 3;
136 		fwspec.param[0] = 0;
137 		fwspec.param[1] = hwirq - 32;
138 		fwspec.param[2] = IRQ_TYPE_EDGE_RISING;
139 	} else if (is_fwnode_irqchip(domain->parent->fwnode)) {
140 		fwspec.fwnode = domain->parent->fwnode;
141 		fwspec.param_count = 2;
142 		fwspec.param[0] = hwirq;
143 		fwspec.param[1] = IRQ_TYPE_EDGE_RISING;
144 	} else {
145 		return -EINVAL;
146 	}
147 
148 	err = irq_domain_alloc_irqs_parent(domain, virq, 1, &fwspec);
149 	if (err)
150 		return err;
151 
152 	/* Configure the interrupt line to be edge */
153 	d = irq_domain_get_irq_data(domain->parent, virq);
154 	d->chip->irq_set_type(d, IRQ_TYPE_EDGE_RISING);
155 	return 0;
156 }
157 
158 static void gicv2m_unalloc_msi(struct v2m_data *v2m, unsigned int hwirq,
159 			       int nr_irqs)
160 {
161 	spin_lock(&v2m_lock);
162 	bitmap_release_region(v2m->bm, hwirq - v2m->spi_start,
163 			      get_count_order(nr_irqs));
164 	spin_unlock(&v2m_lock);
165 }
166 
167 static int gicv2m_irq_domain_alloc(struct irq_domain *domain, unsigned int virq,
168 				   unsigned int nr_irqs, void *args)
169 {
170 	struct v2m_data *v2m = NULL, *tmp;
171 	int hwirq, offset, i, err = 0;
172 
173 	spin_lock(&v2m_lock);
174 	list_for_each_entry(tmp, &v2m_nodes, entry) {
175 		offset = bitmap_find_free_region(tmp->bm, tmp->nr_spis,
176 						 get_count_order(nr_irqs));
177 		if (offset >= 0) {
178 			v2m = tmp;
179 			break;
180 		}
181 	}
182 	spin_unlock(&v2m_lock);
183 
184 	if (!v2m)
185 		return -ENOSPC;
186 
187 	hwirq = v2m->spi_start + offset;
188 
189 	for (i = 0; i < nr_irqs; i++) {
190 		err = gicv2m_irq_gic_domain_alloc(domain, virq + i, hwirq + i);
191 		if (err)
192 			goto fail;
193 
194 		irq_domain_set_hwirq_and_chip(domain, virq + i, hwirq + i,
195 					      &gicv2m_irq_chip, v2m);
196 	}
197 
198 	return 0;
199 
200 fail:
201 	irq_domain_free_irqs_parent(domain, virq, nr_irqs);
202 	gicv2m_unalloc_msi(v2m, hwirq, get_count_order(nr_irqs));
203 	return err;
204 }
205 
206 static void gicv2m_irq_domain_free(struct irq_domain *domain,
207 				   unsigned int virq, unsigned int nr_irqs)
208 {
209 	struct irq_data *d = irq_domain_get_irq_data(domain, virq);
210 	struct v2m_data *v2m = irq_data_get_irq_chip_data(d);
211 
212 	gicv2m_unalloc_msi(v2m, d->hwirq, nr_irqs);
213 	irq_domain_free_irqs_parent(domain, virq, nr_irqs);
214 }
215 
216 static const struct irq_domain_ops gicv2m_domain_ops = {
217 	.alloc			= gicv2m_irq_domain_alloc,
218 	.free			= gicv2m_irq_domain_free,
219 };
220 
221 static bool is_msi_spi_valid(u32 base, u32 num)
222 {
223 	if (base < V2M_MIN_SPI) {
224 		pr_err("Invalid MSI base SPI (base:%u)\n", base);
225 		return false;
226 	}
227 
228 	if ((num == 0) || (base + num > V2M_MAX_SPI)) {
229 		pr_err("Number of SPIs (%u) exceed maximum (%u)\n",
230 		       num, V2M_MAX_SPI - V2M_MIN_SPI + 1);
231 		return false;
232 	}
233 
234 	return true;
235 }
236 
237 static struct irq_chip gicv2m_pmsi_irq_chip = {
238 	.name			= "pMSI",
239 };
240 
241 static struct msi_domain_ops gicv2m_pmsi_ops = {
242 };
243 
244 static struct msi_domain_info gicv2m_pmsi_domain_info = {
245 	.flags	= (MSI_FLAG_USE_DEF_DOM_OPS | MSI_FLAG_USE_DEF_CHIP_OPS),
246 	.ops	= &gicv2m_pmsi_ops,
247 	.chip	= &gicv2m_pmsi_irq_chip,
248 };
249 
250 static void gicv2m_teardown(void)
251 {
252 	struct v2m_data *v2m, *tmp;
253 
254 	list_for_each_entry_safe(v2m, tmp, &v2m_nodes, entry) {
255 		list_del(&v2m->entry);
256 		kfree(v2m->bm);
257 		iounmap(v2m->base);
258 		of_node_put(to_of_node(v2m->fwnode));
259 		if (is_fwnode_irqchip(v2m->fwnode))
260 			irq_domain_free_fwnode(v2m->fwnode);
261 		kfree(v2m);
262 	}
263 }
264 
265 static int gicv2m_allocate_domains(struct irq_domain *parent)
266 {
267 	struct irq_domain *inner_domain, *pci_domain, *plat_domain;
268 	struct v2m_data *v2m;
269 
270 	v2m = list_first_entry_or_null(&v2m_nodes, struct v2m_data, entry);
271 	if (!v2m)
272 		return 0;
273 
274 	inner_domain = irq_domain_create_tree(v2m->fwnode,
275 					      &gicv2m_domain_ops, v2m);
276 	if (!inner_domain) {
277 		pr_err("Failed to create GICv2m domain\n");
278 		return -ENOMEM;
279 	}
280 
281 	irq_domain_update_bus_token(inner_domain, DOMAIN_BUS_NEXUS);
282 	inner_domain->parent = parent;
283 	pci_domain = pci_msi_create_irq_domain(v2m->fwnode,
284 					       &gicv2m_msi_domain_info,
285 					       inner_domain);
286 	plat_domain = platform_msi_create_irq_domain(v2m->fwnode,
287 						     &gicv2m_pmsi_domain_info,
288 						     inner_domain);
289 	if (!pci_domain || !plat_domain) {
290 		pr_err("Failed to create MSI domains\n");
291 		if (plat_domain)
292 			irq_domain_remove(plat_domain);
293 		if (pci_domain)
294 			irq_domain_remove(pci_domain);
295 		irq_domain_remove(inner_domain);
296 		return -ENOMEM;
297 	}
298 
299 	return 0;
300 }
301 
302 static int __init gicv2m_init_one(struct fwnode_handle *fwnode,
303 				  u32 spi_start, u32 nr_spis,
304 				  struct resource *res)
305 {
306 	int ret;
307 	struct v2m_data *v2m;
308 
309 	v2m = kzalloc(sizeof(struct v2m_data), GFP_KERNEL);
310 	if (!v2m) {
311 		pr_err("Failed to allocate struct v2m_data.\n");
312 		return -ENOMEM;
313 	}
314 
315 	INIT_LIST_HEAD(&v2m->entry);
316 	v2m->fwnode = fwnode;
317 
318 	memcpy(&v2m->res, res, sizeof(struct resource));
319 
320 	v2m->base = ioremap(v2m->res.start, resource_size(&v2m->res));
321 	if (!v2m->base) {
322 		pr_err("Failed to map GICv2m resource\n");
323 		ret = -ENOMEM;
324 		goto err_free_v2m;
325 	}
326 
327 	if (spi_start && nr_spis) {
328 		v2m->spi_start = spi_start;
329 		v2m->nr_spis = nr_spis;
330 	} else {
331 		u32 typer = readl_relaxed(v2m->base + V2M_MSI_TYPER);
332 
333 		v2m->spi_start = V2M_MSI_TYPER_BASE_SPI(typer);
334 		v2m->nr_spis = V2M_MSI_TYPER_NUM_SPI(typer);
335 	}
336 
337 	if (!is_msi_spi_valid(v2m->spi_start, v2m->nr_spis)) {
338 		ret = -EINVAL;
339 		goto err_iounmap;
340 	}
341 
342 	/*
343 	 * APM X-Gene GICv2m implementation has an erratum where
344 	 * the MSI data needs to be the offset from the spi_start
345 	 * in order to trigger the correct MSI interrupt. This is
346 	 * different from the standard GICv2m implementation where
347 	 * the MSI data is the absolute value within the range from
348 	 * spi_start to (spi_start + num_spis).
349 	 *
350 	 * Broadom NS2 GICv2m implementation has an erratum where the MSI data
351 	 * is 'spi_number - 32'
352 	 */
353 	switch (readl_relaxed(v2m->base + V2M_MSI_IIDR)) {
354 	case XGENE_GICV2M_MSI_IIDR:
355 		v2m->flags |= GICV2M_NEEDS_SPI_OFFSET;
356 		v2m->spi_offset = v2m->spi_start;
357 		break;
358 	case BCM_NS2_GICV2M_MSI_IIDR:
359 		v2m->flags |= GICV2M_NEEDS_SPI_OFFSET;
360 		v2m->spi_offset = 32;
361 		break;
362 	}
363 
364 	v2m->bm = kcalloc(BITS_TO_LONGS(v2m->nr_spis), sizeof(long),
365 			  GFP_KERNEL);
366 	if (!v2m->bm) {
367 		ret = -ENOMEM;
368 		goto err_iounmap;
369 	}
370 
371 	list_add_tail(&v2m->entry, &v2m_nodes);
372 
373 	pr_info("range%pR, SPI[%d:%d]\n", res,
374 		v2m->spi_start, (v2m->spi_start + v2m->nr_spis - 1));
375 	return 0;
376 
377 err_iounmap:
378 	iounmap(v2m->base);
379 err_free_v2m:
380 	kfree(v2m);
381 	return ret;
382 }
383 
384 static struct of_device_id gicv2m_device_id[] = {
385 	{	.compatible	= "arm,gic-v2m-frame",	},
386 	{},
387 };
388 
389 static int __init gicv2m_of_init(struct fwnode_handle *parent_handle,
390 				 struct irq_domain *parent)
391 {
392 	int ret = 0;
393 	struct device_node *node = to_of_node(parent_handle);
394 	struct device_node *child;
395 
396 	for (child = of_find_matching_node(node, gicv2m_device_id); child;
397 	     child = of_find_matching_node(child, gicv2m_device_id)) {
398 		u32 spi_start = 0, nr_spis = 0;
399 		struct resource res;
400 
401 		if (!of_find_property(child, "msi-controller", NULL))
402 			continue;
403 
404 		ret = of_address_to_resource(child, 0, &res);
405 		if (ret) {
406 			pr_err("Failed to allocate v2m resource.\n");
407 			break;
408 		}
409 
410 		if (!of_property_read_u32(child, "arm,msi-base-spi",
411 					  &spi_start) &&
412 		    !of_property_read_u32(child, "arm,msi-num-spis", &nr_spis))
413 			pr_info("DT overriding V2M MSI_TYPER (base:%u, num:%u)\n",
414 				spi_start, nr_spis);
415 
416 		ret = gicv2m_init_one(&child->fwnode, spi_start, nr_spis, &res);
417 		if (ret) {
418 			of_node_put(child);
419 			break;
420 		}
421 	}
422 
423 	if (!ret)
424 		ret = gicv2m_allocate_domains(parent);
425 	if (ret)
426 		gicv2m_teardown();
427 	return ret;
428 }
429 
430 #ifdef CONFIG_ACPI
431 static int acpi_num_msi;
432 
433 static struct fwnode_handle *gicv2m_get_fwnode(struct device *dev)
434 {
435 	struct v2m_data *data;
436 
437 	if (WARN_ON(acpi_num_msi <= 0))
438 		return NULL;
439 
440 	/* We only return the fwnode of the first MSI frame. */
441 	data = list_first_entry_or_null(&v2m_nodes, struct v2m_data, entry);
442 	if (!data)
443 		return NULL;
444 
445 	return data->fwnode;
446 }
447 
448 static int __init
449 acpi_parse_madt_msi(struct acpi_subtable_header *header,
450 		    const unsigned long end)
451 {
452 	int ret;
453 	struct resource res;
454 	u32 spi_start = 0, nr_spis = 0;
455 	struct acpi_madt_generic_msi_frame *m;
456 	struct fwnode_handle *fwnode;
457 
458 	m = (struct acpi_madt_generic_msi_frame *)header;
459 	if (BAD_MADT_ENTRY(m, end))
460 		return -EINVAL;
461 
462 	res.start = m->base_address;
463 	res.end = m->base_address + SZ_4K - 1;
464 	res.flags = IORESOURCE_MEM;
465 
466 	if (m->flags & ACPI_MADT_OVERRIDE_SPI_VALUES) {
467 		spi_start = m->spi_base;
468 		nr_spis = m->spi_count;
469 
470 		pr_info("ACPI overriding V2M MSI_TYPER (base:%u, num:%u)\n",
471 			spi_start, nr_spis);
472 	}
473 
474 	fwnode = irq_domain_alloc_fwnode((void *)m->base_address);
475 	if (!fwnode) {
476 		pr_err("Unable to allocate GICv2m domain token\n");
477 		return -EINVAL;
478 	}
479 
480 	ret = gicv2m_init_one(fwnode, spi_start, nr_spis, &res);
481 	if (ret)
482 		irq_domain_free_fwnode(fwnode);
483 
484 	return ret;
485 }
486 
487 static int __init gicv2m_acpi_init(struct irq_domain *parent)
488 {
489 	int ret;
490 
491 	if (acpi_num_msi > 0)
492 		return 0;
493 
494 	acpi_num_msi = acpi_table_parse_madt(ACPI_MADT_TYPE_GENERIC_MSI_FRAME,
495 				      acpi_parse_madt_msi, 0);
496 
497 	if (acpi_num_msi <= 0)
498 		goto err_out;
499 
500 	ret = gicv2m_allocate_domains(parent);
501 	if (ret)
502 		goto err_out;
503 
504 	pci_msi_register_fwnode_provider(&gicv2m_get_fwnode);
505 
506 	return 0;
507 
508 err_out:
509 	gicv2m_teardown();
510 	return -EINVAL;
511 }
512 #else /* CONFIG_ACPI */
513 static int __init gicv2m_acpi_init(struct irq_domain *parent)
514 {
515 	return -EINVAL;
516 }
517 #endif /* CONFIG_ACPI */
518 
519 int __init gicv2m_init(struct fwnode_handle *parent_handle,
520 		       struct irq_domain *parent)
521 {
522 	if (is_of_node(parent_handle))
523 		return gicv2m_of_init(parent_handle, parent);
524 
525 	return gicv2m_acpi_init(parent);
526 }
527