xref: /linux/drivers/soc/dove/pmu.c (revision 2bd1bea5fa6aa79bc563a57919730eb809651b28)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Marvell Dove PMU support
4  */
5 #include <linux/io.h>
6 #include <linux/irq.h>
7 #include <linux/irqdomain.h>
8 #include <linux/of.h>
9 #include <linux/of_irq.h>
10 #include <linux/of_address.h>
11 #include <linux/platform_device.h>
12 #include <linux/pm_domain.h>
13 #include <linux/reset.h>
14 #include <linux/reset-controller.h>
15 #include <linux/sched.h>
16 #include <linux/slab.h>
17 #include <linux/soc/dove/pmu.h>
18 #include <linux/spinlock.h>
19 
20 #define NR_PMU_IRQS		7
21 
22 #define PMC_SW_RST		0x30
23 #define PMC_IRQ_CAUSE		0x50
24 #define PMC_IRQ_MASK		0x54
25 
26 #define PMU_PWR			0x10
27 #define PMU_ISO			0x58
28 
29 struct pmu_data {
30 	spinlock_t lock;
31 	struct device_node *of_node;
32 	void __iomem *pmc_base;
33 	void __iomem *pmu_base;
34 	struct irq_chip_generic *irq_gc;
35 	struct irq_domain *irq_domain;
36 #ifdef CONFIG_RESET_CONTROLLER
37 	struct reset_controller_dev reset;
38 #endif
39 };
40 
41 /*
42  * The PMU contains a register to reset various subsystems within the
43  * SoC.  Export this as a reset controller.
44  */
45 #ifdef CONFIG_RESET_CONTROLLER
46 #define rcdev_to_pmu(rcdev) container_of(rcdev, struct pmu_data, reset)
47 
pmu_reset_reset(struct reset_controller_dev * rc,unsigned long id)48 static int pmu_reset_reset(struct reset_controller_dev *rc, unsigned long id)
49 {
50 	struct pmu_data *pmu = rcdev_to_pmu(rc);
51 	unsigned long flags;
52 	u32 val;
53 
54 	spin_lock_irqsave(&pmu->lock, flags);
55 	val = readl_relaxed(pmu->pmc_base + PMC_SW_RST);
56 	writel_relaxed(val & ~BIT(id), pmu->pmc_base + PMC_SW_RST);
57 	writel_relaxed(val | BIT(id), pmu->pmc_base + PMC_SW_RST);
58 	spin_unlock_irqrestore(&pmu->lock, flags);
59 
60 	return 0;
61 }
62 
pmu_reset_assert(struct reset_controller_dev * rc,unsigned long id)63 static int pmu_reset_assert(struct reset_controller_dev *rc, unsigned long id)
64 {
65 	struct pmu_data *pmu = rcdev_to_pmu(rc);
66 	unsigned long flags;
67 	u32 val = ~BIT(id);
68 
69 	spin_lock_irqsave(&pmu->lock, flags);
70 	val &= readl_relaxed(pmu->pmc_base + PMC_SW_RST);
71 	writel_relaxed(val, pmu->pmc_base + PMC_SW_RST);
72 	spin_unlock_irqrestore(&pmu->lock, flags);
73 
74 	return 0;
75 }
76 
pmu_reset_deassert(struct reset_controller_dev * rc,unsigned long id)77 static int pmu_reset_deassert(struct reset_controller_dev *rc, unsigned long id)
78 {
79 	struct pmu_data *pmu = rcdev_to_pmu(rc);
80 	unsigned long flags;
81 	u32 val = BIT(id);
82 
83 	spin_lock_irqsave(&pmu->lock, flags);
84 	val |= readl_relaxed(pmu->pmc_base + PMC_SW_RST);
85 	writel_relaxed(val, pmu->pmc_base + PMC_SW_RST);
86 	spin_unlock_irqrestore(&pmu->lock, flags);
87 
88 	return 0;
89 }
90 
91 static const struct reset_control_ops pmu_reset_ops = {
92 	.reset = pmu_reset_reset,
93 	.assert = pmu_reset_assert,
94 	.deassert = pmu_reset_deassert,
95 };
96 
97 static struct reset_controller_dev pmu_reset __initdata = {
98 	.ops = &pmu_reset_ops,
99 	.owner = THIS_MODULE,
100 	.nr_resets = 32,
101 };
102 
pmu_reset_init(struct pmu_data * pmu)103 static void __init pmu_reset_init(struct pmu_data *pmu)
104 {
105 	int ret;
106 
107 	pmu->reset = pmu_reset;
108 	pmu->reset.of_node = pmu->of_node;
109 
110 	ret = reset_controller_register(&pmu->reset);
111 	if (ret)
112 		pr_err("pmu: %s failed: %d\n", "reset_controller_register", ret);
113 }
114 #else
pmu_reset_init(struct pmu_data * pmu)115 static void __init pmu_reset_init(struct pmu_data *pmu)
116 {
117 }
118 #endif
119 
120 struct pmu_domain {
121 	struct pmu_data *pmu;
122 	u32 pwr_mask;
123 	u32 rst_mask;
124 	u32 iso_mask;
125 	struct generic_pm_domain base;
126 };
127 
128 #define to_pmu_domain(dom) container_of(dom, struct pmu_domain, base)
129 
130 /*
131  * This deals with the "old" Marvell sequence of bringing a power domain
132  * down/up, which is: apply power, release reset, disable isolators.
133  *
134  * Later devices apparantly use a different sequence: power up, disable
135  * isolators, assert repair signal, enable SRMA clock, enable AXI clock,
136  * enable module clock, deassert reset.
137  *
138  * Note: reading the assembly, it seems that the IO accessors have an
139  * unfortunate side-effect - they cause memory already read into registers
140  * for the if () to be re-read for the bit-set or bit-clear operation.
141  * The code is written to avoid this.
142  */
pmu_domain_power_off(struct generic_pm_domain * domain)143 static int pmu_domain_power_off(struct generic_pm_domain *domain)
144 {
145 	struct pmu_domain *pmu_dom = to_pmu_domain(domain);
146 	struct pmu_data *pmu = pmu_dom->pmu;
147 	unsigned long flags;
148 	unsigned int val;
149 	void __iomem *pmu_base = pmu->pmu_base;
150 	void __iomem *pmc_base = pmu->pmc_base;
151 
152 	spin_lock_irqsave(&pmu->lock, flags);
153 
154 	/* Enable isolators */
155 	if (pmu_dom->iso_mask) {
156 		val = ~pmu_dom->iso_mask;
157 		val &= readl_relaxed(pmu_base + PMU_ISO);
158 		writel_relaxed(val, pmu_base + PMU_ISO);
159 	}
160 
161 	/* Reset unit */
162 	if (pmu_dom->rst_mask) {
163 		val = ~pmu_dom->rst_mask;
164 		val &= readl_relaxed(pmc_base + PMC_SW_RST);
165 		writel_relaxed(val, pmc_base + PMC_SW_RST);
166 	}
167 
168 	/* Power down */
169 	val = readl_relaxed(pmu_base + PMU_PWR) | pmu_dom->pwr_mask;
170 	writel_relaxed(val, pmu_base + PMU_PWR);
171 
172 	spin_unlock_irqrestore(&pmu->lock, flags);
173 
174 	return 0;
175 }
176 
pmu_domain_power_on(struct generic_pm_domain * domain)177 static int pmu_domain_power_on(struct generic_pm_domain *domain)
178 {
179 	struct pmu_domain *pmu_dom = to_pmu_domain(domain);
180 	struct pmu_data *pmu = pmu_dom->pmu;
181 	unsigned long flags;
182 	unsigned int val;
183 	void __iomem *pmu_base = pmu->pmu_base;
184 	void __iomem *pmc_base = pmu->pmc_base;
185 
186 	spin_lock_irqsave(&pmu->lock, flags);
187 
188 	/* Power on */
189 	val = ~pmu_dom->pwr_mask & readl_relaxed(pmu_base + PMU_PWR);
190 	writel_relaxed(val, pmu_base + PMU_PWR);
191 
192 	/* Release reset */
193 	if (pmu_dom->rst_mask) {
194 		val = pmu_dom->rst_mask;
195 		val |= readl_relaxed(pmc_base + PMC_SW_RST);
196 		writel_relaxed(val, pmc_base + PMC_SW_RST);
197 	}
198 
199 	/* Disable isolators */
200 	if (pmu_dom->iso_mask) {
201 		val = pmu_dom->iso_mask;
202 		val |= readl_relaxed(pmu_base + PMU_ISO);
203 		writel_relaxed(val, pmu_base + PMU_ISO);
204 	}
205 
206 	spin_unlock_irqrestore(&pmu->lock, flags);
207 
208 	return 0;
209 }
210 
__pmu_domain_register(struct pmu_domain * domain,struct device_node * np)211 static void __pmu_domain_register(struct pmu_domain *domain,
212 	struct device_node *np)
213 {
214 	unsigned int val = readl_relaxed(domain->pmu->pmu_base + PMU_PWR);
215 
216 	domain->base.power_off = pmu_domain_power_off;
217 	domain->base.power_on = pmu_domain_power_on;
218 
219 	pm_genpd_init(&domain->base, NULL, !(val & domain->pwr_mask));
220 
221 	if (np)
222 		of_genpd_add_provider_simple(np, &domain->base);
223 }
224 
225 /* PMU IRQ controller */
pmu_irq_handler(struct irq_desc * desc)226 static void pmu_irq_handler(struct irq_desc *desc)
227 {
228 	struct pmu_data *pmu = irq_desc_get_handler_data(desc);
229 	struct irq_chip_generic *gc = pmu->irq_gc;
230 	struct irq_domain *domain = pmu->irq_domain;
231 	void __iomem *base = gc->reg_base;
232 	u32 stat = readl_relaxed(base + PMC_IRQ_CAUSE) & gc->mask_cache;
233 	u32 done = ~0;
234 
235 	if (stat == 0) {
236 		handle_bad_irq(desc);
237 		return;
238 	}
239 
240 	while (stat) {
241 		u32 hwirq = fls(stat) - 1;
242 
243 		stat &= ~(1 << hwirq);
244 		done &= ~(1 << hwirq);
245 
246 		generic_handle_irq(irq_find_mapping(domain, hwirq));
247 	}
248 
249 	/*
250 	 * The PMU mask register is not RW0C: it is RW.  This means that
251 	 * the bits take whatever value is written to them; if you write
252 	 * a '1', you will set the interrupt.
253 	 *
254 	 * Unfortunately this means there is NO race free way to clear
255 	 * these interrupts.
256 	 *
257 	 * So, let's structure the code so that the window is as small as
258 	 * possible.
259 	 */
260 	guard(raw_spinlock)(&gc->lock);
261 	done &= readl_relaxed(base + PMC_IRQ_CAUSE);
262 	writel_relaxed(done, base + PMC_IRQ_CAUSE);
263 }
264 
dove_init_pmu_irq(struct pmu_data * pmu,int irq)265 static int __init dove_init_pmu_irq(struct pmu_data *pmu, int irq)
266 {
267 	const char *name = "pmu_irq";
268 	struct irq_chip_generic *gc;
269 	struct irq_domain *domain;
270 	int ret;
271 
272 	/* mask and clear all interrupts */
273 	writel(0, pmu->pmc_base + PMC_IRQ_MASK);
274 	writel(0, pmu->pmc_base + PMC_IRQ_CAUSE);
275 
276 	domain = irq_domain_create_linear(of_fwnode_handle(pmu->of_node), NR_PMU_IRQS,
277 					  &irq_generic_chip_ops, NULL);
278 	if (!domain) {
279 		pr_err("%s: unable to add irq domain\n", name);
280 		return -ENOMEM;
281 	}
282 
283 	ret = irq_alloc_domain_generic_chips(domain, NR_PMU_IRQS, 1, name,
284 					     handle_level_irq,
285 					     IRQ_NOREQUEST | IRQ_NOPROBE, 0,
286 					     IRQ_GC_INIT_MASK_CACHE);
287 	if (ret) {
288 		pr_err("%s: unable to alloc irq domain gc: %d\n", name, ret);
289 		irq_domain_remove(domain);
290 		return ret;
291 	}
292 
293 	gc = irq_get_domain_generic_chip(domain, 0);
294 	gc->reg_base = pmu->pmc_base;
295 	gc->chip_types[0].regs.mask = PMC_IRQ_MASK;
296 	gc->chip_types[0].chip.irq_mask = irq_gc_mask_clr_bit;
297 	gc->chip_types[0].chip.irq_unmask = irq_gc_mask_set_bit;
298 
299 	pmu->irq_domain = domain;
300 	pmu->irq_gc = gc;
301 
302 	irq_set_handler_data(irq, pmu);
303 	irq_set_chained_handler(irq, pmu_irq_handler);
304 
305 	return 0;
306 }
307 
dove_init_pmu_legacy(const struct dove_pmu_initdata * initdata)308 int __init dove_init_pmu_legacy(const struct dove_pmu_initdata *initdata)
309 {
310 	const struct dove_pmu_domain_initdata *domain_initdata;
311 	struct pmu_data *pmu;
312 	int ret;
313 
314 	pmu = kzalloc(sizeof(*pmu), GFP_KERNEL);
315 	if (!pmu)
316 		return -ENOMEM;
317 
318 	spin_lock_init(&pmu->lock);
319 	pmu->pmc_base = initdata->pmc_base;
320 	pmu->pmu_base = initdata->pmu_base;
321 
322 	pmu_reset_init(pmu);
323 	for (domain_initdata = initdata->domains; domain_initdata->name;
324 	     domain_initdata++) {
325 		struct pmu_domain *domain;
326 
327 		domain = kzalloc(sizeof(*domain), GFP_KERNEL);
328 		if (domain) {
329 			domain->pmu = pmu;
330 			domain->pwr_mask = domain_initdata->pwr_mask;
331 			domain->rst_mask = domain_initdata->rst_mask;
332 			domain->iso_mask = domain_initdata->iso_mask;
333 			domain->base.name = domain_initdata->name;
334 
335 			__pmu_domain_register(domain, NULL);
336 		}
337 	}
338 
339 	ret = dove_init_pmu_irq(pmu, initdata->irq);
340 	if (ret)
341 		pr_err("dove_init_pmu_irq() failed: %d\n", ret);
342 
343 	if (pmu->irq_domain)
344 		irq_domain_associate_many(pmu->irq_domain,
345 					  initdata->irq_domain_start,
346 					  0, NR_PMU_IRQS);
347 
348 	return 0;
349 }
350 
351 /*
352  * pmu: power-manager@d0000 {
353  *	compatible = "marvell,dove-pmu";
354  *	reg = <0xd0000 0x8000> <0xd8000 0x8000>;
355  *	interrupts = <33>;
356  *	interrupt-controller;
357  *	#reset-cells = 1;
358  *	vpu_domain: vpu-domain {
359  *		#power-domain-cells = <0>;
360  *		marvell,pmu_pwr_mask = <0x00000008>;
361  *		marvell,pmu_iso_mask = <0x00000001>;
362  *		resets = <&pmu 16>;
363  *	};
364  *	gpu_domain: gpu-domain {
365  *		#power-domain-cells = <0>;
366  *		marvell,pmu_pwr_mask = <0x00000004>;
367  *		marvell,pmu_iso_mask = <0x00000002>;
368  *		resets = <&pmu 18>;
369  *	};
370  * };
371  */
dove_init_pmu(void)372 int __init dove_init_pmu(void)
373 {
374 	struct device_node *np_pmu, *domains_node, *np;
375 	struct pmu_data *pmu;
376 	int ret, parent_irq;
377 
378 	/* Lookup the PMU node */
379 	np_pmu = of_find_compatible_node(NULL, NULL, "marvell,dove-pmu");
380 	if (!np_pmu)
381 		return 0;
382 
383 	domains_node = of_get_child_by_name(np_pmu, "domains");
384 	if (!domains_node) {
385 		pr_err("%pOFn: failed to find domains sub-node\n", np_pmu);
386 		return 0;
387 	}
388 
389 	pmu = kzalloc(sizeof(*pmu), GFP_KERNEL);
390 	if (!pmu)
391 		return -ENOMEM;
392 
393 	spin_lock_init(&pmu->lock);
394 	pmu->of_node = np_pmu;
395 	pmu->pmc_base = of_iomap(pmu->of_node, 0);
396 	pmu->pmu_base = of_iomap(pmu->of_node, 1);
397 	if (!pmu->pmc_base || !pmu->pmu_base) {
398 		pr_err("%pOFn: failed to map PMU\n", np_pmu);
399 		iounmap(pmu->pmu_base);
400 		iounmap(pmu->pmc_base);
401 		kfree(pmu);
402 		return -ENOMEM;
403 	}
404 
405 	pmu_reset_init(pmu);
406 
407 	for_each_available_child_of_node(domains_node, np) {
408 		struct of_phandle_args args;
409 		struct pmu_domain *domain;
410 
411 		domain = kzalloc(sizeof(*domain), GFP_KERNEL);
412 		if (!domain) {
413 			of_node_put(np);
414 			break;
415 		}
416 
417 		domain->pmu = pmu;
418 		domain->base.name = kasprintf(GFP_KERNEL, "%pOFn", np);
419 		if (!domain->base.name) {
420 			kfree(domain);
421 			of_node_put(np);
422 			break;
423 		}
424 
425 		of_property_read_u32(np, "marvell,pmu_pwr_mask",
426 				     &domain->pwr_mask);
427 		of_property_read_u32(np, "marvell,pmu_iso_mask",
428 				     &domain->iso_mask);
429 
430 		/*
431 		 * We parse the reset controller property directly here
432 		 * to ensure that we can operate when the reset controller
433 		 * support is not configured into the kernel.
434 		 */
435 		ret = of_parse_phandle_with_args(np, "resets", "#reset-cells",
436 						 0, &args);
437 		if (ret == 0) {
438 			if (args.np == pmu->of_node)
439 				domain->rst_mask = BIT(args.args[0]);
440 			of_node_put(args.np);
441 		}
442 
443 		__pmu_domain_register(domain, np);
444 	}
445 
446 	/* Loss of the interrupt controller is not a fatal error. */
447 	parent_irq = irq_of_parse_and_map(pmu->of_node, 0);
448 	if (!parent_irq) {
449 		pr_err("%pOFn: no interrupt specified\n", np_pmu);
450 	} else {
451 		ret = dove_init_pmu_irq(pmu, parent_irq);
452 		if (ret)
453 			pr_err("dove_init_pmu_irq() failed: %d\n", ret);
454 	}
455 
456 	return 0;
457 }
458