1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * ti-sysc.c - Texas Instruments sysc interconnect target driver
4 *
5 * TI SoCs have an interconnect target wrapper IP for many devices. The wrapper
6 * IP manages clock gating, resets, and PM capabilities for the connected devices.
7 *
8 * Copyright (C) 2017-2024 Texas Instruments Incorporated - https://www.ti.com/
9 *
10 * Many features are based on the earlier omap_hwmod arch code with thanks to all
11 * the people who developed and debugged the code over the years:
12 *
13 * Copyright (C) 2009-2011 Nokia Corporation
14 * Copyright (C) 2011-2021 Texas Instruments Incorporated - https://www.ti.com/
15 */
16
17 #include <linux/io.h>
18 #include <linux/clk.h>
19 #include <linux/clkdev.h>
20 #include <linux/cpu_pm.h>
21 #include <linux/delay.h>
22 #include <linux/list.h>
23 #include <linux/module.h>
24 #include <linux/platform_device.h>
25 #include <linux/pm_domain.h>
26 #include <linux/pm_runtime.h>
27 #include <linux/reset.h>
28 #include <linux/of_address.h>
29 #include <linux/of_platform.h>
30 #include <linux/slab.h>
31 #include <linux/sys_soc.h>
32 #include <linux/timekeeping.h>
33 #include <linux/iopoll.h>
34
35 #include <linux/platform_data/ti-sysc.h>
36
37 #include <dt-bindings/bus/ti-sysc.h>
38
39 #define DIS_ISP BIT(2)
40 #define DIS_IVA BIT(1)
41 #define DIS_SGX BIT(0)
42
43 #define SOC_FLAG(match, flag) { .machine = match, .data = (void *)(flag), }
44
45 #define MAX_MODULE_SOFTRESET_WAIT 10000
46
47 enum sysc_soc {
48 SOC_UNKNOWN,
49 SOC_2420,
50 SOC_2430,
51 SOC_3430,
52 SOC_AM35,
53 SOC_3630,
54 SOC_4430,
55 SOC_4460,
56 SOC_4470,
57 SOC_5430,
58 SOC_AM3,
59 SOC_AM4,
60 SOC_DRA7,
61 };
62
63 struct sysc_address {
64 unsigned long base;
65 struct list_head node;
66 };
67
68 struct sysc_module {
69 struct sysc *ddata;
70 struct list_head node;
71 };
72
73 struct sysc_soc_info {
74 unsigned long general_purpose:1;
75 enum sysc_soc soc;
76 struct mutex list_lock; /* disabled and restored modules list lock */
77 struct list_head disabled_modules;
78 struct list_head restored_modules;
79 struct notifier_block nb;
80 };
81
82 enum sysc_clocks {
83 SYSC_FCK,
84 SYSC_ICK,
85 SYSC_OPTFCK0,
86 SYSC_OPTFCK1,
87 SYSC_OPTFCK2,
88 SYSC_OPTFCK3,
89 SYSC_OPTFCK4,
90 SYSC_OPTFCK5,
91 SYSC_OPTFCK6,
92 SYSC_OPTFCK7,
93 SYSC_MAX_CLOCKS,
94 };
95
96 static struct sysc_soc_info *sysc_soc;
97 static const char * const reg_names[] = { "rev", "sysc", "syss", };
98 static const char * const clock_names[SYSC_MAX_CLOCKS] = {
99 "fck", "ick", "opt0", "opt1", "opt2", "opt3", "opt4",
100 "opt5", "opt6", "opt7",
101 };
102
103 #define SYSC_IDLEMODE_MASK 3
104 #define SYSC_CLOCKACTIVITY_MASK 3
105
106 /**
107 * struct sysc - TI sysc interconnect target module registers and capabilities
108 * @dev: struct device pointer
109 * @module_pa: physical address of the interconnect target module
110 * @module_size: size of the interconnect target module
111 * @module_va: virtual address of the interconnect target module
112 * @offsets: register offsets from module base
113 * @mdata: ti-sysc to hwmod translation data for a module
114 * @clocks: clocks used by the interconnect target module
115 * @clock_roles: clock role names for the found clocks
116 * @nr_clocks: number of clocks used by the interconnect target module
117 * @rsts: resets used by the interconnect target module
118 * @legacy_mode: configured for legacy mode if set
119 * @cap: interconnect target module capabilities
120 * @cfg: interconnect target module configuration
121 * @cookie: data used by legacy platform callbacks
122 * @name: name if available
123 * @revision: interconnect target module revision
124 * @sysconfig: saved sysconfig register value
125 * @reserved: target module is reserved and already in use
126 * @enabled: sysc runtime enabled status
127 * @needs_resume: runtime resume needed on resume from suspend
128 * @child_needs_resume: runtime resume needed for child on resume from suspend
129 * @idle_work: work structure used to perform delayed idle on a module
130 * @pre_reset_quirk: module specific pre-reset quirk
131 * @post_reset_quirk: module specific post-reset quirk
132 * @reset_done_quirk: module specific reset done quirk
133 * @module_enable_quirk: module specific enable quirk
134 * @module_disable_quirk: module specific disable quirk
135 * @module_unlock_quirk: module specific sysconfig unlock quirk
136 * @module_lock_quirk: module specific sysconfig lock quirk
137 */
138 struct sysc {
139 struct device *dev;
140 u64 module_pa;
141 u32 module_size;
142 void __iomem *module_va;
143 int offsets[SYSC_MAX_REGS];
144 struct ti_sysc_module_data *mdata;
145 struct clk **clocks;
146 const char **clock_roles;
147 int nr_clocks;
148 struct reset_control *rsts;
149 const char *legacy_mode;
150 const struct sysc_capabilities *cap;
151 struct sysc_config cfg;
152 struct ti_sysc_cookie cookie;
153 const char *name;
154 u32 revision;
155 u32 sysconfig;
156 unsigned int reserved:1;
157 unsigned int enabled:1;
158 unsigned int needs_resume:1;
159 unsigned int child_needs_resume:1;
160 struct delayed_work idle_work;
161 void (*pre_reset_quirk)(struct sysc *sysc);
162 void (*post_reset_quirk)(struct sysc *sysc);
163 void (*reset_done_quirk)(struct sysc *sysc);
164 void (*module_enable_quirk)(struct sysc *sysc);
165 void (*module_disable_quirk)(struct sysc *sysc);
166 void (*module_unlock_quirk)(struct sysc *sysc);
167 void (*module_lock_quirk)(struct sysc *sysc);
168 };
169
170 static void sysc_parse_dts_quirks(struct sysc *ddata, struct device_node *np,
171 bool is_child);
172 static int sysc_reset(struct sysc *ddata);
173
sysc_write(struct sysc * ddata,int offset,u32 value)174 static void sysc_write(struct sysc *ddata, int offset, u32 value)
175 {
176 if (ddata->cfg.quirks & SYSC_QUIRK_16BIT) {
177 writew_relaxed(value & 0xffff, ddata->module_va + offset);
178
179 /* Only i2c revision has LO and HI register with stride of 4 */
180 if (ddata->offsets[SYSC_REVISION] >= 0 &&
181 offset == ddata->offsets[SYSC_REVISION]) {
182 u16 hi = value >> 16;
183
184 writew_relaxed(hi, ddata->module_va + offset + 4);
185 }
186
187 return;
188 }
189
190 writel_relaxed(value, ddata->module_va + offset);
191 }
192
sysc_read(struct sysc * ddata,int offset)193 static u32 sysc_read(struct sysc *ddata, int offset)
194 {
195 if (ddata->cfg.quirks & SYSC_QUIRK_16BIT) {
196 u32 val;
197
198 val = readw_relaxed(ddata->module_va + offset);
199
200 /* Only i2c revision has LO and HI register with stride of 4 */
201 if (ddata->offsets[SYSC_REVISION] >= 0 &&
202 offset == ddata->offsets[SYSC_REVISION]) {
203 u16 tmp = readw_relaxed(ddata->module_va + offset + 4);
204
205 val |= tmp << 16;
206 }
207
208 return val;
209 }
210
211 return readl_relaxed(ddata->module_va + offset);
212 }
213
sysc_opt_clks_needed(struct sysc * ddata)214 static bool sysc_opt_clks_needed(struct sysc *ddata)
215 {
216 return !!(ddata->cfg.quirks & SYSC_QUIRK_OPT_CLKS_NEEDED);
217 }
218
sysc_read_revision(struct sysc * ddata)219 static u32 sysc_read_revision(struct sysc *ddata)
220 {
221 int offset = ddata->offsets[SYSC_REVISION];
222
223 if (offset < 0)
224 return 0;
225
226 return sysc_read(ddata, offset);
227 }
228
sysc_read_sysconfig(struct sysc * ddata)229 static u32 sysc_read_sysconfig(struct sysc *ddata)
230 {
231 int offset = ddata->offsets[SYSC_SYSCONFIG];
232
233 if (offset < 0)
234 return 0;
235
236 return sysc_read(ddata, offset);
237 }
238
sysc_read_sysstatus(struct sysc * ddata)239 static u32 sysc_read_sysstatus(struct sysc *ddata)
240 {
241 int offset = ddata->offsets[SYSC_SYSSTATUS];
242
243 if (offset < 0)
244 return 0;
245
246 return sysc_read(ddata, offset);
247 }
248
sysc_poll_reset_sysstatus(struct sysc * ddata)249 static int sysc_poll_reset_sysstatus(struct sysc *ddata)
250 {
251 int error, retries;
252 u32 syss_done, rstval;
253
254 if (ddata->cfg.quirks & SYSS_QUIRK_RESETDONE_INVERTED)
255 syss_done = 0;
256 else
257 syss_done = ddata->cfg.syss_mask;
258
259 if (likely(!timekeeping_suspended)) {
260 error = readx_poll_timeout_atomic(sysc_read_sysstatus, ddata,
261 rstval, (rstval & ddata->cfg.syss_mask) ==
262 syss_done, 100, MAX_MODULE_SOFTRESET_WAIT);
263 } else {
264 retries = MAX_MODULE_SOFTRESET_WAIT;
265 while (retries--) {
266 rstval = sysc_read_sysstatus(ddata);
267 if ((rstval & ddata->cfg.syss_mask) == syss_done)
268 return 0;
269 udelay(2); /* Account for udelay flakeyness */
270 }
271 error = -ETIMEDOUT;
272 }
273
274 return error;
275 }
276
sysc_poll_reset_sysconfig(struct sysc * ddata)277 static int sysc_poll_reset_sysconfig(struct sysc *ddata)
278 {
279 int error, retries;
280 u32 sysc_mask, rstval;
281
282 sysc_mask = BIT(ddata->cap->regbits->srst_shift);
283
284 if (likely(!timekeeping_suspended)) {
285 error = readx_poll_timeout_atomic(sysc_read_sysconfig, ddata,
286 rstval, !(rstval & sysc_mask),
287 100, MAX_MODULE_SOFTRESET_WAIT);
288 } else {
289 retries = MAX_MODULE_SOFTRESET_WAIT;
290 while (retries--) {
291 rstval = sysc_read_sysconfig(ddata);
292 if (!(rstval & sysc_mask))
293 return 0;
294 udelay(2); /* Account for udelay flakeyness */
295 }
296 error = -ETIMEDOUT;
297 }
298
299 return error;
300 }
301
302 /* Poll on reset status */
sysc_wait_softreset(struct sysc * ddata)303 static int sysc_wait_softreset(struct sysc *ddata)
304 {
305 int syss_offset, error = 0;
306
307 if (ddata->cap->regbits->srst_shift < 0)
308 return 0;
309
310 syss_offset = ddata->offsets[SYSC_SYSSTATUS];
311
312 if (syss_offset >= 0)
313 error = sysc_poll_reset_sysstatus(ddata);
314 else if (ddata->cfg.quirks & SYSC_QUIRK_RESET_STATUS)
315 error = sysc_poll_reset_sysconfig(ddata);
316
317 return error;
318 }
319
sysc_add_named_clock_from_child(struct sysc * ddata,const char * name,const char * optfck_name)320 static int sysc_add_named_clock_from_child(struct sysc *ddata,
321 const char *name,
322 const char *optfck_name)
323 {
324 struct device_node *np = ddata->dev->of_node;
325 struct device_node *child;
326 struct clk_lookup *cl;
327 struct clk *clock;
328 const char *n;
329
330 if (name)
331 n = name;
332 else
333 n = optfck_name;
334
335 /* Does the clock alias already exist? */
336 clock = of_clk_get_by_name(np, n);
337 if (!IS_ERR(clock)) {
338 clk_put(clock);
339
340 return 0;
341 }
342
343 child = of_get_next_available_child(np, NULL);
344 if (!child)
345 return -ENODEV;
346
347 clock = devm_get_clk_from_child(ddata->dev, child, name);
348 if (IS_ERR(clock))
349 return PTR_ERR(clock);
350
351 /*
352 * Use clkdev_add() instead of clkdev_alloc() to avoid the MAX_DEV_ID
353 * limit for clk_get(). If cl ever needs to be freed, it should be done
354 * with clkdev_drop().
355 */
356 cl = kzalloc(sizeof(*cl), GFP_KERNEL);
357 if (!cl)
358 return -ENOMEM;
359
360 cl->con_id = n;
361 cl->dev_id = dev_name(ddata->dev);
362 cl->clk = clock;
363 clkdev_add(cl);
364
365 clk_put(clock);
366
367 return 0;
368 }
369
sysc_init_ext_opt_clock(struct sysc * ddata,const char * name)370 static int sysc_init_ext_opt_clock(struct sysc *ddata, const char *name)
371 {
372 const char *optfck_name;
373 int error, index;
374
375 if (ddata->nr_clocks < SYSC_OPTFCK0)
376 index = SYSC_OPTFCK0;
377 else
378 index = ddata->nr_clocks;
379
380 if (name)
381 optfck_name = name;
382 else
383 optfck_name = clock_names[index];
384
385 error = sysc_add_named_clock_from_child(ddata, name, optfck_name);
386 if (error)
387 return error;
388
389 ddata->clock_roles[index] = optfck_name;
390 ddata->nr_clocks++;
391
392 return 0;
393 }
394
sysc_get_one_clock(struct sysc * ddata,const char * name)395 static int sysc_get_one_clock(struct sysc *ddata, const char *name)
396 {
397 int error, i, index = -ENODEV;
398
399 if (!strncmp(clock_names[SYSC_FCK], name, 3))
400 index = SYSC_FCK;
401 else if (!strncmp(clock_names[SYSC_ICK], name, 3))
402 index = SYSC_ICK;
403
404 if (index < 0) {
405 for (i = SYSC_OPTFCK0; i < SYSC_MAX_CLOCKS; i++) {
406 if (!ddata->clocks[i]) {
407 index = i;
408 break;
409 }
410 }
411 }
412
413 if (index < 0) {
414 dev_err(ddata->dev, "clock %s not added\n", name);
415 return index;
416 }
417
418 ddata->clocks[index] = devm_clk_get(ddata->dev, name);
419 if (IS_ERR(ddata->clocks[index])) {
420 dev_err(ddata->dev, "clock get error for %s: %li\n",
421 name, PTR_ERR(ddata->clocks[index]));
422
423 return PTR_ERR(ddata->clocks[index]);
424 }
425
426 error = clk_prepare(ddata->clocks[index]);
427 if (error) {
428 dev_err(ddata->dev, "clock prepare error for %s: %i\n",
429 name, error);
430
431 return error;
432 }
433
434 return 0;
435 }
436
sysc_get_clocks(struct sysc * ddata)437 static int sysc_get_clocks(struct sysc *ddata)
438 {
439 struct device_node *np = ddata->dev->of_node;
440 struct property *prop;
441 const char *name;
442 int nr_fck = 0, nr_ick = 0, i, error = 0;
443
444 ddata->clock_roles = devm_kcalloc(ddata->dev,
445 SYSC_MAX_CLOCKS,
446 sizeof(*ddata->clock_roles),
447 GFP_KERNEL);
448 if (!ddata->clock_roles)
449 return -ENOMEM;
450
451 of_property_for_each_string(np, "clock-names", prop, name) {
452 if (!strncmp(clock_names[SYSC_FCK], name, 3))
453 nr_fck++;
454 if (!strncmp(clock_names[SYSC_ICK], name, 3))
455 nr_ick++;
456 ddata->clock_roles[ddata->nr_clocks] = name;
457 ddata->nr_clocks++;
458 }
459
460 if (ddata->nr_clocks < 1)
461 return 0;
462
463 if ((ddata->cfg.quirks & SYSC_QUIRK_EXT_OPT_CLOCK)) {
464 error = sysc_init_ext_opt_clock(ddata, NULL);
465 if (error)
466 return error;
467 }
468
469 if (ddata->nr_clocks > SYSC_MAX_CLOCKS) {
470 dev_err(ddata->dev, "too many clocks for %pOF\n", np);
471
472 return -EINVAL;
473 }
474
475 if (nr_fck > 1 || nr_ick > 1) {
476 dev_err(ddata->dev, "max one fck and ick for %pOF\n", np);
477
478 return -EINVAL;
479 }
480
481 /* Always add a slot for main clocks fck and ick even if unused */
482 if (!nr_fck)
483 ddata->nr_clocks++;
484 if (!nr_ick)
485 ddata->nr_clocks++;
486
487 ddata->clocks = devm_kcalloc(ddata->dev,
488 ddata->nr_clocks, sizeof(*ddata->clocks),
489 GFP_KERNEL);
490 if (!ddata->clocks)
491 return -ENOMEM;
492
493 for (i = 0; i < SYSC_MAX_CLOCKS; i++) {
494 const char *name = ddata->clock_roles[i];
495
496 if (!name)
497 continue;
498
499 error = sysc_get_one_clock(ddata, name);
500 if (error)
501 return error;
502 }
503
504 return 0;
505 }
506
sysc_enable_main_clocks(struct sysc * ddata)507 static int sysc_enable_main_clocks(struct sysc *ddata)
508 {
509 struct clk *clock;
510 int i, error;
511
512 if (!ddata->clocks)
513 return 0;
514
515 for (i = 0; i < SYSC_OPTFCK0; i++) {
516 clock = ddata->clocks[i];
517
518 /* Main clocks may not have ick */
519 if (IS_ERR_OR_NULL(clock))
520 continue;
521
522 error = clk_enable(clock);
523 if (error)
524 goto err_disable;
525 }
526
527 return 0;
528
529 err_disable:
530 for (i--; i >= 0; i--) {
531 clock = ddata->clocks[i];
532
533 /* Main clocks may not have ick */
534 if (IS_ERR_OR_NULL(clock))
535 continue;
536
537 clk_disable(clock);
538 }
539
540 return error;
541 }
542
sysc_disable_main_clocks(struct sysc * ddata)543 static void sysc_disable_main_clocks(struct sysc *ddata)
544 {
545 struct clk *clock;
546 int i;
547
548 if (!ddata->clocks)
549 return;
550
551 for (i = 0; i < SYSC_OPTFCK0; i++) {
552 clock = ddata->clocks[i];
553 if (IS_ERR_OR_NULL(clock))
554 continue;
555
556 clk_disable(clock);
557 }
558 }
559
sysc_enable_opt_clocks(struct sysc * ddata)560 static int sysc_enable_opt_clocks(struct sysc *ddata)
561 {
562 struct clk *clock;
563 int i, error;
564
565 if (!ddata->clocks || ddata->nr_clocks < SYSC_OPTFCK0 + 1)
566 return 0;
567
568 for (i = SYSC_OPTFCK0; i < SYSC_MAX_CLOCKS; i++) {
569 clock = ddata->clocks[i];
570
571 /* Assume no holes for opt clocks */
572 if (IS_ERR_OR_NULL(clock))
573 return 0;
574
575 error = clk_enable(clock);
576 if (error)
577 goto err_disable;
578 }
579
580 return 0;
581
582 err_disable:
583 for (i--; i >= 0; i--) {
584 clock = ddata->clocks[i];
585 if (IS_ERR_OR_NULL(clock))
586 continue;
587
588 clk_disable(clock);
589 }
590
591 return error;
592 }
593
sysc_disable_opt_clocks(struct sysc * ddata)594 static void sysc_disable_opt_clocks(struct sysc *ddata)
595 {
596 struct clk *clock;
597 int i;
598
599 if (!ddata->clocks || ddata->nr_clocks < SYSC_OPTFCK0 + 1)
600 return;
601
602 for (i = SYSC_OPTFCK0; i < SYSC_MAX_CLOCKS; i++) {
603 clock = ddata->clocks[i];
604
605 /* Assume no holes for opt clocks */
606 if (IS_ERR_OR_NULL(clock))
607 return;
608
609 clk_disable(clock);
610 }
611 }
612
sysc_clkdm_deny_idle(struct sysc * ddata)613 static void sysc_clkdm_deny_idle(struct sysc *ddata)
614 {
615 struct ti_sysc_platform_data *pdata;
616
617 if (ddata->legacy_mode || (ddata->cfg.quirks & SYSC_QUIRK_CLKDM_NOAUTO))
618 return;
619
620 pdata = dev_get_platdata(ddata->dev);
621 if (pdata && pdata->clkdm_deny_idle)
622 pdata->clkdm_deny_idle(ddata->dev, &ddata->cookie);
623 }
624
sysc_clkdm_allow_idle(struct sysc * ddata)625 static void sysc_clkdm_allow_idle(struct sysc *ddata)
626 {
627 struct ti_sysc_platform_data *pdata;
628
629 if (ddata->legacy_mode || (ddata->cfg.quirks & SYSC_QUIRK_CLKDM_NOAUTO))
630 return;
631
632 pdata = dev_get_platdata(ddata->dev);
633 if (pdata && pdata->clkdm_allow_idle)
634 pdata->clkdm_allow_idle(ddata->dev, &ddata->cookie);
635 }
636
637 /**
638 * sysc_init_resets - init rstctrl reset line if configured
639 * @ddata: device driver data
640 *
641 * See sysc_rstctrl_reset_deassert().
642 */
sysc_init_resets(struct sysc * ddata)643 static int sysc_init_resets(struct sysc *ddata)
644 {
645 ddata->rsts =
646 devm_reset_control_get_optional_shared(ddata->dev, "rstctrl");
647
648 return PTR_ERR_OR_ZERO(ddata->rsts);
649 }
650
651 /**
652 * sysc_parse_and_check_child_range - parses module IO region from ranges
653 * @ddata: device driver data
654 *
655 * In general we only need rev, syss, and sysc registers and not the whole
656 * module range. But we do want the offsets for these registers from the
657 * module base. This allows us to check them against the legacy hwmod
658 * platform data. Let's also check the ranges are configured properly.
659 */
sysc_parse_and_check_child_range(struct sysc * ddata)660 static int sysc_parse_and_check_child_range(struct sysc *ddata)
661 {
662 struct device_node *np = ddata->dev->of_node;
663 struct of_range_parser parser;
664 struct of_range range;
665 int error;
666
667 error = of_range_parser_init(&parser, np);
668 if (error)
669 return error;
670
671 for_each_of_range(&parser, &range) {
672 ddata->module_pa = range.cpu_addr;
673 ddata->module_size = range.size;
674 break;
675 }
676
677 return 0;
678 }
679
680 static struct device_node *stdout_path;
681
sysc_init_stdout_path(struct sysc * ddata)682 static void sysc_init_stdout_path(struct sysc *ddata)
683 {
684 struct device_node *np = NULL;
685 const char *uart;
686
687 if (IS_ERR(stdout_path))
688 return;
689
690 if (stdout_path)
691 return;
692
693 np = of_find_node_by_path("/chosen");
694 if (!np)
695 goto err;
696
697 uart = of_get_property(np, "stdout-path", NULL);
698 if (!uart)
699 goto err;
700
701 np = of_find_node_by_path(uart);
702 if (!np)
703 goto err;
704
705 stdout_path = np;
706
707 return;
708
709 err:
710 stdout_path = ERR_PTR(-ENODEV);
711 }
712
sysc_check_quirk_stdout(struct sysc * ddata,struct device_node * np)713 static void sysc_check_quirk_stdout(struct sysc *ddata,
714 struct device_node *np)
715 {
716 sysc_init_stdout_path(ddata);
717 if (np != stdout_path)
718 return;
719
720 ddata->cfg.quirks |= SYSC_QUIRK_NO_IDLE_ON_INIT |
721 SYSC_QUIRK_NO_RESET_ON_INIT;
722 }
723
724 /**
725 * sysc_check_one_child - check child configuration
726 * @ddata: device driver data
727 * @np: child device node
728 *
729 * Let's avoid messy situations where we have new interconnect target
730 * node but children have "ti,hwmods". These belong to the interconnect
731 * target node and are managed by this driver.
732 */
sysc_check_one_child(struct sysc * ddata,struct device_node * np)733 static void sysc_check_one_child(struct sysc *ddata,
734 struct device_node *np)
735 {
736 const char *name;
737
738 name = of_get_property(np, "ti,hwmods", NULL);
739 if (name && !of_device_is_compatible(np, "ti,sysc"))
740 dev_warn(ddata->dev, "really a child ti,hwmods property?");
741
742 sysc_check_quirk_stdout(ddata, np);
743 sysc_parse_dts_quirks(ddata, np, true);
744 }
745
sysc_check_children(struct sysc * ddata)746 static void sysc_check_children(struct sysc *ddata)
747 {
748 struct device_node *child;
749
750 for_each_child_of_node(ddata->dev->of_node, child)
751 sysc_check_one_child(ddata, child);
752 }
753
754 /*
755 * So far only I2C uses 16-bit read access with clockactivity with revision
756 * in two registers with stride of 4. We can detect this based on the rev
757 * register size to configure things far enough to be able to properly read
758 * the revision register.
759 */
sysc_check_quirk_16bit(struct sysc * ddata,struct resource * res)760 static void sysc_check_quirk_16bit(struct sysc *ddata, struct resource *res)
761 {
762 if (resource_size(res) == 8)
763 ddata->cfg.quirks |= SYSC_QUIRK_16BIT | SYSC_QUIRK_USE_CLOCKACT;
764 }
765
766 /**
767 * sysc_parse_one - parses the interconnect target module registers
768 * @ddata: device driver data
769 * @reg: register to parse
770 */
sysc_parse_one(struct sysc * ddata,enum sysc_registers reg)771 static int sysc_parse_one(struct sysc *ddata, enum sysc_registers reg)
772 {
773 struct resource *res;
774 const char *name;
775
776 switch (reg) {
777 case SYSC_REVISION:
778 case SYSC_SYSCONFIG:
779 case SYSC_SYSSTATUS:
780 name = reg_names[reg];
781 break;
782 default:
783 return -EINVAL;
784 }
785
786 res = platform_get_resource_byname(to_platform_device(ddata->dev),
787 IORESOURCE_MEM, name);
788 if (!res) {
789 ddata->offsets[reg] = -ENODEV;
790
791 return 0;
792 }
793
794 ddata->offsets[reg] = res->start - ddata->module_pa;
795 if (reg == SYSC_REVISION)
796 sysc_check_quirk_16bit(ddata, res);
797
798 return 0;
799 }
800
sysc_parse_registers(struct sysc * ddata)801 static int sysc_parse_registers(struct sysc *ddata)
802 {
803 int i, error;
804
805 for (i = 0; i < SYSC_MAX_REGS; i++) {
806 error = sysc_parse_one(ddata, i);
807 if (error)
808 return error;
809 }
810
811 return 0;
812 }
813
814 /**
815 * sysc_check_registers - check for misconfigured register overlaps
816 * @ddata: device driver data
817 */
sysc_check_registers(struct sysc * ddata)818 static int sysc_check_registers(struct sysc *ddata)
819 {
820 int i, j, nr_regs = 0, nr_matches = 0;
821
822 for (i = 0; i < SYSC_MAX_REGS; i++) {
823 if (ddata->offsets[i] < 0)
824 continue;
825
826 if (ddata->offsets[i] > (ddata->module_size - 4)) {
827 dev_err(ddata->dev, "register outside module range");
828
829 return -EINVAL;
830 }
831
832 for (j = 0; j < SYSC_MAX_REGS; j++) {
833 if (ddata->offsets[j] < 0)
834 continue;
835
836 if (ddata->offsets[i] == ddata->offsets[j])
837 nr_matches++;
838 }
839 nr_regs++;
840 }
841
842 if (nr_matches > nr_regs) {
843 dev_err(ddata->dev, "overlapping registers: (%i/%i)",
844 nr_regs, nr_matches);
845
846 return -EINVAL;
847 }
848
849 return 0;
850 }
851
852 /**
853 * sysc_ioremap - ioremap register space for the interconnect target module
854 * @ddata: device driver data
855 *
856 * Note that the interconnect target module registers can be anywhere
857 * within the interconnect target module range. For example, SGX has
858 * them at offset 0x1fc00 in the 32MB module address space. And cpsw
859 * has them at offset 0x1200 in the CPSW_WR child. Usually the
860 * interconnect target module registers are at the beginning of
861 * the module range though.
862 */
sysc_ioremap(struct sysc * ddata)863 static int sysc_ioremap(struct sysc *ddata)
864 {
865 int size;
866
867 if (ddata->offsets[SYSC_REVISION] < 0 &&
868 ddata->offsets[SYSC_SYSCONFIG] < 0 &&
869 ddata->offsets[SYSC_SYSSTATUS] < 0) {
870 size = ddata->module_size;
871 } else {
872 size = max3(ddata->offsets[SYSC_REVISION],
873 ddata->offsets[SYSC_SYSCONFIG],
874 ddata->offsets[SYSC_SYSSTATUS]);
875
876 if (size < SZ_1K)
877 size = SZ_1K;
878
879 if ((size + sizeof(u32)) > ddata->module_size)
880 size = ddata->module_size;
881 }
882
883 ddata->module_va = devm_ioremap(ddata->dev,
884 ddata->module_pa,
885 size + sizeof(u32));
886 if (!ddata->module_va)
887 return -EIO;
888
889 return 0;
890 }
891
892 /**
893 * sysc_map_and_check_registers - ioremap and check device registers
894 * @ddata: device driver data
895 */
sysc_map_and_check_registers(struct sysc * ddata)896 static int sysc_map_and_check_registers(struct sysc *ddata)
897 {
898 struct device_node *np = ddata->dev->of_node;
899 int error;
900
901 error = sysc_parse_and_check_child_range(ddata);
902 if (error)
903 return error;
904
905 sysc_check_children(ddata);
906
907 if (!of_property_present(np, "reg"))
908 return 0;
909
910 error = sysc_parse_registers(ddata);
911 if (error)
912 return error;
913
914 error = sysc_ioremap(ddata);
915 if (error)
916 return error;
917
918 error = sysc_check_registers(ddata);
919 if (error)
920 return error;
921
922 return 0;
923 }
924
925 /**
926 * sysc_show_rev - read and show interconnect target module revision
927 * @bufp: buffer to print the information to
928 * @ddata: device driver data
929 */
sysc_show_rev(char * bufp,struct sysc * ddata)930 static int sysc_show_rev(char *bufp, struct sysc *ddata)
931 {
932 int len;
933
934 if (ddata->offsets[SYSC_REVISION] < 0)
935 return sprintf(bufp, ":NA");
936
937 len = sprintf(bufp, ":%08x", ddata->revision);
938
939 return len;
940 }
941
sysc_show_reg(struct sysc * ddata,char * bufp,enum sysc_registers reg)942 static int sysc_show_reg(struct sysc *ddata,
943 char *bufp, enum sysc_registers reg)
944 {
945 if (ddata->offsets[reg] < 0)
946 return sprintf(bufp, ":NA");
947
948 return sprintf(bufp, ":%x", ddata->offsets[reg]);
949 }
950
sysc_show_name(char * bufp,struct sysc * ddata)951 static int sysc_show_name(char *bufp, struct sysc *ddata)
952 {
953 if (!ddata->name)
954 return 0;
955
956 return sprintf(bufp, ":%s", ddata->name);
957 }
958
959 /**
960 * sysc_show_registers - show information about interconnect target module
961 * @ddata: device driver data
962 */
sysc_show_registers(struct sysc * ddata)963 static void sysc_show_registers(struct sysc *ddata)
964 {
965 char buf[128];
966 char *bufp = buf;
967 int i;
968
969 for (i = 0; i < SYSC_MAX_REGS; i++)
970 bufp += sysc_show_reg(ddata, bufp, i);
971
972 bufp += sysc_show_rev(bufp, ddata);
973 bufp += sysc_show_name(bufp, ddata);
974
975 dev_dbg(ddata->dev, "%llx:%x%s\n",
976 ddata->module_pa, ddata->module_size,
977 buf);
978 }
979
980 /**
981 * sysc_write_sysconfig - handle sysconfig quirks for register write
982 * @ddata: device driver data
983 * @value: register value
984 */
sysc_write_sysconfig(struct sysc * ddata,u32 value)985 static void sysc_write_sysconfig(struct sysc *ddata, u32 value)
986 {
987 if (ddata->module_unlock_quirk)
988 ddata->module_unlock_quirk(ddata);
989
990 sysc_write(ddata, ddata->offsets[SYSC_SYSCONFIG], value);
991
992 if (ddata->module_lock_quirk)
993 ddata->module_lock_quirk(ddata);
994 }
995
996 #define SYSC_IDLE_MASK (SYSC_NR_IDLEMODES - 1)
997 #define SYSC_CLOCACT_ICK 2
998
999 /* Caller needs to manage sysc_clkdm_deny_idle() and sysc_clkdm_allow_idle() */
sysc_enable_module(struct device * dev)1000 static int sysc_enable_module(struct device *dev)
1001 {
1002 struct sysc *ddata;
1003 const struct sysc_regbits *regbits;
1004 u32 reg, idlemodes, best_mode;
1005 int error;
1006
1007 ddata = dev_get_drvdata(dev);
1008
1009 /*
1010 * Some modules like DSS reset automatically on idle. Enable optional
1011 * reset clocks and wait for OCP softreset to complete.
1012 */
1013 if (ddata->cfg.quirks & SYSC_QUIRK_OPT_CLKS_IN_RESET) {
1014 error = sysc_enable_opt_clocks(ddata);
1015 if (error) {
1016 dev_err(ddata->dev,
1017 "Optional clocks failed for enable: %i\n",
1018 error);
1019 return error;
1020 }
1021 }
1022 /*
1023 * Some modules like i2c and hdq1w have unusable reset status unless
1024 * the module reset quirk is enabled. Skip status check on enable.
1025 */
1026 if (!(ddata->cfg.quirks & SYSC_MODULE_QUIRK_ENA_RESETDONE)) {
1027 error = sysc_wait_softreset(ddata);
1028 if (error)
1029 dev_warn(ddata->dev, "OCP softreset timed out\n");
1030 }
1031 if (ddata->cfg.quirks & SYSC_QUIRK_OPT_CLKS_IN_RESET)
1032 sysc_disable_opt_clocks(ddata);
1033
1034 /*
1035 * Some subsystem private interconnects, like DSS top level module,
1036 * need only the automatic OCP softreset handling with no sysconfig
1037 * register bits to configure.
1038 */
1039 if (ddata->offsets[SYSC_SYSCONFIG] == -ENODEV)
1040 return 0;
1041
1042 regbits = ddata->cap->regbits;
1043 reg = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1044
1045 /*
1046 * Set CLOCKACTIVITY, we only use it for ick. And we only configure it
1047 * based on the SYSC_QUIRK_USE_CLOCKACT flag, not based on the hardware
1048 * capabilities. See the old HWMOD_SET_DEFAULT_CLOCKACT flag.
1049 */
1050 if (regbits->clkact_shift >= 0 &&
1051 (ddata->cfg.quirks & SYSC_QUIRK_USE_CLOCKACT))
1052 reg |= SYSC_CLOCACT_ICK << regbits->clkact_shift;
1053
1054 /* Set SIDLE mode */
1055 idlemodes = ddata->cfg.sidlemodes;
1056 if (!idlemodes || regbits->sidle_shift < 0)
1057 goto set_midle;
1058
1059 if (ddata->cfg.quirks & (SYSC_QUIRK_SWSUP_SIDLE |
1060 SYSC_QUIRK_SWSUP_SIDLE_ACT)) {
1061 best_mode = SYSC_IDLE_NO;
1062
1063 /* Clear WAKEUP */
1064 if (regbits->enwkup_shift >= 0 &&
1065 ddata->cfg.sysc_val & BIT(regbits->enwkup_shift))
1066 reg &= ~BIT(regbits->enwkup_shift);
1067 } else {
1068 best_mode = fls(ddata->cfg.sidlemodes) - 1;
1069 if (best_mode > SYSC_IDLE_MASK) {
1070 dev_err(dev, "%s: invalid sidlemode\n", __func__);
1071 return -EINVAL;
1072 }
1073
1074 /* Set WAKEUP */
1075 if (regbits->enwkup_shift >= 0 &&
1076 ddata->cfg.sysc_val & BIT(regbits->enwkup_shift))
1077 reg |= BIT(regbits->enwkup_shift);
1078 }
1079
1080 reg &= ~(SYSC_IDLE_MASK << regbits->sidle_shift);
1081 reg |= best_mode << regbits->sidle_shift;
1082 sysc_write_sysconfig(ddata, reg);
1083
1084 set_midle:
1085 /* Set MIDLE mode */
1086 idlemodes = ddata->cfg.midlemodes;
1087 if (!idlemodes || regbits->midle_shift < 0)
1088 goto set_autoidle;
1089
1090 best_mode = fls(ddata->cfg.midlemodes) - 1;
1091 if (best_mode > SYSC_IDLE_MASK) {
1092 dev_err(dev, "%s: invalid midlemode\n", __func__);
1093 error = -EINVAL;
1094 goto save_context;
1095 }
1096
1097 if (ddata->cfg.quirks & SYSC_QUIRK_SWSUP_MSTANDBY)
1098 best_mode = SYSC_IDLE_NO;
1099
1100 reg &= ~(SYSC_IDLE_MASK << regbits->midle_shift);
1101 reg |= best_mode << regbits->midle_shift;
1102 sysc_write_sysconfig(ddata, reg);
1103
1104 set_autoidle:
1105 /* Autoidle bit must enabled separately if available */
1106 if (regbits->autoidle_shift >= 0 &&
1107 ddata->cfg.sysc_val & BIT(regbits->autoidle_shift)) {
1108 reg |= 1 << regbits->autoidle_shift;
1109 sysc_write_sysconfig(ddata, reg);
1110 }
1111
1112 error = 0;
1113
1114 save_context:
1115 /* Save context and flush posted write */
1116 ddata->sysconfig = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1117
1118 if (ddata->module_enable_quirk)
1119 ddata->module_enable_quirk(ddata);
1120
1121 return error;
1122 }
1123
sysc_best_idle_mode(u32 idlemodes,u32 * best_mode)1124 static int sysc_best_idle_mode(u32 idlemodes, u32 *best_mode)
1125 {
1126 if (idlemodes & BIT(SYSC_IDLE_SMART_WKUP))
1127 *best_mode = SYSC_IDLE_SMART_WKUP;
1128 else if (idlemodes & BIT(SYSC_IDLE_SMART))
1129 *best_mode = SYSC_IDLE_SMART;
1130 else if (idlemodes & BIT(SYSC_IDLE_FORCE))
1131 *best_mode = SYSC_IDLE_FORCE;
1132 else
1133 return -EINVAL;
1134
1135 return 0;
1136 }
1137
1138 /* Caller needs to manage sysc_clkdm_deny_idle() and sysc_clkdm_allow_idle() */
sysc_disable_module(struct device * dev)1139 static int sysc_disable_module(struct device *dev)
1140 {
1141 struct sysc *ddata;
1142 const struct sysc_regbits *regbits;
1143 u32 reg, idlemodes, best_mode;
1144 int ret;
1145
1146 ddata = dev_get_drvdata(dev);
1147 if (ddata->offsets[SYSC_SYSCONFIG] == -ENODEV)
1148 return 0;
1149
1150 if (ddata->module_disable_quirk)
1151 ddata->module_disable_quirk(ddata);
1152
1153 regbits = ddata->cap->regbits;
1154 reg = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1155
1156 /* Set MIDLE mode */
1157 idlemodes = ddata->cfg.midlemodes;
1158 if (!idlemodes || regbits->midle_shift < 0)
1159 goto set_sidle;
1160
1161 ret = sysc_best_idle_mode(idlemodes, &best_mode);
1162 if (ret) {
1163 dev_err(dev, "%s: invalid midlemode\n", __func__);
1164 return ret;
1165 }
1166
1167 if (ddata->cfg.quirks & (SYSC_QUIRK_SWSUP_MSTANDBY) ||
1168 ddata->cfg.quirks & (SYSC_QUIRK_FORCE_MSTANDBY))
1169 best_mode = SYSC_IDLE_FORCE;
1170
1171 reg &= ~(SYSC_IDLE_MASK << regbits->midle_shift);
1172 reg |= best_mode << regbits->midle_shift;
1173 sysc_write_sysconfig(ddata, reg);
1174
1175 set_sidle:
1176 /* Set SIDLE mode */
1177 idlemodes = ddata->cfg.sidlemodes;
1178 if (!idlemodes || regbits->sidle_shift < 0) {
1179 ret = 0;
1180 goto save_context;
1181 }
1182
1183 if (ddata->cfg.quirks & SYSC_QUIRK_SWSUP_SIDLE) {
1184 best_mode = SYSC_IDLE_FORCE;
1185 } else {
1186 ret = sysc_best_idle_mode(idlemodes, &best_mode);
1187 if (ret) {
1188 dev_err(dev, "%s: invalid sidlemode\n", __func__);
1189 ret = -EINVAL;
1190 goto save_context;
1191 }
1192 }
1193
1194 if (ddata->cfg.quirks & SYSC_QUIRK_SWSUP_SIDLE_ACT) {
1195 /* Set WAKEUP */
1196 if (regbits->enwkup_shift >= 0 &&
1197 ddata->cfg.sysc_val & BIT(regbits->enwkup_shift))
1198 reg |= BIT(regbits->enwkup_shift);
1199 }
1200
1201 reg &= ~(SYSC_IDLE_MASK << regbits->sidle_shift);
1202 reg |= best_mode << regbits->sidle_shift;
1203 if (regbits->autoidle_shift >= 0 &&
1204 ddata->cfg.sysc_val & BIT(regbits->autoidle_shift))
1205 reg |= 1 << regbits->autoidle_shift;
1206 sysc_write_sysconfig(ddata, reg);
1207
1208 ret = 0;
1209
1210 save_context:
1211 /* Save context and flush posted write */
1212 ddata->sysconfig = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1213
1214 return ret;
1215 }
1216
sysc_runtime_suspend_legacy(struct device * dev,struct sysc * ddata)1217 static int __maybe_unused sysc_runtime_suspend_legacy(struct device *dev,
1218 struct sysc *ddata)
1219 {
1220 struct ti_sysc_platform_data *pdata;
1221 int error;
1222
1223 pdata = dev_get_platdata(ddata->dev);
1224 if (!pdata)
1225 return 0;
1226
1227 if (!pdata->idle_module)
1228 return -ENODEV;
1229
1230 error = pdata->idle_module(dev, &ddata->cookie);
1231 if (error)
1232 dev_err(dev, "%s: could not idle: %i\n",
1233 __func__, error);
1234
1235 reset_control_assert(ddata->rsts);
1236
1237 return 0;
1238 }
1239
sysc_runtime_resume_legacy(struct device * dev,struct sysc * ddata)1240 static int __maybe_unused sysc_runtime_resume_legacy(struct device *dev,
1241 struct sysc *ddata)
1242 {
1243 struct ti_sysc_platform_data *pdata;
1244 int error;
1245
1246 pdata = dev_get_platdata(ddata->dev);
1247 if (!pdata)
1248 return 0;
1249
1250 if (!pdata->enable_module)
1251 return -ENODEV;
1252
1253 error = pdata->enable_module(dev, &ddata->cookie);
1254 if (error)
1255 dev_err(dev, "%s: could not enable: %i\n",
1256 __func__, error);
1257
1258 reset_control_deassert(ddata->rsts);
1259
1260 return 0;
1261 }
1262
sysc_runtime_suspend(struct device * dev)1263 static int __maybe_unused sysc_runtime_suspend(struct device *dev)
1264 {
1265 struct sysc *ddata;
1266 int error = 0;
1267
1268 ddata = dev_get_drvdata(dev);
1269
1270 if (!ddata->enabled)
1271 return 0;
1272
1273 sysc_clkdm_deny_idle(ddata);
1274
1275 if (ddata->legacy_mode) {
1276 error = sysc_runtime_suspend_legacy(dev, ddata);
1277 if (error)
1278 goto err_allow_idle;
1279 } else {
1280 error = sysc_disable_module(dev);
1281 if (error)
1282 goto err_allow_idle;
1283 }
1284
1285 sysc_disable_main_clocks(ddata);
1286
1287 if (sysc_opt_clks_needed(ddata))
1288 sysc_disable_opt_clocks(ddata);
1289
1290 ddata->enabled = false;
1291
1292 err_allow_idle:
1293 sysc_clkdm_allow_idle(ddata);
1294
1295 reset_control_assert(ddata->rsts);
1296
1297 return error;
1298 }
1299
sysc_runtime_resume(struct device * dev)1300 static int __maybe_unused sysc_runtime_resume(struct device *dev)
1301 {
1302 struct sysc *ddata;
1303 int error = 0;
1304
1305 ddata = dev_get_drvdata(dev);
1306
1307 if (ddata->enabled)
1308 return 0;
1309
1310
1311 sysc_clkdm_deny_idle(ddata);
1312
1313 if (sysc_opt_clks_needed(ddata)) {
1314 error = sysc_enable_opt_clocks(ddata);
1315 if (error)
1316 goto err_allow_idle;
1317 }
1318
1319 error = sysc_enable_main_clocks(ddata);
1320 if (error)
1321 goto err_opt_clocks;
1322
1323 reset_control_deassert(ddata->rsts);
1324
1325 if (ddata->legacy_mode) {
1326 error = sysc_runtime_resume_legacy(dev, ddata);
1327 if (error)
1328 goto err_main_clocks;
1329 } else {
1330 error = sysc_enable_module(dev);
1331 if (error)
1332 goto err_main_clocks;
1333 }
1334
1335 ddata->enabled = true;
1336
1337 sysc_clkdm_allow_idle(ddata);
1338
1339 return 0;
1340
1341 err_main_clocks:
1342 sysc_disable_main_clocks(ddata);
1343 err_opt_clocks:
1344 if (sysc_opt_clks_needed(ddata))
1345 sysc_disable_opt_clocks(ddata);
1346 err_allow_idle:
1347 sysc_clkdm_allow_idle(ddata);
1348
1349 return error;
1350 }
1351
1352 /*
1353 * Checks if device context was lost. Assumes the sysconfig register value
1354 * after lost context is different from the configured value. Only works for
1355 * enabled devices.
1356 *
1357 * Eventually we may want to also add support to using the context lost
1358 * registers that some SoCs have.
1359 */
sysc_check_context(struct sysc * ddata)1360 static int sysc_check_context(struct sysc *ddata)
1361 {
1362 u32 reg;
1363
1364 if (!ddata->enabled)
1365 return -ENODATA;
1366
1367 reg = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1368 if (reg == ddata->sysconfig)
1369 return 0;
1370
1371 return -EACCES;
1372 }
1373
sysc_reinit_module(struct sysc * ddata,bool leave_enabled)1374 static int sysc_reinit_module(struct sysc *ddata, bool leave_enabled)
1375 {
1376 struct device *dev = ddata->dev;
1377 int error;
1378
1379 if (ddata->enabled) {
1380 /* Nothing to do if enabled and context not lost */
1381 error = sysc_check_context(ddata);
1382 if (!error)
1383 return 0;
1384
1385 /* Disable target module if it is enabled */
1386 error = sysc_runtime_suspend(dev);
1387 if (error)
1388 dev_warn(dev, "reinit suspend failed: %i\n", error);
1389 }
1390
1391 /* Enable target module */
1392 error = sysc_runtime_resume(dev);
1393 if (error)
1394 dev_warn(dev, "reinit resume failed: %i\n", error);
1395
1396 /* Some modules like am335x gpmc need reset and restore of sysconfig */
1397 if (ddata->cfg.quirks & SYSC_QUIRK_RESET_ON_CTX_LOST) {
1398 error = sysc_reset(ddata);
1399 if (error)
1400 dev_warn(dev, "reinit reset failed: %i\n", error);
1401
1402 sysc_write_sysconfig(ddata, ddata->sysconfig);
1403 }
1404
1405 if (leave_enabled)
1406 return error;
1407
1408 /* Disable target module if no leave_enabled was set */
1409 error = sysc_runtime_suspend(dev);
1410 if (error)
1411 dev_warn(dev, "reinit suspend failed: %i\n", error);
1412
1413 return error;
1414 }
1415
sysc_noirq_suspend(struct device * dev)1416 static int __maybe_unused sysc_noirq_suspend(struct device *dev)
1417 {
1418 struct sysc *ddata;
1419
1420 ddata = dev_get_drvdata(dev);
1421
1422 if (ddata->cfg.quirks & SYSC_QUIRK_NO_IDLE)
1423 return 0;
1424
1425 if (!ddata->enabled)
1426 return 0;
1427
1428 ddata->needs_resume = 1;
1429
1430 return sysc_runtime_suspend(dev);
1431 }
1432
sysc_noirq_resume(struct device * dev)1433 static int __maybe_unused sysc_noirq_resume(struct device *dev)
1434 {
1435 struct sysc *ddata;
1436 int error = 0;
1437
1438 ddata = dev_get_drvdata(dev);
1439
1440 if (ddata->cfg.quirks & SYSC_QUIRK_NO_IDLE)
1441 return 0;
1442
1443 if (ddata->cfg.quirks & SYSC_QUIRK_REINIT_ON_RESUME) {
1444 error = sysc_reinit_module(ddata, ddata->needs_resume);
1445 if (error)
1446 dev_warn(dev, "noirq_resume failed: %i\n", error);
1447 } else if (ddata->needs_resume) {
1448 error = sysc_runtime_resume(dev);
1449 if (error)
1450 dev_warn(dev, "noirq_resume failed: %i\n", error);
1451 }
1452
1453 ddata->needs_resume = 0;
1454
1455 return error;
1456 }
1457
1458 static const struct dev_pm_ops sysc_pm_ops = {
1459 SET_NOIRQ_SYSTEM_SLEEP_PM_OPS(sysc_noirq_suspend, sysc_noirq_resume)
1460 SET_RUNTIME_PM_OPS(sysc_runtime_suspend,
1461 sysc_runtime_resume,
1462 NULL)
1463 };
1464
1465 /* Module revision register based quirks */
1466 struct sysc_revision_quirk {
1467 const char *name;
1468 u32 base;
1469 int rev_offset;
1470 int sysc_offset;
1471 int syss_offset;
1472 u32 revision;
1473 u32 revision_mask;
1474 u32 quirks;
1475 };
1476
1477 #define SYSC_QUIRK(optname, optbase, optrev, optsysc, optsyss, \
1478 optrev_val, optrevmask, optquirkmask) \
1479 { \
1480 .name = (optname), \
1481 .base = (optbase), \
1482 .rev_offset = (optrev), \
1483 .sysc_offset = (optsysc), \
1484 .syss_offset = (optsyss), \
1485 .revision = (optrev_val), \
1486 .revision_mask = (optrevmask), \
1487 .quirks = (optquirkmask), \
1488 }
1489
1490 static const struct sysc_revision_quirk sysc_revision_quirks[] = {
1491 /* Quirks that need to be set based on the module address */
1492 SYSC_QUIRK("mcpdm", 0x40132000, 0, 0x10, -ENODEV, 0x50000800, 0xffffffff,
1493 SYSC_QUIRK_EXT_OPT_CLOCK | SYSC_QUIRK_NO_RESET_ON_INIT |
1494 SYSC_QUIRK_SWSUP_SIDLE),
1495
1496 /* Quirks that need to be set based on detected module */
1497 SYSC_QUIRK("aess", 0, 0, 0x10, -ENODEV, 0x40000000, 0xffffffff,
1498 SYSC_MODULE_QUIRK_AESS),
1499 /* Errata i893 handling for dra7 dcan1 and 2 */
1500 SYSC_QUIRK("dcan", 0x4ae3c000, 0x20, -ENODEV, -ENODEV, 0xa3170504, 0xffffffff,
1501 SYSC_QUIRK_CLKDM_NOAUTO),
1502 SYSC_QUIRK("dcan", 0x48480000, 0x20, -ENODEV, -ENODEV, 0xa3170504, 0xffffffff,
1503 SYSC_QUIRK_CLKDM_NOAUTO),
1504 SYSC_QUIRK("dss", 0x4832a000, 0, 0x10, 0x14, 0x00000020, 0xffffffff,
1505 SYSC_QUIRK_OPT_CLKS_IN_RESET | SYSC_MODULE_QUIRK_DSS_RESET),
1506 SYSC_QUIRK("dss", 0x58000000, 0, -ENODEV, 0x14, 0x00000040, 0xffffffff,
1507 SYSC_QUIRK_OPT_CLKS_IN_RESET | SYSC_MODULE_QUIRK_DSS_RESET),
1508 SYSC_QUIRK("dss", 0x58000000, 0, -ENODEV, 0x14, 0x00000061, 0xffffffff,
1509 SYSC_QUIRK_OPT_CLKS_IN_RESET | SYSC_MODULE_QUIRK_DSS_RESET),
1510 SYSC_QUIRK("dwc3", 0x48880000, 0, 0x10, -ENODEV, 0x500a0200, 0xffffffff,
1511 SYSC_QUIRK_CLKDM_NOAUTO),
1512 SYSC_QUIRK("dwc3", 0x488c0000, 0, 0x10, -ENODEV, 0x500a0200, 0xffffffff,
1513 SYSC_QUIRK_CLKDM_NOAUTO),
1514 SYSC_QUIRK("gpio", 0, 0, 0x10, 0x114, 0x50600801, 0xffff00ff,
1515 SYSC_QUIRK_OPT_CLKS_IN_RESET),
1516 SYSC_QUIRK("gpmc", 0, 0, 0x10, 0x14, 0x00000060, 0xffffffff,
1517 SYSC_QUIRK_REINIT_ON_CTX_LOST | SYSC_QUIRK_RESET_ON_CTX_LOST |
1518 SYSC_QUIRK_GPMC_DEBUG),
1519 SYSC_QUIRK("hdmi", 0, 0, 0x10, -ENODEV, 0x50030200, 0xffffffff,
1520 SYSC_QUIRK_OPT_CLKS_NEEDED),
1521 SYSC_QUIRK("hdq1w", 0, 0, 0x14, 0x18, 0x00000006, 0xffffffff,
1522 SYSC_MODULE_QUIRK_HDQ1W | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1523 SYSC_QUIRK("hdq1w", 0, 0, 0x14, 0x18, 0x0000000a, 0xffffffff,
1524 SYSC_MODULE_QUIRK_HDQ1W | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1525 SYSC_QUIRK("i2c", 0, 0, 0x20, 0x10, 0x00000036, 0x000000ff,
1526 SYSC_MODULE_QUIRK_I2C | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1527 SYSC_QUIRK("i2c", 0, 0, 0x20, 0x10, 0x0000003c, 0x000000ff,
1528 SYSC_MODULE_QUIRK_I2C | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1529 SYSC_QUIRK("i2c", 0, 0, 0x20, 0x10, 0x00000040, 0x000000ff,
1530 SYSC_MODULE_QUIRK_I2C | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1531 SYSC_QUIRK("i2c", 0, 0, 0x10, 0x90, 0x5040000a, 0xfffff0f0,
1532 SYSC_MODULE_QUIRK_I2C | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1533 SYSC_QUIRK("gpu", 0x50000000, 0x14, -ENODEV, -ENODEV, 0x00010201, 0xffffffff, 0),
1534 SYSC_QUIRK("gpu", 0x50000000, 0xfe00, 0xfe10, -ENODEV, 0x40000000 , 0xffffffff,
1535 SYSC_MODULE_QUIRK_SGX),
1536 SYSC_QUIRK("lcdc", 0, 0, 0x54, -ENODEV, 0x4f201000, 0xffffffff,
1537 SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1538 SYSC_QUIRK("mcasp", 0, 0, 0x4, -ENODEV, 0x44306302, 0xffffffff,
1539 SYSC_QUIRK_SWSUP_SIDLE),
1540 SYSC_QUIRK("rtc", 0, 0x74, 0x78, -ENODEV, 0x4eb01908, 0xffff00f0,
1541 SYSC_MODULE_QUIRK_RTC_UNLOCK),
1542 SYSC_QUIRK("tptc", 0, 0, 0x10, -ENODEV, 0x40006c00, 0xffffefff,
1543 SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1544 SYSC_QUIRK("tptc", 0, 0, -ENODEV, -ENODEV, 0x40007c00, 0xffffffff,
1545 SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1546 SYSC_QUIRK("sata", 0, 0xfc, 0x1100, -ENODEV, 0x5e412000, 0xffffffff,
1547 SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1548 SYSC_QUIRK("uart", 0, 0x50, 0x54, 0x58, 0x00000046, 0xffffffff,
1549 SYSC_QUIRK_SWSUP_SIDLE_ACT),
1550 SYSC_QUIRK("uart", 0, 0x50, 0x54, 0x58, 0x00000052, 0xffffffff,
1551 SYSC_QUIRK_SWSUP_SIDLE_ACT),
1552 /* Uarts on omap4 and later */
1553 SYSC_QUIRK("uart", 0, 0x50, 0x54, 0x58, 0x50411e03, 0xffff00ff,
1554 SYSC_QUIRK_SWSUP_SIDLE_ACT),
1555 SYSC_QUIRK("uart", 0, 0x50, 0x54, 0x58, 0x47422e03, 0xffffffff,
1556 SYSC_QUIRK_SWSUP_SIDLE_ACT),
1557 SYSC_QUIRK("uart", 0, 0x50, 0x54, 0x58, 0x47424e03, 0xffffffff,
1558 SYSC_QUIRK_SWSUP_SIDLE_ACT),
1559 SYSC_QUIRK("usb_host_hs", 0, 0, 0x10, 0x14, 0x50700100, 0xffffffff,
1560 SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1561 SYSC_QUIRK("usb_host_hs", 0, 0, 0x10, -ENODEV, 0x50700101, 0xffffffff,
1562 SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1563 SYSC_QUIRK("usb_otg_hs", 0, 0x400, 0x404, 0x408, 0x00000033,
1564 0xffffffff, SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY |
1565 SYSC_MODULE_QUIRK_OTG),
1566 SYSC_QUIRK("usb_otg_hs", 0, 0x400, 0x404, 0x408, 0x00000040,
1567 0xffffffff, SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY |
1568 SYSC_MODULE_QUIRK_OTG),
1569 SYSC_QUIRK("usb_otg_hs", 0, 0x400, 0x404, 0x408, 0x00000050,
1570 0xffffffff, SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY |
1571 SYSC_MODULE_QUIRK_OTG),
1572 SYSC_QUIRK("usb_otg_hs", 0, 0, 0x10, -ENODEV, 0x4ea2080d, 0xffffffff,
1573 SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY |
1574 SYSC_QUIRK_REINIT_ON_CTX_LOST),
1575 SYSC_QUIRK("wdt", 0, 0, 0x10, 0x14, 0x502a0500, 0xfffff0f0,
1576 SYSC_MODULE_QUIRK_WDT),
1577 /* PRUSS on am3, am4 and am5 */
1578 SYSC_QUIRK("pruss", 0, 0x26000, 0x26004, -ENODEV, 0x47000000, 0xff000000,
1579 SYSC_MODULE_QUIRK_PRUSS),
1580 /* Watchdog on am3 and am4 */
1581 SYSC_QUIRK("wdt", 0x44e35000, 0, 0x10, 0x14, 0x502a0500, 0xfffff0f0,
1582 SYSC_MODULE_QUIRK_WDT | SYSC_QUIRK_SWSUP_SIDLE),
1583
1584 #ifdef DEBUG
1585 SYSC_QUIRK("adc", 0, 0, 0x10, -ENODEV, 0x47300001, 0xffffffff, 0),
1586 SYSC_QUIRK("atl", 0, 0, -ENODEV, -ENODEV, 0x0a070100, 0xffffffff, 0),
1587 SYSC_QUIRK("cm", 0, 0, -ENODEV, -ENODEV, 0x40000301, 0xffffffff, 0),
1588 SYSC_QUIRK("control", 0, 0, 0x10, -ENODEV, 0x40000900, 0xffffffff, 0),
1589 SYSC_QUIRK("cpgmac", 0, 0x1200, 0x1208, 0x1204, 0x4edb1902,
1590 0xffff00f0, 0),
1591 SYSC_QUIRK("dcan", 0, 0x20, -ENODEV, -ENODEV, 0xa3170504, 0xffffffff, 0),
1592 SYSC_QUIRK("dcan", 0, 0x20, -ENODEV, -ENODEV, 0x4edb1902, 0xffffffff, 0),
1593 SYSC_QUIRK("dispc", 0x4832a400, 0, 0x10, 0x14, 0x00000030, 0xffffffff, 0),
1594 SYSC_QUIRK("dispc", 0x58001000, 0, 0x10, 0x14, 0x00000040, 0xffffffff, 0),
1595 SYSC_QUIRK("dispc", 0x58001000, 0, 0x10, 0x14, 0x00000051, 0xffffffff, 0),
1596 SYSC_QUIRK("dmic", 0, 0, 0x10, -ENODEV, 0x50010000, 0xffffffff, 0),
1597 SYSC_QUIRK("dsi", 0x58004000, 0, 0x10, 0x14, 0x00000030, 0xffffffff, 0),
1598 SYSC_QUIRK("dsi", 0x58005000, 0, 0x10, 0x14, 0x00000030, 0xffffffff, 0),
1599 SYSC_QUIRK("dsi", 0x58005000, 0, 0x10, 0x14, 0x00000040, 0xffffffff, 0),
1600 SYSC_QUIRK("dsi", 0x58009000, 0, 0x10, 0x14, 0x00000040, 0xffffffff, 0),
1601 SYSC_QUIRK("dwc3", 0, 0, 0x10, -ENODEV, 0x500a0200, 0xffffffff, 0),
1602 SYSC_QUIRK("d2d", 0x4a0b6000, 0, 0x10, 0x14, 0x00000010, 0xffffffff, 0),
1603 SYSC_QUIRK("d2d", 0x4a0cd000, 0, 0x10, 0x14, 0x00000010, 0xffffffff, 0),
1604 SYSC_QUIRK("elm", 0x48080000, 0, 0x10, 0x14, 0x00000020, 0xffffffff, 0),
1605 SYSC_QUIRK("emif", 0, 0, -ENODEV, -ENODEV, 0x40441403, 0xffff0fff, 0),
1606 SYSC_QUIRK("emif", 0, 0, -ENODEV, -ENODEV, 0x50440500, 0xffffffff, 0),
1607 SYSC_QUIRK("epwmss", 0, 0, 0x4, -ENODEV, 0x47400001, 0xffffffff, 0),
1608 SYSC_QUIRK("gpu", 0, 0x1fc00, 0x1fc10, -ENODEV, 0, 0, 0),
1609 SYSC_QUIRK("gpu", 0, 0xfe00, 0xfe10, -ENODEV, 0x40000000 , 0xffffffff, 0),
1610 SYSC_QUIRK("hdmi", 0, 0, 0x10, -ENODEV, 0x50031d00, 0xffffffff, 0),
1611 SYSC_QUIRK("hsi", 0, 0, 0x10, 0x14, 0x50043101, 0xffffffff, 0),
1612 SYSC_QUIRK("iss", 0, 0, 0x10, -ENODEV, 0x40000101, 0xffffffff, 0),
1613 SYSC_QUIRK("keypad", 0x4a31c000, 0, 0x10, 0x14, 0x00000020, 0xffffffff, 0),
1614 SYSC_QUIRK("mcasp", 0, 0, 0x4, -ENODEV, 0x44307b02, 0xffffffff, 0),
1615 SYSC_QUIRK("mcbsp", 0, -ENODEV, 0x8c, -ENODEV, 0, 0, 0),
1616 SYSC_QUIRK("mcspi", 0, 0, 0x10, -ENODEV, 0x40300a0b, 0xffff00ff, 0),
1617 SYSC_QUIRK("mcspi", 0, 0, 0x110, 0x114, 0x40300a0b, 0xffffffff, 0),
1618 SYSC_QUIRK("mailbox", 0, 0, 0x10, -ENODEV, 0x00000400, 0xffffffff, 0),
1619 SYSC_QUIRK("m3", 0, 0, -ENODEV, -ENODEV, 0x5f580105, 0x0fff0f00, 0),
1620 SYSC_QUIRK("ocp2scp", 0, 0, 0x10, 0x14, 0x50060005, 0xfffffff0, 0),
1621 SYSC_QUIRK("ocp2scp", 0, 0, -ENODEV, -ENODEV, 0x50060007, 0xffffffff, 0),
1622 SYSC_QUIRK("padconf", 0, 0, 0x10, -ENODEV, 0x4fff0800, 0xffffffff, 0),
1623 SYSC_QUIRK("padconf", 0, 0, -ENODEV, -ENODEV, 0x40001100, 0xffffffff, 0),
1624 SYSC_QUIRK("pcie", 0x51000000, -ENODEV, -ENODEV, -ENODEV, 0, 0, 0),
1625 SYSC_QUIRK("pcie", 0x51800000, -ENODEV, -ENODEV, -ENODEV, 0, 0, 0),
1626 SYSC_QUIRK("prcm", 0, 0, -ENODEV, -ENODEV, 0x40000100, 0xffffffff, 0),
1627 SYSC_QUIRK("prcm", 0, 0, -ENODEV, -ENODEV, 0x00004102, 0xffffffff, 0),
1628 SYSC_QUIRK("prcm", 0, 0, -ENODEV, -ENODEV, 0x40000400, 0xffffffff, 0),
1629 SYSC_QUIRK("rfbi", 0x4832a800, 0, 0x10, 0x14, 0x00000010, 0xffffffff, 0),
1630 SYSC_QUIRK("rfbi", 0x58002000, 0, 0x10, 0x14, 0x00000010, 0xffffffff, 0),
1631 SYSC_QUIRK("scm", 0, 0, 0x10, -ENODEV, 0x40000900, 0xffffffff, 0),
1632 SYSC_QUIRK("scm", 0, 0, -ENODEV, -ENODEV, 0x4e8b0100, 0xffffffff, 0),
1633 SYSC_QUIRK("scm", 0, 0, -ENODEV, -ENODEV, 0x4f000100, 0xffffffff, 0),
1634 SYSC_QUIRK("scm", 0, 0, -ENODEV, -ENODEV, 0x40000900, 0xffffffff, 0),
1635 SYSC_QUIRK("scrm", 0, 0, -ENODEV, -ENODEV, 0x00000010, 0xffffffff, 0),
1636 SYSC_QUIRK("sdio", 0, 0, 0x10, -ENODEV, 0x40202301, 0xffff0ff0, 0),
1637 SYSC_QUIRK("sdio", 0, 0x2fc, 0x110, 0x114, 0x31010000, 0xffffffff, 0),
1638 SYSC_QUIRK("sdma", 0, 0, 0x2c, 0x28, 0x00010900, 0xffffffff, 0),
1639 SYSC_QUIRK("sham", 0, 0x100, 0x110, 0x114, 0x40000c03, 0xffffffff, 0),
1640 SYSC_QUIRK("slimbus", 0, 0, 0x10, -ENODEV, 0x40000902, 0xffffffff, 0),
1641 SYSC_QUIRK("slimbus", 0, 0, 0x10, -ENODEV, 0x40002903, 0xffffffff, 0),
1642 SYSC_QUIRK("smartreflex", 0, -ENODEV, 0x24, -ENODEV, 0x00000000, 0xffffffff, 0),
1643 SYSC_QUIRK("smartreflex", 0, -ENODEV, 0x38, -ENODEV, 0x00000000, 0xffffffff, 0),
1644 SYSC_QUIRK("spinlock", 0, 0, 0x10, -ENODEV, 0x50020000, 0xffffffff, 0),
1645 SYSC_QUIRK("rng", 0, 0x1fe0, 0x1fe4, -ENODEV, 0x00000020, 0xffffffff, 0),
1646 SYSC_QUIRK("timer", 0, 0, 0x10, 0x14, 0x00000013, 0xffffffff, 0),
1647 SYSC_QUIRK("timer", 0, 0, 0x10, 0x14, 0x00000015, 0xffffffff, 0),
1648 /* Some timers on omap4 and later */
1649 SYSC_QUIRK("timer", 0, 0, 0x10, -ENODEV, 0x50002100, 0xffffffff, 0),
1650 SYSC_QUIRK("timer", 0, 0, 0x10, -ENODEV, 0x4fff1301, 0xffff00ff, 0),
1651 SYSC_QUIRK("timer32k", 0, 0, 0x4, -ENODEV, 0x00000040, 0xffffffff, 0),
1652 SYSC_QUIRK("timer32k", 0, 0, 0x4, -ENODEV, 0x00000011, 0xffffffff, 0),
1653 SYSC_QUIRK("timer32k", 0, 0, 0x4, -ENODEV, 0x00000060, 0xffffffff, 0),
1654 SYSC_QUIRK("tpcc", 0, 0, -ENODEV, -ENODEV, 0x40014c00, 0xffffffff, 0),
1655 SYSC_QUIRK("usbhstll", 0, 0, 0x10, 0x14, 0x00000004, 0xffffffff, 0),
1656 SYSC_QUIRK("usbhstll", 0, 0, 0x10, 0x14, 0x00000008, 0xffffffff, 0),
1657 SYSC_QUIRK("venc", 0x58003000, 0, -ENODEV, -ENODEV, 0x00000002, 0xffffffff, 0),
1658 SYSC_QUIRK("vfpe", 0, 0, 0x104, -ENODEV, 0x4d001200, 0xffffffff, 0),
1659 #endif
1660 };
1661
1662 /*
1663 * Early quirks based on module base and register offsets only that are
1664 * needed before the module revision can be read
1665 */
sysc_init_early_quirks(struct sysc * ddata)1666 static void sysc_init_early_quirks(struct sysc *ddata)
1667 {
1668 const struct sysc_revision_quirk *q;
1669 int i;
1670
1671 for (i = 0; i < ARRAY_SIZE(sysc_revision_quirks); i++) {
1672 q = &sysc_revision_quirks[i];
1673
1674 if (!q->base)
1675 continue;
1676
1677 if (q->base != ddata->module_pa)
1678 continue;
1679
1680 if (q->rev_offset != ddata->offsets[SYSC_REVISION])
1681 continue;
1682
1683 if (q->sysc_offset != ddata->offsets[SYSC_SYSCONFIG])
1684 continue;
1685
1686 if (q->syss_offset != ddata->offsets[SYSC_SYSSTATUS])
1687 continue;
1688
1689 ddata->name = q->name;
1690 ddata->cfg.quirks |= q->quirks;
1691 }
1692 }
1693
1694 /* Quirks that also consider the revision register value */
sysc_init_revision_quirks(struct sysc * ddata)1695 static void sysc_init_revision_quirks(struct sysc *ddata)
1696 {
1697 const struct sysc_revision_quirk *q;
1698 int i;
1699
1700 for (i = 0; i < ARRAY_SIZE(sysc_revision_quirks); i++) {
1701 q = &sysc_revision_quirks[i];
1702
1703 if (q->base && q->base != ddata->module_pa)
1704 continue;
1705
1706 if (q->rev_offset != ddata->offsets[SYSC_REVISION])
1707 continue;
1708
1709 if (q->sysc_offset != ddata->offsets[SYSC_SYSCONFIG])
1710 continue;
1711
1712 if (q->syss_offset != ddata->offsets[SYSC_SYSSTATUS])
1713 continue;
1714
1715 if (q->revision == ddata->revision ||
1716 (q->revision & q->revision_mask) ==
1717 (ddata->revision & q->revision_mask)) {
1718 ddata->name = q->name;
1719 ddata->cfg.quirks |= q->quirks;
1720 }
1721 }
1722 }
1723
1724 /*
1725 * DSS needs dispc outputs disabled to reset modules. Returns mask of
1726 * enabled DSS interrupts. Eventually we may be able to do this on
1727 * dispc init rather than top-level DSS init.
1728 */
sysc_quirk_dispc(struct sysc * ddata,int dispc_offset,bool disable)1729 static u32 sysc_quirk_dispc(struct sysc *ddata, int dispc_offset,
1730 bool disable)
1731 {
1732 bool lcd_en, digit_en, lcd2_en = false, lcd3_en = false;
1733 const int lcd_en_mask = BIT(0), digit_en_mask = BIT(1);
1734 int manager_count;
1735 bool framedonetv_irq = true;
1736 u32 val, irq_mask = 0;
1737
1738 switch (sysc_soc->soc) {
1739 case SOC_2420 ... SOC_3630:
1740 manager_count = 2;
1741 framedonetv_irq = false;
1742 break;
1743 case SOC_4430 ... SOC_4470:
1744 manager_count = 3;
1745 break;
1746 case SOC_5430:
1747 case SOC_DRA7:
1748 manager_count = 4;
1749 break;
1750 case SOC_AM4:
1751 manager_count = 1;
1752 framedonetv_irq = false;
1753 break;
1754 case SOC_UNKNOWN:
1755 default:
1756 return 0;
1757 }
1758
1759 /* Remap the whole module range to be able to reset dispc outputs */
1760 devm_iounmap(ddata->dev, ddata->module_va);
1761 ddata->module_va = devm_ioremap(ddata->dev,
1762 ddata->module_pa,
1763 ddata->module_size);
1764 if (!ddata->module_va)
1765 return -EIO;
1766
1767 /* DISP_CONTROL, shut down lcd and digit on disable if enabled */
1768 val = sysc_read(ddata, dispc_offset + 0x40);
1769 lcd_en = val & lcd_en_mask;
1770 digit_en = val & digit_en_mask;
1771 if (lcd_en)
1772 irq_mask |= BIT(0); /* FRAMEDONE */
1773 if (digit_en) {
1774 if (framedonetv_irq)
1775 irq_mask |= BIT(24); /* FRAMEDONETV */
1776 else
1777 irq_mask |= BIT(2) | BIT(3); /* EVSYNC bits */
1778 }
1779 if (disable && (lcd_en || digit_en))
1780 sysc_write(ddata, dispc_offset + 0x40,
1781 val & ~(lcd_en_mask | digit_en_mask));
1782
1783 if (manager_count <= 2)
1784 return irq_mask;
1785
1786 /* DISPC_CONTROL2 */
1787 val = sysc_read(ddata, dispc_offset + 0x238);
1788 lcd2_en = val & lcd_en_mask;
1789 if (lcd2_en)
1790 irq_mask |= BIT(22); /* FRAMEDONE2 */
1791 if (disable && lcd2_en)
1792 sysc_write(ddata, dispc_offset + 0x238,
1793 val & ~lcd_en_mask);
1794
1795 if (manager_count <= 3)
1796 return irq_mask;
1797
1798 /* DISPC_CONTROL3 */
1799 val = sysc_read(ddata, dispc_offset + 0x848);
1800 lcd3_en = val & lcd_en_mask;
1801 if (lcd3_en)
1802 irq_mask |= BIT(30); /* FRAMEDONE3 */
1803 if (disable && lcd3_en)
1804 sysc_write(ddata, dispc_offset + 0x848,
1805 val & ~lcd_en_mask);
1806
1807 return irq_mask;
1808 }
1809
1810 /* DSS needs child outputs disabled and SDI registers cleared for reset */
sysc_pre_reset_quirk_dss(struct sysc * ddata)1811 static void sysc_pre_reset_quirk_dss(struct sysc *ddata)
1812 {
1813 const int dispc_offset = 0x1000;
1814 int error;
1815 u32 irq_mask, val;
1816
1817 /* Get enabled outputs */
1818 irq_mask = sysc_quirk_dispc(ddata, dispc_offset, false);
1819 if (!irq_mask)
1820 return;
1821
1822 /* Clear IRQSTATUS */
1823 sysc_write(ddata, dispc_offset + 0x18, irq_mask);
1824
1825 /* Disable outputs */
1826 val = sysc_quirk_dispc(ddata, dispc_offset, true);
1827
1828 /* Poll IRQSTATUS */
1829 error = readl_poll_timeout(ddata->module_va + dispc_offset + 0x18,
1830 val, val != irq_mask, 100, 50);
1831 if (error)
1832 dev_warn(ddata->dev, "%s: timed out %08x !+ %08x\n",
1833 __func__, val, irq_mask);
1834
1835 if (sysc_soc->soc == SOC_3430 || sysc_soc->soc == SOC_AM35) {
1836 /* Clear DSS_SDI_CONTROL */
1837 sysc_write(ddata, 0x44, 0);
1838
1839 /* Clear DSS_PLL_CONTROL */
1840 sysc_write(ddata, 0x48, 0);
1841 }
1842
1843 /* Clear DSS_CONTROL to switch DSS clock sources to PRCM if not */
1844 sysc_write(ddata, 0x40, 0);
1845 }
1846
1847 /* 1-wire needs module's internal clocks enabled for reset */
sysc_pre_reset_quirk_hdq1w(struct sysc * ddata)1848 static void sysc_pre_reset_quirk_hdq1w(struct sysc *ddata)
1849 {
1850 int offset = 0x0c; /* HDQ_CTRL_STATUS */
1851 u16 val;
1852
1853 val = sysc_read(ddata, offset);
1854 val |= BIT(5);
1855 sysc_write(ddata, offset, val);
1856 }
1857
1858 /* AESS (Audio Engine SubSystem) needs autogating set after enable */
sysc_module_enable_quirk_aess(struct sysc * ddata)1859 static void sysc_module_enable_quirk_aess(struct sysc *ddata)
1860 {
1861 int offset = 0x7c; /* AESS_AUTO_GATING_ENABLE */
1862
1863 sysc_write(ddata, offset, 1);
1864 }
1865
1866 /* I2C needs to be disabled for reset */
sysc_clk_quirk_i2c(struct sysc * ddata,bool enable)1867 static void sysc_clk_quirk_i2c(struct sysc *ddata, bool enable)
1868 {
1869 int offset;
1870 u16 val;
1871
1872 /* I2C_CON, omap2/3 is different from omap4 and later */
1873 if ((ddata->revision & 0xffffff00) == 0x001f0000)
1874 offset = 0x24;
1875 else
1876 offset = 0xa4;
1877
1878 /* I2C_EN */
1879 val = sysc_read(ddata, offset);
1880 if (enable)
1881 val |= BIT(15);
1882 else
1883 val &= ~BIT(15);
1884 sysc_write(ddata, offset, val);
1885 }
1886
sysc_pre_reset_quirk_i2c(struct sysc * ddata)1887 static void sysc_pre_reset_quirk_i2c(struct sysc *ddata)
1888 {
1889 sysc_clk_quirk_i2c(ddata, false);
1890 }
1891
sysc_post_reset_quirk_i2c(struct sysc * ddata)1892 static void sysc_post_reset_quirk_i2c(struct sysc *ddata)
1893 {
1894 sysc_clk_quirk_i2c(ddata, true);
1895 }
1896
1897 /* RTC on am3 and 4 needs to be unlocked and locked for sysconfig */
sysc_quirk_rtc(struct sysc * ddata,bool lock)1898 static void sysc_quirk_rtc(struct sysc *ddata, bool lock)
1899 {
1900 u32 val, kick0_val = 0, kick1_val = 0;
1901 unsigned long flags;
1902 int error;
1903
1904 if (!lock) {
1905 kick0_val = 0x83e70b13;
1906 kick1_val = 0x95a4f1e0;
1907 }
1908
1909 local_irq_save(flags);
1910 /* RTC_STATUS BUSY bit may stay active for 1/32768 seconds (~30 usec) */
1911 error = readl_poll_timeout_atomic(ddata->module_va + 0x44, val,
1912 !(val & BIT(0)), 100, 50);
1913 if (error)
1914 dev_warn(ddata->dev, "rtc busy timeout\n");
1915 /* Now we have ~15 microseconds to read/write various registers */
1916 sysc_write(ddata, 0x6c, kick0_val);
1917 sysc_write(ddata, 0x70, kick1_val);
1918 local_irq_restore(flags);
1919 }
1920
sysc_module_unlock_quirk_rtc(struct sysc * ddata)1921 static void sysc_module_unlock_quirk_rtc(struct sysc *ddata)
1922 {
1923 sysc_quirk_rtc(ddata, false);
1924 }
1925
sysc_module_lock_quirk_rtc(struct sysc * ddata)1926 static void sysc_module_lock_quirk_rtc(struct sysc *ddata)
1927 {
1928 sysc_quirk_rtc(ddata, true);
1929 }
1930
1931 /* OTG omap2430 glue layer up to omap4 needs OTG_FORCESTDBY configured */
sysc_module_enable_quirk_otg(struct sysc * ddata)1932 static void sysc_module_enable_quirk_otg(struct sysc *ddata)
1933 {
1934 int offset = 0x414; /* OTG_FORCESTDBY */
1935
1936 sysc_write(ddata, offset, 0);
1937 }
1938
sysc_module_disable_quirk_otg(struct sysc * ddata)1939 static void sysc_module_disable_quirk_otg(struct sysc *ddata)
1940 {
1941 int offset = 0x414; /* OTG_FORCESTDBY */
1942 u32 val = BIT(0); /* ENABLEFORCE */
1943
1944 sysc_write(ddata, offset, val);
1945 }
1946
1947 /* 36xx SGX needs a quirk for to bypass OCP IPG interrupt logic */
sysc_module_enable_quirk_sgx(struct sysc * ddata)1948 static void sysc_module_enable_quirk_sgx(struct sysc *ddata)
1949 {
1950 int offset = 0xff08; /* OCP_DEBUG_CONFIG */
1951 u32 val = BIT(31); /* THALIA_INT_BYPASS */
1952
1953 sysc_write(ddata, offset, val);
1954 }
1955
1956 /* Watchdog timer needs a disable sequence after reset */
sysc_reset_done_quirk_wdt(struct sysc * ddata)1957 static void sysc_reset_done_quirk_wdt(struct sysc *ddata)
1958 {
1959 int wps, spr, error;
1960 u32 val;
1961
1962 wps = 0x34;
1963 spr = 0x48;
1964
1965 sysc_write(ddata, spr, 0xaaaa);
1966 error = readl_poll_timeout(ddata->module_va + wps, val,
1967 !(val & 0x10), 100,
1968 MAX_MODULE_SOFTRESET_WAIT);
1969 if (error)
1970 dev_warn(ddata->dev, "wdt disable step1 failed\n");
1971
1972 sysc_write(ddata, spr, 0x5555);
1973 error = readl_poll_timeout(ddata->module_va + wps, val,
1974 !(val & 0x10), 100,
1975 MAX_MODULE_SOFTRESET_WAIT);
1976 if (error)
1977 dev_warn(ddata->dev, "wdt disable step2 failed\n");
1978 }
1979
1980 /* PRUSS needs to set MSTANDBY_INIT inorder to idle properly */
sysc_module_disable_quirk_pruss(struct sysc * ddata)1981 static void sysc_module_disable_quirk_pruss(struct sysc *ddata)
1982 {
1983 u32 reg;
1984
1985 reg = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1986 reg |= SYSC_PRUSS_STANDBY_INIT;
1987 sysc_write(ddata, ddata->offsets[SYSC_SYSCONFIG], reg);
1988 }
1989
sysc_module_enable_quirk_pruss(struct sysc * ddata)1990 static void sysc_module_enable_quirk_pruss(struct sysc *ddata)
1991 {
1992 u32 reg;
1993
1994 reg = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1995
1996 /*
1997 * Clearing the SYSC_PRUSS_STANDBY_INIT bit - Updates OCP master
1998 * port configuration to enable memory access outside of the
1999 * PRU-ICSS subsystem.
2000 */
2001 reg &= (~SYSC_PRUSS_STANDBY_INIT);
2002 sysc_write(ddata, ddata->offsets[SYSC_SYSCONFIG], reg);
2003 }
2004
sysc_init_module_quirks(struct sysc * ddata)2005 static void sysc_init_module_quirks(struct sysc *ddata)
2006 {
2007 if (ddata->legacy_mode || !ddata->name)
2008 return;
2009
2010 if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_HDQ1W) {
2011 ddata->pre_reset_quirk = sysc_pre_reset_quirk_hdq1w;
2012
2013 return;
2014 }
2015
2016 #ifdef CONFIG_OMAP_GPMC_DEBUG
2017 if (ddata->cfg.quirks & SYSC_QUIRK_GPMC_DEBUG) {
2018 ddata->cfg.quirks |= SYSC_QUIRK_NO_RESET_ON_INIT;
2019
2020 return;
2021 }
2022 #endif
2023
2024 if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_I2C) {
2025 ddata->pre_reset_quirk = sysc_pre_reset_quirk_i2c;
2026 ddata->post_reset_quirk = sysc_post_reset_quirk_i2c;
2027
2028 return;
2029 }
2030
2031 if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_AESS)
2032 ddata->module_enable_quirk = sysc_module_enable_quirk_aess;
2033
2034 if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_DSS_RESET)
2035 ddata->pre_reset_quirk = sysc_pre_reset_quirk_dss;
2036
2037 if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_RTC_UNLOCK) {
2038 ddata->module_unlock_quirk = sysc_module_unlock_quirk_rtc;
2039 ddata->module_lock_quirk = sysc_module_lock_quirk_rtc;
2040
2041 return;
2042 }
2043
2044 if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_OTG) {
2045 ddata->module_enable_quirk = sysc_module_enable_quirk_otg;
2046 ddata->module_disable_quirk = sysc_module_disable_quirk_otg;
2047 }
2048
2049 if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_SGX)
2050 ddata->module_enable_quirk = sysc_module_enable_quirk_sgx;
2051
2052 if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_WDT) {
2053 ddata->reset_done_quirk = sysc_reset_done_quirk_wdt;
2054 ddata->module_disable_quirk = sysc_reset_done_quirk_wdt;
2055 }
2056
2057 if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_PRUSS) {
2058 ddata->module_enable_quirk = sysc_module_enable_quirk_pruss;
2059 ddata->module_disable_quirk = sysc_module_disable_quirk_pruss;
2060 }
2061 }
2062
sysc_clockdomain_init(struct sysc * ddata)2063 static int sysc_clockdomain_init(struct sysc *ddata)
2064 {
2065 struct ti_sysc_platform_data *pdata = dev_get_platdata(ddata->dev);
2066 struct clk *fck = NULL, *ick = NULL;
2067 int error;
2068
2069 if (!pdata || !pdata->init_clockdomain)
2070 return 0;
2071
2072 switch (ddata->nr_clocks) {
2073 case 2:
2074 ick = ddata->clocks[SYSC_ICK];
2075 fallthrough;
2076 case 1:
2077 fck = ddata->clocks[SYSC_FCK];
2078 break;
2079 case 0:
2080 return 0;
2081 }
2082
2083 error = pdata->init_clockdomain(ddata->dev, fck, ick, &ddata->cookie);
2084 if (!error || error == -ENODEV)
2085 return 0;
2086
2087 return error;
2088 }
2089
2090 /*
2091 * Note that pdata->init_module() typically does a reset first. After
2092 * pdata->init_module() is done, PM runtime can be used for the interconnect
2093 * target module.
2094 */
sysc_legacy_init(struct sysc * ddata)2095 static int sysc_legacy_init(struct sysc *ddata)
2096 {
2097 struct ti_sysc_platform_data *pdata = dev_get_platdata(ddata->dev);
2098 int error;
2099
2100 if (!pdata || !pdata->init_module)
2101 return 0;
2102
2103 error = pdata->init_module(ddata->dev, ddata->mdata, &ddata->cookie);
2104 if (error == -EEXIST)
2105 error = 0;
2106
2107 return error;
2108 }
2109
2110 /*
2111 * Note that the caller must ensure the interconnect target module is enabled
2112 * before calling reset. Otherwise reset will not complete.
2113 */
sysc_reset(struct sysc * ddata)2114 static int sysc_reset(struct sysc *ddata)
2115 {
2116 int sysc_offset, sysc_val, error;
2117 u32 sysc_mask;
2118
2119 sysc_offset = ddata->offsets[SYSC_SYSCONFIG];
2120
2121 if (ddata->legacy_mode ||
2122 ddata->cap->regbits->srst_shift < 0)
2123 return 0;
2124
2125 sysc_mask = BIT(ddata->cap->regbits->srst_shift);
2126
2127 if (ddata->pre_reset_quirk)
2128 ddata->pre_reset_quirk(ddata);
2129
2130 if (sysc_offset >= 0) {
2131 sysc_val = sysc_read_sysconfig(ddata);
2132 sysc_val |= sysc_mask;
2133 sysc_write(ddata, sysc_offset, sysc_val);
2134
2135 /*
2136 * Some devices need a delay before reading registers
2137 * after reset. Presumably a srst_udelay is not needed
2138 * for devices that use a rstctrl register reset.
2139 */
2140 if (ddata->cfg.srst_udelay)
2141 fsleep(ddata->cfg.srst_udelay);
2142
2143 /*
2144 * Flush posted write. For devices needing srst_udelay
2145 * this should trigger an interconnect error if the
2146 * srst_udelay value is needed but not configured.
2147 */
2148 sysc_val = sysc_read_sysconfig(ddata);
2149 }
2150
2151 if (ddata->post_reset_quirk)
2152 ddata->post_reset_quirk(ddata);
2153
2154 error = sysc_wait_softreset(ddata);
2155 if (error)
2156 dev_warn(ddata->dev, "OCP softreset timed out\n");
2157
2158 if (ddata->reset_done_quirk)
2159 ddata->reset_done_quirk(ddata);
2160
2161 return error;
2162 }
2163
2164 /*
2165 * At this point the module is configured enough to read the revision but
2166 * module may not be completely configured yet to use PM runtime. Enable
2167 * all clocks directly during init to configure the quirks needed for PM
2168 * runtime based on the revision register.
2169 */
sysc_init_module(struct sysc * ddata)2170 static int sysc_init_module(struct sysc *ddata)
2171 {
2172 bool rstctrl_deasserted = false;
2173 int error = sysc_clockdomain_init(ddata);
2174
2175 if (error)
2176 return error;
2177
2178 sysc_clkdm_deny_idle(ddata);
2179
2180 /*
2181 * Always enable clocks. The bootloader may or may not have enabled
2182 * the related clocks.
2183 */
2184 error = sysc_enable_opt_clocks(ddata);
2185 if (error)
2186 return error;
2187
2188 error = sysc_enable_main_clocks(ddata);
2189 if (error)
2190 goto err_opt_clocks;
2191
2192 if (!(ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT)) {
2193 error = reset_control_deassert(ddata->rsts);
2194 if (error)
2195 goto err_main_clocks;
2196 rstctrl_deasserted = true;
2197 }
2198
2199 ddata->revision = sysc_read_revision(ddata);
2200 sysc_init_revision_quirks(ddata);
2201 sysc_init_module_quirks(ddata);
2202
2203 if (ddata->legacy_mode) {
2204 error = sysc_legacy_init(ddata);
2205 if (error)
2206 goto err_main_clocks;
2207 }
2208
2209 if (!ddata->legacy_mode) {
2210 error = sysc_enable_module(ddata->dev);
2211 if (error)
2212 goto err_main_clocks;
2213 }
2214
2215 if (!(ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT)) {
2216 error = sysc_reset(ddata);
2217 if (error)
2218 dev_err(ddata->dev, "Reset failed with %d\n", error);
2219
2220 if (error && !ddata->legacy_mode)
2221 sysc_disable_module(ddata->dev);
2222 }
2223
2224 err_main_clocks:
2225 if (error)
2226 sysc_disable_main_clocks(ddata);
2227 err_opt_clocks:
2228 /* No re-enable of clockdomain autoidle to prevent module autoidle */
2229 if (error) {
2230 sysc_disable_opt_clocks(ddata);
2231 sysc_clkdm_allow_idle(ddata);
2232 }
2233
2234 if (error && rstctrl_deasserted &&
2235 !(ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT))
2236 reset_control_assert(ddata->rsts);
2237
2238 return error;
2239 }
2240
sysc_init_sysc_mask(struct sysc * ddata)2241 static int sysc_init_sysc_mask(struct sysc *ddata)
2242 {
2243 struct device_node *np = ddata->dev->of_node;
2244 int error;
2245 u32 val;
2246
2247 error = of_property_read_u32(np, "ti,sysc-mask", &val);
2248 if (error)
2249 return 0;
2250
2251 ddata->cfg.sysc_val = val & ddata->cap->sysc_mask;
2252
2253 return 0;
2254 }
2255
sysc_init_idlemode(struct sysc * ddata,u8 * idlemodes,const char * name)2256 static int sysc_init_idlemode(struct sysc *ddata, u8 *idlemodes,
2257 const char *name)
2258 {
2259 struct device_node *np = ddata->dev->of_node;
2260 u32 val;
2261
2262 of_property_for_each_u32(np, name, val) {
2263 if (val >= SYSC_NR_IDLEMODES) {
2264 dev_err(ddata->dev, "invalid idlemode: %i\n", val);
2265 return -EINVAL;
2266 }
2267 *idlemodes |= (1 << val);
2268 }
2269
2270 return 0;
2271 }
2272
sysc_init_idlemodes(struct sysc * ddata)2273 static int sysc_init_idlemodes(struct sysc *ddata)
2274 {
2275 int error;
2276
2277 error = sysc_init_idlemode(ddata, &ddata->cfg.midlemodes,
2278 "ti,sysc-midle");
2279 if (error)
2280 return error;
2281
2282 error = sysc_init_idlemode(ddata, &ddata->cfg.sidlemodes,
2283 "ti,sysc-sidle");
2284 if (error)
2285 return error;
2286
2287 return 0;
2288 }
2289
2290 /*
2291 * Only some devices on omap4 and later have SYSCONFIG reset done
2292 * bit. We can detect this if there is no SYSSTATUS at all, or the
2293 * SYSTATUS bit 0 is not used. Note that some SYSSTATUS registers
2294 * have multiple bits for the child devices like OHCI and EHCI.
2295 * Depends on SYSC being parsed first.
2296 */
sysc_init_syss_mask(struct sysc * ddata)2297 static int sysc_init_syss_mask(struct sysc *ddata)
2298 {
2299 struct device_node *np = ddata->dev->of_node;
2300 int error;
2301 u32 val;
2302
2303 error = of_property_read_u32(np, "ti,syss-mask", &val);
2304 if (error) {
2305 if ((ddata->cap->type == TI_SYSC_OMAP4 ||
2306 ddata->cap->type == TI_SYSC_OMAP4_TIMER) &&
2307 (ddata->cfg.sysc_val & SYSC_OMAP4_SOFTRESET))
2308 ddata->cfg.quirks |= SYSC_QUIRK_RESET_STATUS;
2309
2310 return 0;
2311 }
2312
2313 if (!(val & 1) && (ddata->cfg.sysc_val & SYSC_OMAP4_SOFTRESET))
2314 ddata->cfg.quirks |= SYSC_QUIRK_RESET_STATUS;
2315
2316 ddata->cfg.syss_mask = val;
2317
2318 return 0;
2319 }
2320
2321 /*
2322 * Many child device drivers need to have fck and opt clocks available
2323 * to get the clock rate for device internal configuration etc.
2324 */
sysc_child_add_named_clock(struct sysc * ddata,struct device * child,const char * name)2325 static int sysc_child_add_named_clock(struct sysc *ddata,
2326 struct device *child,
2327 const char *name)
2328 {
2329 struct clk *clk;
2330 struct clk_lookup *l;
2331 int error = 0;
2332
2333 if (!name)
2334 return 0;
2335
2336 clk = clk_get(child, name);
2337 if (!IS_ERR(clk)) {
2338 error = -EEXIST;
2339 goto put_clk;
2340 }
2341
2342 clk = clk_get(ddata->dev, name);
2343 if (IS_ERR(clk))
2344 return -ENODEV;
2345
2346 l = clkdev_create(clk, name, dev_name(child));
2347 if (!l)
2348 error = -ENOMEM;
2349 put_clk:
2350 clk_put(clk);
2351
2352 return error;
2353 }
2354
sysc_child_add_clocks(struct sysc * ddata,struct device * child)2355 static int sysc_child_add_clocks(struct sysc *ddata,
2356 struct device *child)
2357 {
2358 int i, error;
2359
2360 for (i = 0; i < ddata->nr_clocks; i++) {
2361 error = sysc_child_add_named_clock(ddata,
2362 child,
2363 ddata->clock_roles[i]);
2364 if (error && error != -EEXIST) {
2365 dev_err(ddata->dev, "could not add child clock %s: %i\n",
2366 ddata->clock_roles[i], error);
2367
2368 return error;
2369 }
2370 }
2371
2372 return 0;
2373 }
2374
2375 static const struct device_type sysc_device_type = {
2376 };
2377
sysc_child_to_parent(struct device * dev)2378 static struct sysc *sysc_child_to_parent(struct device *dev)
2379 {
2380 struct device *parent = dev->parent;
2381
2382 if (!parent || parent->type != &sysc_device_type)
2383 return NULL;
2384
2385 return dev_get_drvdata(parent);
2386 }
2387
sysc_child_runtime_suspend(struct device * dev)2388 static int __maybe_unused sysc_child_runtime_suspend(struct device *dev)
2389 {
2390 struct sysc *ddata;
2391 int error;
2392
2393 ddata = sysc_child_to_parent(dev);
2394
2395 error = pm_generic_runtime_suspend(dev);
2396 if (error)
2397 return error;
2398
2399 if (!ddata->enabled)
2400 return 0;
2401
2402 return sysc_runtime_suspend(ddata->dev);
2403 }
2404
sysc_child_runtime_resume(struct device * dev)2405 static int __maybe_unused sysc_child_runtime_resume(struct device *dev)
2406 {
2407 struct sysc *ddata;
2408 int error;
2409
2410 ddata = sysc_child_to_parent(dev);
2411
2412 if (!ddata->enabled) {
2413 error = sysc_runtime_resume(ddata->dev);
2414 if (error < 0)
2415 dev_err(ddata->dev,
2416 "%s error: %i\n", __func__, error);
2417 }
2418
2419 return pm_generic_runtime_resume(dev);
2420 }
2421
2422 /* Caller needs to take list_lock if ever used outside of cpu_pm */
sysc_reinit_modules(struct sysc_soc_info * soc)2423 static void sysc_reinit_modules(struct sysc_soc_info *soc)
2424 {
2425 struct sysc_module *module;
2426 struct sysc *ddata;
2427
2428 list_for_each_entry(module, &sysc_soc->restored_modules, node) {
2429 ddata = module->ddata;
2430 sysc_reinit_module(ddata, ddata->enabled);
2431 }
2432 }
2433
2434 /**
2435 * sysc_context_notifier - optionally reset and restore module after idle
2436 * @nb: notifier block
2437 * @cmd: unused
2438 * @v: unused
2439 *
2440 * Some interconnect target modules need to be restored, or reset and restored
2441 * on CPU_PM CPU_PM_CLUSTER_EXIT notifier. This is needed at least for am335x
2442 * OTG and GPMC target modules even if the modules are unused.
2443 */
sysc_context_notifier(struct notifier_block * nb,unsigned long cmd,void * v)2444 static int sysc_context_notifier(struct notifier_block *nb, unsigned long cmd,
2445 void *v)
2446 {
2447 struct sysc_soc_info *soc;
2448
2449 soc = container_of(nb, struct sysc_soc_info, nb);
2450
2451 switch (cmd) {
2452 case CPU_CLUSTER_PM_ENTER:
2453 break;
2454 case CPU_CLUSTER_PM_ENTER_FAILED: /* No need to restore context */
2455 break;
2456 case CPU_CLUSTER_PM_EXIT:
2457 sysc_reinit_modules(soc);
2458 break;
2459 }
2460
2461 return NOTIFY_OK;
2462 }
2463
2464 /**
2465 * sysc_add_restored - optionally add reset and restore quirk hanlling
2466 * @ddata: device data
2467 */
sysc_add_restored(struct sysc * ddata)2468 static void sysc_add_restored(struct sysc *ddata)
2469 {
2470 struct sysc_module *restored_module;
2471
2472 restored_module = kzalloc(sizeof(*restored_module), GFP_KERNEL);
2473 if (!restored_module)
2474 return;
2475
2476 restored_module->ddata = ddata;
2477
2478 mutex_lock(&sysc_soc->list_lock);
2479
2480 list_add(&restored_module->node, &sysc_soc->restored_modules);
2481
2482 if (sysc_soc->nb.notifier_call)
2483 goto out_unlock;
2484
2485 sysc_soc->nb.notifier_call = sysc_context_notifier;
2486 cpu_pm_register_notifier(&sysc_soc->nb);
2487
2488 out_unlock:
2489 mutex_unlock(&sysc_soc->list_lock);
2490 }
2491
sysc_notifier_call(struct notifier_block * nb,unsigned long event,void * device)2492 static int sysc_notifier_call(struct notifier_block *nb,
2493 unsigned long event, void *device)
2494 {
2495 struct device *dev = device;
2496 struct sysc *ddata;
2497 int error;
2498
2499 ddata = sysc_child_to_parent(dev);
2500 if (!ddata)
2501 return NOTIFY_DONE;
2502
2503 switch (event) {
2504 case BUS_NOTIFY_ADD_DEVICE:
2505 error = sysc_child_add_clocks(ddata, dev);
2506 if (error)
2507 return error;
2508 break;
2509 default:
2510 break;
2511 }
2512
2513 return NOTIFY_DONE;
2514 }
2515
2516 static struct notifier_block sysc_nb = {
2517 .notifier_call = sysc_notifier_call,
2518 };
2519
2520 /* Device tree configured quirks */
2521 struct sysc_dts_quirk {
2522 const char *name;
2523 u32 mask;
2524 };
2525
2526 static const struct sysc_dts_quirk sysc_dts_quirks[] = {
2527 { .name = "ti,no-idle-on-init",
2528 .mask = SYSC_QUIRK_NO_IDLE_ON_INIT, },
2529 { .name = "ti,no-reset-on-init",
2530 .mask = SYSC_QUIRK_NO_RESET_ON_INIT, },
2531 { .name = "ti,no-idle",
2532 .mask = SYSC_QUIRK_NO_IDLE, },
2533 };
2534
sysc_parse_dts_quirks(struct sysc * ddata,struct device_node * np,bool is_child)2535 static void sysc_parse_dts_quirks(struct sysc *ddata, struct device_node *np,
2536 bool is_child)
2537 {
2538 int i;
2539
2540 for (i = 0; i < ARRAY_SIZE(sysc_dts_quirks); i++) {
2541 const char *name = sysc_dts_quirks[i].name;
2542
2543 if (!of_property_present(np, name))
2544 continue;
2545
2546 ddata->cfg.quirks |= sysc_dts_quirks[i].mask;
2547 if (is_child) {
2548 dev_warn(ddata->dev,
2549 "dts flag should be at module level for %s\n",
2550 name);
2551 }
2552 }
2553 }
2554
sysc_init_dts_quirks(struct sysc * ddata)2555 static int sysc_init_dts_quirks(struct sysc *ddata)
2556 {
2557 struct device_node *np = ddata->dev->of_node;
2558 int error;
2559 u32 val;
2560
2561 ddata->legacy_mode = of_get_property(np, "ti,hwmods", NULL);
2562
2563 sysc_parse_dts_quirks(ddata, np, false);
2564 error = of_property_read_u32(np, "ti,sysc-delay-us", &val);
2565 if (!error) {
2566 if (val > 255) {
2567 dev_warn(ddata->dev, "bad ti,sysc-delay-us: %i\n",
2568 val);
2569 }
2570
2571 ddata->cfg.srst_udelay = (u8)val;
2572 }
2573
2574 return 0;
2575 }
2576
sysc_unprepare(struct sysc * ddata)2577 static void sysc_unprepare(struct sysc *ddata)
2578 {
2579 int i;
2580
2581 if (!ddata->clocks)
2582 return;
2583
2584 for (i = 0; i < SYSC_MAX_CLOCKS; i++) {
2585 if (!IS_ERR_OR_NULL(ddata->clocks[i]))
2586 clk_unprepare(ddata->clocks[i]);
2587 }
2588 }
2589
2590 /*
2591 * Common sysc register bits found on omap2, also known as type1
2592 */
2593 static const struct sysc_regbits sysc_regbits_omap2 = {
2594 .dmadisable_shift = -ENODEV,
2595 .midle_shift = 12,
2596 .sidle_shift = 3,
2597 .clkact_shift = 8,
2598 .emufree_shift = 5,
2599 .enwkup_shift = 2,
2600 .srst_shift = 1,
2601 .autoidle_shift = 0,
2602 };
2603
2604 static const struct sysc_capabilities sysc_omap2 = {
2605 .type = TI_SYSC_OMAP2,
2606 .sysc_mask = SYSC_OMAP2_CLOCKACTIVITY | SYSC_OMAP2_EMUFREE |
2607 SYSC_OMAP2_ENAWAKEUP | SYSC_OMAP2_SOFTRESET |
2608 SYSC_OMAP2_AUTOIDLE,
2609 .regbits = &sysc_regbits_omap2,
2610 };
2611
2612 /* All omap2 and 3 timers, and timers 1, 2 & 10 on omap 4 and 5 */
2613 static const struct sysc_capabilities sysc_omap2_timer = {
2614 .type = TI_SYSC_OMAP2_TIMER,
2615 .sysc_mask = SYSC_OMAP2_CLOCKACTIVITY | SYSC_OMAP2_EMUFREE |
2616 SYSC_OMAP2_ENAWAKEUP | SYSC_OMAP2_SOFTRESET |
2617 SYSC_OMAP2_AUTOIDLE,
2618 .regbits = &sysc_regbits_omap2,
2619 .mod_quirks = SYSC_QUIRK_USE_CLOCKACT,
2620 };
2621
2622 /*
2623 * SHAM2 (SHA1/MD5) sysc found on omap3, a variant of sysc_regbits_omap2
2624 * with different sidle position
2625 */
2626 static const struct sysc_regbits sysc_regbits_omap3_sham = {
2627 .dmadisable_shift = -ENODEV,
2628 .midle_shift = -ENODEV,
2629 .sidle_shift = 4,
2630 .clkact_shift = -ENODEV,
2631 .enwkup_shift = -ENODEV,
2632 .srst_shift = 1,
2633 .autoidle_shift = 0,
2634 .emufree_shift = -ENODEV,
2635 };
2636
2637 static const struct sysc_capabilities sysc_omap3_sham = {
2638 .type = TI_SYSC_OMAP3_SHAM,
2639 .sysc_mask = SYSC_OMAP2_SOFTRESET | SYSC_OMAP2_AUTOIDLE,
2640 .regbits = &sysc_regbits_omap3_sham,
2641 };
2642
2643 /*
2644 * AES register bits found on omap3 and later, a variant of
2645 * sysc_regbits_omap2 with different sidle position
2646 */
2647 static const struct sysc_regbits sysc_regbits_omap3_aes = {
2648 .dmadisable_shift = -ENODEV,
2649 .midle_shift = -ENODEV,
2650 .sidle_shift = 6,
2651 .clkact_shift = -ENODEV,
2652 .enwkup_shift = -ENODEV,
2653 .srst_shift = 1,
2654 .autoidle_shift = 0,
2655 .emufree_shift = -ENODEV,
2656 };
2657
2658 static const struct sysc_capabilities sysc_omap3_aes = {
2659 .type = TI_SYSC_OMAP3_AES,
2660 .sysc_mask = SYSC_OMAP2_SOFTRESET | SYSC_OMAP2_AUTOIDLE,
2661 .regbits = &sysc_regbits_omap3_aes,
2662 };
2663
2664 /*
2665 * Common sysc register bits found on omap4, also known as type2
2666 */
2667 static const struct sysc_regbits sysc_regbits_omap4 = {
2668 .dmadisable_shift = 16,
2669 .midle_shift = 4,
2670 .sidle_shift = 2,
2671 .clkact_shift = -ENODEV,
2672 .enwkup_shift = -ENODEV,
2673 .emufree_shift = 1,
2674 .srst_shift = 0,
2675 .autoidle_shift = -ENODEV,
2676 };
2677
2678 static const struct sysc_capabilities sysc_omap4 = {
2679 .type = TI_SYSC_OMAP4,
2680 .sysc_mask = SYSC_OMAP4_DMADISABLE | SYSC_OMAP4_FREEEMU |
2681 SYSC_OMAP4_SOFTRESET,
2682 .regbits = &sysc_regbits_omap4,
2683 };
2684
2685 static const struct sysc_capabilities sysc_omap4_timer = {
2686 .type = TI_SYSC_OMAP4_TIMER,
2687 .sysc_mask = SYSC_OMAP4_DMADISABLE | SYSC_OMAP4_FREEEMU |
2688 SYSC_OMAP4_SOFTRESET,
2689 .regbits = &sysc_regbits_omap4,
2690 };
2691
2692 /*
2693 * Common sysc register bits found on omap4, also known as type3
2694 */
2695 static const struct sysc_regbits sysc_regbits_omap4_simple = {
2696 .dmadisable_shift = -ENODEV,
2697 .midle_shift = 2,
2698 .sidle_shift = 0,
2699 .clkact_shift = -ENODEV,
2700 .enwkup_shift = -ENODEV,
2701 .srst_shift = -ENODEV,
2702 .emufree_shift = -ENODEV,
2703 .autoidle_shift = -ENODEV,
2704 };
2705
2706 static const struct sysc_capabilities sysc_omap4_simple = {
2707 .type = TI_SYSC_OMAP4_SIMPLE,
2708 .regbits = &sysc_regbits_omap4_simple,
2709 };
2710
2711 /*
2712 * SmartReflex sysc found on omap34xx
2713 */
2714 static const struct sysc_regbits sysc_regbits_omap34xx_sr = {
2715 .dmadisable_shift = -ENODEV,
2716 .midle_shift = -ENODEV,
2717 .sidle_shift = -ENODEV,
2718 .clkact_shift = 20,
2719 .enwkup_shift = -ENODEV,
2720 .srst_shift = -ENODEV,
2721 .emufree_shift = -ENODEV,
2722 .autoidle_shift = -ENODEV,
2723 };
2724
2725 static const struct sysc_capabilities sysc_34xx_sr = {
2726 .type = TI_SYSC_OMAP34XX_SR,
2727 .sysc_mask = SYSC_OMAP2_CLOCKACTIVITY,
2728 .regbits = &sysc_regbits_omap34xx_sr,
2729 .mod_quirks = SYSC_QUIRK_USE_CLOCKACT | SYSC_QUIRK_UNCACHED,
2730 };
2731
2732 /*
2733 * SmartReflex sysc found on omap36xx and later
2734 */
2735 static const struct sysc_regbits sysc_regbits_omap36xx_sr = {
2736 .dmadisable_shift = -ENODEV,
2737 .midle_shift = -ENODEV,
2738 .sidle_shift = 24,
2739 .clkact_shift = -ENODEV,
2740 .enwkup_shift = 26,
2741 .srst_shift = -ENODEV,
2742 .emufree_shift = -ENODEV,
2743 .autoidle_shift = -ENODEV,
2744 };
2745
2746 static const struct sysc_capabilities sysc_36xx_sr = {
2747 .type = TI_SYSC_OMAP36XX_SR,
2748 .sysc_mask = SYSC_OMAP3_SR_ENAWAKEUP,
2749 .regbits = &sysc_regbits_omap36xx_sr,
2750 .mod_quirks = SYSC_QUIRK_UNCACHED,
2751 };
2752
2753 static const struct sysc_capabilities sysc_omap4_sr = {
2754 .type = TI_SYSC_OMAP4_SR,
2755 .regbits = &sysc_regbits_omap36xx_sr,
2756 };
2757
2758 /*
2759 * McASP register bits found on omap4 and later
2760 */
2761 static const struct sysc_regbits sysc_regbits_omap4_mcasp = {
2762 .dmadisable_shift = -ENODEV,
2763 .midle_shift = -ENODEV,
2764 .sidle_shift = 0,
2765 .clkact_shift = -ENODEV,
2766 .enwkup_shift = -ENODEV,
2767 .srst_shift = -ENODEV,
2768 .emufree_shift = -ENODEV,
2769 .autoidle_shift = -ENODEV,
2770 };
2771
2772 static const struct sysc_capabilities sysc_omap4_mcasp = {
2773 .type = TI_SYSC_OMAP4_MCASP,
2774 .regbits = &sysc_regbits_omap4_mcasp,
2775 .mod_quirks = SYSC_QUIRK_OPT_CLKS_NEEDED,
2776 };
2777
2778 /*
2779 * McASP found on dra7 and later
2780 */
2781 static const struct sysc_capabilities sysc_dra7_mcasp = {
2782 .type = TI_SYSC_OMAP4_SIMPLE,
2783 .regbits = &sysc_regbits_omap4_simple,
2784 .mod_quirks = SYSC_QUIRK_OPT_CLKS_NEEDED,
2785 };
2786
2787 /*
2788 * FS USB host found on omap4 and later
2789 */
2790 static const struct sysc_regbits sysc_regbits_omap4_usb_host_fs = {
2791 .dmadisable_shift = -ENODEV,
2792 .midle_shift = -ENODEV,
2793 .sidle_shift = 24,
2794 .clkact_shift = -ENODEV,
2795 .enwkup_shift = 26,
2796 .srst_shift = -ENODEV,
2797 .emufree_shift = -ENODEV,
2798 .autoidle_shift = -ENODEV,
2799 };
2800
2801 static const struct sysc_capabilities sysc_omap4_usb_host_fs = {
2802 .type = TI_SYSC_OMAP4_USB_HOST_FS,
2803 .sysc_mask = SYSC_OMAP2_ENAWAKEUP,
2804 .regbits = &sysc_regbits_omap4_usb_host_fs,
2805 };
2806
2807 static const struct sysc_regbits sysc_regbits_dra7_mcan = {
2808 .dmadisable_shift = -ENODEV,
2809 .midle_shift = -ENODEV,
2810 .sidle_shift = -ENODEV,
2811 .clkact_shift = -ENODEV,
2812 .enwkup_shift = 4,
2813 .srst_shift = 0,
2814 .emufree_shift = -ENODEV,
2815 .autoidle_shift = -ENODEV,
2816 };
2817
2818 static const struct sysc_capabilities sysc_dra7_mcan = {
2819 .type = TI_SYSC_DRA7_MCAN,
2820 .sysc_mask = SYSC_DRA7_MCAN_ENAWAKEUP | SYSC_OMAP4_SOFTRESET,
2821 .regbits = &sysc_regbits_dra7_mcan,
2822 .mod_quirks = SYSS_QUIRK_RESETDONE_INVERTED,
2823 };
2824
2825 /*
2826 * PRUSS found on some AM33xx, AM437x and AM57xx SoCs
2827 */
2828 static const struct sysc_capabilities sysc_pruss = {
2829 .type = TI_SYSC_PRUSS,
2830 .sysc_mask = SYSC_PRUSS_STANDBY_INIT | SYSC_PRUSS_SUB_MWAIT,
2831 .regbits = &sysc_regbits_omap4_simple,
2832 .mod_quirks = SYSC_MODULE_QUIRK_PRUSS,
2833 };
2834
sysc_init_pdata(struct sysc * ddata)2835 static int sysc_init_pdata(struct sysc *ddata)
2836 {
2837 struct ti_sysc_platform_data *pdata = dev_get_platdata(ddata->dev);
2838 struct ti_sysc_module_data *mdata;
2839
2840 if (!pdata)
2841 return 0;
2842
2843 mdata = devm_kzalloc(ddata->dev, sizeof(*mdata), GFP_KERNEL);
2844 if (!mdata)
2845 return -ENOMEM;
2846
2847 if (ddata->legacy_mode) {
2848 mdata->name = ddata->legacy_mode;
2849 mdata->module_pa = ddata->module_pa;
2850 mdata->module_size = ddata->module_size;
2851 mdata->offsets = ddata->offsets;
2852 mdata->nr_offsets = SYSC_MAX_REGS;
2853 mdata->cap = ddata->cap;
2854 mdata->cfg = &ddata->cfg;
2855 }
2856
2857 ddata->mdata = mdata;
2858
2859 return 0;
2860 }
2861
sysc_init_match(struct sysc * ddata)2862 static int sysc_init_match(struct sysc *ddata)
2863 {
2864 const struct sysc_capabilities *cap;
2865
2866 cap = of_device_get_match_data(ddata->dev);
2867 if (!cap)
2868 return -EINVAL;
2869
2870 ddata->cap = cap;
2871 if (ddata->cap)
2872 ddata->cfg.quirks |= ddata->cap->mod_quirks;
2873
2874 return 0;
2875 }
2876
ti_sysc_idle(struct work_struct * work)2877 static void ti_sysc_idle(struct work_struct *work)
2878 {
2879 struct sysc *ddata;
2880
2881 ddata = container_of(work, struct sysc, idle_work.work);
2882
2883 /*
2884 * One time decrement of clock usage counts if left on from init.
2885 * Note that we disable opt clocks unconditionally in this case
2886 * as they are enabled unconditionally during init without
2887 * considering sysc_opt_clks_needed() at that point.
2888 */
2889 if (ddata->cfg.quirks & (SYSC_QUIRK_NO_IDLE |
2890 SYSC_QUIRK_NO_IDLE_ON_INIT)) {
2891 sysc_disable_main_clocks(ddata);
2892 sysc_disable_opt_clocks(ddata);
2893 sysc_clkdm_allow_idle(ddata);
2894 }
2895
2896 /* Keep permanent PM runtime usage count for SYSC_QUIRK_NO_IDLE */
2897 if (ddata->cfg.quirks & SYSC_QUIRK_NO_IDLE)
2898 return;
2899
2900 /*
2901 * Decrement PM runtime usage count for SYSC_QUIRK_NO_IDLE_ON_INIT
2902 * and SYSC_QUIRK_NO_RESET_ON_INIT
2903 */
2904 if (pm_runtime_active(ddata->dev))
2905 pm_runtime_put_sync(ddata->dev);
2906 }
2907
2908 /*
2909 * SoC model and features detection. Only needed for SoCs that need
2910 * special handling for quirks, no need to list others.
2911 */
2912 static const struct soc_device_attribute sysc_soc_match[] = {
2913 SOC_FLAG("OMAP242*", SOC_2420),
2914 SOC_FLAG("OMAP243*", SOC_2430),
2915 SOC_FLAG("AM35*", SOC_AM35),
2916 SOC_FLAG("OMAP3[45]*", SOC_3430),
2917 SOC_FLAG("OMAP3[67]*", SOC_3630),
2918 SOC_FLAG("OMAP443*", SOC_4430),
2919 SOC_FLAG("OMAP446*", SOC_4460),
2920 SOC_FLAG("OMAP447*", SOC_4470),
2921 SOC_FLAG("OMAP54*", SOC_5430),
2922 SOC_FLAG("AM433", SOC_AM3),
2923 SOC_FLAG("AM43*", SOC_AM4),
2924 SOC_FLAG("DRA7*", SOC_DRA7),
2925
2926 { /* sentinel */ }
2927 };
2928
2929 /*
2930 * List of SoCs variants with disabled features. By default we assume all
2931 * devices in the device tree are available so no need to list those SoCs.
2932 */
2933 static const struct soc_device_attribute sysc_soc_feat_match[] = {
2934 /* OMAP3430/3530 and AM3517 variants with some accelerators disabled */
2935 SOC_FLAG("AM3505", DIS_SGX),
2936 SOC_FLAG("OMAP3525", DIS_SGX),
2937 SOC_FLAG("OMAP3515", DIS_IVA | DIS_SGX),
2938 SOC_FLAG("OMAP3503", DIS_ISP | DIS_IVA | DIS_SGX),
2939
2940 /* OMAP3630/DM3730 variants with some accelerators disabled */
2941 SOC_FLAG("AM3703", DIS_IVA | DIS_SGX),
2942 SOC_FLAG("DM3725", DIS_SGX),
2943 SOC_FLAG("OMAP3611", DIS_ISP | DIS_IVA | DIS_SGX),
2944 SOC_FLAG("OMAP3615/AM3715", DIS_IVA),
2945 SOC_FLAG("OMAP3621", DIS_ISP),
2946
2947 { /* sentinel */ }
2948 };
2949
sysc_add_disabled(unsigned long base)2950 static int sysc_add_disabled(unsigned long base)
2951 {
2952 struct sysc_address *disabled_module;
2953
2954 disabled_module = kzalloc(sizeof(*disabled_module), GFP_KERNEL);
2955 if (!disabled_module)
2956 return -ENOMEM;
2957
2958 disabled_module->base = base;
2959
2960 mutex_lock(&sysc_soc->list_lock);
2961 list_add(&disabled_module->node, &sysc_soc->disabled_modules);
2962 mutex_unlock(&sysc_soc->list_lock);
2963
2964 return 0;
2965 }
2966
2967 /*
2968 * One time init to detect the booted SoC, disable unavailable features
2969 * and initialize list for optional cpu_pm notifier.
2970 *
2971 * Note that we initialize static data shared across all ti-sysc instances
2972 * so ddata is only used for SoC type. This can be called from module_init
2973 * once we no longer need to rely on platform data.
2974 */
sysc_init_static_data(struct sysc * ddata)2975 static int sysc_init_static_data(struct sysc *ddata)
2976 {
2977 const struct soc_device_attribute *match;
2978 struct ti_sysc_platform_data *pdata;
2979 unsigned long features = 0;
2980 struct device_node *np;
2981
2982 if (sysc_soc)
2983 return 0;
2984
2985 sysc_soc = kzalloc(sizeof(*sysc_soc), GFP_KERNEL);
2986 if (!sysc_soc)
2987 return -ENOMEM;
2988
2989 mutex_init(&sysc_soc->list_lock);
2990 INIT_LIST_HEAD(&sysc_soc->disabled_modules);
2991 INIT_LIST_HEAD(&sysc_soc->restored_modules);
2992 sysc_soc->general_purpose = true;
2993
2994 pdata = dev_get_platdata(ddata->dev);
2995 if (pdata && pdata->soc_type_gp)
2996 sysc_soc->general_purpose = pdata->soc_type_gp();
2997
2998 match = soc_device_match(sysc_soc_match);
2999 if (match && match->data)
3000 sysc_soc->soc = (enum sysc_soc)(uintptr_t)match->data;
3001
3002 /*
3003 * Check and warn about possible old incomplete dtb. We now want to see
3004 * simple-pm-bus instead of simple-bus in the dtb for genpd using SoCs.
3005 */
3006 switch (sysc_soc->soc) {
3007 case SOC_AM3:
3008 case SOC_AM4:
3009 case SOC_4430 ... SOC_4470:
3010 case SOC_5430:
3011 case SOC_DRA7:
3012 np = of_find_node_by_path("/ocp");
3013 WARN_ONCE(np && of_device_is_compatible(np, "simple-bus"),
3014 "ti-sysc: Incomplete old dtb, please update\n");
3015 break;
3016 default:
3017 break;
3018 }
3019
3020 /* Ignore devices that are not available on HS and EMU SoCs */
3021 if (!sysc_soc->general_purpose) {
3022 switch (sysc_soc->soc) {
3023 case SOC_3430 ... SOC_3630:
3024 sysc_add_disabled(0x48304000); /* timer12 */
3025 break;
3026 case SOC_AM3:
3027 sysc_add_disabled(0x48310000); /* rng */
3028 break;
3029 default:
3030 break;
3031 }
3032 }
3033
3034 match = soc_device_match(sysc_soc_feat_match);
3035 if (!match)
3036 return 0;
3037
3038 if (match->data)
3039 features = (unsigned long)match->data;
3040
3041 /*
3042 * Add disabled devices to the list based on the module base.
3043 * Note that this must be done before we attempt to access the
3044 * device and have module revision checks working.
3045 */
3046 if (features & DIS_ISP)
3047 sysc_add_disabled(0x480bd400);
3048 if (features & DIS_IVA)
3049 sysc_add_disabled(0x5d000000);
3050 if (features & DIS_SGX)
3051 sysc_add_disabled(0x50000000);
3052
3053 return 0;
3054 }
3055
sysc_cleanup_static_data(void)3056 static void sysc_cleanup_static_data(void)
3057 {
3058 struct sysc_module *restored_module;
3059 struct sysc_address *disabled_module;
3060 struct list_head *pos, *tmp;
3061
3062 if (!sysc_soc)
3063 return;
3064
3065 if (sysc_soc->nb.notifier_call)
3066 cpu_pm_unregister_notifier(&sysc_soc->nb);
3067
3068 mutex_lock(&sysc_soc->list_lock);
3069 list_for_each_safe(pos, tmp, &sysc_soc->restored_modules) {
3070 restored_module = list_entry(pos, struct sysc_module, node);
3071 list_del(pos);
3072 kfree(restored_module);
3073 }
3074 list_for_each_safe(pos, tmp, &sysc_soc->disabled_modules) {
3075 disabled_module = list_entry(pos, struct sysc_address, node);
3076 list_del(pos);
3077 kfree(disabled_module);
3078 }
3079 mutex_unlock(&sysc_soc->list_lock);
3080 }
3081
sysc_check_disabled_devices(struct sysc * ddata)3082 static int sysc_check_disabled_devices(struct sysc *ddata)
3083 {
3084 struct sysc_address *disabled_module;
3085 int error = 0;
3086
3087 mutex_lock(&sysc_soc->list_lock);
3088 list_for_each_entry(disabled_module, &sysc_soc->disabled_modules, node) {
3089 if (ddata->module_pa == disabled_module->base) {
3090 dev_dbg(ddata->dev, "module disabled for this SoC\n");
3091 error = -ENODEV;
3092 break;
3093 }
3094 }
3095 mutex_unlock(&sysc_soc->list_lock);
3096
3097 return error;
3098 }
3099
3100 /*
3101 * Ignore timers tagged with no-reset and no-idle. These are likely in use,
3102 * for example by drivers/clocksource/timer-ti-dm-systimer.c. If more checks
3103 * are needed, we could also look at the timer register configuration.
3104 */
sysc_check_active_timer(struct sysc * ddata)3105 static int sysc_check_active_timer(struct sysc *ddata)
3106 {
3107 int error;
3108
3109 if (ddata->cap->type != TI_SYSC_OMAP2_TIMER &&
3110 ddata->cap->type != TI_SYSC_OMAP4_TIMER)
3111 return 0;
3112
3113 /*
3114 * Quirk for omap3 beagleboard revision A to B4 to use gpt12.
3115 * Revision C and later are fixed with commit 23885389dbbb ("ARM:
3116 * dts: Fix timer regression for beagleboard revision c"). This all
3117 * can be dropped if we stop supporting old beagleboard revisions
3118 * A to B4 at some point.
3119 */
3120 if (sysc_soc->soc == SOC_3430 || sysc_soc->soc == SOC_AM35)
3121 error = -ENXIO;
3122 else
3123 error = -EBUSY;
3124
3125 if ((ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT) &&
3126 (ddata->cfg.quirks & SYSC_QUIRK_NO_IDLE))
3127 return error;
3128
3129 return 0;
3130 }
3131
3132 static const struct of_device_id sysc_match_table[] = {
3133 { .compatible = "simple-bus", },
3134 { /* sentinel */ },
3135 };
3136
sysc_probe(struct platform_device * pdev)3137 static int sysc_probe(struct platform_device *pdev)
3138 {
3139 struct ti_sysc_platform_data *pdata = dev_get_platdata(&pdev->dev);
3140 struct sysc *ddata;
3141 int error;
3142
3143 ddata = devm_kzalloc(&pdev->dev, sizeof(*ddata), GFP_KERNEL);
3144 if (!ddata)
3145 return -ENOMEM;
3146
3147 ddata->offsets[SYSC_REVISION] = -ENODEV;
3148 ddata->offsets[SYSC_SYSCONFIG] = -ENODEV;
3149 ddata->offsets[SYSC_SYSSTATUS] = -ENODEV;
3150 ddata->dev = &pdev->dev;
3151 platform_set_drvdata(pdev, ddata);
3152
3153 error = sysc_init_static_data(ddata);
3154 if (error)
3155 return error;
3156
3157 error = sysc_init_match(ddata);
3158 if (error)
3159 return error;
3160
3161 error = sysc_init_dts_quirks(ddata);
3162 if (error)
3163 return error;
3164
3165 error = sysc_map_and_check_registers(ddata);
3166 if (error)
3167 return error;
3168
3169 error = sysc_init_sysc_mask(ddata);
3170 if (error)
3171 return error;
3172
3173 error = sysc_init_idlemodes(ddata);
3174 if (error)
3175 return error;
3176
3177 error = sysc_init_syss_mask(ddata);
3178 if (error)
3179 return error;
3180
3181 error = sysc_init_pdata(ddata);
3182 if (error)
3183 return error;
3184
3185 sysc_init_early_quirks(ddata);
3186
3187 error = sysc_check_disabled_devices(ddata);
3188 if (error)
3189 return error;
3190
3191 error = sysc_check_active_timer(ddata);
3192 if (error == -ENXIO)
3193 ddata->reserved = true;
3194 else if (error)
3195 return error;
3196
3197 error = sysc_get_clocks(ddata);
3198 if (error)
3199 return error;
3200
3201 error = sysc_init_resets(ddata);
3202 if (error)
3203 goto unprepare;
3204
3205 error = sysc_init_module(ddata);
3206 if (error)
3207 goto unprepare;
3208
3209 pm_runtime_enable(ddata->dev);
3210 error = pm_runtime_resume_and_get(ddata->dev);
3211 if (error < 0) {
3212 pm_runtime_disable(ddata->dev);
3213 goto unprepare;
3214 }
3215
3216 /* Balance use counts as PM runtime should have enabled these all */
3217 if (!(ddata->cfg.quirks &
3218 (SYSC_QUIRK_NO_IDLE | SYSC_QUIRK_NO_IDLE_ON_INIT))) {
3219 sysc_disable_main_clocks(ddata);
3220 sysc_disable_opt_clocks(ddata);
3221 sysc_clkdm_allow_idle(ddata);
3222 }
3223
3224 if (!(ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT))
3225 reset_control_assert(ddata->rsts);
3226
3227 sysc_show_registers(ddata);
3228
3229 ddata->dev->type = &sysc_device_type;
3230
3231 if (!ddata->reserved) {
3232 error = of_platform_populate(ddata->dev->of_node,
3233 sysc_match_table,
3234 pdata ? pdata->auxdata : NULL,
3235 ddata->dev);
3236 if (error)
3237 goto err;
3238 }
3239
3240 INIT_DELAYED_WORK(&ddata->idle_work, ti_sysc_idle);
3241
3242 /* At least earlycon won't survive without deferred idle */
3243 if (ddata->cfg.quirks & (SYSC_QUIRK_NO_IDLE |
3244 SYSC_QUIRK_NO_IDLE_ON_INIT |
3245 SYSC_QUIRK_NO_RESET_ON_INIT)) {
3246 schedule_delayed_work(&ddata->idle_work, 3000);
3247 } else {
3248 pm_runtime_put(&pdev->dev);
3249 }
3250
3251 if (ddata->cfg.quirks & SYSC_QUIRK_REINIT_ON_CTX_LOST)
3252 sysc_add_restored(ddata);
3253
3254 return 0;
3255
3256 err:
3257 pm_runtime_put_sync(&pdev->dev);
3258 pm_runtime_disable(&pdev->dev);
3259 unprepare:
3260 sysc_unprepare(ddata);
3261
3262 return error;
3263 }
3264
sysc_remove(struct platform_device * pdev)3265 static void sysc_remove(struct platform_device *pdev)
3266 {
3267 struct sysc *ddata = platform_get_drvdata(pdev);
3268 int error;
3269
3270 /* Device can still be enabled, see deferred idle quirk in probe */
3271 if (cancel_delayed_work_sync(&ddata->idle_work))
3272 ti_sysc_idle(&ddata->idle_work.work);
3273
3274 error = pm_runtime_resume_and_get(ddata->dev);
3275 if (error < 0) {
3276 pm_runtime_disable(ddata->dev);
3277 goto unprepare;
3278 }
3279
3280 of_platform_depopulate(&pdev->dev);
3281
3282 pm_runtime_put_sync(&pdev->dev);
3283 pm_runtime_disable(&pdev->dev);
3284
3285 if (!reset_control_status(ddata->rsts))
3286 reset_control_assert(ddata->rsts);
3287
3288 unprepare:
3289 sysc_unprepare(ddata);
3290 }
3291
3292 static const struct of_device_id sysc_match[] = {
3293 { .compatible = "ti,sysc-omap2", .data = &sysc_omap2, },
3294 { .compatible = "ti,sysc-omap2-timer", .data = &sysc_omap2_timer, },
3295 { .compatible = "ti,sysc-omap4", .data = &sysc_omap4, },
3296 { .compatible = "ti,sysc-omap4-timer", .data = &sysc_omap4_timer, },
3297 { .compatible = "ti,sysc-omap4-simple", .data = &sysc_omap4_simple, },
3298 { .compatible = "ti,sysc-omap3430-sr", .data = &sysc_34xx_sr, },
3299 { .compatible = "ti,sysc-omap3630-sr", .data = &sysc_36xx_sr, },
3300 { .compatible = "ti,sysc-omap4-sr", .data = &sysc_omap4_sr, },
3301 { .compatible = "ti,sysc-omap3-sham", .data = &sysc_omap3_sham, },
3302 { .compatible = "ti,sysc-omap-aes", .data = &sysc_omap3_aes, },
3303 { .compatible = "ti,sysc-mcasp", .data = &sysc_omap4_mcasp, },
3304 { .compatible = "ti,sysc-dra7-mcasp", .data = &sysc_dra7_mcasp, },
3305 { .compatible = "ti,sysc-usb-host-fs",
3306 .data = &sysc_omap4_usb_host_fs, },
3307 { .compatible = "ti,sysc-dra7-mcan", .data = &sysc_dra7_mcan, },
3308 { .compatible = "ti,sysc-pruss", .data = &sysc_pruss, },
3309 { },
3310 };
3311 MODULE_DEVICE_TABLE(of, sysc_match);
3312
3313 static struct platform_driver sysc_driver = {
3314 .probe = sysc_probe,
3315 .remove = sysc_remove,
3316 .driver = {
3317 .name = "ti-sysc",
3318 .of_match_table = sysc_match,
3319 .pm = &sysc_pm_ops,
3320 },
3321 };
3322
sysc_init(void)3323 static int __init sysc_init(void)
3324 {
3325 bus_register_notifier(&platform_bus_type, &sysc_nb);
3326
3327 return platform_driver_register(&sysc_driver);
3328 }
3329 module_init(sysc_init);
3330
sysc_exit(void)3331 static void __exit sysc_exit(void)
3332 {
3333 bus_unregister_notifier(&platform_bus_type, &sysc_nb);
3334 platform_driver_unregister(&sysc_driver);
3335 sysc_cleanup_static_data();
3336 }
3337 module_exit(sysc_exit);
3338
3339 MODULE_DESCRIPTION("TI sysc interconnect target driver");
3340 MODULE_LICENSE("GPL v2");
3341