• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * ti-sysc.c - Texas Instruments sysc interconnect target driver
4  */
5 
6 #include <linux/io.h>
7 #include <linux/clk.h>
8 #include <linux/clkdev.h>
9 #include <linux/cpu_pm.h>
10 #include <linux/delay.h>
11 #include <linux/list.h>
12 #include <linux/module.h>
13 #include <linux/platform_device.h>
14 #include <linux/pm_domain.h>
15 #include <linux/pm_runtime.h>
16 #include <linux/reset.h>
17 #include <linux/of_address.h>
18 #include <linux/of_platform.h>
19 #include <linux/slab.h>
20 #include <linux/sys_soc.h>
21 #include <linux/timekeeping.h>
22 #include <linux/iopoll.h>
23 
24 #include <linux/platform_data/ti-sysc.h>
25 
26 #include <dt-bindings/bus/ti-sysc.h>
27 
28 #define DIS_ISP		BIT(2)
29 #define DIS_IVA		BIT(1)
30 #define DIS_SGX		BIT(0)
31 
32 #define SOC_FLAG(match, flag)	{ .machine = match, .data = (void *)(flag), }
33 
34 #define MAX_MODULE_SOFTRESET_WAIT		10000
35 
36 enum sysc_soc {
37 	SOC_UNKNOWN,
38 	SOC_2420,
39 	SOC_2430,
40 	SOC_3430,
41 	SOC_3630,
42 	SOC_4430,
43 	SOC_4460,
44 	SOC_4470,
45 	SOC_5430,
46 	SOC_AM3,
47 	SOC_AM4,
48 	SOC_DRA7,
49 };
50 
51 struct sysc_address {
52 	unsigned long base;
53 	struct list_head node;
54 };
55 
56 struct sysc_module {
57 	struct sysc *ddata;
58 	struct list_head node;
59 };
60 
61 struct sysc_soc_info {
62 	unsigned long general_purpose:1;
63 	enum sysc_soc soc;
64 	struct mutex list_lock;	/* disabled and restored modules list lock */
65 	struct list_head disabled_modules;
66 	struct list_head restored_modules;
67 	struct notifier_block nb;
68 };
69 
70 enum sysc_clocks {
71 	SYSC_FCK,
72 	SYSC_ICK,
73 	SYSC_OPTFCK0,
74 	SYSC_OPTFCK1,
75 	SYSC_OPTFCK2,
76 	SYSC_OPTFCK3,
77 	SYSC_OPTFCK4,
78 	SYSC_OPTFCK5,
79 	SYSC_OPTFCK6,
80 	SYSC_OPTFCK7,
81 	SYSC_MAX_CLOCKS,
82 };
83 
84 static struct sysc_soc_info *sysc_soc;
85 static const char * const reg_names[] = { "rev", "sysc", "syss", };
86 static const char * const clock_names[SYSC_MAX_CLOCKS] = {
87 	"fck", "ick", "opt0", "opt1", "opt2", "opt3", "opt4",
88 	"opt5", "opt6", "opt7",
89 };
90 
91 #define SYSC_IDLEMODE_MASK		3
92 #define SYSC_CLOCKACTIVITY_MASK		3
93 
94 /**
95  * struct sysc - TI sysc interconnect target module registers and capabilities
96  * @dev: struct device pointer
97  * @module_pa: physical address of the interconnect target module
98  * @module_size: size of the interconnect target module
99  * @module_va: virtual address of the interconnect target module
100  * @offsets: register offsets from module base
101  * @mdata: ti-sysc to hwmod translation data for a module
102  * @clocks: clocks used by the interconnect target module
103  * @clock_roles: clock role names for the found clocks
104  * @nr_clocks: number of clocks used by the interconnect target module
105  * @rsts: resets used by the interconnect target module
106  * @legacy_mode: configured for legacy mode if set
107  * @cap: interconnect target module capabilities
108  * @cfg: interconnect target module configuration
109  * @cookie: data used by legacy platform callbacks
110  * @name: name if available
111  * @revision: interconnect target module revision
112  * @reserved: target module is reserved and already in use
113  * @enabled: sysc runtime enabled status
114  * @needs_resume: runtime resume needed on resume from suspend
115  * @child_needs_resume: runtime resume needed for child on resume from suspend
116  * @disable_on_idle: status flag used for disabling modules with resets
117  * @idle_work: work structure used to perform delayed idle on a module
118  * @pre_reset_quirk: module specific pre-reset quirk
119  * @post_reset_quirk: module specific post-reset quirk
120  * @reset_done_quirk: module specific reset done quirk
121  * @module_enable_quirk: module specific enable quirk
122  * @module_disable_quirk: module specific disable quirk
123  * @module_unlock_quirk: module specific sysconfig unlock quirk
124  * @module_lock_quirk: module specific sysconfig lock quirk
125  */
126 struct sysc {
127 	struct device *dev;
128 	u64 module_pa;
129 	u32 module_size;
130 	void __iomem *module_va;
131 	int offsets[SYSC_MAX_REGS];
132 	struct ti_sysc_module_data *mdata;
133 	struct clk **clocks;
134 	const char **clock_roles;
135 	int nr_clocks;
136 	struct reset_control *rsts;
137 	const char *legacy_mode;
138 	const struct sysc_capabilities *cap;
139 	struct sysc_config cfg;
140 	struct ti_sysc_cookie cookie;
141 	const char *name;
142 	u32 revision;
143 	unsigned int reserved:1;
144 	unsigned int enabled:1;
145 	unsigned int needs_resume:1;
146 	unsigned int child_needs_resume:1;
147 	struct delayed_work idle_work;
148 	void (*pre_reset_quirk)(struct sysc *sysc);
149 	void (*post_reset_quirk)(struct sysc *sysc);
150 	void (*reset_done_quirk)(struct sysc *sysc);
151 	void (*module_enable_quirk)(struct sysc *sysc);
152 	void (*module_disable_quirk)(struct sysc *sysc);
153 	void (*module_unlock_quirk)(struct sysc *sysc);
154 	void (*module_lock_quirk)(struct sysc *sysc);
155 };
156 
157 static void sysc_parse_dts_quirks(struct sysc *ddata, struct device_node *np,
158 				  bool is_child);
159 
sysc_write(struct sysc * ddata,int offset,u32 value)160 static void sysc_write(struct sysc *ddata, int offset, u32 value)
161 {
162 	if (ddata->cfg.quirks & SYSC_QUIRK_16BIT) {
163 		writew_relaxed(value & 0xffff, ddata->module_va + offset);
164 
165 		/* Only i2c revision has LO and HI register with stride of 4 */
166 		if (ddata->offsets[SYSC_REVISION] >= 0 &&
167 		    offset == ddata->offsets[SYSC_REVISION]) {
168 			u16 hi = value >> 16;
169 
170 			writew_relaxed(hi, ddata->module_va + offset + 4);
171 		}
172 
173 		return;
174 	}
175 
176 	writel_relaxed(value, ddata->module_va + offset);
177 }
178 
sysc_read(struct sysc * ddata,int offset)179 static u32 sysc_read(struct sysc *ddata, int offset)
180 {
181 	if (ddata->cfg.quirks & SYSC_QUIRK_16BIT) {
182 		u32 val;
183 
184 		val = readw_relaxed(ddata->module_va + offset);
185 
186 		/* Only i2c revision has LO and HI register with stride of 4 */
187 		if (ddata->offsets[SYSC_REVISION] >= 0 &&
188 		    offset == ddata->offsets[SYSC_REVISION]) {
189 			u16 tmp = readw_relaxed(ddata->module_va + offset + 4);
190 
191 			val |= tmp << 16;
192 		}
193 
194 		return val;
195 	}
196 
197 	return readl_relaxed(ddata->module_va + offset);
198 }
199 
sysc_opt_clks_needed(struct sysc * ddata)200 static bool sysc_opt_clks_needed(struct sysc *ddata)
201 {
202 	return !!(ddata->cfg.quirks & SYSC_QUIRK_OPT_CLKS_NEEDED);
203 }
204 
sysc_read_revision(struct sysc * ddata)205 static u32 sysc_read_revision(struct sysc *ddata)
206 {
207 	int offset = ddata->offsets[SYSC_REVISION];
208 
209 	if (offset < 0)
210 		return 0;
211 
212 	return sysc_read(ddata, offset);
213 }
214 
sysc_read_sysconfig(struct sysc * ddata)215 static u32 sysc_read_sysconfig(struct sysc *ddata)
216 {
217 	int offset = ddata->offsets[SYSC_SYSCONFIG];
218 
219 	if (offset < 0)
220 		return 0;
221 
222 	return sysc_read(ddata, offset);
223 }
224 
sysc_read_sysstatus(struct sysc * ddata)225 static u32 sysc_read_sysstatus(struct sysc *ddata)
226 {
227 	int offset = ddata->offsets[SYSC_SYSSTATUS];
228 
229 	if (offset < 0)
230 		return 0;
231 
232 	return sysc_read(ddata, offset);
233 }
234 
sysc_poll_reset_sysstatus(struct sysc * ddata)235 static int sysc_poll_reset_sysstatus(struct sysc *ddata)
236 {
237 	int error, retries;
238 	u32 syss_done, rstval;
239 
240 	if (ddata->cfg.quirks & SYSS_QUIRK_RESETDONE_INVERTED)
241 		syss_done = 0;
242 	else
243 		syss_done = ddata->cfg.syss_mask;
244 
245 	if (likely(!timekeeping_suspended)) {
246 		error = readx_poll_timeout_atomic(sysc_read_sysstatus, ddata,
247 				rstval, (rstval & ddata->cfg.syss_mask) ==
248 				syss_done, 100, MAX_MODULE_SOFTRESET_WAIT);
249 	} else {
250 		retries = MAX_MODULE_SOFTRESET_WAIT;
251 		while (retries--) {
252 			rstval = sysc_read_sysstatus(ddata);
253 			if ((rstval & ddata->cfg.syss_mask) == syss_done)
254 				return 0;
255 			udelay(2); /* Account for udelay flakeyness */
256 		}
257 		error = -ETIMEDOUT;
258 	}
259 
260 	return error;
261 }
262 
sysc_poll_reset_sysconfig(struct sysc * ddata)263 static int sysc_poll_reset_sysconfig(struct sysc *ddata)
264 {
265 	int error, retries;
266 	u32 sysc_mask, rstval;
267 
268 	sysc_mask = BIT(ddata->cap->regbits->srst_shift);
269 
270 	if (likely(!timekeeping_suspended)) {
271 		error = readx_poll_timeout_atomic(sysc_read_sysconfig, ddata,
272 				rstval, !(rstval & sysc_mask),
273 				100, MAX_MODULE_SOFTRESET_WAIT);
274 	} else {
275 		retries = MAX_MODULE_SOFTRESET_WAIT;
276 		while (retries--) {
277 			rstval = sysc_read_sysconfig(ddata);
278 			if (!(rstval & sysc_mask))
279 				return 0;
280 			udelay(2); /* Account for udelay flakeyness */
281 		}
282 		error = -ETIMEDOUT;
283 	}
284 
285 	return error;
286 }
287 
288 /* Poll on reset status */
sysc_wait_softreset(struct sysc * ddata)289 static int sysc_wait_softreset(struct sysc *ddata)
290 {
291 	int syss_offset, error = 0;
292 
293 	if (ddata->cap->regbits->srst_shift < 0)
294 		return 0;
295 
296 	syss_offset = ddata->offsets[SYSC_SYSSTATUS];
297 
298 	if (syss_offset >= 0)
299 		error = sysc_poll_reset_sysstatus(ddata);
300 	else if (ddata->cfg.quirks & SYSC_QUIRK_RESET_STATUS)
301 		error = sysc_poll_reset_sysconfig(ddata);
302 
303 	return error;
304 }
305 
sysc_add_named_clock_from_child(struct sysc * ddata,const char * name,const char * optfck_name)306 static int sysc_add_named_clock_from_child(struct sysc *ddata,
307 					   const char *name,
308 					   const char *optfck_name)
309 {
310 	struct device_node *np = ddata->dev->of_node;
311 	struct device_node *child;
312 	struct clk_lookup *cl;
313 	struct clk *clock;
314 	const char *n;
315 
316 	if (name)
317 		n = name;
318 	else
319 		n = optfck_name;
320 
321 	/* Does the clock alias already exist? */
322 	clock = of_clk_get_by_name(np, n);
323 	if (!IS_ERR(clock)) {
324 		clk_put(clock);
325 
326 		return 0;
327 	}
328 
329 	child = of_get_next_available_child(np, NULL);
330 	if (!child)
331 		return -ENODEV;
332 
333 	clock = devm_get_clk_from_child(ddata->dev, child, name);
334 	if (IS_ERR(clock))
335 		return PTR_ERR(clock);
336 
337 	/*
338 	 * Use clkdev_add() instead of clkdev_alloc() to avoid the MAX_DEV_ID
339 	 * limit for clk_get(). If cl ever needs to be freed, it should be done
340 	 * with clkdev_drop().
341 	 */
342 	cl = kcalloc(1, sizeof(*cl), GFP_KERNEL);
343 	if (!cl)
344 		return -ENOMEM;
345 
346 	cl->con_id = n;
347 	cl->dev_id = dev_name(ddata->dev);
348 	cl->clk = clock;
349 	clkdev_add(cl);
350 
351 	clk_put(clock);
352 
353 	return 0;
354 }
355 
sysc_init_ext_opt_clock(struct sysc * ddata,const char * name)356 static int sysc_init_ext_opt_clock(struct sysc *ddata, const char *name)
357 {
358 	const char *optfck_name;
359 	int error, index;
360 
361 	if (ddata->nr_clocks < SYSC_OPTFCK0)
362 		index = SYSC_OPTFCK0;
363 	else
364 		index = ddata->nr_clocks;
365 
366 	if (name)
367 		optfck_name = name;
368 	else
369 		optfck_name = clock_names[index];
370 
371 	error = sysc_add_named_clock_from_child(ddata, name, optfck_name);
372 	if (error)
373 		return error;
374 
375 	ddata->clock_roles[index] = optfck_name;
376 	ddata->nr_clocks++;
377 
378 	return 0;
379 }
380 
sysc_get_one_clock(struct sysc * ddata,const char * name)381 static int sysc_get_one_clock(struct sysc *ddata, const char *name)
382 {
383 	int error, i, index = -ENODEV;
384 
385 	if (!strncmp(clock_names[SYSC_FCK], name, 3))
386 		index = SYSC_FCK;
387 	else if (!strncmp(clock_names[SYSC_ICK], name, 3))
388 		index = SYSC_ICK;
389 
390 	if (index < 0) {
391 		for (i = SYSC_OPTFCK0; i < SYSC_MAX_CLOCKS; i++) {
392 			if (!ddata->clocks[i]) {
393 				index = i;
394 				break;
395 			}
396 		}
397 	}
398 
399 	if (index < 0) {
400 		dev_err(ddata->dev, "clock %s not added\n", name);
401 		return index;
402 	}
403 
404 	ddata->clocks[index] = devm_clk_get(ddata->dev, name);
405 	if (IS_ERR(ddata->clocks[index])) {
406 		dev_err(ddata->dev, "clock get error for %s: %li\n",
407 			name, PTR_ERR(ddata->clocks[index]));
408 
409 		return PTR_ERR(ddata->clocks[index]);
410 	}
411 
412 	error = clk_prepare(ddata->clocks[index]);
413 	if (error) {
414 		dev_err(ddata->dev, "clock prepare error for %s: %i\n",
415 			name, error);
416 
417 		return error;
418 	}
419 
420 	return 0;
421 }
422 
sysc_get_clocks(struct sysc * ddata)423 static int sysc_get_clocks(struct sysc *ddata)
424 {
425 	struct device_node *np = ddata->dev->of_node;
426 	struct property *prop;
427 	const char *name;
428 	int nr_fck = 0, nr_ick = 0, i, error = 0;
429 
430 	ddata->clock_roles = devm_kcalloc(ddata->dev,
431 					  SYSC_MAX_CLOCKS,
432 					  sizeof(*ddata->clock_roles),
433 					  GFP_KERNEL);
434 	if (!ddata->clock_roles)
435 		return -ENOMEM;
436 
437 	of_property_for_each_string(np, "clock-names", prop, name) {
438 		if (!strncmp(clock_names[SYSC_FCK], name, 3))
439 			nr_fck++;
440 		if (!strncmp(clock_names[SYSC_ICK], name, 3))
441 			nr_ick++;
442 		ddata->clock_roles[ddata->nr_clocks] = name;
443 		ddata->nr_clocks++;
444 	}
445 
446 	if (ddata->nr_clocks < 1)
447 		return 0;
448 
449 	if ((ddata->cfg.quirks & SYSC_QUIRK_EXT_OPT_CLOCK)) {
450 		error = sysc_init_ext_opt_clock(ddata, NULL);
451 		if (error)
452 			return error;
453 	}
454 
455 	if (ddata->nr_clocks > SYSC_MAX_CLOCKS) {
456 		dev_err(ddata->dev, "too many clocks for %pOF\n", np);
457 
458 		return -EINVAL;
459 	}
460 
461 	if (nr_fck > 1 || nr_ick > 1) {
462 		dev_err(ddata->dev, "max one fck and ick for %pOF\n", np);
463 
464 		return -EINVAL;
465 	}
466 
467 	/* Always add a slot for main clocks fck and ick even if unused */
468 	if (!nr_fck)
469 		ddata->nr_clocks++;
470 	if (!nr_ick)
471 		ddata->nr_clocks++;
472 
473 	ddata->clocks = devm_kcalloc(ddata->dev,
474 				     ddata->nr_clocks, sizeof(*ddata->clocks),
475 				     GFP_KERNEL);
476 	if (!ddata->clocks)
477 		return -ENOMEM;
478 
479 	for (i = 0; i < SYSC_MAX_CLOCKS; i++) {
480 		const char *name = ddata->clock_roles[i];
481 
482 		if (!name)
483 			continue;
484 
485 		error = sysc_get_one_clock(ddata, name);
486 		if (error)
487 			return error;
488 	}
489 
490 	return 0;
491 }
492 
sysc_enable_main_clocks(struct sysc * ddata)493 static int sysc_enable_main_clocks(struct sysc *ddata)
494 {
495 	struct clk *clock;
496 	int i, error;
497 
498 	if (!ddata->clocks)
499 		return 0;
500 
501 	for (i = 0; i < SYSC_OPTFCK0; i++) {
502 		clock = ddata->clocks[i];
503 
504 		/* Main clocks may not have ick */
505 		if (IS_ERR_OR_NULL(clock))
506 			continue;
507 
508 		error = clk_enable(clock);
509 		if (error)
510 			goto err_disable;
511 	}
512 
513 	return 0;
514 
515 err_disable:
516 	for (i--; i >= 0; i--) {
517 		clock = ddata->clocks[i];
518 
519 		/* Main clocks may not have ick */
520 		if (IS_ERR_OR_NULL(clock))
521 			continue;
522 
523 		clk_disable(clock);
524 	}
525 
526 	return error;
527 }
528 
sysc_disable_main_clocks(struct sysc * ddata)529 static void sysc_disable_main_clocks(struct sysc *ddata)
530 {
531 	struct clk *clock;
532 	int i;
533 
534 	if (!ddata->clocks)
535 		return;
536 
537 	for (i = 0; i < SYSC_OPTFCK0; i++) {
538 		clock = ddata->clocks[i];
539 		if (IS_ERR_OR_NULL(clock))
540 			continue;
541 
542 		clk_disable(clock);
543 	}
544 }
545 
sysc_enable_opt_clocks(struct sysc * ddata)546 static int sysc_enable_opt_clocks(struct sysc *ddata)
547 {
548 	struct clk *clock;
549 	int i, error;
550 
551 	if (!ddata->clocks || ddata->nr_clocks < SYSC_OPTFCK0 + 1)
552 		return 0;
553 
554 	for (i = SYSC_OPTFCK0; i < SYSC_MAX_CLOCKS; i++) {
555 		clock = ddata->clocks[i];
556 
557 		/* Assume no holes for opt clocks */
558 		if (IS_ERR_OR_NULL(clock))
559 			return 0;
560 
561 		error = clk_enable(clock);
562 		if (error)
563 			goto err_disable;
564 	}
565 
566 	return 0;
567 
568 err_disable:
569 	for (i--; i >= 0; i--) {
570 		clock = ddata->clocks[i];
571 		if (IS_ERR_OR_NULL(clock))
572 			continue;
573 
574 		clk_disable(clock);
575 	}
576 
577 	return error;
578 }
579 
sysc_disable_opt_clocks(struct sysc * ddata)580 static void sysc_disable_opt_clocks(struct sysc *ddata)
581 {
582 	struct clk *clock;
583 	int i;
584 
585 	if (!ddata->clocks || ddata->nr_clocks < SYSC_OPTFCK0 + 1)
586 		return;
587 
588 	for (i = SYSC_OPTFCK0; i < SYSC_MAX_CLOCKS; i++) {
589 		clock = ddata->clocks[i];
590 
591 		/* Assume no holes for opt clocks */
592 		if (IS_ERR_OR_NULL(clock))
593 			return;
594 
595 		clk_disable(clock);
596 	}
597 }
598 
sysc_clkdm_deny_idle(struct sysc * ddata)599 static void sysc_clkdm_deny_idle(struct sysc *ddata)
600 {
601 	struct ti_sysc_platform_data *pdata;
602 
603 	if (ddata->legacy_mode || (ddata->cfg.quirks & SYSC_QUIRK_CLKDM_NOAUTO))
604 		return;
605 
606 	pdata = dev_get_platdata(ddata->dev);
607 	if (pdata && pdata->clkdm_deny_idle)
608 		pdata->clkdm_deny_idle(ddata->dev, &ddata->cookie);
609 }
610 
sysc_clkdm_allow_idle(struct sysc * ddata)611 static void sysc_clkdm_allow_idle(struct sysc *ddata)
612 {
613 	struct ti_sysc_platform_data *pdata;
614 
615 	if (ddata->legacy_mode || (ddata->cfg.quirks & SYSC_QUIRK_CLKDM_NOAUTO))
616 		return;
617 
618 	pdata = dev_get_platdata(ddata->dev);
619 	if (pdata && pdata->clkdm_allow_idle)
620 		pdata->clkdm_allow_idle(ddata->dev, &ddata->cookie);
621 }
622 
623 /**
624  * sysc_init_resets - init rstctrl reset line if configured
625  * @ddata: device driver data
626  *
627  * See sysc_rstctrl_reset_deassert().
628  */
sysc_init_resets(struct sysc * ddata)629 static int sysc_init_resets(struct sysc *ddata)
630 {
631 	ddata->rsts =
632 		devm_reset_control_get_optional_shared(ddata->dev, "rstctrl");
633 
634 	return PTR_ERR_OR_ZERO(ddata->rsts);
635 }
636 
637 /**
638  * sysc_parse_and_check_child_range - parses module IO region from ranges
639  * @ddata: device driver data
640  *
641  * In general we only need rev, syss, and sysc registers and not the whole
642  * module range. But we do want the offsets for these registers from the
643  * module base. This allows us to check them against the legacy hwmod
644  * platform data. Let's also check the ranges are configured properly.
645  */
sysc_parse_and_check_child_range(struct sysc * ddata)646 static int sysc_parse_and_check_child_range(struct sysc *ddata)
647 {
648 	struct device_node *np = ddata->dev->of_node;
649 	const __be32 *ranges;
650 	u32 nr_addr, nr_size;
651 	int len, error;
652 
653 	ranges = of_get_property(np, "ranges", &len);
654 	if (!ranges) {
655 		dev_err(ddata->dev, "missing ranges for %pOF\n", np);
656 
657 		return -ENOENT;
658 	}
659 
660 	len /= sizeof(*ranges);
661 
662 	if (len < 3) {
663 		dev_err(ddata->dev, "incomplete ranges for %pOF\n", np);
664 
665 		return -EINVAL;
666 	}
667 
668 	error = of_property_read_u32(np, "#address-cells", &nr_addr);
669 	if (error)
670 		return -ENOENT;
671 
672 	error = of_property_read_u32(np, "#size-cells", &nr_size);
673 	if (error)
674 		return -ENOENT;
675 
676 	if (nr_addr != 1 || nr_size != 1) {
677 		dev_err(ddata->dev, "invalid ranges for %pOF\n", np);
678 
679 		return -EINVAL;
680 	}
681 
682 	ranges++;
683 	ddata->module_pa = of_translate_address(np, ranges++);
684 	ddata->module_size = be32_to_cpup(ranges);
685 
686 	return 0;
687 }
688 
689 /* Interconnect instances to probe before l4_per instances */
690 static struct resource early_bus_ranges[] = {
691 	/* am3/4 l4_wkup */
692 	{ .start = 0x44c00000, .end = 0x44c00000 + 0x300000, },
693 	/* omap4/5 and dra7 l4_cfg */
694 	{ .start = 0x4a000000, .end = 0x4a000000 + 0x300000, },
695 	/* omap4 l4_wkup */
696 	{ .start = 0x4a300000, .end = 0x4a300000 + 0x30000,  },
697 	/* omap5 and dra7 l4_wkup without dra7 dcan segment */
698 	{ .start = 0x4ae00000, .end = 0x4ae00000 + 0x30000,  },
699 };
700 
701 static atomic_t sysc_defer = ATOMIC_INIT(10);
702 
703 /**
704  * sysc_defer_non_critical - defer non_critical interconnect probing
705  * @ddata: device driver data
706  *
707  * We want to probe l4_cfg and l4_wkup interconnect instances before any
708  * l4_per instances as l4_per instances depend on resources on l4_cfg and
709  * l4_wkup interconnects.
710  */
sysc_defer_non_critical(struct sysc * ddata)711 static int sysc_defer_non_critical(struct sysc *ddata)
712 {
713 	struct resource *res;
714 	int i;
715 
716 	if (!atomic_read(&sysc_defer))
717 		return 0;
718 
719 	for (i = 0; i < ARRAY_SIZE(early_bus_ranges); i++) {
720 		res = &early_bus_ranges[i];
721 		if (ddata->module_pa >= res->start &&
722 		    ddata->module_pa <= res->end) {
723 			atomic_set(&sysc_defer, 0);
724 
725 			return 0;
726 		}
727 	}
728 
729 	atomic_dec_if_positive(&sysc_defer);
730 
731 	return -EPROBE_DEFER;
732 }
733 
734 static struct device_node *stdout_path;
735 
sysc_init_stdout_path(struct sysc * ddata)736 static void sysc_init_stdout_path(struct sysc *ddata)
737 {
738 	struct device_node *np = NULL;
739 	const char *uart;
740 
741 	if (IS_ERR(stdout_path))
742 		return;
743 
744 	if (stdout_path)
745 		return;
746 
747 	np = of_find_node_by_path("/chosen");
748 	if (!np)
749 		goto err;
750 
751 	uart = of_get_property(np, "stdout-path", NULL);
752 	if (!uart)
753 		goto err;
754 
755 	np = of_find_node_by_path(uart);
756 	if (!np)
757 		goto err;
758 
759 	stdout_path = np;
760 
761 	return;
762 
763 err:
764 	stdout_path = ERR_PTR(-ENODEV);
765 }
766 
sysc_check_quirk_stdout(struct sysc * ddata,struct device_node * np)767 static void sysc_check_quirk_stdout(struct sysc *ddata,
768 				    struct device_node *np)
769 {
770 	sysc_init_stdout_path(ddata);
771 	if (np != stdout_path)
772 		return;
773 
774 	ddata->cfg.quirks |= SYSC_QUIRK_NO_IDLE_ON_INIT |
775 				SYSC_QUIRK_NO_RESET_ON_INIT;
776 }
777 
778 /**
779  * sysc_check_one_child - check child configuration
780  * @ddata: device driver data
781  * @np: child device node
782  *
783  * Let's avoid messy situations where we have new interconnect target
784  * node but children have "ti,hwmods". These belong to the interconnect
785  * target node and are managed by this driver.
786  */
sysc_check_one_child(struct sysc * ddata,struct device_node * np)787 static void sysc_check_one_child(struct sysc *ddata,
788 				 struct device_node *np)
789 {
790 	const char *name;
791 
792 	name = of_get_property(np, "ti,hwmods", NULL);
793 	if (name && !of_device_is_compatible(np, "ti,sysc"))
794 		dev_warn(ddata->dev, "really a child ti,hwmods property?");
795 
796 	sysc_check_quirk_stdout(ddata, np);
797 	sysc_parse_dts_quirks(ddata, np, true);
798 }
799 
sysc_check_children(struct sysc * ddata)800 static void sysc_check_children(struct sysc *ddata)
801 {
802 	struct device_node *child;
803 
804 	for_each_child_of_node(ddata->dev->of_node, child)
805 		sysc_check_one_child(ddata, child);
806 }
807 
808 /*
809  * So far only I2C uses 16-bit read access with clockactivity with revision
810  * in two registers with stride of 4. We can detect this based on the rev
811  * register size to configure things far enough to be able to properly read
812  * the revision register.
813  */
sysc_check_quirk_16bit(struct sysc * ddata,struct resource * res)814 static void sysc_check_quirk_16bit(struct sysc *ddata, struct resource *res)
815 {
816 	if (resource_size(res) == 8)
817 		ddata->cfg.quirks |= SYSC_QUIRK_16BIT | SYSC_QUIRK_USE_CLOCKACT;
818 }
819 
820 /**
821  * sysc_parse_one - parses the interconnect target module registers
822  * @ddata: device driver data
823  * @reg: register to parse
824  */
sysc_parse_one(struct sysc * ddata,enum sysc_registers reg)825 static int sysc_parse_one(struct sysc *ddata, enum sysc_registers reg)
826 {
827 	struct resource *res;
828 	const char *name;
829 
830 	switch (reg) {
831 	case SYSC_REVISION:
832 	case SYSC_SYSCONFIG:
833 	case SYSC_SYSSTATUS:
834 		name = reg_names[reg];
835 		break;
836 	default:
837 		return -EINVAL;
838 	}
839 
840 	res = platform_get_resource_byname(to_platform_device(ddata->dev),
841 					   IORESOURCE_MEM, name);
842 	if (!res) {
843 		ddata->offsets[reg] = -ENODEV;
844 
845 		return 0;
846 	}
847 
848 	ddata->offsets[reg] = res->start - ddata->module_pa;
849 	if (reg == SYSC_REVISION)
850 		sysc_check_quirk_16bit(ddata, res);
851 
852 	return 0;
853 }
854 
sysc_parse_registers(struct sysc * ddata)855 static int sysc_parse_registers(struct sysc *ddata)
856 {
857 	int i, error;
858 
859 	for (i = 0; i < SYSC_MAX_REGS; i++) {
860 		error = sysc_parse_one(ddata, i);
861 		if (error)
862 			return error;
863 	}
864 
865 	return 0;
866 }
867 
868 /**
869  * sysc_check_registers - check for misconfigured register overlaps
870  * @ddata: device driver data
871  */
sysc_check_registers(struct sysc * ddata)872 static int sysc_check_registers(struct sysc *ddata)
873 {
874 	int i, j, nr_regs = 0, nr_matches = 0;
875 
876 	for (i = 0; i < SYSC_MAX_REGS; i++) {
877 		if (ddata->offsets[i] < 0)
878 			continue;
879 
880 		if (ddata->offsets[i] > (ddata->module_size - 4)) {
881 			dev_err(ddata->dev, "register outside module range");
882 
883 				return -EINVAL;
884 		}
885 
886 		for (j = 0; j < SYSC_MAX_REGS; j++) {
887 			if (ddata->offsets[j] < 0)
888 				continue;
889 
890 			if (ddata->offsets[i] == ddata->offsets[j])
891 				nr_matches++;
892 		}
893 		nr_regs++;
894 	}
895 
896 	if (nr_matches > nr_regs) {
897 		dev_err(ddata->dev, "overlapping registers: (%i/%i)",
898 			nr_regs, nr_matches);
899 
900 		return -EINVAL;
901 	}
902 
903 	return 0;
904 }
905 
906 /**
907  * syc_ioremap - ioremap register space for the interconnect target module
908  * @ddata: device driver data
909  *
910  * Note that the interconnect target module registers can be anywhere
911  * within the interconnect target module range. For example, SGX has
912  * them at offset 0x1fc00 in the 32MB module address space. And cpsw
913  * has them at offset 0x1200 in the CPSW_WR child. Usually the
914  * the interconnect target module registers are at the beginning of
915  * the module range though.
916  */
sysc_ioremap(struct sysc * ddata)917 static int sysc_ioremap(struct sysc *ddata)
918 {
919 	int size;
920 
921 	if (ddata->offsets[SYSC_REVISION] < 0 &&
922 	    ddata->offsets[SYSC_SYSCONFIG] < 0 &&
923 	    ddata->offsets[SYSC_SYSSTATUS] < 0) {
924 		size = ddata->module_size;
925 	} else {
926 		size = max3(ddata->offsets[SYSC_REVISION],
927 			    ddata->offsets[SYSC_SYSCONFIG],
928 			    ddata->offsets[SYSC_SYSSTATUS]);
929 
930 		if (size < SZ_1K)
931 			size = SZ_1K;
932 
933 		if ((size + sizeof(u32)) > ddata->module_size)
934 			size = ddata->module_size;
935 	}
936 
937 	ddata->module_va = devm_ioremap(ddata->dev,
938 					ddata->module_pa,
939 					size + sizeof(u32));
940 	if (!ddata->module_va)
941 		return -EIO;
942 
943 	return 0;
944 }
945 
946 /**
947  * sysc_map_and_check_registers - ioremap and check device registers
948  * @ddata: device driver data
949  */
sysc_map_and_check_registers(struct sysc * ddata)950 static int sysc_map_and_check_registers(struct sysc *ddata)
951 {
952 	int error;
953 
954 	error = sysc_parse_and_check_child_range(ddata);
955 	if (error)
956 		return error;
957 
958 	error = sysc_defer_non_critical(ddata);
959 	if (error)
960 		return error;
961 
962 	sysc_check_children(ddata);
963 
964 	error = sysc_parse_registers(ddata);
965 	if (error)
966 		return error;
967 
968 	error = sysc_ioremap(ddata);
969 	if (error)
970 		return error;
971 
972 	error = sysc_check_registers(ddata);
973 	if (error)
974 		return error;
975 
976 	return 0;
977 }
978 
979 /**
980  * sysc_show_rev - read and show interconnect target module revision
981  * @bufp: buffer to print the information to
982  * @ddata: device driver data
983  */
sysc_show_rev(char * bufp,struct sysc * ddata)984 static int sysc_show_rev(char *bufp, struct sysc *ddata)
985 {
986 	int len;
987 
988 	if (ddata->offsets[SYSC_REVISION] < 0)
989 		return sprintf(bufp, ":NA");
990 
991 	len = sprintf(bufp, ":%08x", ddata->revision);
992 
993 	return len;
994 }
995 
sysc_show_reg(struct sysc * ddata,char * bufp,enum sysc_registers reg)996 static int sysc_show_reg(struct sysc *ddata,
997 			 char *bufp, enum sysc_registers reg)
998 {
999 	if (ddata->offsets[reg] < 0)
1000 		return sprintf(bufp, ":NA");
1001 
1002 	return sprintf(bufp, ":%x", ddata->offsets[reg]);
1003 }
1004 
sysc_show_name(char * bufp,struct sysc * ddata)1005 static int sysc_show_name(char *bufp, struct sysc *ddata)
1006 {
1007 	if (!ddata->name)
1008 		return 0;
1009 
1010 	return sprintf(bufp, ":%s", ddata->name);
1011 }
1012 
1013 /**
1014  * sysc_show_registers - show information about interconnect target module
1015  * @ddata: device driver data
1016  */
sysc_show_registers(struct sysc * ddata)1017 static void sysc_show_registers(struct sysc *ddata)
1018 {
1019 	char buf[128];
1020 	char *bufp = buf;
1021 	int i;
1022 
1023 	for (i = 0; i < SYSC_MAX_REGS; i++)
1024 		bufp += sysc_show_reg(ddata, bufp, i);
1025 
1026 	bufp += sysc_show_rev(bufp, ddata);
1027 	bufp += sysc_show_name(bufp, ddata);
1028 
1029 	dev_dbg(ddata->dev, "%llx:%x%s\n",
1030 		ddata->module_pa, ddata->module_size,
1031 		buf);
1032 }
1033 
1034 /**
1035  * sysc_write_sysconfig - handle sysconfig quirks for register write
1036  * @ddata: device driver data
1037  * @value: register value
1038  */
sysc_write_sysconfig(struct sysc * ddata,u32 value)1039 static void sysc_write_sysconfig(struct sysc *ddata, u32 value)
1040 {
1041 	if (ddata->module_unlock_quirk)
1042 		ddata->module_unlock_quirk(ddata);
1043 
1044 	sysc_write(ddata, ddata->offsets[SYSC_SYSCONFIG], value);
1045 
1046 	if (ddata->module_lock_quirk)
1047 		ddata->module_lock_quirk(ddata);
1048 }
1049 
1050 #define SYSC_IDLE_MASK	(SYSC_NR_IDLEMODES - 1)
1051 #define SYSC_CLOCACT_ICK	2
1052 
1053 /* Caller needs to manage sysc_clkdm_deny_idle() and sysc_clkdm_allow_idle() */
sysc_enable_module(struct device * dev)1054 static int sysc_enable_module(struct device *dev)
1055 {
1056 	struct sysc *ddata;
1057 	const struct sysc_regbits *regbits;
1058 	u32 reg, idlemodes, best_mode;
1059 	int error;
1060 
1061 	ddata = dev_get_drvdata(dev);
1062 
1063 	/*
1064 	 * Some modules like DSS reset automatically on idle. Enable optional
1065 	 * reset clocks and wait for OCP softreset to complete.
1066 	 */
1067 	if (ddata->cfg.quirks & SYSC_QUIRK_OPT_CLKS_IN_RESET) {
1068 		error = sysc_enable_opt_clocks(ddata);
1069 		if (error) {
1070 			dev_err(ddata->dev,
1071 				"Optional clocks failed for enable: %i\n",
1072 				error);
1073 			return error;
1074 		}
1075 	}
1076 	/*
1077 	 * Some modules like i2c and hdq1w have unusable reset status unless
1078 	 * the module reset quirk is enabled. Skip status check on enable.
1079 	 */
1080 	if (!(ddata->cfg.quirks & SYSC_MODULE_QUIRK_ENA_RESETDONE)) {
1081 		error = sysc_wait_softreset(ddata);
1082 		if (error)
1083 			dev_warn(ddata->dev, "OCP softreset timed out\n");
1084 	}
1085 	if (ddata->cfg.quirks & SYSC_QUIRK_OPT_CLKS_IN_RESET)
1086 		sysc_disable_opt_clocks(ddata);
1087 
1088 	/*
1089 	 * Some subsystem private interconnects, like DSS top level module,
1090 	 * need only the automatic OCP softreset handling with no sysconfig
1091 	 * register bits to configure.
1092 	 */
1093 	if (ddata->offsets[SYSC_SYSCONFIG] == -ENODEV)
1094 		return 0;
1095 
1096 	regbits = ddata->cap->regbits;
1097 	reg = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1098 
1099 	/*
1100 	 * Set CLOCKACTIVITY, we only use it for ick. And we only configure it
1101 	 * based on the SYSC_QUIRK_USE_CLOCKACT flag, not based on the hardware
1102 	 * capabilities. See the old HWMOD_SET_DEFAULT_CLOCKACT flag.
1103 	 */
1104 	if (regbits->clkact_shift >= 0 &&
1105 	    (ddata->cfg.quirks & SYSC_QUIRK_USE_CLOCKACT))
1106 		reg |= SYSC_CLOCACT_ICK << regbits->clkact_shift;
1107 
1108 	/* Set SIDLE mode */
1109 	idlemodes = ddata->cfg.sidlemodes;
1110 	if (!idlemodes || regbits->sidle_shift < 0)
1111 		goto set_midle;
1112 
1113 	if (ddata->cfg.quirks & (SYSC_QUIRK_SWSUP_SIDLE |
1114 				 SYSC_QUIRK_SWSUP_SIDLE_ACT)) {
1115 		best_mode = SYSC_IDLE_NO;
1116 	} else {
1117 		best_mode = fls(ddata->cfg.sidlemodes) - 1;
1118 		if (best_mode > SYSC_IDLE_MASK) {
1119 			dev_err(dev, "%s: invalid sidlemode\n", __func__);
1120 			return -EINVAL;
1121 		}
1122 
1123 		/* Set WAKEUP */
1124 		if (regbits->enwkup_shift >= 0 &&
1125 		    ddata->cfg.sysc_val & BIT(regbits->enwkup_shift))
1126 			reg |= BIT(regbits->enwkup_shift);
1127 	}
1128 
1129 	reg &= ~(SYSC_IDLE_MASK << regbits->sidle_shift);
1130 	reg |= best_mode << regbits->sidle_shift;
1131 	sysc_write_sysconfig(ddata, reg);
1132 
1133 set_midle:
1134 	/* Set MIDLE mode */
1135 	idlemodes = ddata->cfg.midlemodes;
1136 	if (!idlemodes || regbits->midle_shift < 0)
1137 		goto set_autoidle;
1138 
1139 	best_mode = fls(ddata->cfg.midlemodes) - 1;
1140 	if (best_mode > SYSC_IDLE_MASK) {
1141 		dev_err(dev, "%s: invalid midlemode\n", __func__);
1142 		return -EINVAL;
1143 	}
1144 
1145 	if (ddata->cfg.quirks & SYSC_QUIRK_SWSUP_MSTANDBY)
1146 		best_mode = SYSC_IDLE_NO;
1147 
1148 	reg &= ~(SYSC_IDLE_MASK << regbits->midle_shift);
1149 	reg |= best_mode << regbits->midle_shift;
1150 	sysc_write_sysconfig(ddata, reg);
1151 
1152 set_autoidle:
1153 	/* Autoidle bit must enabled separately if available */
1154 	if (regbits->autoidle_shift >= 0 &&
1155 	    ddata->cfg.sysc_val & BIT(regbits->autoidle_shift)) {
1156 		reg |= 1 << regbits->autoidle_shift;
1157 		sysc_write_sysconfig(ddata, reg);
1158 	}
1159 
1160 	/* Flush posted write */
1161 	sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1162 
1163 	if (ddata->module_enable_quirk)
1164 		ddata->module_enable_quirk(ddata);
1165 
1166 	return 0;
1167 }
1168 
sysc_best_idle_mode(u32 idlemodes,u32 * best_mode)1169 static int sysc_best_idle_mode(u32 idlemodes, u32 *best_mode)
1170 {
1171 	if (idlemodes & BIT(SYSC_IDLE_SMART_WKUP))
1172 		*best_mode = SYSC_IDLE_SMART_WKUP;
1173 	else if (idlemodes & BIT(SYSC_IDLE_SMART))
1174 		*best_mode = SYSC_IDLE_SMART;
1175 	else if (idlemodes & BIT(SYSC_IDLE_FORCE))
1176 		*best_mode = SYSC_IDLE_FORCE;
1177 	else
1178 		return -EINVAL;
1179 
1180 	return 0;
1181 }
1182 
1183 /* Caller needs to manage sysc_clkdm_deny_idle() and sysc_clkdm_allow_idle() */
sysc_disable_module(struct device * dev)1184 static int sysc_disable_module(struct device *dev)
1185 {
1186 	struct sysc *ddata;
1187 	const struct sysc_regbits *regbits;
1188 	u32 reg, idlemodes, best_mode;
1189 	int ret;
1190 
1191 	ddata = dev_get_drvdata(dev);
1192 	if (ddata->offsets[SYSC_SYSCONFIG] == -ENODEV)
1193 		return 0;
1194 
1195 	if (ddata->module_disable_quirk)
1196 		ddata->module_disable_quirk(ddata);
1197 
1198 	regbits = ddata->cap->regbits;
1199 	reg = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1200 
1201 	/* Set MIDLE mode */
1202 	idlemodes = ddata->cfg.midlemodes;
1203 	if (!idlemodes || regbits->midle_shift < 0)
1204 		goto set_sidle;
1205 
1206 	ret = sysc_best_idle_mode(idlemodes, &best_mode);
1207 	if (ret) {
1208 		dev_err(dev, "%s: invalid midlemode\n", __func__);
1209 		return ret;
1210 	}
1211 
1212 	if (ddata->cfg.quirks & (SYSC_QUIRK_SWSUP_MSTANDBY) ||
1213 	    ddata->cfg.quirks & (SYSC_QUIRK_FORCE_MSTANDBY))
1214 		best_mode = SYSC_IDLE_FORCE;
1215 
1216 	reg &= ~(SYSC_IDLE_MASK << regbits->midle_shift);
1217 	reg |= best_mode << regbits->midle_shift;
1218 	sysc_write_sysconfig(ddata, reg);
1219 
1220 set_sidle:
1221 	/* Set SIDLE mode */
1222 	idlemodes = ddata->cfg.sidlemodes;
1223 	if (!idlemodes || regbits->sidle_shift < 0)
1224 		return 0;
1225 
1226 	if (ddata->cfg.quirks & SYSC_QUIRK_SWSUP_SIDLE) {
1227 		best_mode = SYSC_IDLE_FORCE;
1228 	} else {
1229 		ret = sysc_best_idle_mode(idlemodes, &best_mode);
1230 		if (ret) {
1231 			dev_err(dev, "%s: invalid sidlemode\n", __func__);
1232 			return ret;
1233 		}
1234 	}
1235 
1236 	reg &= ~(SYSC_IDLE_MASK << regbits->sidle_shift);
1237 	reg |= best_mode << regbits->sidle_shift;
1238 	if (regbits->autoidle_shift >= 0 &&
1239 	    ddata->cfg.sysc_val & BIT(regbits->autoidle_shift))
1240 		reg |= 1 << regbits->autoidle_shift;
1241 	sysc_write_sysconfig(ddata, reg);
1242 
1243 	/* Flush posted write */
1244 	sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1245 
1246 	return 0;
1247 }
1248 
sysc_runtime_suspend_legacy(struct device * dev,struct sysc * ddata)1249 static int __maybe_unused sysc_runtime_suspend_legacy(struct device *dev,
1250 						      struct sysc *ddata)
1251 {
1252 	struct ti_sysc_platform_data *pdata;
1253 	int error;
1254 
1255 	pdata = dev_get_platdata(ddata->dev);
1256 	if (!pdata)
1257 		return 0;
1258 
1259 	if (!pdata->idle_module)
1260 		return -ENODEV;
1261 
1262 	error = pdata->idle_module(dev, &ddata->cookie);
1263 	if (error)
1264 		dev_err(dev, "%s: could not idle: %i\n",
1265 			__func__, error);
1266 
1267 	reset_control_assert(ddata->rsts);
1268 
1269 	return 0;
1270 }
1271 
sysc_runtime_resume_legacy(struct device * dev,struct sysc * ddata)1272 static int __maybe_unused sysc_runtime_resume_legacy(struct device *dev,
1273 						     struct sysc *ddata)
1274 {
1275 	struct ti_sysc_platform_data *pdata;
1276 	int error;
1277 
1278 	pdata = dev_get_platdata(ddata->dev);
1279 	if (!pdata)
1280 		return 0;
1281 
1282 	if (!pdata->enable_module)
1283 		return -ENODEV;
1284 
1285 	error = pdata->enable_module(dev, &ddata->cookie);
1286 	if (error)
1287 		dev_err(dev, "%s: could not enable: %i\n",
1288 			__func__, error);
1289 
1290 	reset_control_deassert(ddata->rsts);
1291 
1292 	return 0;
1293 }
1294 
sysc_runtime_suspend(struct device * dev)1295 static int __maybe_unused sysc_runtime_suspend(struct device *dev)
1296 {
1297 	struct sysc *ddata;
1298 	int error = 0;
1299 
1300 	ddata = dev_get_drvdata(dev);
1301 
1302 	if (!ddata->enabled)
1303 		return 0;
1304 
1305 	sysc_clkdm_deny_idle(ddata);
1306 
1307 	if (ddata->legacy_mode) {
1308 		error = sysc_runtime_suspend_legacy(dev, ddata);
1309 		if (error)
1310 			goto err_allow_idle;
1311 	} else {
1312 		error = sysc_disable_module(dev);
1313 		if (error)
1314 			goto err_allow_idle;
1315 	}
1316 
1317 	sysc_disable_main_clocks(ddata);
1318 
1319 	if (sysc_opt_clks_needed(ddata))
1320 		sysc_disable_opt_clocks(ddata);
1321 
1322 	ddata->enabled = false;
1323 
1324 err_allow_idle:
1325 	reset_control_assert(ddata->rsts);
1326 
1327 	sysc_clkdm_allow_idle(ddata);
1328 
1329 	return error;
1330 }
1331 
sysc_runtime_resume(struct device * dev)1332 static int __maybe_unused sysc_runtime_resume(struct device *dev)
1333 {
1334 	struct sysc *ddata;
1335 	int error = 0;
1336 
1337 	ddata = dev_get_drvdata(dev);
1338 
1339 	if (ddata->enabled)
1340 		return 0;
1341 
1342 
1343 	sysc_clkdm_deny_idle(ddata);
1344 
1345 	if (sysc_opt_clks_needed(ddata)) {
1346 		error = sysc_enable_opt_clocks(ddata);
1347 		if (error)
1348 			goto err_allow_idle;
1349 	}
1350 
1351 	error = sysc_enable_main_clocks(ddata);
1352 	if (error)
1353 		goto err_opt_clocks;
1354 
1355 	reset_control_deassert(ddata->rsts);
1356 
1357 	if (ddata->legacy_mode) {
1358 		error = sysc_runtime_resume_legacy(dev, ddata);
1359 		if (error)
1360 			goto err_main_clocks;
1361 	} else {
1362 		error = sysc_enable_module(dev);
1363 		if (error)
1364 			goto err_main_clocks;
1365 	}
1366 
1367 	ddata->enabled = true;
1368 
1369 	sysc_clkdm_allow_idle(ddata);
1370 
1371 	return 0;
1372 
1373 err_main_clocks:
1374 	sysc_disable_main_clocks(ddata);
1375 err_opt_clocks:
1376 	if (sysc_opt_clks_needed(ddata))
1377 		sysc_disable_opt_clocks(ddata);
1378 err_allow_idle:
1379 	sysc_clkdm_allow_idle(ddata);
1380 
1381 	return error;
1382 }
1383 
sysc_reinit_module(struct sysc * ddata,bool leave_enabled)1384 static int sysc_reinit_module(struct sysc *ddata, bool leave_enabled)
1385 {
1386 	struct device *dev = ddata->dev;
1387 	int error;
1388 
1389 	/* Disable target module if it is enabled */
1390 	if (ddata->enabled) {
1391 		error = sysc_runtime_suspend(dev);
1392 		if (error)
1393 			dev_warn(dev, "reinit suspend failed: %i\n", error);
1394 	}
1395 
1396 	/* Enable target module */
1397 	error = sysc_runtime_resume(dev);
1398 	if (error)
1399 		dev_warn(dev, "reinit resume failed: %i\n", error);
1400 
1401 	if (leave_enabled)
1402 		return error;
1403 
1404 	/* Disable target module if no leave_enabled was set */
1405 	error = sysc_runtime_suspend(dev);
1406 	if (error)
1407 		dev_warn(dev, "reinit suspend failed: %i\n", error);
1408 
1409 	return error;
1410 }
1411 
sysc_noirq_suspend(struct device * dev)1412 static int __maybe_unused sysc_noirq_suspend(struct device *dev)
1413 {
1414 	struct sysc *ddata;
1415 
1416 	ddata = dev_get_drvdata(dev);
1417 
1418 	if (ddata->cfg.quirks &
1419 	    (SYSC_QUIRK_LEGACY_IDLE | SYSC_QUIRK_NO_IDLE))
1420 		return 0;
1421 
1422 	if (!ddata->enabled)
1423 		return 0;
1424 
1425 	ddata->needs_resume = 1;
1426 
1427 	return sysc_runtime_suspend(dev);
1428 }
1429 
sysc_noirq_resume(struct device * dev)1430 static int __maybe_unused sysc_noirq_resume(struct device *dev)
1431 {
1432 	struct sysc *ddata;
1433 	int error = 0;
1434 
1435 	ddata = dev_get_drvdata(dev);
1436 
1437 	if (ddata->cfg.quirks &
1438 	    (SYSC_QUIRK_LEGACY_IDLE | SYSC_QUIRK_NO_IDLE))
1439 		return 0;
1440 
1441 	if (ddata->cfg.quirks & SYSC_QUIRK_REINIT_ON_RESUME) {
1442 		error = sysc_reinit_module(ddata, ddata->needs_resume);
1443 		if (error)
1444 			dev_warn(dev, "noirq_resume failed: %i\n", error);
1445 	} else if (ddata->needs_resume) {
1446 		error = sysc_runtime_resume(dev);
1447 		if (error)
1448 			dev_warn(dev, "noirq_resume failed: %i\n", error);
1449 	}
1450 
1451 	ddata->needs_resume = 0;
1452 
1453 	return error;
1454 }
1455 
1456 static const struct dev_pm_ops sysc_pm_ops = {
1457 	SET_NOIRQ_SYSTEM_SLEEP_PM_OPS(sysc_noirq_suspend, sysc_noirq_resume)
1458 	SET_RUNTIME_PM_OPS(sysc_runtime_suspend,
1459 			   sysc_runtime_resume,
1460 			   NULL)
1461 };
1462 
1463 /* Module revision register based quirks */
1464 struct sysc_revision_quirk {
1465 	const char *name;
1466 	u32 base;
1467 	int rev_offset;
1468 	int sysc_offset;
1469 	int syss_offset;
1470 	u32 revision;
1471 	u32 revision_mask;
1472 	u32 quirks;
1473 };
1474 
1475 #define SYSC_QUIRK(optname, optbase, optrev, optsysc, optsyss,		\
1476 		   optrev_val, optrevmask, optquirkmask)		\
1477 	{								\
1478 		.name = (optname),					\
1479 		.base = (optbase),					\
1480 		.rev_offset = (optrev),					\
1481 		.sysc_offset = (optsysc),				\
1482 		.syss_offset = (optsyss),				\
1483 		.revision = (optrev_val),				\
1484 		.revision_mask = (optrevmask),				\
1485 		.quirks = (optquirkmask),				\
1486 	}
1487 
1488 static const struct sysc_revision_quirk sysc_revision_quirks[] = {
1489 	/* These drivers need to be fixed to not use pm_runtime_irq_safe() */
1490 	SYSC_QUIRK("gpio", 0, 0, 0x10, 0x114, 0x50600801, 0xffff00ff,
1491 		   SYSC_QUIRK_LEGACY_IDLE | SYSC_QUIRK_OPT_CLKS_IN_RESET),
1492 	SYSC_QUIRK("sham", 0, 0x100, 0x110, 0x114, 0x40000c03, 0xffffffff,
1493 		   SYSC_QUIRK_LEGACY_IDLE),
1494 	SYSC_QUIRK("smartreflex", 0, -ENODEV, 0x24, -ENODEV, 0x00000000, 0xffffffff,
1495 		   SYSC_QUIRK_LEGACY_IDLE),
1496 	SYSC_QUIRK("smartreflex", 0, -ENODEV, 0x38, -ENODEV, 0x00000000, 0xffffffff,
1497 		   SYSC_QUIRK_LEGACY_IDLE),
1498 	SYSC_QUIRK("uart", 0, 0x50, 0x54, 0x58, 0x00000046, 0xffffffff,
1499 		   SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_LEGACY_IDLE),
1500 	SYSC_QUIRK("uart", 0, 0x50, 0x54, 0x58, 0x00000052, 0xffffffff,
1501 		   SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_LEGACY_IDLE),
1502 	/* Uarts on omap4 and later */
1503 	SYSC_QUIRK("uart", 0, 0x50, 0x54, 0x58, 0x50411e03, 0xffff00ff,
1504 		   SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_LEGACY_IDLE),
1505 	SYSC_QUIRK("uart", 0, 0x50, 0x54, 0x58, 0x47422e03, 0xffffffff,
1506 		   SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_LEGACY_IDLE),
1507 
1508 	/* Quirks that need to be set based on the module address */
1509 	SYSC_QUIRK("mcpdm", 0x40132000, 0, 0x10, -ENODEV, 0x50000800, 0xffffffff,
1510 		   SYSC_QUIRK_EXT_OPT_CLOCK | SYSC_QUIRK_NO_RESET_ON_INIT |
1511 		   SYSC_QUIRK_SWSUP_SIDLE),
1512 
1513 	/* Quirks that need to be set based on detected module */
1514 	SYSC_QUIRK("aess", 0, 0, 0x10, -ENODEV, 0x40000000, 0xffffffff,
1515 		   SYSC_MODULE_QUIRK_AESS),
1516 	/* Errata i893 handling for dra7 dcan1 and 2 */
1517 	SYSC_QUIRK("dcan", 0x4ae3c000, 0x20, -ENODEV, -ENODEV, 0xa3170504, 0xffffffff,
1518 		   SYSC_QUIRK_CLKDM_NOAUTO),
1519 	SYSC_QUIRK("dcan", 0x48480000, 0x20, -ENODEV, -ENODEV, 0xa3170504, 0xffffffff,
1520 		   SYSC_QUIRK_CLKDM_NOAUTO),
1521 	SYSC_QUIRK("dss", 0x4832a000, 0, 0x10, 0x14, 0x00000020, 0xffffffff,
1522 		   SYSC_QUIRK_OPT_CLKS_IN_RESET | SYSC_MODULE_QUIRK_DSS_RESET),
1523 	SYSC_QUIRK("dss", 0x58000000, 0, -ENODEV, 0x14, 0x00000040, 0xffffffff,
1524 		   SYSC_QUIRK_OPT_CLKS_IN_RESET | SYSC_MODULE_QUIRK_DSS_RESET),
1525 	SYSC_QUIRK("dss", 0x58000000, 0, -ENODEV, 0x14, 0x00000061, 0xffffffff,
1526 		   SYSC_QUIRK_OPT_CLKS_IN_RESET | SYSC_MODULE_QUIRK_DSS_RESET),
1527 	SYSC_QUIRK("dwc3", 0x48880000, 0, 0x10, -ENODEV, 0x500a0200, 0xffffffff,
1528 		   SYSC_QUIRK_CLKDM_NOAUTO),
1529 	SYSC_QUIRK("dwc3", 0x488c0000, 0, 0x10, -ENODEV, 0x500a0200, 0xffffffff,
1530 		   SYSC_QUIRK_CLKDM_NOAUTO),
1531 	SYSC_QUIRK("gpmc", 0, 0, 0x10, 0x14, 0x00000060, 0xffffffff,
1532 		   SYSC_QUIRK_GPMC_DEBUG),
1533 	SYSC_QUIRK("hdmi", 0, 0, 0x10, -ENODEV, 0x50030200, 0xffffffff,
1534 		   SYSC_QUIRK_OPT_CLKS_NEEDED),
1535 	SYSC_QUIRK("hdq1w", 0, 0, 0x14, 0x18, 0x00000006, 0xffffffff,
1536 		   SYSC_MODULE_QUIRK_HDQ1W | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1537 	SYSC_QUIRK("hdq1w", 0, 0, 0x14, 0x18, 0x0000000a, 0xffffffff,
1538 		   SYSC_MODULE_QUIRK_HDQ1W | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1539 	SYSC_QUIRK("i2c", 0, 0, 0x20, 0x10, 0x00000036, 0x000000ff,
1540 		   SYSC_MODULE_QUIRK_I2C | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1541 	SYSC_QUIRK("i2c", 0, 0, 0x20, 0x10, 0x0000003c, 0x000000ff,
1542 		   SYSC_MODULE_QUIRK_I2C | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1543 	SYSC_QUIRK("i2c", 0, 0, 0x20, 0x10, 0x00000040, 0x000000ff,
1544 		   SYSC_MODULE_QUIRK_I2C | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1545 	SYSC_QUIRK("i2c", 0, 0, 0x10, 0x90, 0x5040000a, 0xfffff0f0,
1546 		   SYSC_MODULE_QUIRK_I2C | SYSC_MODULE_QUIRK_ENA_RESETDONE),
1547 	SYSC_QUIRK("gpu", 0x50000000, 0x14, -ENODEV, -ENODEV, 0x00010201, 0xffffffff, 0),
1548 	SYSC_QUIRK("gpu", 0x50000000, 0xfe00, 0xfe10, -ENODEV, 0x40000000 , 0xffffffff,
1549 		   SYSC_MODULE_QUIRK_SGX),
1550 	SYSC_QUIRK("lcdc", 0, 0, 0x54, -ENODEV, 0x4f201000, 0xffffffff,
1551 		   SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1552 	SYSC_QUIRK("rtc", 0, 0x74, 0x78, -ENODEV, 0x4eb01908, 0xffff00f0,
1553 		   SYSC_MODULE_QUIRK_RTC_UNLOCK),
1554 	SYSC_QUIRK("tptc", 0, 0, 0x10, -ENODEV, 0x40006c00, 0xffffefff,
1555 		   SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1556 	SYSC_QUIRK("tptc", 0, 0, -ENODEV, -ENODEV, 0x40007c00, 0xffffffff,
1557 		   SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1558 	SYSC_QUIRK("usb_host_hs", 0, 0, 0x10, 0x14, 0x50700100, 0xffffffff,
1559 		   SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1560 	SYSC_QUIRK("usb_host_hs", 0, 0, 0x10, -ENODEV, 0x50700101, 0xffffffff,
1561 		   SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1562 	SYSC_QUIRK("usb_otg_hs", 0, 0x400, 0x404, 0x408, 0x00000050,
1563 		   0xffffffff, SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY),
1564 	SYSC_QUIRK("usb_otg_hs", 0, 0, 0x10, -ENODEV, 0x4ea2080d, 0xffffffff,
1565 		   SYSC_QUIRK_SWSUP_SIDLE | SYSC_QUIRK_SWSUP_MSTANDBY |
1566 		   SYSC_QUIRK_REINIT_ON_CTX_LOST),
1567 	SYSC_QUIRK("wdt", 0, 0, 0x10, 0x14, 0x502a0500, 0xfffff0f0,
1568 		   SYSC_MODULE_QUIRK_WDT),
1569 	/* PRUSS on am3, am4 and am5 */
1570 	SYSC_QUIRK("pruss", 0, 0x26000, 0x26004, -ENODEV, 0x47000000, 0xff000000,
1571 		   SYSC_MODULE_QUIRK_PRUSS),
1572 	/* Watchdog on am3 and am4 */
1573 	SYSC_QUIRK("wdt", 0x44e35000, 0, 0x10, 0x14, 0x502a0500, 0xfffff0f0,
1574 		   SYSC_MODULE_QUIRK_WDT | SYSC_QUIRK_SWSUP_SIDLE),
1575 
1576 #ifdef DEBUG
1577 	SYSC_QUIRK("adc", 0, 0, 0x10, -ENODEV, 0x47300001, 0xffffffff, 0),
1578 	SYSC_QUIRK("atl", 0, 0, -ENODEV, -ENODEV, 0x0a070100, 0xffffffff, 0),
1579 	SYSC_QUIRK("cm", 0, 0, -ENODEV, -ENODEV, 0x40000301, 0xffffffff, 0),
1580 	SYSC_QUIRK("control", 0, 0, 0x10, -ENODEV, 0x40000900, 0xffffffff, 0),
1581 	SYSC_QUIRK("cpgmac", 0, 0x1200, 0x1208, 0x1204, 0x4edb1902,
1582 		   0xffff00f0, 0),
1583 	SYSC_QUIRK("dcan", 0, 0x20, -ENODEV, -ENODEV, 0xa3170504, 0xffffffff, 0),
1584 	SYSC_QUIRK("dcan", 0, 0x20, -ENODEV, -ENODEV, 0x4edb1902, 0xffffffff, 0),
1585 	SYSC_QUIRK("dispc", 0x4832a400, 0, 0x10, 0x14, 0x00000030, 0xffffffff, 0),
1586 	SYSC_QUIRK("dispc", 0x58001000, 0, 0x10, 0x14, 0x00000040, 0xffffffff, 0),
1587 	SYSC_QUIRK("dispc", 0x58001000, 0, 0x10, 0x14, 0x00000051, 0xffffffff, 0),
1588 	SYSC_QUIRK("dmic", 0, 0, 0x10, -ENODEV, 0x50010000, 0xffffffff, 0),
1589 	SYSC_QUIRK("dsi", 0x58004000, 0, 0x10, 0x14, 0x00000030, 0xffffffff, 0),
1590 	SYSC_QUIRK("dsi", 0x58005000, 0, 0x10, 0x14, 0x00000030, 0xffffffff, 0),
1591 	SYSC_QUIRK("dsi", 0x58005000, 0, 0x10, 0x14, 0x00000040, 0xffffffff, 0),
1592 	SYSC_QUIRK("dsi", 0x58009000, 0, 0x10, 0x14, 0x00000040, 0xffffffff, 0),
1593 	SYSC_QUIRK("dwc3", 0, 0, 0x10, -ENODEV, 0x500a0200, 0xffffffff, 0),
1594 	SYSC_QUIRK("d2d", 0x4a0b6000, 0, 0x10, 0x14, 0x00000010, 0xffffffff, 0),
1595 	SYSC_QUIRK("d2d", 0x4a0cd000, 0, 0x10, 0x14, 0x00000010, 0xffffffff, 0),
1596 	SYSC_QUIRK("epwmss", 0, 0, 0x4, -ENODEV, 0x47400001, 0xffffffff, 0),
1597 	SYSC_QUIRK("gpu", 0, 0x1fc00, 0x1fc10, -ENODEV, 0, 0, 0),
1598 	SYSC_QUIRK("gpu", 0, 0xfe00, 0xfe10, -ENODEV, 0x40000000 , 0xffffffff, 0),
1599 	SYSC_QUIRK("hdmi", 0, 0, 0x10, -ENODEV, 0x50031d00, 0xffffffff, 0),
1600 	SYSC_QUIRK("hsi", 0, 0, 0x10, 0x14, 0x50043101, 0xffffffff, 0),
1601 	SYSC_QUIRK("iss", 0, 0, 0x10, -ENODEV, 0x40000101, 0xffffffff, 0),
1602 	SYSC_QUIRK("mcasp", 0, 0, 0x4, -ENODEV, 0x44306302, 0xffffffff, 0),
1603 	SYSC_QUIRK("mcasp", 0, 0, 0x4, -ENODEV, 0x44307b02, 0xffffffff, 0),
1604 	SYSC_QUIRK("mcbsp", 0, -ENODEV, 0x8c, -ENODEV, 0, 0, 0),
1605 	SYSC_QUIRK("mcspi", 0, 0, 0x10, -ENODEV, 0x40300a0b, 0xffff00ff, 0),
1606 	SYSC_QUIRK("mcspi", 0, 0, 0x110, 0x114, 0x40300a0b, 0xffffffff, 0),
1607 	SYSC_QUIRK("mailbox", 0, 0, 0x10, -ENODEV, 0x00000400, 0xffffffff, 0),
1608 	SYSC_QUIRK("m3", 0, 0, -ENODEV, -ENODEV, 0x5f580105, 0x0fff0f00, 0),
1609 	SYSC_QUIRK("ocp2scp", 0, 0, 0x10, 0x14, 0x50060005, 0xfffffff0, 0),
1610 	SYSC_QUIRK("ocp2scp", 0, 0, -ENODEV, -ENODEV, 0x50060007, 0xffffffff, 0),
1611 	SYSC_QUIRK("padconf", 0, 0, 0x10, -ENODEV, 0x4fff0800, 0xffffffff, 0),
1612 	SYSC_QUIRK("padconf", 0, 0, -ENODEV, -ENODEV, 0x40001100, 0xffffffff, 0),
1613 	SYSC_QUIRK("prcm", 0, 0, -ENODEV, -ENODEV, 0x40000100, 0xffffffff, 0),
1614 	SYSC_QUIRK("prcm", 0, 0, -ENODEV, -ENODEV, 0x00004102, 0xffffffff, 0),
1615 	SYSC_QUIRK("prcm", 0, 0, -ENODEV, -ENODEV, 0x40000400, 0xffffffff, 0),
1616 	SYSC_QUIRK("rfbi", 0x4832a800, 0, 0x10, 0x14, 0x00000010, 0xffffffff, 0),
1617 	SYSC_QUIRK("rfbi", 0x58002000, 0, 0x10, 0x14, 0x00000010, 0xffffffff, 0),
1618 	SYSC_QUIRK("scm", 0, 0, 0x10, -ENODEV, 0x40000900, 0xffffffff, 0),
1619 	SYSC_QUIRK("scm", 0, 0, -ENODEV, -ENODEV, 0x4e8b0100, 0xffffffff, 0),
1620 	SYSC_QUIRK("scm", 0, 0, -ENODEV, -ENODEV, 0x4f000100, 0xffffffff, 0),
1621 	SYSC_QUIRK("scm", 0, 0, -ENODEV, -ENODEV, 0x40000900, 0xffffffff, 0),
1622 	SYSC_QUIRK("scrm", 0, 0, -ENODEV, -ENODEV, 0x00000010, 0xffffffff, 0),
1623 	SYSC_QUIRK("sdio", 0, 0, 0x10, -ENODEV, 0x40202301, 0xffff0ff0, 0),
1624 	SYSC_QUIRK("sdio", 0, 0x2fc, 0x110, 0x114, 0x31010000, 0xffffffff, 0),
1625 	SYSC_QUIRK("sdma", 0, 0, 0x2c, 0x28, 0x00010900, 0xffffffff, 0),
1626 	SYSC_QUIRK("slimbus", 0, 0, 0x10, -ENODEV, 0x40000902, 0xffffffff, 0),
1627 	SYSC_QUIRK("slimbus", 0, 0, 0x10, -ENODEV, 0x40002903, 0xffffffff, 0),
1628 	SYSC_QUIRK("spinlock", 0, 0, 0x10, -ENODEV, 0x50020000, 0xffffffff, 0),
1629 	SYSC_QUIRK("rng", 0, 0x1fe0, 0x1fe4, -ENODEV, 0x00000020, 0xffffffff, 0),
1630 	SYSC_QUIRK("timer", 0, 0, 0x10, 0x14, 0x00000013, 0xffffffff, 0),
1631 	SYSC_QUIRK("timer", 0, 0, 0x10, 0x14, 0x00000015, 0xffffffff, 0),
1632 	/* Some timers on omap4 and later */
1633 	SYSC_QUIRK("timer", 0, 0, 0x10, -ENODEV, 0x50002100, 0xffffffff, 0),
1634 	SYSC_QUIRK("timer", 0, 0, 0x10, -ENODEV, 0x4fff1301, 0xffff00ff, 0),
1635 	SYSC_QUIRK("timer32k", 0, 0, 0x4, -ENODEV, 0x00000040, 0xffffffff, 0),
1636 	SYSC_QUIRK("timer32k", 0, 0, 0x4, -ENODEV, 0x00000011, 0xffffffff, 0),
1637 	SYSC_QUIRK("timer32k", 0, 0, 0x4, -ENODEV, 0x00000060, 0xffffffff, 0),
1638 	SYSC_QUIRK("tpcc", 0, 0, -ENODEV, -ENODEV, 0x40014c00, 0xffffffff, 0),
1639 	SYSC_QUIRK("usbhstll", 0, 0, 0x10, 0x14, 0x00000004, 0xffffffff, 0),
1640 	SYSC_QUIRK("usbhstll", 0, 0, 0x10, 0x14, 0x00000008, 0xffffffff, 0),
1641 	SYSC_QUIRK("venc", 0x58003000, 0, -ENODEV, -ENODEV, 0x00000002, 0xffffffff, 0),
1642 	SYSC_QUIRK("vfpe", 0, 0, 0x104, -ENODEV, 0x4d001200, 0xffffffff, 0),
1643 #endif
1644 };
1645 
1646 /*
1647  * Early quirks based on module base and register offsets only that are
1648  * needed before the module revision can be read
1649  */
sysc_init_early_quirks(struct sysc * ddata)1650 static void sysc_init_early_quirks(struct sysc *ddata)
1651 {
1652 	const struct sysc_revision_quirk *q;
1653 	int i;
1654 
1655 	for (i = 0; i < ARRAY_SIZE(sysc_revision_quirks); i++) {
1656 		q = &sysc_revision_quirks[i];
1657 
1658 		if (!q->base)
1659 			continue;
1660 
1661 		if (q->base != ddata->module_pa)
1662 			continue;
1663 
1664 		if (q->rev_offset != ddata->offsets[SYSC_REVISION])
1665 			continue;
1666 
1667 		if (q->sysc_offset != ddata->offsets[SYSC_SYSCONFIG])
1668 			continue;
1669 
1670 		if (q->syss_offset != ddata->offsets[SYSC_SYSSTATUS])
1671 			continue;
1672 
1673 		ddata->name = q->name;
1674 		ddata->cfg.quirks |= q->quirks;
1675 	}
1676 }
1677 
1678 /* Quirks that also consider the revision register value */
sysc_init_revision_quirks(struct sysc * ddata)1679 static void sysc_init_revision_quirks(struct sysc *ddata)
1680 {
1681 	const struct sysc_revision_quirk *q;
1682 	int i;
1683 
1684 	for (i = 0; i < ARRAY_SIZE(sysc_revision_quirks); i++) {
1685 		q = &sysc_revision_quirks[i];
1686 
1687 		if (q->base && q->base != ddata->module_pa)
1688 			continue;
1689 
1690 		if (q->rev_offset != ddata->offsets[SYSC_REVISION])
1691 			continue;
1692 
1693 		if (q->sysc_offset != ddata->offsets[SYSC_SYSCONFIG])
1694 			continue;
1695 
1696 		if (q->syss_offset != ddata->offsets[SYSC_SYSSTATUS])
1697 			continue;
1698 
1699 		if (q->revision == ddata->revision ||
1700 		    (q->revision & q->revision_mask) ==
1701 		    (ddata->revision & q->revision_mask)) {
1702 			ddata->name = q->name;
1703 			ddata->cfg.quirks |= q->quirks;
1704 		}
1705 	}
1706 }
1707 
1708 /*
1709  * DSS needs dispc outputs disabled to reset modules. Returns mask of
1710  * enabled DSS interrupts. Eventually we may be able to do this on
1711  * dispc init rather than top-level DSS init.
1712  */
sysc_quirk_dispc(struct sysc * ddata,int dispc_offset,bool disable)1713 static u32 sysc_quirk_dispc(struct sysc *ddata, int dispc_offset,
1714 			    bool disable)
1715 {
1716 	bool lcd_en, digit_en, lcd2_en = false, lcd3_en = false;
1717 	const int lcd_en_mask = BIT(0), digit_en_mask = BIT(1);
1718 	int manager_count;
1719 	bool framedonetv_irq = true;
1720 	u32 val, irq_mask = 0;
1721 
1722 	switch (sysc_soc->soc) {
1723 	case SOC_2420 ... SOC_3630:
1724 		manager_count = 2;
1725 		framedonetv_irq = false;
1726 		break;
1727 	case SOC_4430 ... SOC_4470:
1728 		manager_count = 3;
1729 		break;
1730 	case SOC_5430:
1731 	case SOC_DRA7:
1732 		manager_count = 4;
1733 		break;
1734 	case SOC_AM4:
1735 		manager_count = 1;
1736 		framedonetv_irq = false;
1737 		break;
1738 	case SOC_UNKNOWN:
1739 	default:
1740 		return 0;
1741 	};
1742 
1743 	/* Remap the whole module range to be able to reset dispc outputs */
1744 	devm_iounmap(ddata->dev, ddata->module_va);
1745 	ddata->module_va = devm_ioremap(ddata->dev,
1746 					ddata->module_pa,
1747 					ddata->module_size);
1748 	if (!ddata->module_va)
1749 		return -EIO;
1750 
1751 	/* DISP_CONTROL */
1752 	val = sysc_read(ddata, dispc_offset + 0x40);
1753 	lcd_en = val & lcd_en_mask;
1754 	digit_en = val & digit_en_mask;
1755 	if (lcd_en)
1756 		irq_mask |= BIT(0);			/* FRAMEDONE */
1757 	if (digit_en) {
1758 		if (framedonetv_irq)
1759 			irq_mask |= BIT(24);		/* FRAMEDONETV */
1760 		else
1761 			irq_mask |= BIT(2) | BIT(3);	/* EVSYNC bits */
1762 	}
1763 	if (disable & (lcd_en | digit_en))
1764 		sysc_write(ddata, dispc_offset + 0x40,
1765 			   val & ~(lcd_en_mask | digit_en_mask));
1766 
1767 	if (manager_count <= 2)
1768 		return irq_mask;
1769 
1770 	/* DISPC_CONTROL2 */
1771 	val = sysc_read(ddata, dispc_offset + 0x238);
1772 	lcd2_en = val & lcd_en_mask;
1773 	if (lcd2_en)
1774 		irq_mask |= BIT(22);			/* FRAMEDONE2 */
1775 	if (disable && lcd2_en)
1776 		sysc_write(ddata, dispc_offset + 0x238,
1777 			   val & ~lcd_en_mask);
1778 
1779 	if (manager_count <= 3)
1780 		return irq_mask;
1781 
1782 	/* DISPC_CONTROL3 */
1783 	val = sysc_read(ddata, dispc_offset + 0x848);
1784 	lcd3_en = val & lcd_en_mask;
1785 	if (lcd3_en)
1786 		irq_mask |= BIT(30);			/* FRAMEDONE3 */
1787 	if (disable && lcd3_en)
1788 		sysc_write(ddata, dispc_offset + 0x848,
1789 			   val & ~lcd_en_mask);
1790 
1791 	return irq_mask;
1792 }
1793 
1794 /* DSS needs child outputs disabled and SDI registers cleared for reset */
sysc_pre_reset_quirk_dss(struct sysc * ddata)1795 static void sysc_pre_reset_quirk_dss(struct sysc *ddata)
1796 {
1797 	const int dispc_offset = 0x1000;
1798 	int error;
1799 	u32 irq_mask, val;
1800 
1801 	/* Get enabled outputs */
1802 	irq_mask = sysc_quirk_dispc(ddata, dispc_offset, false);
1803 	if (!irq_mask)
1804 		return;
1805 
1806 	/* Clear IRQSTATUS */
1807 	sysc_write(ddata, dispc_offset + 0x18, irq_mask);
1808 
1809 	/* Disable outputs */
1810 	val = sysc_quirk_dispc(ddata, dispc_offset, true);
1811 
1812 	/* Poll IRQSTATUS */
1813 	error = readl_poll_timeout(ddata->module_va + dispc_offset + 0x18,
1814 				   val, val != irq_mask, 100, 50);
1815 	if (error)
1816 		dev_warn(ddata->dev, "%s: timed out %08x !+ %08x\n",
1817 			 __func__, val, irq_mask);
1818 
1819 	if (sysc_soc->soc == SOC_3430) {
1820 		/* Clear DSS_SDI_CONTROL */
1821 		sysc_write(ddata, 0x44, 0);
1822 
1823 		/* Clear DSS_PLL_CONTROL */
1824 		sysc_write(ddata, 0x48, 0);
1825 	}
1826 
1827 	/* Clear DSS_CONTROL to switch DSS clock sources to PRCM if not */
1828 	sysc_write(ddata, 0x40, 0);
1829 }
1830 
1831 /* 1-wire needs module's internal clocks enabled for reset */
sysc_pre_reset_quirk_hdq1w(struct sysc * ddata)1832 static void sysc_pre_reset_quirk_hdq1w(struct sysc *ddata)
1833 {
1834 	int offset = 0x0c;	/* HDQ_CTRL_STATUS */
1835 	u16 val;
1836 
1837 	val = sysc_read(ddata, offset);
1838 	val |= BIT(5);
1839 	sysc_write(ddata, offset, val);
1840 }
1841 
1842 /* AESS (Audio Engine SubSystem) needs autogating set after enable */
sysc_module_enable_quirk_aess(struct sysc * ddata)1843 static void sysc_module_enable_quirk_aess(struct sysc *ddata)
1844 {
1845 	int offset = 0x7c;	/* AESS_AUTO_GATING_ENABLE */
1846 
1847 	sysc_write(ddata, offset, 1);
1848 }
1849 
1850 /* I2C needs to be disabled for reset */
sysc_clk_quirk_i2c(struct sysc * ddata,bool enable)1851 static void sysc_clk_quirk_i2c(struct sysc *ddata, bool enable)
1852 {
1853 	int offset;
1854 	u16 val;
1855 
1856 	/* I2C_CON, omap2/3 is different from omap4 and later */
1857 	if ((ddata->revision & 0xffffff00) == 0x001f0000)
1858 		offset = 0x24;
1859 	else
1860 		offset = 0xa4;
1861 
1862 	/* I2C_EN */
1863 	val = sysc_read(ddata, offset);
1864 	if (enable)
1865 		val |= BIT(15);
1866 	else
1867 		val &= ~BIT(15);
1868 	sysc_write(ddata, offset, val);
1869 }
1870 
sysc_pre_reset_quirk_i2c(struct sysc * ddata)1871 static void sysc_pre_reset_quirk_i2c(struct sysc *ddata)
1872 {
1873 	sysc_clk_quirk_i2c(ddata, false);
1874 }
1875 
sysc_post_reset_quirk_i2c(struct sysc * ddata)1876 static void sysc_post_reset_quirk_i2c(struct sysc *ddata)
1877 {
1878 	sysc_clk_quirk_i2c(ddata, true);
1879 }
1880 
1881 /* RTC on am3 and 4 needs to be unlocked and locked for sysconfig */
sysc_quirk_rtc(struct sysc * ddata,bool lock)1882 static void sysc_quirk_rtc(struct sysc *ddata, bool lock)
1883 {
1884 	u32 val, kick0_val = 0, kick1_val = 0;
1885 	unsigned long flags;
1886 	int error;
1887 
1888 	if (!lock) {
1889 		kick0_val = 0x83e70b13;
1890 		kick1_val = 0x95a4f1e0;
1891 	}
1892 
1893 	local_irq_save(flags);
1894 	/* RTC_STATUS BUSY bit may stay active for 1/32768 seconds (~30 usec) */
1895 	error = readl_poll_timeout_atomic(ddata->module_va + 0x44, val,
1896 					  !(val & BIT(0)), 100, 50);
1897 	if (error)
1898 		dev_warn(ddata->dev, "rtc busy timeout\n");
1899 	/* Now we have ~15 microseconds to read/write various registers */
1900 	sysc_write(ddata, 0x6c, kick0_val);
1901 	sysc_write(ddata, 0x70, kick1_val);
1902 	local_irq_restore(flags);
1903 }
1904 
sysc_module_unlock_quirk_rtc(struct sysc * ddata)1905 static void sysc_module_unlock_quirk_rtc(struct sysc *ddata)
1906 {
1907 	sysc_quirk_rtc(ddata, false);
1908 }
1909 
sysc_module_lock_quirk_rtc(struct sysc * ddata)1910 static void sysc_module_lock_quirk_rtc(struct sysc *ddata)
1911 {
1912 	sysc_quirk_rtc(ddata, true);
1913 }
1914 
1915 /* 36xx SGX needs a quirk for to bypass OCP IPG interrupt logic */
sysc_module_enable_quirk_sgx(struct sysc * ddata)1916 static void sysc_module_enable_quirk_sgx(struct sysc *ddata)
1917 {
1918 	int offset = 0xff08;	/* OCP_DEBUG_CONFIG */
1919 	u32 val = BIT(31);	/* THALIA_INT_BYPASS */
1920 
1921 	sysc_write(ddata, offset, val);
1922 }
1923 
1924 /* Watchdog timer needs a disable sequence after reset */
sysc_reset_done_quirk_wdt(struct sysc * ddata)1925 static void sysc_reset_done_quirk_wdt(struct sysc *ddata)
1926 {
1927 	int wps, spr, error;
1928 	u32 val;
1929 
1930 	wps = 0x34;
1931 	spr = 0x48;
1932 
1933 	sysc_write(ddata, spr, 0xaaaa);
1934 	error = readl_poll_timeout(ddata->module_va + wps, val,
1935 				   !(val & 0x10), 100,
1936 				   MAX_MODULE_SOFTRESET_WAIT);
1937 	if (error)
1938 		dev_warn(ddata->dev, "wdt disable step1 failed\n");
1939 
1940 	sysc_write(ddata, spr, 0x5555);
1941 	error = readl_poll_timeout(ddata->module_va + wps, val,
1942 				   !(val & 0x10), 100,
1943 				   MAX_MODULE_SOFTRESET_WAIT);
1944 	if (error)
1945 		dev_warn(ddata->dev, "wdt disable step2 failed\n");
1946 }
1947 
1948 /* PRUSS needs to set MSTANDBY_INIT inorder to idle properly */
sysc_module_disable_quirk_pruss(struct sysc * ddata)1949 static void sysc_module_disable_quirk_pruss(struct sysc *ddata)
1950 {
1951 	u32 reg;
1952 
1953 	reg = sysc_read(ddata, ddata->offsets[SYSC_SYSCONFIG]);
1954 	reg |= SYSC_PRUSS_STANDBY_INIT;
1955 	sysc_write(ddata, ddata->offsets[SYSC_SYSCONFIG], reg);
1956 }
1957 
sysc_init_module_quirks(struct sysc * ddata)1958 static void sysc_init_module_quirks(struct sysc *ddata)
1959 {
1960 	if (ddata->legacy_mode || !ddata->name)
1961 		return;
1962 
1963 	if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_HDQ1W) {
1964 		ddata->pre_reset_quirk = sysc_pre_reset_quirk_hdq1w;
1965 
1966 		return;
1967 	}
1968 
1969 #ifdef CONFIG_OMAP_GPMC_DEBUG
1970 	if (ddata->cfg.quirks & SYSC_QUIRK_GPMC_DEBUG) {
1971 		ddata->cfg.quirks |= SYSC_QUIRK_NO_RESET_ON_INIT;
1972 
1973 		return;
1974 	}
1975 #endif
1976 
1977 	if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_I2C) {
1978 		ddata->pre_reset_quirk = sysc_pre_reset_quirk_i2c;
1979 		ddata->post_reset_quirk = sysc_post_reset_quirk_i2c;
1980 
1981 		return;
1982 	}
1983 
1984 	if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_AESS)
1985 		ddata->module_enable_quirk = sysc_module_enable_quirk_aess;
1986 
1987 	if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_DSS_RESET)
1988 		ddata->pre_reset_quirk = sysc_pre_reset_quirk_dss;
1989 
1990 	if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_RTC_UNLOCK) {
1991 		ddata->module_unlock_quirk = sysc_module_unlock_quirk_rtc;
1992 		ddata->module_lock_quirk = sysc_module_lock_quirk_rtc;
1993 
1994 		return;
1995 	}
1996 
1997 	if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_SGX)
1998 		ddata->module_enable_quirk = sysc_module_enable_quirk_sgx;
1999 
2000 	if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_WDT) {
2001 		ddata->reset_done_quirk = sysc_reset_done_quirk_wdt;
2002 		ddata->module_disable_quirk = sysc_reset_done_quirk_wdt;
2003 	}
2004 
2005 	if (ddata->cfg.quirks & SYSC_MODULE_QUIRK_PRUSS)
2006 		ddata->module_disable_quirk = sysc_module_disable_quirk_pruss;
2007 }
2008 
sysc_clockdomain_init(struct sysc * ddata)2009 static int sysc_clockdomain_init(struct sysc *ddata)
2010 {
2011 	struct ti_sysc_platform_data *pdata = dev_get_platdata(ddata->dev);
2012 	struct clk *fck = NULL, *ick = NULL;
2013 	int error;
2014 
2015 	if (!pdata || !pdata->init_clockdomain)
2016 		return 0;
2017 
2018 	switch (ddata->nr_clocks) {
2019 	case 2:
2020 		ick = ddata->clocks[SYSC_ICK];
2021 		fallthrough;
2022 	case 1:
2023 		fck = ddata->clocks[SYSC_FCK];
2024 		break;
2025 	case 0:
2026 		return 0;
2027 	}
2028 
2029 	error = pdata->init_clockdomain(ddata->dev, fck, ick, &ddata->cookie);
2030 	if (!error || error == -ENODEV)
2031 		return 0;
2032 
2033 	return error;
2034 }
2035 
2036 /*
2037  * Note that pdata->init_module() typically does a reset first. After
2038  * pdata->init_module() is done, PM runtime can be used for the interconnect
2039  * target module.
2040  */
sysc_legacy_init(struct sysc * ddata)2041 static int sysc_legacy_init(struct sysc *ddata)
2042 {
2043 	struct ti_sysc_platform_data *pdata = dev_get_platdata(ddata->dev);
2044 	int error;
2045 
2046 	if (!pdata || !pdata->init_module)
2047 		return 0;
2048 
2049 	error = pdata->init_module(ddata->dev, ddata->mdata, &ddata->cookie);
2050 	if (error == -EEXIST)
2051 		error = 0;
2052 
2053 	return error;
2054 }
2055 
2056 /*
2057  * Note that the caller must ensure the interconnect target module is enabled
2058  * before calling reset. Otherwise reset will not complete.
2059  */
sysc_reset(struct sysc * ddata)2060 static int sysc_reset(struct sysc *ddata)
2061 {
2062 	int sysc_offset, sysc_val, error;
2063 	u32 sysc_mask;
2064 
2065 	sysc_offset = ddata->offsets[SYSC_SYSCONFIG];
2066 
2067 	if (ddata->legacy_mode ||
2068 	    ddata->cap->regbits->srst_shift < 0 ||
2069 	    ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT)
2070 		return 0;
2071 
2072 	sysc_mask = BIT(ddata->cap->regbits->srst_shift);
2073 
2074 	if (ddata->pre_reset_quirk)
2075 		ddata->pre_reset_quirk(ddata);
2076 
2077 	if (sysc_offset >= 0) {
2078 		sysc_val = sysc_read_sysconfig(ddata);
2079 		sysc_val |= sysc_mask;
2080 		sysc_write(ddata, sysc_offset, sysc_val);
2081 	}
2082 
2083 	if (ddata->cfg.srst_udelay)
2084 		usleep_range(ddata->cfg.srst_udelay,
2085 			     ddata->cfg.srst_udelay * 2);
2086 
2087 	if (ddata->post_reset_quirk)
2088 		ddata->post_reset_quirk(ddata);
2089 
2090 	error = sysc_wait_softreset(ddata);
2091 	if (error)
2092 		dev_warn(ddata->dev, "OCP softreset timed out\n");
2093 
2094 	if (ddata->reset_done_quirk)
2095 		ddata->reset_done_quirk(ddata);
2096 
2097 	return error;
2098 }
2099 
2100 /*
2101  * At this point the module is configured enough to read the revision but
2102  * module may not be completely configured yet to use PM runtime. Enable
2103  * all clocks directly during init to configure the quirks needed for PM
2104  * runtime based on the revision register.
2105  */
sysc_init_module(struct sysc * ddata)2106 static int sysc_init_module(struct sysc *ddata)
2107 {
2108 	int error = 0;
2109 
2110 	error = sysc_clockdomain_init(ddata);
2111 	if (error)
2112 		return error;
2113 
2114 	sysc_clkdm_deny_idle(ddata);
2115 
2116 	/*
2117 	 * Always enable clocks. The bootloader may or may not have enabled
2118 	 * the related clocks.
2119 	 */
2120 	error = sysc_enable_opt_clocks(ddata);
2121 	if (error)
2122 		return error;
2123 
2124 	error = sysc_enable_main_clocks(ddata);
2125 	if (error)
2126 		goto err_opt_clocks;
2127 
2128 	if (!(ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT)) {
2129 		error = reset_control_deassert(ddata->rsts);
2130 		if (error)
2131 			goto err_main_clocks;
2132 	}
2133 
2134 	ddata->revision = sysc_read_revision(ddata);
2135 	sysc_init_revision_quirks(ddata);
2136 	sysc_init_module_quirks(ddata);
2137 
2138 	if (ddata->legacy_mode) {
2139 		error = sysc_legacy_init(ddata);
2140 		if (error)
2141 			goto err_reset;
2142 	}
2143 
2144 	if (!ddata->legacy_mode) {
2145 		error = sysc_enable_module(ddata->dev);
2146 		if (error)
2147 			goto err_reset;
2148 	}
2149 
2150 	error = sysc_reset(ddata);
2151 	if (error)
2152 		dev_err(ddata->dev, "Reset failed with %d\n", error);
2153 
2154 	if (error && !ddata->legacy_mode)
2155 		sysc_disable_module(ddata->dev);
2156 
2157 err_reset:
2158 	if (error && !(ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT))
2159 		reset_control_assert(ddata->rsts);
2160 
2161 err_main_clocks:
2162 	if (error)
2163 		sysc_disable_main_clocks(ddata);
2164 err_opt_clocks:
2165 	/* No re-enable of clockdomain autoidle to prevent module autoidle */
2166 	if (error) {
2167 		sysc_disable_opt_clocks(ddata);
2168 		sysc_clkdm_allow_idle(ddata);
2169 	}
2170 
2171 	return error;
2172 }
2173 
sysc_init_sysc_mask(struct sysc * ddata)2174 static int sysc_init_sysc_mask(struct sysc *ddata)
2175 {
2176 	struct device_node *np = ddata->dev->of_node;
2177 	int error;
2178 	u32 val;
2179 
2180 	error = of_property_read_u32(np, "ti,sysc-mask", &val);
2181 	if (error)
2182 		return 0;
2183 
2184 	ddata->cfg.sysc_val = val & ddata->cap->sysc_mask;
2185 
2186 	return 0;
2187 }
2188 
sysc_init_idlemode(struct sysc * ddata,u8 * idlemodes,const char * name)2189 static int sysc_init_idlemode(struct sysc *ddata, u8 *idlemodes,
2190 			      const char *name)
2191 {
2192 	struct device_node *np = ddata->dev->of_node;
2193 	struct property *prop;
2194 	const __be32 *p;
2195 	u32 val;
2196 
2197 	of_property_for_each_u32(np, name, prop, p, val) {
2198 		if (val >= SYSC_NR_IDLEMODES) {
2199 			dev_err(ddata->dev, "invalid idlemode: %i\n", val);
2200 			return -EINVAL;
2201 		}
2202 		*idlemodes |=  (1 << val);
2203 	}
2204 
2205 	return 0;
2206 }
2207 
sysc_init_idlemodes(struct sysc * ddata)2208 static int sysc_init_idlemodes(struct sysc *ddata)
2209 {
2210 	int error;
2211 
2212 	error = sysc_init_idlemode(ddata, &ddata->cfg.midlemodes,
2213 				   "ti,sysc-midle");
2214 	if (error)
2215 		return error;
2216 
2217 	error = sysc_init_idlemode(ddata, &ddata->cfg.sidlemodes,
2218 				   "ti,sysc-sidle");
2219 	if (error)
2220 		return error;
2221 
2222 	return 0;
2223 }
2224 
2225 /*
2226  * Only some devices on omap4 and later have SYSCONFIG reset done
2227  * bit. We can detect this if there is no SYSSTATUS at all, or the
2228  * SYSTATUS bit 0 is not used. Note that some SYSSTATUS registers
2229  * have multiple bits for the child devices like OHCI and EHCI.
2230  * Depends on SYSC being parsed first.
2231  */
sysc_init_syss_mask(struct sysc * ddata)2232 static int sysc_init_syss_mask(struct sysc *ddata)
2233 {
2234 	struct device_node *np = ddata->dev->of_node;
2235 	int error;
2236 	u32 val;
2237 
2238 	error = of_property_read_u32(np, "ti,syss-mask", &val);
2239 	if (error) {
2240 		if ((ddata->cap->type == TI_SYSC_OMAP4 ||
2241 		     ddata->cap->type == TI_SYSC_OMAP4_TIMER) &&
2242 		    (ddata->cfg.sysc_val & SYSC_OMAP4_SOFTRESET))
2243 			ddata->cfg.quirks |= SYSC_QUIRK_RESET_STATUS;
2244 
2245 		return 0;
2246 	}
2247 
2248 	if (!(val & 1) && (ddata->cfg.sysc_val & SYSC_OMAP4_SOFTRESET))
2249 		ddata->cfg.quirks |= SYSC_QUIRK_RESET_STATUS;
2250 
2251 	ddata->cfg.syss_mask = val;
2252 
2253 	return 0;
2254 }
2255 
2256 /*
2257  * Many child device drivers need to have fck and opt clocks available
2258  * to get the clock rate for device internal configuration etc.
2259  */
sysc_child_add_named_clock(struct sysc * ddata,struct device * child,const char * name)2260 static int sysc_child_add_named_clock(struct sysc *ddata,
2261 				      struct device *child,
2262 				      const char *name)
2263 {
2264 	struct clk *clk;
2265 	struct clk_lookup *l;
2266 	int error = 0;
2267 
2268 	if (!name)
2269 		return 0;
2270 
2271 	clk = clk_get(child, name);
2272 	if (!IS_ERR(clk)) {
2273 		error = -EEXIST;
2274 		goto put_clk;
2275 	}
2276 
2277 	clk = clk_get(ddata->dev, name);
2278 	if (IS_ERR(clk))
2279 		return -ENODEV;
2280 
2281 	l = clkdev_create(clk, name, dev_name(child));
2282 	if (!l)
2283 		error = -ENOMEM;
2284 put_clk:
2285 	clk_put(clk);
2286 
2287 	return error;
2288 }
2289 
sysc_child_add_clocks(struct sysc * ddata,struct device * child)2290 static int sysc_child_add_clocks(struct sysc *ddata,
2291 				 struct device *child)
2292 {
2293 	int i, error;
2294 
2295 	for (i = 0; i < ddata->nr_clocks; i++) {
2296 		error = sysc_child_add_named_clock(ddata,
2297 						   child,
2298 						   ddata->clock_roles[i]);
2299 		if (error && error != -EEXIST) {
2300 			dev_err(ddata->dev, "could not add child clock %s: %i\n",
2301 				ddata->clock_roles[i], error);
2302 
2303 			return error;
2304 		}
2305 	}
2306 
2307 	return 0;
2308 }
2309 
2310 static struct device_type sysc_device_type = {
2311 };
2312 
sysc_child_to_parent(struct device * dev)2313 static struct sysc *sysc_child_to_parent(struct device *dev)
2314 {
2315 	struct device *parent = dev->parent;
2316 
2317 	if (!parent || parent->type != &sysc_device_type)
2318 		return NULL;
2319 
2320 	return dev_get_drvdata(parent);
2321 }
2322 
sysc_child_runtime_suspend(struct device * dev)2323 static int __maybe_unused sysc_child_runtime_suspend(struct device *dev)
2324 {
2325 	struct sysc *ddata;
2326 	int error;
2327 
2328 	ddata = sysc_child_to_parent(dev);
2329 
2330 	error = pm_generic_runtime_suspend(dev);
2331 	if (error)
2332 		return error;
2333 
2334 	if (!ddata->enabled)
2335 		return 0;
2336 
2337 	return sysc_runtime_suspend(ddata->dev);
2338 }
2339 
sysc_child_runtime_resume(struct device * dev)2340 static int __maybe_unused sysc_child_runtime_resume(struct device *dev)
2341 {
2342 	struct sysc *ddata;
2343 	int error;
2344 
2345 	ddata = sysc_child_to_parent(dev);
2346 
2347 	if (!ddata->enabled) {
2348 		error = sysc_runtime_resume(ddata->dev);
2349 		if (error < 0)
2350 			dev_err(ddata->dev,
2351 				"%s error: %i\n", __func__, error);
2352 	}
2353 
2354 	return pm_generic_runtime_resume(dev);
2355 }
2356 
2357 #ifdef CONFIG_PM_SLEEP
sysc_child_suspend_noirq(struct device * dev)2358 static int sysc_child_suspend_noirq(struct device *dev)
2359 {
2360 	struct sysc *ddata;
2361 	int error;
2362 
2363 	ddata = sysc_child_to_parent(dev);
2364 
2365 	dev_dbg(ddata->dev, "%s %s\n", __func__,
2366 		ddata->name ? ddata->name : "");
2367 
2368 	error = pm_generic_suspend_noirq(dev);
2369 	if (error) {
2370 		dev_err(dev, "%s error at %i: %i\n",
2371 			__func__, __LINE__, error);
2372 
2373 		return error;
2374 	}
2375 
2376 	if (!pm_runtime_status_suspended(dev)) {
2377 		error = pm_generic_runtime_suspend(dev);
2378 		if (error) {
2379 			dev_dbg(dev, "%s busy at %i: %i\n",
2380 				__func__, __LINE__, error);
2381 
2382 			return 0;
2383 		}
2384 
2385 		error = sysc_runtime_suspend(ddata->dev);
2386 		if (error) {
2387 			dev_err(dev, "%s error at %i: %i\n",
2388 				__func__, __LINE__, error);
2389 
2390 			return error;
2391 		}
2392 
2393 		ddata->child_needs_resume = true;
2394 	}
2395 
2396 	return 0;
2397 }
2398 
sysc_child_resume_noirq(struct device * dev)2399 static int sysc_child_resume_noirq(struct device *dev)
2400 {
2401 	struct sysc *ddata;
2402 	int error;
2403 
2404 	ddata = sysc_child_to_parent(dev);
2405 
2406 	dev_dbg(ddata->dev, "%s %s\n", __func__,
2407 		ddata->name ? ddata->name : "");
2408 
2409 	if (ddata->child_needs_resume) {
2410 		ddata->child_needs_resume = false;
2411 
2412 		error = sysc_runtime_resume(ddata->dev);
2413 		if (error)
2414 			dev_err(ddata->dev,
2415 				"%s runtime resume error: %i\n",
2416 				__func__, error);
2417 
2418 		error = pm_generic_runtime_resume(dev);
2419 		if (error)
2420 			dev_err(ddata->dev,
2421 				"%s generic runtime resume: %i\n",
2422 				__func__, error);
2423 	}
2424 
2425 	return pm_generic_resume_noirq(dev);
2426 }
2427 #endif
2428 
2429 static struct dev_pm_domain sysc_child_pm_domain = {
2430 	.ops = {
2431 		SET_RUNTIME_PM_OPS(sysc_child_runtime_suspend,
2432 				   sysc_child_runtime_resume,
2433 				   NULL)
2434 		USE_PLATFORM_PM_SLEEP_OPS
2435 		SET_NOIRQ_SYSTEM_SLEEP_PM_OPS(sysc_child_suspend_noirq,
2436 					      sysc_child_resume_noirq)
2437 	}
2438 };
2439 
2440 /* Caller needs to take list_lock if ever used outside of cpu_pm */
sysc_reinit_modules(struct sysc_soc_info * soc)2441 static void sysc_reinit_modules(struct sysc_soc_info *soc)
2442 {
2443 	struct sysc_module *module;
2444 	struct list_head *pos;
2445 	struct sysc *ddata;
2446 
2447 	list_for_each(pos, &sysc_soc->restored_modules) {
2448 		module = list_entry(pos, struct sysc_module, node);
2449 		ddata = module->ddata;
2450 		sysc_reinit_module(ddata, ddata->enabled);
2451 	}
2452 }
2453 
2454 /**
2455  * sysc_context_notifier - optionally reset and restore module after idle
2456  * @nb: notifier block
2457  * @cmd: unused
2458  * @v: unused
2459  *
2460  * Some interconnect target modules need to be restored, or reset and restored
2461  * on CPU_PM CPU_PM_CLUSTER_EXIT notifier. This is needed at least for am335x
2462  * OTG and GPMC target modules even if the modules are unused.
2463  */
sysc_context_notifier(struct notifier_block * nb,unsigned long cmd,void * v)2464 static int sysc_context_notifier(struct notifier_block *nb, unsigned long cmd,
2465 				 void *v)
2466 {
2467 	struct sysc_soc_info *soc;
2468 
2469 	soc = container_of(nb, struct sysc_soc_info, nb);
2470 
2471 	switch (cmd) {
2472 	case CPU_CLUSTER_PM_ENTER:
2473 		break;
2474 	case CPU_CLUSTER_PM_ENTER_FAILED:	/* No need to restore context */
2475 		break;
2476 	case CPU_CLUSTER_PM_EXIT:
2477 		sysc_reinit_modules(soc);
2478 		break;
2479 	}
2480 
2481 	return NOTIFY_OK;
2482 }
2483 
2484 /**
2485  * sysc_add_restored - optionally add reset and restore quirk hanlling
2486  * @ddata: device data
2487  */
sysc_add_restored(struct sysc * ddata)2488 static void sysc_add_restored(struct sysc *ddata)
2489 {
2490 	struct sysc_module *restored_module;
2491 
2492 	restored_module = kzalloc(sizeof(*restored_module), GFP_KERNEL);
2493 	if (!restored_module)
2494 		return;
2495 
2496 	restored_module->ddata = ddata;
2497 
2498 	mutex_lock(&sysc_soc->list_lock);
2499 
2500 	list_add(&restored_module->node, &sysc_soc->restored_modules);
2501 
2502 	if (sysc_soc->nb.notifier_call)
2503 		goto out_unlock;
2504 
2505 	sysc_soc->nb.notifier_call = sysc_context_notifier;
2506 	cpu_pm_register_notifier(&sysc_soc->nb);
2507 
2508 out_unlock:
2509 	mutex_unlock(&sysc_soc->list_lock);
2510 }
2511 
2512 /**
2513  * sysc_legacy_idle_quirk - handle children in omap_device compatible way
2514  * @ddata: device driver data
2515  * @child: child device driver
2516  *
2517  * Allow idle for child devices as done with _od_runtime_suspend().
2518  * Otherwise many child devices will not idle because of the permanent
2519  * parent usecount set in pm_runtime_irq_safe().
2520  *
2521  * Note that the long term solution is to just modify the child device
2522  * drivers to not set pm_runtime_irq_safe() and then this can be just
2523  * dropped.
2524  */
sysc_legacy_idle_quirk(struct sysc * ddata,struct device * child)2525 static void sysc_legacy_idle_quirk(struct sysc *ddata, struct device *child)
2526 {
2527 	if (ddata->cfg.quirks & SYSC_QUIRK_LEGACY_IDLE)
2528 		dev_pm_domain_set(child, &sysc_child_pm_domain);
2529 }
2530 
sysc_notifier_call(struct notifier_block * nb,unsigned long event,void * device)2531 static int sysc_notifier_call(struct notifier_block *nb,
2532 			      unsigned long event, void *device)
2533 {
2534 	struct device *dev = device;
2535 	struct sysc *ddata;
2536 	int error;
2537 
2538 	ddata = sysc_child_to_parent(dev);
2539 	if (!ddata)
2540 		return NOTIFY_DONE;
2541 
2542 	switch (event) {
2543 	case BUS_NOTIFY_ADD_DEVICE:
2544 		error = sysc_child_add_clocks(ddata, dev);
2545 		if (error)
2546 			return error;
2547 		sysc_legacy_idle_quirk(ddata, dev);
2548 		break;
2549 	default:
2550 		break;
2551 	}
2552 
2553 	return NOTIFY_DONE;
2554 }
2555 
2556 static struct notifier_block sysc_nb = {
2557 	.notifier_call = sysc_notifier_call,
2558 };
2559 
2560 /* Device tree configured quirks */
2561 struct sysc_dts_quirk {
2562 	const char *name;
2563 	u32 mask;
2564 };
2565 
2566 static const struct sysc_dts_quirk sysc_dts_quirks[] = {
2567 	{ .name = "ti,no-idle-on-init",
2568 	  .mask = SYSC_QUIRK_NO_IDLE_ON_INIT, },
2569 	{ .name = "ti,no-reset-on-init",
2570 	  .mask = SYSC_QUIRK_NO_RESET_ON_INIT, },
2571 	{ .name = "ti,no-idle",
2572 	  .mask = SYSC_QUIRK_NO_IDLE, },
2573 };
2574 
sysc_parse_dts_quirks(struct sysc * ddata,struct device_node * np,bool is_child)2575 static void sysc_parse_dts_quirks(struct sysc *ddata, struct device_node *np,
2576 				  bool is_child)
2577 {
2578 	const struct property *prop;
2579 	int i, len;
2580 
2581 	for (i = 0; i < ARRAY_SIZE(sysc_dts_quirks); i++) {
2582 		const char *name = sysc_dts_quirks[i].name;
2583 
2584 		prop = of_get_property(np, name, &len);
2585 		if (!prop)
2586 			continue;
2587 
2588 		ddata->cfg.quirks |= sysc_dts_quirks[i].mask;
2589 		if (is_child) {
2590 			dev_warn(ddata->dev,
2591 				 "dts flag should be at module level for %s\n",
2592 				 name);
2593 		}
2594 	}
2595 }
2596 
sysc_init_dts_quirks(struct sysc * ddata)2597 static int sysc_init_dts_quirks(struct sysc *ddata)
2598 {
2599 	struct device_node *np = ddata->dev->of_node;
2600 	int error;
2601 	u32 val;
2602 
2603 	ddata->legacy_mode = of_get_property(np, "ti,hwmods", NULL);
2604 
2605 	sysc_parse_dts_quirks(ddata, np, false);
2606 	error = of_property_read_u32(np, "ti,sysc-delay-us", &val);
2607 	if (!error) {
2608 		if (val > 255) {
2609 			dev_warn(ddata->dev, "bad ti,sysc-delay-us: %i\n",
2610 				 val);
2611 		}
2612 
2613 		ddata->cfg.srst_udelay = (u8)val;
2614 	}
2615 
2616 	return 0;
2617 }
2618 
sysc_unprepare(struct sysc * ddata)2619 static void sysc_unprepare(struct sysc *ddata)
2620 {
2621 	int i;
2622 
2623 	if (!ddata->clocks)
2624 		return;
2625 
2626 	for (i = 0; i < SYSC_MAX_CLOCKS; i++) {
2627 		if (!IS_ERR_OR_NULL(ddata->clocks[i]))
2628 			clk_unprepare(ddata->clocks[i]);
2629 	}
2630 }
2631 
2632 /*
2633  * Common sysc register bits found on omap2, also known as type1
2634  */
2635 static const struct sysc_regbits sysc_regbits_omap2 = {
2636 	.dmadisable_shift = -ENODEV,
2637 	.midle_shift = 12,
2638 	.sidle_shift = 3,
2639 	.clkact_shift = 8,
2640 	.emufree_shift = 5,
2641 	.enwkup_shift = 2,
2642 	.srst_shift = 1,
2643 	.autoidle_shift = 0,
2644 };
2645 
2646 static const struct sysc_capabilities sysc_omap2 = {
2647 	.type = TI_SYSC_OMAP2,
2648 	.sysc_mask = SYSC_OMAP2_CLOCKACTIVITY | SYSC_OMAP2_EMUFREE |
2649 		     SYSC_OMAP2_ENAWAKEUP | SYSC_OMAP2_SOFTRESET |
2650 		     SYSC_OMAP2_AUTOIDLE,
2651 	.regbits = &sysc_regbits_omap2,
2652 };
2653 
2654 /* All omap2 and 3 timers, and timers 1, 2 & 10 on omap 4 and 5 */
2655 static const struct sysc_capabilities sysc_omap2_timer = {
2656 	.type = TI_SYSC_OMAP2_TIMER,
2657 	.sysc_mask = SYSC_OMAP2_CLOCKACTIVITY | SYSC_OMAP2_EMUFREE |
2658 		     SYSC_OMAP2_ENAWAKEUP | SYSC_OMAP2_SOFTRESET |
2659 		     SYSC_OMAP2_AUTOIDLE,
2660 	.regbits = &sysc_regbits_omap2,
2661 	.mod_quirks = SYSC_QUIRK_USE_CLOCKACT,
2662 };
2663 
2664 /*
2665  * SHAM2 (SHA1/MD5) sysc found on omap3, a variant of sysc_regbits_omap2
2666  * with different sidle position
2667  */
2668 static const struct sysc_regbits sysc_regbits_omap3_sham = {
2669 	.dmadisable_shift = -ENODEV,
2670 	.midle_shift = -ENODEV,
2671 	.sidle_shift = 4,
2672 	.clkact_shift = -ENODEV,
2673 	.enwkup_shift = -ENODEV,
2674 	.srst_shift = 1,
2675 	.autoidle_shift = 0,
2676 	.emufree_shift = -ENODEV,
2677 };
2678 
2679 static const struct sysc_capabilities sysc_omap3_sham = {
2680 	.type = TI_SYSC_OMAP3_SHAM,
2681 	.sysc_mask = SYSC_OMAP2_SOFTRESET | SYSC_OMAP2_AUTOIDLE,
2682 	.regbits = &sysc_regbits_omap3_sham,
2683 };
2684 
2685 /*
2686  * AES register bits found on omap3 and later, a variant of
2687  * sysc_regbits_omap2 with different sidle position
2688  */
2689 static const struct sysc_regbits sysc_regbits_omap3_aes = {
2690 	.dmadisable_shift = -ENODEV,
2691 	.midle_shift = -ENODEV,
2692 	.sidle_shift = 6,
2693 	.clkact_shift = -ENODEV,
2694 	.enwkup_shift = -ENODEV,
2695 	.srst_shift = 1,
2696 	.autoidle_shift = 0,
2697 	.emufree_shift = -ENODEV,
2698 };
2699 
2700 static const struct sysc_capabilities sysc_omap3_aes = {
2701 	.type = TI_SYSC_OMAP3_AES,
2702 	.sysc_mask = SYSC_OMAP2_SOFTRESET | SYSC_OMAP2_AUTOIDLE,
2703 	.regbits = &sysc_regbits_omap3_aes,
2704 };
2705 
2706 /*
2707  * Common sysc register bits found on omap4, also known as type2
2708  */
2709 static const struct sysc_regbits sysc_regbits_omap4 = {
2710 	.dmadisable_shift = 16,
2711 	.midle_shift = 4,
2712 	.sidle_shift = 2,
2713 	.clkact_shift = -ENODEV,
2714 	.enwkup_shift = -ENODEV,
2715 	.emufree_shift = 1,
2716 	.srst_shift = 0,
2717 	.autoidle_shift = -ENODEV,
2718 };
2719 
2720 static const struct sysc_capabilities sysc_omap4 = {
2721 	.type = TI_SYSC_OMAP4,
2722 	.sysc_mask = SYSC_OMAP4_DMADISABLE | SYSC_OMAP4_FREEEMU |
2723 		     SYSC_OMAP4_SOFTRESET,
2724 	.regbits = &sysc_regbits_omap4,
2725 };
2726 
2727 static const struct sysc_capabilities sysc_omap4_timer = {
2728 	.type = TI_SYSC_OMAP4_TIMER,
2729 	.sysc_mask = SYSC_OMAP4_DMADISABLE | SYSC_OMAP4_FREEEMU |
2730 		     SYSC_OMAP4_SOFTRESET,
2731 	.regbits = &sysc_regbits_omap4,
2732 };
2733 
2734 /*
2735  * Common sysc register bits found on omap4, also known as type3
2736  */
2737 static const struct sysc_regbits sysc_regbits_omap4_simple = {
2738 	.dmadisable_shift = -ENODEV,
2739 	.midle_shift = 2,
2740 	.sidle_shift = 0,
2741 	.clkact_shift = -ENODEV,
2742 	.enwkup_shift = -ENODEV,
2743 	.srst_shift = -ENODEV,
2744 	.emufree_shift = -ENODEV,
2745 	.autoidle_shift = -ENODEV,
2746 };
2747 
2748 static const struct sysc_capabilities sysc_omap4_simple = {
2749 	.type = TI_SYSC_OMAP4_SIMPLE,
2750 	.regbits = &sysc_regbits_omap4_simple,
2751 };
2752 
2753 /*
2754  * SmartReflex sysc found on omap34xx
2755  */
2756 static const struct sysc_regbits sysc_regbits_omap34xx_sr = {
2757 	.dmadisable_shift = -ENODEV,
2758 	.midle_shift = -ENODEV,
2759 	.sidle_shift = -ENODEV,
2760 	.clkact_shift = 20,
2761 	.enwkup_shift = -ENODEV,
2762 	.srst_shift = -ENODEV,
2763 	.emufree_shift = -ENODEV,
2764 	.autoidle_shift = -ENODEV,
2765 };
2766 
2767 static const struct sysc_capabilities sysc_34xx_sr = {
2768 	.type = TI_SYSC_OMAP34XX_SR,
2769 	.sysc_mask = SYSC_OMAP2_CLOCKACTIVITY,
2770 	.regbits = &sysc_regbits_omap34xx_sr,
2771 	.mod_quirks = SYSC_QUIRK_USE_CLOCKACT | SYSC_QUIRK_UNCACHED |
2772 		      SYSC_QUIRK_LEGACY_IDLE,
2773 };
2774 
2775 /*
2776  * SmartReflex sysc found on omap36xx and later
2777  */
2778 static const struct sysc_regbits sysc_regbits_omap36xx_sr = {
2779 	.dmadisable_shift = -ENODEV,
2780 	.midle_shift = -ENODEV,
2781 	.sidle_shift = 24,
2782 	.clkact_shift = -ENODEV,
2783 	.enwkup_shift = 26,
2784 	.srst_shift = -ENODEV,
2785 	.emufree_shift = -ENODEV,
2786 	.autoidle_shift = -ENODEV,
2787 };
2788 
2789 static const struct sysc_capabilities sysc_36xx_sr = {
2790 	.type = TI_SYSC_OMAP36XX_SR,
2791 	.sysc_mask = SYSC_OMAP3_SR_ENAWAKEUP,
2792 	.regbits = &sysc_regbits_omap36xx_sr,
2793 	.mod_quirks = SYSC_QUIRK_UNCACHED | SYSC_QUIRK_LEGACY_IDLE,
2794 };
2795 
2796 static const struct sysc_capabilities sysc_omap4_sr = {
2797 	.type = TI_SYSC_OMAP4_SR,
2798 	.regbits = &sysc_regbits_omap36xx_sr,
2799 	.mod_quirks = SYSC_QUIRK_LEGACY_IDLE,
2800 };
2801 
2802 /*
2803  * McASP register bits found on omap4 and later
2804  */
2805 static const struct sysc_regbits sysc_regbits_omap4_mcasp = {
2806 	.dmadisable_shift = -ENODEV,
2807 	.midle_shift = -ENODEV,
2808 	.sidle_shift = 0,
2809 	.clkact_shift = -ENODEV,
2810 	.enwkup_shift = -ENODEV,
2811 	.srst_shift = -ENODEV,
2812 	.emufree_shift = -ENODEV,
2813 	.autoidle_shift = -ENODEV,
2814 };
2815 
2816 static const struct sysc_capabilities sysc_omap4_mcasp = {
2817 	.type = TI_SYSC_OMAP4_MCASP,
2818 	.regbits = &sysc_regbits_omap4_mcasp,
2819 	.mod_quirks = SYSC_QUIRK_OPT_CLKS_NEEDED,
2820 };
2821 
2822 /*
2823  * McASP found on dra7 and later
2824  */
2825 static const struct sysc_capabilities sysc_dra7_mcasp = {
2826 	.type = TI_SYSC_OMAP4_SIMPLE,
2827 	.regbits = &sysc_regbits_omap4_simple,
2828 	.mod_quirks = SYSC_QUIRK_OPT_CLKS_NEEDED,
2829 };
2830 
2831 /*
2832  * FS USB host found on omap4 and later
2833  */
2834 static const struct sysc_regbits sysc_regbits_omap4_usb_host_fs = {
2835 	.dmadisable_shift = -ENODEV,
2836 	.midle_shift = -ENODEV,
2837 	.sidle_shift = 24,
2838 	.clkact_shift = -ENODEV,
2839 	.enwkup_shift = 26,
2840 	.srst_shift = -ENODEV,
2841 	.emufree_shift = -ENODEV,
2842 	.autoidle_shift = -ENODEV,
2843 };
2844 
2845 static const struct sysc_capabilities sysc_omap4_usb_host_fs = {
2846 	.type = TI_SYSC_OMAP4_USB_HOST_FS,
2847 	.sysc_mask = SYSC_OMAP2_ENAWAKEUP,
2848 	.regbits = &sysc_regbits_omap4_usb_host_fs,
2849 };
2850 
2851 static const struct sysc_regbits sysc_regbits_dra7_mcan = {
2852 	.dmadisable_shift = -ENODEV,
2853 	.midle_shift = -ENODEV,
2854 	.sidle_shift = -ENODEV,
2855 	.clkact_shift = -ENODEV,
2856 	.enwkup_shift = 4,
2857 	.srst_shift = 0,
2858 	.emufree_shift = -ENODEV,
2859 	.autoidle_shift = -ENODEV,
2860 };
2861 
2862 static const struct sysc_capabilities sysc_dra7_mcan = {
2863 	.type = TI_SYSC_DRA7_MCAN,
2864 	.sysc_mask = SYSC_DRA7_MCAN_ENAWAKEUP | SYSC_OMAP4_SOFTRESET,
2865 	.regbits = &sysc_regbits_dra7_mcan,
2866 	.mod_quirks = SYSS_QUIRK_RESETDONE_INVERTED,
2867 };
2868 
2869 /*
2870  * PRUSS found on some AM33xx, AM437x and AM57xx SoCs
2871  */
2872 static const struct sysc_capabilities sysc_pruss = {
2873 	.type = TI_SYSC_PRUSS,
2874 	.sysc_mask = SYSC_PRUSS_STANDBY_INIT | SYSC_PRUSS_SUB_MWAIT,
2875 	.regbits = &sysc_regbits_omap4_simple,
2876 	.mod_quirks = SYSC_MODULE_QUIRK_PRUSS,
2877 };
2878 
sysc_init_pdata(struct sysc * ddata)2879 static int sysc_init_pdata(struct sysc *ddata)
2880 {
2881 	struct ti_sysc_platform_data *pdata = dev_get_platdata(ddata->dev);
2882 	struct ti_sysc_module_data *mdata;
2883 
2884 	if (!pdata)
2885 		return 0;
2886 
2887 	mdata = devm_kzalloc(ddata->dev, sizeof(*mdata), GFP_KERNEL);
2888 	if (!mdata)
2889 		return -ENOMEM;
2890 
2891 	if (ddata->legacy_mode) {
2892 		mdata->name = ddata->legacy_mode;
2893 		mdata->module_pa = ddata->module_pa;
2894 		mdata->module_size = ddata->module_size;
2895 		mdata->offsets = ddata->offsets;
2896 		mdata->nr_offsets = SYSC_MAX_REGS;
2897 		mdata->cap = ddata->cap;
2898 		mdata->cfg = &ddata->cfg;
2899 	}
2900 
2901 	ddata->mdata = mdata;
2902 
2903 	return 0;
2904 }
2905 
sysc_init_match(struct sysc * ddata)2906 static int sysc_init_match(struct sysc *ddata)
2907 {
2908 	const struct sysc_capabilities *cap;
2909 
2910 	cap = of_device_get_match_data(ddata->dev);
2911 	if (!cap)
2912 		return -EINVAL;
2913 
2914 	ddata->cap = cap;
2915 	if (ddata->cap)
2916 		ddata->cfg.quirks |= ddata->cap->mod_quirks;
2917 
2918 	return 0;
2919 }
2920 
ti_sysc_idle(struct work_struct * work)2921 static void ti_sysc_idle(struct work_struct *work)
2922 {
2923 	struct sysc *ddata;
2924 
2925 	ddata = container_of(work, struct sysc, idle_work.work);
2926 
2927 	/*
2928 	 * One time decrement of clock usage counts if left on from init.
2929 	 * Note that we disable opt clocks unconditionally in this case
2930 	 * as they are enabled unconditionally during init without
2931 	 * considering sysc_opt_clks_needed() at that point.
2932 	 */
2933 	if (ddata->cfg.quirks & (SYSC_QUIRK_NO_IDLE |
2934 				 SYSC_QUIRK_NO_IDLE_ON_INIT)) {
2935 		sysc_disable_main_clocks(ddata);
2936 		sysc_disable_opt_clocks(ddata);
2937 		sysc_clkdm_allow_idle(ddata);
2938 	}
2939 
2940 	/* Keep permanent PM runtime usage count for SYSC_QUIRK_NO_IDLE */
2941 	if (ddata->cfg.quirks & SYSC_QUIRK_NO_IDLE)
2942 		return;
2943 
2944 	/*
2945 	 * Decrement PM runtime usage count for SYSC_QUIRK_NO_IDLE_ON_INIT
2946 	 * and SYSC_QUIRK_NO_RESET_ON_INIT
2947 	 */
2948 	if (pm_runtime_active(ddata->dev))
2949 		pm_runtime_put_sync(ddata->dev);
2950 }
2951 
2952 /*
2953  * SoC model and features detection. Only needed for SoCs that need
2954  * special handling for quirks, no need to list others.
2955  */
2956 static const struct soc_device_attribute sysc_soc_match[] = {
2957 	SOC_FLAG("OMAP242*", SOC_2420),
2958 	SOC_FLAG("OMAP243*", SOC_2430),
2959 	SOC_FLAG("OMAP3[45]*", SOC_3430),
2960 	SOC_FLAG("OMAP3[67]*", SOC_3630),
2961 	SOC_FLAG("OMAP443*", SOC_4430),
2962 	SOC_FLAG("OMAP446*", SOC_4460),
2963 	SOC_FLAG("OMAP447*", SOC_4470),
2964 	SOC_FLAG("OMAP54*", SOC_5430),
2965 	SOC_FLAG("AM433", SOC_AM3),
2966 	SOC_FLAG("AM43*", SOC_AM4),
2967 	SOC_FLAG("DRA7*", SOC_DRA7),
2968 
2969 	{ /* sentinel */ },
2970 };
2971 
2972 /*
2973  * List of SoCs variants with disabled features. By default we assume all
2974  * devices in the device tree are available so no need to list those SoCs.
2975  */
2976 static const struct soc_device_attribute sysc_soc_feat_match[] = {
2977 	/* OMAP3430/3530 and AM3517 variants with some accelerators disabled */
2978 	SOC_FLAG("AM3505", DIS_SGX),
2979 	SOC_FLAG("OMAP3525", DIS_SGX),
2980 	SOC_FLAG("OMAP3515", DIS_IVA | DIS_SGX),
2981 	SOC_FLAG("OMAP3503", DIS_ISP | DIS_IVA | DIS_SGX),
2982 
2983 	/* OMAP3630/DM3730 variants with some accelerators disabled */
2984 	SOC_FLAG("AM3703", DIS_IVA | DIS_SGX),
2985 	SOC_FLAG("DM3725", DIS_SGX),
2986 	SOC_FLAG("OMAP3611", DIS_ISP | DIS_IVA | DIS_SGX),
2987 	SOC_FLAG("OMAP3615/AM3715", DIS_IVA),
2988 	SOC_FLAG("OMAP3621", DIS_ISP),
2989 
2990 	{ /* sentinel */ },
2991 };
2992 
sysc_add_disabled(unsigned long base)2993 static int sysc_add_disabled(unsigned long base)
2994 {
2995 	struct sysc_address *disabled_module;
2996 
2997 	disabled_module = kzalloc(sizeof(*disabled_module), GFP_KERNEL);
2998 	if (!disabled_module)
2999 		return -ENOMEM;
3000 
3001 	disabled_module->base = base;
3002 
3003 	mutex_lock(&sysc_soc->list_lock);
3004 	list_add(&disabled_module->node, &sysc_soc->disabled_modules);
3005 	mutex_unlock(&sysc_soc->list_lock);
3006 
3007 	return 0;
3008 }
3009 
3010 /*
3011  * One time init to detect the booted SoC, disable unavailable features
3012  * and initialize list for optional cpu_pm notifier.
3013  *
3014  * Note that we initialize static data shared across all ti-sysc instances
3015  * so ddata is only used for SoC type. This can be called from module_init
3016  * once we no longer need to rely on platform data.
3017  */
sysc_init_static_data(struct sysc * ddata)3018 static int sysc_init_static_data(struct sysc *ddata)
3019 {
3020 	const struct soc_device_attribute *match;
3021 	struct ti_sysc_platform_data *pdata;
3022 	unsigned long features = 0;
3023 
3024 	if (sysc_soc)
3025 		return 0;
3026 
3027 	sysc_soc = kzalloc(sizeof(*sysc_soc), GFP_KERNEL);
3028 	if (!sysc_soc)
3029 		return -ENOMEM;
3030 
3031 	mutex_init(&sysc_soc->list_lock);
3032 	INIT_LIST_HEAD(&sysc_soc->disabled_modules);
3033 	INIT_LIST_HEAD(&sysc_soc->restored_modules);
3034 	sysc_soc->general_purpose = true;
3035 
3036 	pdata = dev_get_platdata(ddata->dev);
3037 	if (pdata && pdata->soc_type_gp)
3038 		sysc_soc->general_purpose = pdata->soc_type_gp();
3039 
3040 	match = soc_device_match(sysc_soc_match);
3041 	if (match && match->data)
3042 		sysc_soc->soc = (int)match->data;
3043 
3044 	/* Ignore devices that are not available on HS and EMU SoCs */
3045 	if (!sysc_soc->general_purpose) {
3046 		switch (sysc_soc->soc) {
3047 		case SOC_3430 ... SOC_3630:
3048 			sysc_add_disabled(0x48304000);	/* timer12 */
3049 			break;
3050 		case SOC_AM3:
3051 			sysc_add_disabled(0x48310000);  /* rng */
3052 			break;
3053 		default:
3054 			break;
3055 		};
3056 	}
3057 
3058 	match = soc_device_match(sysc_soc_feat_match);
3059 	if (!match)
3060 		return 0;
3061 
3062 	if (match->data)
3063 		features = (unsigned long)match->data;
3064 
3065 	/*
3066 	 * Add disabled devices to the list based on the module base.
3067 	 * Note that this must be done before we attempt to access the
3068 	 * device and have module revision checks working.
3069 	 */
3070 	if (features & DIS_ISP)
3071 		sysc_add_disabled(0x480bd400);
3072 	if (features & DIS_IVA)
3073 		sysc_add_disabled(0x5d000000);
3074 	if (features & DIS_SGX)
3075 		sysc_add_disabled(0x50000000);
3076 
3077 	return 0;
3078 }
3079 
sysc_cleanup_static_data(void)3080 static void sysc_cleanup_static_data(void)
3081 {
3082 	struct sysc_module *restored_module;
3083 	struct sysc_address *disabled_module;
3084 	struct list_head *pos, *tmp;
3085 
3086 	if (!sysc_soc)
3087 		return;
3088 
3089 	if (sysc_soc->nb.notifier_call)
3090 		cpu_pm_unregister_notifier(&sysc_soc->nb);
3091 
3092 	mutex_lock(&sysc_soc->list_lock);
3093 	list_for_each_safe(pos, tmp, &sysc_soc->restored_modules) {
3094 		restored_module = list_entry(pos, struct sysc_module, node);
3095 		list_del(pos);
3096 		kfree(restored_module);
3097 	}
3098 	list_for_each_safe(pos, tmp, &sysc_soc->disabled_modules) {
3099 		disabled_module = list_entry(pos, struct sysc_address, node);
3100 		list_del(pos);
3101 		kfree(disabled_module);
3102 	}
3103 	mutex_unlock(&sysc_soc->list_lock);
3104 }
3105 
sysc_check_disabled_devices(struct sysc * ddata)3106 static int sysc_check_disabled_devices(struct sysc *ddata)
3107 {
3108 	struct sysc_address *disabled_module;
3109 	struct list_head *pos;
3110 	int error = 0;
3111 
3112 	mutex_lock(&sysc_soc->list_lock);
3113 	list_for_each(pos, &sysc_soc->disabled_modules) {
3114 		disabled_module = list_entry(pos, struct sysc_address, node);
3115 		if (ddata->module_pa == disabled_module->base) {
3116 			dev_dbg(ddata->dev, "module disabled for this SoC\n");
3117 			error = -ENODEV;
3118 			break;
3119 		}
3120 	}
3121 	mutex_unlock(&sysc_soc->list_lock);
3122 
3123 	return error;
3124 }
3125 
3126 /*
3127  * Ignore timers tagged with no-reset and no-idle. These are likely in use,
3128  * for example by drivers/clocksource/timer-ti-dm-systimer.c. If more checks
3129  * are needed, we could also look at the timer register configuration.
3130  */
sysc_check_active_timer(struct sysc * ddata)3131 static int sysc_check_active_timer(struct sysc *ddata)
3132 {
3133 	if (ddata->cap->type != TI_SYSC_OMAP2_TIMER &&
3134 	    ddata->cap->type != TI_SYSC_OMAP4_TIMER)
3135 		return 0;
3136 
3137 	if ((ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT) &&
3138 	    (ddata->cfg.quirks & SYSC_QUIRK_NO_IDLE))
3139 		return -ENXIO;
3140 
3141 	return 0;
3142 }
3143 
3144 static const struct of_device_id sysc_match_table[] = {
3145 	{ .compatible = "simple-bus", },
3146 	{ /* sentinel */ },
3147 };
3148 
sysc_probe(struct platform_device * pdev)3149 static int sysc_probe(struct platform_device *pdev)
3150 {
3151 	struct ti_sysc_platform_data *pdata = dev_get_platdata(&pdev->dev);
3152 	struct sysc *ddata;
3153 	int error;
3154 
3155 	ddata = devm_kzalloc(&pdev->dev, sizeof(*ddata), GFP_KERNEL);
3156 	if (!ddata)
3157 		return -ENOMEM;
3158 
3159 	ddata->dev = &pdev->dev;
3160 	platform_set_drvdata(pdev, ddata);
3161 
3162 	error = sysc_init_static_data(ddata);
3163 	if (error)
3164 		return error;
3165 
3166 	error = sysc_init_match(ddata);
3167 	if (error)
3168 		return error;
3169 
3170 	error = sysc_init_dts_quirks(ddata);
3171 	if (error)
3172 		return error;
3173 
3174 	error = sysc_map_and_check_registers(ddata);
3175 	if (error)
3176 		return error;
3177 
3178 	error = sysc_init_sysc_mask(ddata);
3179 	if (error)
3180 		return error;
3181 
3182 	error = sysc_init_idlemodes(ddata);
3183 	if (error)
3184 		return error;
3185 
3186 	error = sysc_init_syss_mask(ddata);
3187 	if (error)
3188 		return error;
3189 
3190 	error = sysc_init_pdata(ddata);
3191 	if (error)
3192 		return error;
3193 
3194 	sysc_init_early_quirks(ddata);
3195 
3196 	error = sysc_check_disabled_devices(ddata);
3197 	if (error)
3198 		return error;
3199 
3200 	error = sysc_check_active_timer(ddata);
3201 	if (error == -ENXIO)
3202 		ddata->reserved = true;
3203 	else if (error)
3204 		return error;
3205 
3206 	error = sysc_get_clocks(ddata);
3207 	if (error)
3208 		return error;
3209 
3210 	error = sysc_init_resets(ddata);
3211 	if (error)
3212 		goto unprepare;
3213 
3214 	error = sysc_init_module(ddata);
3215 	if (error)
3216 		goto unprepare;
3217 
3218 	pm_runtime_enable(ddata->dev);
3219 	error = pm_runtime_get_sync(ddata->dev);
3220 	if (error < 0) {
3221 		pm_runtime_put_noidle(ddata->dev);
3222 		pm_runtime_disable(ddata->dev);
3223 		goto unprepare;
3224 	}
3225 
3226 	/* Balance use counts as PM runtime should have enabled these all */
3227 	if (!(ddata->cfg.quirks & SYSC_QUIRK_NO_RESET_ON_INIT))
3228 		reset_control_assert(ddata->rsts);
3229 
3230 	if (!(ddata->cfg.quirks &
3231 	      (SYSC_QUIRK_NO_IDLE | SYSC_QUIRK_NO_IDLE_ON_INIT))) {
3232 		sysc_disable_main_clocks(ddata);
3233 		sysc_disable_opt_clocks(ddata);
3234 		sysc_clkdm_allow_idle(ddata);
3235 	}
3236 
3237 	sysc_show_registers(ddata);
3238 
3239 	ddata->dev->type = &sysc_device_type;
3240 
3241 	if (!ddata->reserved) {
3242 		error = of_platform_populate(ddata->dev->of_node,
3243 					     sysc_match_table,
3244 					     pdata ? pdata->auxdata : NULL,
3245 					     ddata->dev);
3246 		if (error)
3247 			goto err;
3248 	}
3249 
3250 	INIT_DELAYED_WORK(&ddata->idle_work, ti_sysc_idle);
3251 
3252 	/* At least earlycon won't survive without deferred idle */
3253 	if (ddata->cfg.quirks & (SYSC_QUIRK_NO_IDLE |
3254 				 SYSC_QUIRK_NO_IDLE_ON_INIT |
3255 				 SYSC_QUIRK_NO_RESET_ON_INIT)) {
3256 		schedule_delayed_work(&ddata->idle_work, 3000);
3257 	} else {
3258 		pm_runtime_put(&pdev->dev);
3259 	}
3260 
3261 	if (ddata->cfg.quirks & SYSC_QUIRK_REINIT_ON_CTX_LOST)
3262 		sysc_add_restored(ddata);
3263 
3264 	return 0;
3265 
3266 err:
3267 	pm_runtime_put_sync(&pdev->dev);
3268 	pm_runtime_disable(&pdev->dev);
3269 unprepare:
3270 	sysc_unprepare(ddata);
3271 
3272 	return error;
3273 }
3274 
sysc_remove(struct platform_device * pdev)3275 static int sysc_remove(struct platform_device *pdev)
3276 {
3277 	struct sysc *ddata = platform_get_drvdata(pdev);
3278 	int error;
3279 
3280 	cancel_delayed_work_sync(&ddata->idle_work);
3281 
3282 	error = pm_runtime_get_sync(ddata->dev);
3283 	if (error < 0) {
3284 		pm_runtime_put_noidle(ddata->dev);
3285 		pm_runtime_disable(ddata->dev);
3286 		goto unprepare;
3287 	}
3288 
3289 	of_platform_depopulate(&pdev->dev);
3290 
3291 	pm_runtime_put_sync(&pdev->dev);
3292 	pm_runtime_disable(&pdev->dev);
3293 
3294 	if (!reset_control_status(ddata->rsts))
3295 		reset_control_assert(ddata->rsts);
3296 
3297 unprepare:
3298 	sysc_unprepare(ddata);
3299 
3300 	return 0;
3301 }
3302 
3303 static const struct of_device_id sysc_match[] = {
3304 	{ .compatible = "ti,sysc-omap2", .data = &sysc_omap2, },
3305 	{ .compatible = "ti,sysc-omap2-timer", .data = &sysc_omap2_timer, },
3306 	{ .compatible = "ti,sysc-omap4", .data = &sysc_omap4, },
3307 	{ .compatible = "ti,sysc-omap4-timer", .data = &sysc_omap4_timer, },
3308 	{ .compatible = "ti,sysc-omap4-simple", .data = &sysc_omap4_simple, },
3309 	{ .compatible = "ti,sysc-omap3430-sr", .data = &sysc_34xx_sr, },
3310 	{ .compatible = "ti,sysc-omap3630-sr", .data = &sysc_36xx_sr, },
3311 	{ .compatible = "ti,sysc-omap4-sr", .data = &sysc_omap4_sr, },
3312 	{ .compatible = "ti,sysc-omap3-sham", .data = &sysc_omap3_sham, },
3313 	{ .compatible = "ti,sysc-omap-aes", .data = &sysc_omap3_aes, },
3314 	{ .compatible = "ti,sysc-mcasp", .data = &sysc_omap4_mcasp, },
3315 	{ .compatible = "ti,sysc-dra7-mcasp", .data = &sysc_dra7_mcasp, },
3316 	{ .compatible = "ti,sysc-usb-host-fs",
3317 	  .data = &sysc_omap4_usb_host_fs, },
3318 	{ .compatible = "ti,sysc-dra7-mcan", .data = &sysc_dra7_mcan, },
3319 	{ .compatible = "ti,sysc-pruss", .data = &sysc_pruss, },
3320 	{  },
3321 };
3322 MODULE_DEVICE_TABLE(of, sysc_match);
3323 
3324 static struct platform_driver sysc_driver = {
3325 	.probe		= sysc_probe,
3326 	.remove		= sysc_remove,
3327 	.driver         = {
3328 		.name   = "ti-sysc",
3329 		.of_match_table	= sysc_match,
3330 		.pm = &sysc_pm_ops,
3331 	},
3332 };
3333 
sysc_init(void)3334 static int __init sysc_init(void)
3335 {
3336 	bus_register_notifier(&platform_bus_type, &sysc_nb);
3337 
3338 	return platform_driver_register(&sysc_driver);
3339 }
3340 module_init(sysc_init);
3341 
sysc_exit(void)3342 static void __exit sysc_exit(void)
3343 {
3344 	bus_unregister_notifier(&platform_bus_type, &sysc_nb);
3345 	platform_driver_unregister(&sysc_driver);
3346 	sysc_cleanup_static_data();
3347 }
3348 module_exit(sysc_exit);
3349 
3350 MODULE_DESCRIPTION("TI sysc interconnect target driver");
3351 MODULE_LICENSE("GPL v2");
3352