Linux Audio

Check our new training course

Loading...
Note: File does not exist in v3.15.
  1/*
  2 * Marvell Dove PMU support
  3 */
  4#include <linux/io.h>
  5#include <linux/irq.h>
  6#include <linux/irqdomain.h>
  7#include <linux/of.h>
  8#include <linux/of_irq.h>
  9#include <linux/of_address.h>
 10#include <linux/platform_device.h>
 11#include <linux/pm_domain.h>
 12#include <linux/reset.h>
 13#include <linux/reset-controller.h>
 14#include <linux/sched.h>
 15#include <linux/slab.h>
 16#include <linux/soc/dove/pmu.h>
 17#include <linux/spinlock.h>
 18
 19#define NR_PMU_IRQS		7
 20
 21#define PMC_SW_RST		0x30
 22#define PMC_IRQ_CAUSE		0x50
 23#define PMC_IRQ_MASK		0x54
 24
 25#define PMU_PWR			0x10
 26#define PMU_ISO			0x58
 27
 28struct pmu_data {
 29	spinlock_t lock;
 30	struct device_node *of_node;
 31	void __iomem *pmc_base;
 32	void __iomem *pmu_base;
 33	struct irq_chip_generic *irq_gc;
 34	struct irq_domain *irq_domain;
 35#ifdef CONFIG_RESET_CONTROLLER
 36	struct reset_controller_dev reset;
 37#endif
 38};
 39
 40/*
 41 * The PMU contains a register to reset various subsystems within the
 42 * SoC.  Export this as a reset controller.
 43 */
 44#ifdef CONFIG_RESET_CONTROLLER
 45#define rcdev_to_pmu(rcdev) container_of(rcdev, struct pmu_data, reset)
 46
 47static int pmu_reset_reset(struct reset_controller_dev *rc, unsigned long id)
 48{
 49	struct pmu_data *pmu = rcdev_to_pmu(rc);
 50	unsigned long flags;
 51	u32 val;
 52
 53	spin_lock_irqsave(&pmu->lock, flags);
 54	val = readl_relaxed(pmu->pmc_base + PMC_SW_RST);
 55	writel_relaxed(val & ~BIT(id), pmu->pmc_base + PMC_SW_RST);
 56	writel_relaxed(val | BIT(id), pmu->pmc_base + PMC_SW_RST);
 57	spin_unlock_irqrestore(&pmu->lock, flags);
 58
 59	return 0;
 60}
 61
 62static int pmu_reset_assert(struct reset_controller_dev *rc, unsigned long id)
 63{
 64	struct pmu_data *pmu = rcdev_to_pmu(rc);
 65	unsigned long flags;
 66	u32 val = ~BIT(id);
 67
 68	spin_lock_irqsave(&pmu->lock, flags);
 69	val &= readl_relaxed(pmu->pmc_base + PMC_SW_RST);
 70	writel_relaxed(val, pmu->pmc_base + PMC_SW_RST);
 71	spin_unlock_irqrestore(&pmu->lock, flags);
 72
 73	return 0;
 74}
 75
 76static int pmu_reset_deassert(struct reset_controller_dev *rc, unsigned long id)
 77{
 78	struct pmu_data *pmu = rcdev_to_pmu(rc);
 79	unsigned long flags;
 80	u32 val = BIT(id);
 81
 82	spin_lock_irqsave(&pmu->lock, flags);
 83	val |= readl_relaxed(pmu->pmc_base + PMC_SW_RST);
 84	writel_relaxed(val, pmu->pmc_base + PMC_SW_RST);
 85	spin_unlock_irqrestore(&pmu->lock, flags);
 86
 87	return 0;
 88}
 89
 90static struct reset_control_ops pmu_reset_ops = {
 91	.reset = pmu_reset_reset,
 92	.assert = pmu_reset_assert,
 93	.deassert = pmu_reset_deassert,
 94};
 95
 96static struct reset_controller_dev pmu_reset __initdata = {
 97	.ops = &pmu_reset_ops,
 98	.owner = THIS_MODULE,
 99	.nr_resets = 32,
100};
101
102static void __init pmu_reset_init(struct pmu_data *pmu)
103{
104	int ret;
105
106	pmu->reset = pmu_reset;
107	pmu->reset.of_node = pmu->of_node;
108
109	ret = reset_controller_register(&pmu->reset);
110	if (ret)
111		pr_err("pmu: %s failed: %d\n", "reset_controller_register", ret);
112}
113#else
114static void __init pmu_reset_init(struct pmu_data *pmu)
115{
116}
117#endif
118
119struct pmu_domain {
120	struct pmu_data *pmu;
121	u32 pwr_mask;
122	u32 rst_mask;
123	u32 iso_mask;
124	struct generic_pm_domain base;
125};
126
127#define to_pmu_domain(dom) container_of(dom, struct pmu_domain, base)
128
129/*
130 * This deals with the "old" Marvell sequence of bringing a power domain
131 * down/up, which is: apply power, release reset, disable isolators.
132 *
133 * Later devices apparantly use a different sequence: power up, disable
134 * isolators, assert repair signal, enable SRMA clock, enable AXI clock,
135 * enable module clock, deassert reset.
136 *
137 * Note: reading the assembly, it seems that the IO accessors have an
138 * unfortunate side-effect - they cause memory already read into registers
139 * for the if () to be re-read for the bit-set or bit-clear operation.
140 * The code is written to avoid this.
141 */
142static int pmu_domain_power_off(struct generic_pm_domain *domain)
143{
144	struct pmu_domain *pmu_dom = to_pmu_domain(domain);
145	struct pmu_data *pmu = pmu_dom->pmu;
146	unsigned long flags;
147	unsigned int val;
148	void __iomem *pmu_base = pmu->pmu_base;
149	void __iomem *pmc_base = pmu->pmc_base;
150
151	spin_lock_irqsave(&pmu->lock, flags);
152
153	/* Enable isolators */
154	if (pmu_dom->iso_mask) {
155		val = ~pmu_dom->iso_mask;
156		val &= readl_relaxed(pmu_base + PMU_ISO);
157		writel_relaxed(val, pmu_base + PMU_ISO);
158	}
159
160	/* Reset unit */
161	if (pmu_dom->rst_mask) {
162		val = ~pmu_dom->rst_mask;
163		val &= readl_relaxed(pmc_base + PMC_SW_RST);
164		writel_relaxed(val, pmc_base + PMC_SW_RST);
165	}
166
167	/* Power down */
168	val = readl_relaxed(pmu_base + PMU_PWR) | pmu_dom->pwr_mask;
169	writel_relaxed(val, pmu_base + PMU_PWR);
170
171	spin_unlock_irqrestore(&pmu->lock, flags);
172
173	return 0;
174}
175
176static int pmu_domain_power_on(struct generic_pm_domain *domain)
177{
178	struct pmu_domain *pmu_dom = to_pmu_domain(domain);
179	struct pmu_data *pmu = pmu_dom->pmu;
180	unsigned long flags;
181	unsigned int val;
182	void __iomem *pmu_base = pmu->pmu_base;
183	void __iomem *pmc_base = pmu->pmc_base;
184
185	spin_lock_irqsave(&pmu->lock, flags);
186
187	/* Power on */
188	val = ~pmu_dom->pwr_mask & readl_relaxed(pmu_base + PMU_PWR);
189	writel_relaxed(val, pmu_base + PMU_PWR);
190
191	/* Release reset */
192	if (pmu_dom->rst_mask) {
193		val = pmu_dom->rst_mask;
194		val |= readl_relaxed(pmc_base + PMC_SW_RST);
195		writel_relaxed(val, pmc_base + PMC_SW_RST);
196	}
197
198	/* Disable isolators */
199	if (pmu_dom->iso_mask) {
200		val = pmu_dom->iso_mask;
201		val |= readl_relaxed(pmu_base + PMU_ISO);
202		writel_relaxed(val, pmu_base + PMU_ISO);
203	}
204
205	spin_unlock_irqrestore(&pmu->lock, flags);
206
207	return 0;
208}
209
210static void __pmu_domain_register(struct pmu_domain *domain,
211	struct device_node *np)
212{
213	unsigned int val = readl_relaxed(domain->pmu->pmu_base + PMU_PWR);
214
215	domain->base.power_off = pmu_domain_power_off;
216	domain->base.power_on = pmu_domain_power_on;
217
218	pm_genpd_init(&domain->base, NULL, !(val & domain->pwr_mask));
219
220	if (np)
221		of_genpd_add_provider_simple(np, &domain->base);
222}
223
224/* PMU IRQ controller */
225static void pmu_irq_handler(struct irq_desc *desc)
226{
227	struct pmu_data *pmu = irq_desc_get_handler_data(desc);
228	struct irq_chip_generic *gc = pmu->irq_gc;
229	struct irq_domain *domain = pmu->irq_domain;
230	void __iomem *base = gc->reg_base;
231	u32 stat = readl_relaxed(base + PMC_IRQ_CAUSE) & gc->mask_cache;
232	u32 done = ~0;
233
234	if (stat == 0) {
235		handle_bad_irq(desc);
236		return;
237	}
238
239	while (stat) {
240		u32 hwirq = fls(stat) - 1;
241
242		stat &= ~(1 << hwirq);
243		done &= ~(1 << hwirq);
244
245		generic_handle_irq(irq_find_mapping(domain, hwirq));
246	}
247
248	/*
249	 * The PMU mask register is not RW0C: it is RW.  This means that
250	 * the bits take whatever value is written to them; if you write
251	 * a '1', you will set the interrupt.
252	 *
253	 * Unfortunately this means there is NO race free way to clear
254	 * these interrupts.
255	 *
256	 * So, let's structure the code so that the window is as small as
257	 * possible.
258	 */
259	irq_gc_lock(gc);
260	done &= readl_relaxed(base + PMC_IRQ_CAUSE);
261	writel_relaxed(done, base + PMC_IRQ_CAUSE);
262	irq_gc_unlock(gc);
263}
264
265static int __init dove_init_pmu_irq(struct pmu_data *pmu, int irq)
266{
267	const char *name = "pmu_irq";
268	struct irq_chip_generic *gc;
269	struct irq_domain *domain;
270	int ret;
271
272	/* mask and clear all interrupts */
273	writel(0, pmu->pmc_base + PMC_IRQ_MASK);
274	writel(0, pmu->pmc_base + PMC_IRQ_CAUSE);
275
276	domain = irq_domain_add_linear(pmu->of_node, NR_PMU_IRQS,
277				       &irq_generic_chip_ops, NULL);
278	if (!domain) {
279		pr_err("%s: unable to add irq domain\n", name);
280		return -ENOMEM;
281	}
282
283	ret = irq_alloc_domain_generic_chips(domain, NR_PMU_IRQS, 1, name,
284					     handle_level_irq,
285					     IRQ_NOREQUEST | IRQ_NOPROBE, 0,
286					     IRQ_GC_INIT_MASK_CACHE);
287	if (ret) {
288		pr_err("%s: unable to alloc irq domain gc: %d\n", name, ret);
289		irq_domain_remove(domain);
290		return ret;
291	}
292
293	gc = irq_get_domain_generic_chip(domain, 0);
294	gc->reg_base = pmu->pmc_base;
295	gc->chip_types[0].regs.mask = PMC_IRQ_MASK;
296	gc->chip_types[0].chip.irq_mask = irq_gc_mask_clr_bit;
297	gc->chip_types[0].chip.irq_unmask = irq_gc_mask_set_bit;
298
299	pmu->irq_domain = domain;
300	pmu->irq_gc = gc;
301
302	irq_set_handler_data(irq, pmu);
303	irq_set_chained_handler(irq, pmu_irq_handler);
304
305	return 0;
306}
307
308int __init dove_init_pmu_legacy(const struct dove_pmu_initdata *initdata)
309{
310	const struct dove_pmu_domain_initdata *domain_initdata;
311	struct pmu_data *pmu;
312	int ret;
313
314	pmu = kzalloc(sizeof(*pmu), GFP_KERNEL);
315	if (!pmu)
316		return -ENOMEM;
317
318	spin_lock_init(&pmu->lock);
319	pmu->pmc_base = initdata->pmc_base;
320	pmu->pmu_base = initdata->pmu_base;
321
322	pmu_reset_init(pmu);
323	for (domain_initdata = initdata->domains; domain_initdata->name;
324	     domain_initdata++) {
325		struct pmu_domain *domain;
326
327		domain = kzalloc(sizeof(*domain), GFP_KERNEL);
328		if (domain) {
329			domain->pmu = pmu;
330			domain->pwr_mask = domain_initdata->pwr_mask;
331			domain->rst_mask = domain_initdata->rst_mask;
332			domain->iso_mask = domain_initdata->iso_mask;
333			domain->base.name = domain_initdata->name;
334
335			__pmu_domain_register(domain, NULL);
336		}
337	}
338
339	ret = dove_init_pmu_irq(pmu, initdata->irq);
340	if (ret)
341		pr_err("dove_init_pmu_irq() failed: %d\n", ret);
342
343	if (pmu->irq_domain)
344		irq_domain_associate_many(pmu->irq_domain,
345					  initdata->irq_domain_start,
346					  0, NR_PMU_IRQS);
347
348	return 0;
349}
350
351/*
352 * pmu: power-manager@d0000 {
353 *	compatible = "marvell,dove-pmu";
354 *	reg = <0xd0000 0x8000> <0xd8000 0x8000>;
355 *	interrupts = <33>;
356 *	interrupt-controller;
357 *	#reset-cells = 1;
358 *	vpu_domain: vpu-domain {
359 *		#power-domain-cells = <0>;
360 *		marvell,pmu_pwr_mask = <0x00000008>;
361 *		marvell,pmu_iso_mask = <0x00000001>;
362 *		resets = <&pmu 16>;
363 *	};
364 *	gpu_domain: gpu-domain {
365 *		#power-domain-cells = <0>;
366 *		marvell,pmu_pwr_mask = <0x00000004>;
367 *		marvell,pmu_iso_mask = <0x00000002>;
368 *		resets = <&pmu 18>;
369 *	};
370 * };
371 */
372int __init dove_init_pmu(void)
373{
374	struct device_node *np_pmu, *domains_node, *np;
375	struct pmu_data *pmu;
376	int ret, parent_irq;
377
378	/* Lookup the PMU node */
379	np_pmu = of_find_compatible_node(NULL, NULL, "marvell,dove-pmu");
380	if (!np_pmu)
381		return 0;
382
383	domains_node = of_get_child_by_name(np_pmu, "domains");
384	if (!domains_node) {
385		pr_err("%s: failed to find domains sub-node\n", np_pmu->name);
386		return 0;
387	}
388
389	pmu = kzalloc(sizeof(*pmu), GFP_KERNEL);
390	if (!pmu)
391		return -ENOMEM;
392
393	spin_lock_init(&pmu->lock);
394	pmu->of_node = np_pmu;
395	pmu->pmc_base = of_iomap(pmu->of_node, 0);
396	pmu->pmu_base = of_iomap(pmu->of_node, 1);
397	if (!pmu->pmc_base || !pmu->pmu_base) {
398		pr_err("%s: failed to map PMU\n", np_pmu->name);
399		iounmap(pmu->pmu_base);
400		iounmap(pmu->pmc_base);
401		kfree(pmu);
402		return -ENOMEM;
403	}
404
405	pmu_reset_init(pmu);
406
407	for_each_available_child_of_node(domains_node, np) {
408		struct of_phandle_args args;
409		struct pmu_domain *domain;
410
411		domain = kzalloc(sizeof(*domain), GFP_KERNEL);
412		if (!domain)
413			break;
414
415		domain->pmu = pmu;
416		domain->base.name = kstrdup(np->name, GFP_KERNEL);
417		if (!domain->base.name) {
418			kfree(domain);
419			break;
420		}
421
422		of_property_read_u32(np, "marvell,pmu_pwr_mask",
423				     &domain->pwr_mask);
424		of_property_read_u32(np, "marvell,pmu_iso_mask",
425				     &domain->iso_mask);
426
427		/*
428		 * We parse the reset controller property directly here
429		 * to ensure that we can operate when the reset controller
430		 * support is not configured into the kernel.
431		 */
432		ret = of_parse_phandle_with_args(np, "resets", "#reset-cells",
433						 0, &args);
434		if (ret == 0) {
435			if (args.np == pmu->of_node)
436				domain->rst_mask = BIT(args.args[0]);
437			of_node_put(args.np);
438		}
439
440		__pmu_domain_register(domain, np);
441	}
442
443	/* Loss of the interrupt controller is not a fatal error. */
444	parent_irq = irq_of_parse_and_map(pmu->of_node, 0);
445	if (!parent_irq) {
446		pr_err("%s: no interrupt specified\n", np_pmu->name);
447	} else {
448		ret = dove_init_pmu_irq(pmu, parent_irq);
449		if (ret)
450			pr_err("dove_init_pmu_irq() failed: %d\n", ret);
451	}
452
453	return 0;
454}