Loading...
1/*
2 * Copyright (C) 2013 Imagination Technologies
3 * Author: Paul Burton <paul.burton@imgtec.com>
4 *
5 * This program is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License as published by the
7 * Free Software Foundation; either version 2 of the License, or (at your
8 * option) any later version.
9 */
10
11#include <linux/errno.h>
12#include <linux/percpu.h>
13#include <linux/spinlock.h>
14
15#include <asm/mips-cm.h>
16#include <asm/mips-cpc.h>
17
18void __iomem *mips_cpc_base;
19
20static DEFINE_PER_CPU_ALIGNED(spinlock_t, cpc_core_lock);
21
22static DEFINE_PER_CPU_ALIGNED(unsigned long, cpc_core_lock_flags);
23
24/**
25 * mips_cpc_phys_base - retrieve the physical base address of the CPC
26 *
27 * This function returns the physical base address of the Cluster Power
28 * Controller memory mapped registers, or 0 if no Cluster Power Controller
29 * is present.
30 */
31static phys_addr_t mips_cpc_phys_base(void)
32{
33 unsigned long cpc_base;
34
35 if (!mips_cm_present())
36 return 0;
37
38 if (!(read_gcr_cpc_status() & CM_GCR_CPC_STATUS_EX_MSK))
39 return 0;
40
41 /* If the CPC is already enabled, leave it so */
42 cpc_base = read_gcr_cpc_base();
43 if (cpc_base & CM_GCR_CPC_BASE_CPCEN_MSK)
44 return cpc_base & CM_GCR_CPC_BASE_CPCBASE_MSK;
45
46 /* Otherwise, give it the default address & enable it */
47 cpc_base = mips_cpc_default_phys_base();
48 write_gcr_cpc_base(cpc_base | CM_GCR_CPC_BASE_CPCEN_MSK);
49 return cpc_base;
50}
51
52int mips_cpc_probe(void)
53{
54 phys_addr_t addr;
55 unsigned cpu;
56
57 for_each_possible_cpu(cpu)
58 spin_lock_init(&per_cpu(cpc_core_lock, cpu));
59
60 addr = mips_cpc_phys_base();
61 if (!addr)
62 return -ENODEV;
63
64 mips_cpc_base = ioremap_nocache(addr, 0x8000);
65 if (!mips_cpc_base)
66 return -ENXIO;
67
68 return 0;
69}
70
71void mips_cpc_lock_other(unsigned int core)
72{
73 unsigned curr_core;
74 preempt_disable();
75 curr_core = current_cpu_data.core;
76 spin_lock_irqsave(&per_cpu(cpc_core_lock, curr_core),
77 per_cpu(cpc_core_lock_flags, curr_core));
78 write_cpc_cl_other(core << CPC_Cx_OTHER_CORENUM_SHF);
79
80 /*
81 * Ensure the core-other region reflects the appropriate core &
82 * VP before any accesses to it occur.
83 */
84 mb();
85}
86
87void mips_cpc_unlock_other(void)
88{
89 unsigned curr_core = current_cpu_data.core;
90 spin_unlock_irqrestore(&per_cpu(cpc_core_lock, curr_core),
91 per_cpu(cpc_core_lock_flags, curr_core));
92 preempt_enable();
93}
1// SPDX-License-Identifier: GPL-2.0-or-later
2/*
3 * Copyright (C) 2013 Imagination Technologies
4 * Author: Paul Burton <paul.burton@mips.com>
5 */
6
7#include <linux/bitfield.h>
8#include <linux/errno.h>
9#include <linux/percpu.h>
10#include <linux/of.h>
11#include <linux/of_address.h>
12#include <linux/spinlock.h>
13
14#include <asm/mips-cps.h>
15
16void __iomem *mips_cpc_base;
17
18static DEFINE_PER_CPU_ALIGNED(spinlock_t, cpc_core_lock);
19
20static DEFINE_PER_CPU_ALIGNED(unsigned long, cpc_core_lock_flags);
21
22phys_addr_t __weak mips_cpc_default_phys_base(void)
23{
24 struct device_node *cpc_node;
25 struct resource res;
26 int err;
27
28 cpc_node = of_find_compatible_node(of_root, NULL, "mti,mips-cpc");
29 if (cpc_node) {
30 err = of_address_to_resource(cpc_node, 0, &res);
31 of_node_put(cpc_node);
32 if (!err)
33 return res.start;
34 }
35
36 return 0;
37}
38
39/**
40 * mips_cpc_phys_base - retrieve the physical base address of the CPC
41 *
42 * This function returns the physical base address of the Cluster Power
43 * Controller memory mapped registers, or 0 if no Cluster Power Controller
44 * is present.
45 */
46static phys_addr_t mips_cpc_phys_base(void)
47{
48 unsigned long cpc_base;
49
50 if (!mips_cm_present())
51 return 0;
52
53 if (!(read_gcr_cpc_status() & CM_GCR_CPC_STATUS_EX))
54 return 0;
55
56 /* If the CPC is already enabled, leave it so */
57 cpc_base = read_gcr_cpc_base();
58 if (cpc_base & CM_GCR_CPC_BASE_CPCEN)
59 return cpc_base & CM_GCR_CPC_BASE_CPCBASE;
60
61 /* Otherwise, use the default address */
62 cpc_base = mips_cpc_default_phys_base();
63 if (!cpc_base)
64 return cpc_base;
65
66 /* Enable the CPC, mapped at the default address */
67 write_gcr_cpc_base(cpc_base | CM_GCR_CPC_BASE_CPCEN);
68 return cpc_base;
69}
70
71int mips_cpc_probe(void)
72{
73 phys_addr_t addr;
74 unsigned int cpu;
75
76 for_each_possible_cpu(cpu)
77 spin_lock_init(&per_cpu(cpc_core_lock, cpu));
78
79 addr = mips_cpc_phys_base();
80 if (!addr)
81 return -ENODEV;
82
83 mips_cpc_base = ioremap(addr, 0x8000);
84 if (!mips_cpc_base)
85 return -ENXIO;
86
87 return 0;
88}
89
90void mips_cpc_lock_other(unsigned int core)
91{
92 unsigned int curr_core;
93
94 if (mips_cm_revision() >= CM_REV_CM3)
95 /* Systems with CM >= 3 lock the CPC via mips_cm_lock_other */
96 return;
97
98 preempt_disable();
99 curr_core = cpu_core(¤t_cpu_data);
100 spin_lock_irqsave(&per_cpu(cpc_core_lock, curr_core),
101 per_cpu(cpc_core_lock_flags, curr_core));
102 write_cpc_cl_other(FIELD_PREP(CPC_Cx_OTHER_CORENUM, core));
103
104 /*
105 * Ensure the core-other region reflects the appropriate core &
106 * VP before any accesses to it occur.
107 */
108 mb();
109}
110
111void mips_cpc_unlock_other(void)
112{
113 unsigned int curr_core;
114
115 if (mips_cm_revision() >= CM_REV_CM3)
116 /* Systems with CM >= 3 lock the CPC via mips_cm_lock_other */
117 return;
118
119 curr_core = cpu_core(¤t_cpu_data);
120 spin_unlock_irqrestore(&per_cpu(cpc_core_lock, curr_core),
121 per_cpu(cpc_core_lock_flags, curr_core));
122 preempt_enable();
123}