Loading...
1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * Copyright (C) 2017 SiFive
4 */
5
6#include <linux/cacheinfo.h>
7#include <linux/cpu.h>
8#include <linux/of.h>
9#include <linux/of_device.h>
10#include <asm/cacheinfo.h>
11
12static struct riscv_cacheinfo_ops *rv_cache_ops;
13
14void riscv_set_cacheinfo_ops(struct riscv_cacheinfo_ops *ops)
15{
16 rv_cache_ops = ops;
17}
18EXPORT_SYMBOL_GPL(riscv_set_cacheinfo_ops);
19
20const struct attribute_group *
21cache_get_priv_group(struct cacheinfo *this_leaf)
22{
23 if (rv_cache_ops && rv_cache_ops->get_priv_group)
24 return rv_cache_ops->get_priv_group(this_leaf);
25 return NULL;
26}
27
28static void ci_leaf_init(struct cacheinfo *this_leaf,
29 struct device_node *node,
30 enum cache_type type, unsigned int level)
31{
32 this_leaf->level = level;
33 this_leaf->type = type;
34}
35
36static int __init_cache_level(unsigned int cpu)
37{
38 struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu);
39 struct device_node *np = of_cpu_device_node_get(cpu);
40 struct device_node *prev = NULL;
41 int levels = 0, leaves = 0, level;
42
43 if (of_property_read_bool(np, "cache-size"))
44 ++leaves;
45 if (of_property_read_bool(np, "i-cache-size"))
46 ++leaves;
47 if (of_property_read_bool(np, "d-cache-size"))
48 ++leaves;
49 if (leaves > 0)
50 levels = 1;
51
52 prev = np;
53 while ((np = of_find_next_cache_node(np))) {
54 of_node_put(prev);
55 prev = np;
56 if (!of_device_is_compatible(np, "cache"))
57 break;
58 if (of_property_read_u32(np, "cache-level", &level))
59 break;
60 if (level <= levels)
61 break;
62 if (of_property_read_bool(np, "cache-size"))
63 ++leaves;
64 if (of_property_read_bool(np, "i-cache-size"))
65 ++leaves;
66 if (of_property_read_bool(np, "d-cache-size"))
67 ++leaves;
68 levels = level;
69 }
70
71 of_node_put(np);
72 this_cpu_ci->num_levels = levels;
73 this_cpu_ci->num_leaves = leaves;
74
75 return 0;
76}
77
78static int __populate_cache_leaves(unsigned int cpu)
79{
80 struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu);
81 struct cacheinfo *this_leaf = this_cpu_ci->info_list;
82 struct device_node *np = of_cpu_device_node_get(cpu);
83 struct device_node *prev = NULL;
84 int levels = 1, level = 1;
85
86 if (of_property_read_bool(np, "cache-size"))
87 ci_leaf_init(this_leaf++, np, CACHE_TYPE_UNIFIED, level);
88 if (of_property_read_bool(np, "i-cache-size"))
89 ci_leaf_init(this_leaf++, np, CACHE_TYPE_INST, level);
90 if (of_property_read_bool(np, "d-cache-size"))
91 ci_leaf_init(this_leaf++, np, CACHE_TYPE_DATA, level);
92
93 prev = np;
94 while ((np = of_find_next_cache_node(np))) {
95 of_node_put(prev);
96 prev = np;
97 if (!of_device_is_compatible(np, "cache"))
98 break;
99 if (of_property_read_u32(np, "cache-level", &level))
100 break;
101 if (level <= levels)
102 break;
103 if (of_property_read_bool(np, "cache-size"))
104 ci_leaf_init(this_leaf++, np, CACHE_TYPE_UNIFIED, level);
105 if (of_property_read_bool(np, "i-cache-size"))
106 ci_leaf_init(this_leaf++, np, CACHE_TYPE_INST, level);
107 if (of_property_read_bool(np, "d-cache-size"))
108 ci_leaf_init(this_leaf++, np, CACHE_TYPE_DATA, level);
109 levels = level;
110 }
111 of_node_put(np);
112
113 return 0;
114}
115
116DEFINE_SMP_CALL_CACHE_FUNCTION(init_cache_level)
117DEFINE_SMP_CALL_CACHE_FUNCTION(populate_cache_leaves)
1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * Copyright (C) 2017 SiFive
4 */
5
6#include <linux/cpu.h>
7#include <linux/of.h>
8#include <linux/of_device.h>
9#include <asm/cacheinfo.h>
10
11static struct riscv_cacheinfo_ops *rv_cache_ops;
12
13void riscv_set_cacheinfo_ops(struct riscv_cacheinfo_ops *ops)
14{
15 rv_cache_ops = ops;
16}
17EXPORT_SYMBOL_GPL(riscv_set_cacheinfo_ops);
18
19const struct attribute_group *
20cache_get_priv_group(struct cacheinfo *this_leaf)
21{
22 if (rv_cache_ops && rv_cache_ops->get_priv_group)
23 return rv_cache_ops->get_priv_group(this_leaf);
24 return NULL;
25}
26
27static struct cacheinfo *get_cacheinfo(u32 level, enum cache_type type)
28{
29 /*
30 * Using raw_smp_processor_id() elides a preemptability check, but this
31 * is really indicative of a larger problem: the cacheinfo UABI assumes
32 * that cores have a homonogenous view of the cache hierarchy. That
33 * happens to be the case for the current set of RISC-V systems, but
34 * likely won't be true in general. Since there's no way to provide
35 * correct information for these systems via the current UABI we're
36 * just eliding the check for now.
37 */
38 struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(raw_smp_processor_id());
39 struct cacheinfo *this_leaf;
40 int index;
41
42 for (index = 0; index < this_cpu_ci->num_leaves; index++) {
43 this_leaf = this_cpu_ci->info_list + index;
44 if (this_leaf->level == level && this_leaf->type == type)
45 return this_leaf;
46 }
47
48 return NULL;
49}
50
51uintptr_t get_cache_size(u32 level, enum cache_type type)
52{
53 struct cacheinfo *this_leaf = get_cacheinfo(level, type);
54
55 return this_leaf ? this_leaf->size : 0;
56}
57
58uintptr_t get_cache_geometry(u32 level, enum cache_type type)
59{
60 struct cacheinfo *this_leaf = get_cacheinfo(level, type);
61
62 return this_leaf ? (this_leaf->ways_of_associativity << 16 |
63 this_leaf->coherency_line_size) :
64 0;
65}
66
67static void ci_leaf_init(struct cacheinfo *this_leaf, enum cache_type type,
68 unsigned int level, unsigned int size,
69 unsigned int sets, unsigned int line_size)
70{
71 this_leaf->level = level;
72 this_leaf->type = type;
73 this_leaf->size = size;
74 this_leaf->number_of_sets = sets;
75 this_leaf->coherency_line_size = line_size;
76
77 /*
78 * If the cache is fully associative, there is no need to
79 * check the other properties.
80 */
81 if (sets == 1)
82 return;
83
84 /*
85 * Set the ways number for n-ways associative, make sure
86 * all properties are big than zero.
87 */
88 if (sets > 0 && size > 0 && line_size > 0)
89 this_leaf->ways_of_associativity = (size / sets) / line_size;
90}
91
92static void fill_cacheinfo(struct cacheinfo **this_leaf,
93 struct device_node *node, unsigned int level)
94{
95 unsigned int size, sets, line_size;
96
97 if (!of_property_read_u32(node, "cache-size", &size) &&
98 !of_property_read_u32(node, "cache-block-size", &line_size) &&
99 !of_property_read_u32(node, "cache-sets", &sets)) {
100 ci_leaf_init((*this_leaf)++, CACHE_TYPE_UNIFIED, level, size, sets, line_size);
101 }
102
103 if (!of_property_read_u32(node, "i-cache-size", &size) &&
104 !of_property_read_u32(node, "i-cache-sets", &sets) &&
105 !of_property_read_u32(node, "i-cache-block-size", &line_size)) {
106 ci_leaf_init((*this_leaf)++, CACHE_TYPE_INST, level, size, sets, line_size);
107 }
108
109 if (!of_property_read_u32(node, "d-cache-size", &size) &&
110 !of_property_read_u32(node, "d-cache-sets", &sets) &&
111 !of_property_read_u32(node, "d-cache-block-size", &line_size)) {
112 ci_leaf_init((*this_leaf)++, CACHE_TYPE_DATA, level, size, sets, line_size);
113 }
114}
115
116int init_cache_level(unsigned int cpu)
117{
118 struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu);
119 struct device_node *np = of_cpu_device_node_get(cpu);
120 struct device_node *prev = NULL;
121 int levels = 0, leaves = 0, level;
122
123 if (of_property_read_bool(np, "cache-size"))
124 ++leaves;
125 if (of_property_read_bool(np, "i-cache-size"))
126 ++leaves;
127 if (of_property_read_bool(np, "d-cache-size"))
128 ++leaves;
129 if (leaves > 0)
130 levels = 1;
131
132 prev = np;
133 while ((np = of_find_next_cache_node(np))) {
134 of_node_put(prev);
135 prev = np;
136 if (!of_device_is_compatible(np, "cache"))
137 break;
138 if (of_property_read_u32(np, "cache-level", &level))
139 break;
140 if (level <= levels)
141 break;
142 if (of_property_read_bool(np, "cache-size"))
143 ++leaves;
144 if (of_property_read_bool(np, "i-cache-size"))
145 ++leaves;
146 if (of_property_read_bool(np, "d-cache-size"))
147 ++leaves;
148 levels = level;
149 }
150
151 of_node_put(np);
152 this_cpu_ci->num_levels = levels;
153 this_cpu_ci->num_leaves = leaves;
154
155 return 0;
156}
157
158int populate_cache_leaves(unsigned int cpu)
159{
160 struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu);
161 struct cacheinfo *this_leaf = this_cpu_ci->info_list;
162 struct device_node *np = of_cpu_device_node_get(cpu);
163 struct device_node *prev = NULL;
164 int levels = 1, level = 1;
165
166 /* Level 1 caches in cpu node */
167 fill_cacheinfo(&this_leaf, np, level);
168
169 /* Next level caches in cache nodes */
170 prev = np;
171 while ((np = of_find_next_cache_node(np))) {
172 of_node_put(prev);
173 prev = np;
174
175 if (!of_device_is_compatible(np, "cache"))
176 break;
177 if (of_property_read_u32(np, "cache-level", &level))
178 break;
179 if (level <= levels)
180 break;
181
182 fill_cacheinfo(&this_leaf, np, level);
183
184 levels = level;
185 }
186 of_node_put(np);
187
188 return 0;
189}