Loading...
1// SPDX-License-Identifier: GPL-2.0
2/*
3 * System Control and Power Interface (SCMI) Protocol based clock driver
4 *
5 * Copyright (C) 2018-2024 ARM Ltd.
6 */
7
8#include <linux/bits.h>
9#include <linux/clk-provider.h>
10#include <linux/device.h>
11#include <linux/err.h>
12#include <linux/of.h>
13#include <linux/module.h>
14#include <linux/scmi_protocol.h>
15#include <asm/div64.h>
16
17#define NOT_ATOMIC false
18#define ATOMIC true
19
20enum scmi_clk_feats {
21 SCMI_CLK_ATOMIC_SUPPORTED,
22 SCMI_CLK_STATE_CTRL_SUPPORTED,
23 SCMI_CLK_RATE_CTRL_SUPPORTED,
24 SCMI_CLK_PARENT_CTRL_SUPPORTED,
25 SCMI_CLK_DUTY_CYCLE_SUPPORTED,
26 SCMI_CLK_FEATS_COUNT
27};
28
29#define SCMI_MAX_CLK_OPS BIT(SCMI_CLK_FEATS_COUNT)
30
31static const struct scmi_clk_proto_ops *scmi_proto_clk_ops;
32
33struct scmi_clk {
34 u32 id;
35 struct device *dev;
36 struct clk_hw hw;
37 const struct scmi_clock_info *info;
38 const struct scmi_protocol_handle *ph;
39 struct clk_parent_data *parent_data;
40};
41
42#define to_scmi_clk(clk) container_of(clk, struct scmi_clk, hw)
43
44static unsigned long scmi_clk_recalc_rate(struct clk_hw *hw,
45 unsigned long parent_rate)
46{
47 int ret;
48 u64 rate;
49 struct scmi_clk *clk = to_scmi_clk(hw);
50
51 ret = scmi_proto_clk_ops->rate_get(clk->ph, clk->id, &rate);
52 if (ret)
53 return 0;
54 return rate;
55}
56
57static long scmi_clk_round_rate(struct clk_hw *hw, unsigned long rate,
58 unsigned long *parent_rate)
59{
60 u64 fmin, fmax, ftmp;
61 struct scmi_clk *clk = to_scmi_clk(hw);
62
63 /*
64 * We can't figure out what rate it will be, so just return the
65 * rate back to the caller. scmi_clk_recalc_rate() will be called
66 * after the rate is set and we'll know what rate the clock is
67 * running at then.
68 */
69 if (clk->info->rate_discrete)
70 return rate;
71
72 fmin = clk->info->range.min_rate;
73 fmax = clk->info->range.max_rate;
74 if (rate <= fmin)
75 return fmin;
76 else if (rate >= fmax)
77 return fmax;
78
79 ftmp = rate - fmin;
80 ftmp += clk->info->range.step_size - 1; /* to round up */
81 do_div(ftmp, clk->info->range.step_size);
82
83 return ftmp * clk->info->range.step_size + fmin;
84}
85
86static int scmi_clk_set_rate(struct clk_hw *hw, unsigned long rate,
87 unsigned long parent_rate)
88{
89 struct scmi_clk *clk = to_scmi_clk(hw);
90
91 return scmi_proto_clk_ops->rate_set(clk->ph, clk->id, rate);
92}
93
94static int scmi_clk_set_parent(struct clk_hw *hw, u8 parent_index)
95{
96 struct scmi_clk *clk = to_scmi_clk(hw);
97
98 return scmi_proto_clk_ops->parent_set(clk->ph, clk->id, parent_index);
99}
100
101static u8 scmi_clk_get_parent(struct clk_hw *hw)
102{
103 struct scmi_clk *clk = to_scmi_clk(hw);
104 u32 parent_id, p_idx;
105 int ret;
106
107 ret = scmi_proto_clk_ops->parent_get(clk->ph, clk->id, &parent_id);
108 if (ret)
109 return 0;
110
111 for (p_idx = 0; p_idx < clk->info->num_parents; p_idx++) {
112 if (clk->parent_data[p_idx].index == parent_id)
113 break;
114 }
115
116 if (p_idx == clk->info->num_parents)
117 return 0;
118
119 return p_idx;
120}
121
122static int scmi_clk_determine_rate(struct clk_hw *hw, struct clk_rate_request *req)
123{
124 /*
125 * Suppose all the requested rates are supported, and let firmware
126 * to handle the left work.
127 */
128 return 0;
129}
130
131static int scmi_clk_enable(struct clk_hw *hw)
132{
133 struct scmi_clk *clk = to_scmi_clk(hw);
134
135 return scmi_proto_clk_ops->enable(clk->ph, clk->id, NOT_ATOMIC);
136}
137
138static void scmi_clk_disable(struct clk_hw *hw)
139{
140 struct scmi_clk *clk = to_scmi_clk(hw);
141
142 scmi_proto_clk_ops->disable(clk->ph, clk->id, NOT_ATOMIC);
143}
144
145static int scmi_clk_atomic_enable(struct clk_hw *hw)
146{
147 struct scmi_clk *clk = to_scmi_clk(hw);
148
149 return scmi_proto_clk_ops->enable(clk->ph, clk->id, ATOMIC);
150}
151
152static void scmi_clk_atomic_disable(struct clk_hw *hw)
153{
154 struct scmi_clk *clk = to_scmi_clk(hw);
155
156 scmi_proto_clk_ops->disable(clk->ph, clk->id, ATOMIC);
157}
158
159static int __scmi_clk_is_enabled(struct clk_hw *hw, bool atomic)
160{
161 int ret;
162 bool enabled = false;
163 struct scmi_clk *clk = to_scmi_clk(hw);
164
165 ret = scmi_proto_clk_ops->state_get(clk->ph, clk->id, &enabled, atomic);
166 if (ret)
167 dev_warn(clk->dev,
168 "Failed to get state for clock ID %d\n", clk->id);
169
170 return !!enabled;
171}
172
173static int scmi_clk_atomic_is_enabled(struct clk_hw *hw)
174{
175 return __scmi_clk_is_enabled(hw, ATOMIC);
176}
177
178static int scmi_clk_is_enabled(struct clk_hw *hw)
179{
180 return __scmi_clk_is_enabled(hw, NOT_ATOMIC);
181}
182
183static int scmi_clk_get_duty_cycle(struct clk_hw *hw, struct clk_duty *duty)
184{
185 int ret;
186 u32 val;
187 struct scmi_clk *clk = to_scmi_clk(hw);
188
189 ret = scmi_proto_clk_ops->config_oem_get(clk->ph, clk->id,
190 SCMI_CLOCK_CFG_DUTY_CYCLE,
191 &val, NULL, false);
192 if (!ret) {
193 duty->num = val;
194 duty->den = 100;
195 } else {
196 dev_warn(clk->dev,
197 "Failed to get duty cycle for clock ID %d\n", clk->id);
198 }
199
200 return ret;
201}
202
203static int scmi_clk_set_duty_cycle(struct clk_hw *hw, struct clk_duty *duty)
204{
205 int ret;
206 u32 val;
207 struct scmi_clk *clk = to_scmi_clk(hw);
208
209 /* SCMI OEM Duty Cycle is expressed as a percentage */
210 val = (duty->num * 100) / duty->den;
211 ret = scmi_proto_clk_ops->config_oem_set(clk->ph, clk->id,
212 SCMI_CLOCK_CFG_DUTY_CYCLE,
213 val, false);
214 if (ret)
215 dev_warn(clk->dev,
216 "Failed to set duty cycle(%u/%u) for clock ID %d\n",
217 duty->num, duty->den, clk->id);
218
219 return ret;
220}
221
222static int scmi_clk_ops_init(struct device *dev, struct scmi_clk *sclk,
223 const struct clk_ops *scmi_ops)
224{
225 int ret;
226 unsigned long min_rate, max_rate;
227
228 struct clk_init_data init = {
229 .flags = CLK_GET_RATE_NOCACHE,
230 .num_parents = sclk->info->num_parents,
231 .ops = scmi_ops,
232 .name = sclk->info->name,
233 .parent_data = sclk->parent_data,
234 };
235
236 sclk->hw.init = &init;
237 ret = devm_clk_hw_register(dev, &sclk->hw);
238 if (ret)
239 return ret;
240
241 if (sclk->info->rate_discrete) {
242 int num_rates = sclk->info->list.num_rates;
243
244 if (num_rates <= 0)
245 return -EINVAL;
246
247 min_rate = sclk->info->list.rates[0];
248 max_rate = sclk->info->list.rates[num_rates - 1];
249 } else {
250 min_rate = sclk->info->range.min_rate;
251 max_rate = sclk->info->range.max_rate;
252 }
253
254 clk_hw_set_rate_range(&sclk->hw, min_rate, max_rate);
255 return ret;
256}
257
258/**
259 * scmi_clk_ops_alloc() - Alloc and configure clock operations
260 * @dev: A device reference for devres
261 * @feats_key: A bitmap representing the desired clk_ops capabilities
262 *
263 * Allocate and configure a proper set of clock operations depending on the
264 * specifically required SCMI clock features.
265 *
266 * Return: A pointer to the allocated and configured clk_ops on success,
267 * or NULL on allocation failure.
268 */
269static const struct clk_ops *
270scmi_clk_ops_alloc(struct device *dev, unsigned long feats_key)
271{
272 struct clk_ops *ops;
273
274 ops = devm_kzalloc(dev, sizeof(*ops), GFP_KERNEL);
275 if (!ops)
276 return NULL;
277 /*
278 * We can provide enable/disable/is_enabled atomic callbacks only if the
279 * underlying SCMI transport for an SCMI instance is configured to
280 * handle SCMI commands in an atomic manner.
281 *
282 * When no SCMI atomic transport support is available we instead provide
283 * only the prepare/unprepare API, as allowed by the clock framework
284 * when atomic calls are not available.
285 */
286 if (feats_key & BIT(SCMI_CLK_STATE_CTRL_SUPPORTED)) {
287 if (feats_key & BIT(SCMI_CLK_ATOMIC_SUPPORTED)) {
288 ops->enable = scmi_clk_atomic_enable;
289 ops->disable = scmi_clk_atomic_disable;
290 } else {
291 ops->prepare = scmi_clk_enable;
292 ops->unprepare = scmi_clk_disable;
293 }
294 }
295
296 if (feats_key & BIT(SCMI_CLK_ATOMIC_SUPPORTED))
297 ops->is_enabled = scmi_clk_atomic_is_enabled;
298 else
299 ops->is_prepared = scmi_clk_is_enabled;
300
301 /* Rate ops */
302 ops->recalc_rate = scmi_clk_recalc_rate;
303 ops->round_rate = scmi_clk_round_rate;
304 ops->determine_rate = scmi_clk_determine_rate;
305 if (feats_key & BIT(SCMI_CLK_RATE_CTRL_SUPPORTED))
306 ops->set_rate = scmi_clk_set_rate;
307
308 /* Parent ops */
309 ops->get_parent = scmi_clk_get_parent;
310 if (feats_key & BIT(SCMI_CLK_PARENT_CTRL_SUPPORTED))
311 ops->set_parent = scmi_clk_set_parent;
312
313 /* Duty cycle */
314 if (feats_key & BIT(SCMI_CLK_DUTY_CYCLE_SUPPORTED)) {
315 ops->get_duty_cycle = scmi_clk_get_duty_cycle;
316 ops->set_duty_cycle = scmi_clk_set_duty_cycle;
317 }
318
319 return ops;
320}
321
322/**
323 * scmi_clk_ops_select() - Select a proper set of clock operations
324 * @sclk: A reference to an SCMI clock descriptor
325 * @atomic_capable: A flag to indicate if atomic mode is supported by the
326 * transport
327 * @atomic_threshold_us: Platform atomic threshold value in microseconds:
328 * clk_ops are atomic when clock enable latency is less
329 * than this threshold
330 * @clk_ops_db: A reference to the array used as a database to store all the
331 * created clock operations combinations.
332 * @db_size: Maximum number of entries held by @clk_ops_db
333 *
334 * After having built a bitmap descriptor to represent the set of features
335 * needed by this SCMI clock, at first use it to lookup into the set of
336 * previously allocated clk_ops to check if a suitable combination of clock
337 * operations was already created; when no match is found allocate a brand new
338 * set of clk_ops satisfying the required combination of features and save it
339 * for future references.
340 *
341 * In this way only one set of clk_ops is ever created for each different
342 * combination that is effectively needed by a driver instance.
343 *
344 * Return: A pointer to the allocated and configured clk_ops on success, or
345 * NULL otherwise.
346 */
347static const struct clk_ops *
348scmi_clk_ops_select(struct scmi_clk *sclk, bool atomic_capable,
349 unsigned int atomic_threshold_us,
350 const struct clk_ops **clk_ops_db, size_t db_size)
351{
352 const struct scmi_clock_info *ci = sclk->info;
353 unsigned int feats_key = 0;
354 const struct clk_ops *ops;
355
356 /*
357 * Note that when transport is atomic but SCMI protocol did not
358 * specify (or support) an enable_latency associated with a
359 * clock, we default to use atomic operations mode.
360 */
361 if (atomic_capable && ci->enable_latency <= atomic_threshold_us)
362 feats_key |= BIT(SCMI_CLK_ATOMIC_SUPPORTED);
363
364 if (!ci->state_ctrl_forbidden)
365 feats_key |= BIT(SCMI_CLK_STATE_CTRL_SUPPORTED);
366
367 if (!ci->rate_ctrl_forbidden)
368 feats_key |= BIT(SCMI_CLK_RATE_CTRL_SUPPORTED);
369
370 if (!ci->parent_ctrl_forbidden)
371 feats_key |= BIT(SCMI_CLK_PARENT_CTRL_SUPPORTED);
372
373 if (ci->extended_config)
374 feats_key |= BIT(SCMI_CLK_DUTY_CYCLE_SUPPORTED);
375
376 if (WARN_ON(feats_key >= db_size))
377 return NULL;
378
379 /* Lookup previously allocated ops */
380 ops = clk_ops_db[feats_key];
381 if (ops)
382 return ops;
383
384 /* Did not find a pre-allocated clock_ops */
385 ops = scmi_clk_ops_alloc(sclk->dev, feats_key);
386 if (!ops)
387 return NULL;
388
389 /* Store new ops combinations */
390 clk_ops_db[feats_key] = ops;
391
392 return ops;
393}
394
395static int scmi_clocks_probe(struct scmi_device *sdev)
396{
397 int idx, count, err;
398 unsigned int atomic_threshold_us;
399 bool transport_is_atomic;
400 struct clk_hw **hws;
401 struct clk_hw_onecell_data *clk_data;
402 struct device *dev = &sdev->dev;
403 struct device_node *np = dev->of_node;
404 const struct scmi_handle *handle = sdev->handle;
405 struct scmi_protocol_handle *ph;
406 const struct clk_ops *scmi_clk_ops_db[SCMI_MAX_CLK_OPS] = {};
407
408 if (!handle)
409 return -ENODEV;
410
411 scmi_proto_clk_ops =
412 handle->devm_protocol_get(sdev, SCMI_PROTOCOL_CLOCK, &ph);
413 if (IS_ERR(scmi_proto_clk_ops))
414 return PTR_ERR(scmi_proto_clk_ops);
415
416 count = scmi_proto_clk_ops->count_get(ph);
417 if (count < 0) {
418 dev_err(dev, "%pOFn: invalid clock output count\n", np);
419 return -EINVAL;
420 }
421
422 clk_data = devm_kzalloc(dev, struct_size(clk_data, hws, count),
423 GFP_KERNEL);
424 if (!clk_data)
425 return -ENOMEM;
426
427 clk_data->num = count;
428 hws = clk_data->hws;
429
430 transport_is_atomic = handle->is_transport_atomic(handle,
431 &atomic_threshold_us);
432
433 for (idx = 0; idx < count; idx++) {
434 struct scmi_clk *sclk;
435 const struct clk_ops *scmi_ops;
436
437 sclk = devm_kzalloc(dev, sizeof(*sclk), GFP_KERNEL);
438 if (!sclk)
439 return -ENOMEM;
440
441 sclk->info = scmi_proto_clk_ops->info_get(ph, idx);
442 if (!sclk->info) {
443 dev_dbg(dev, "invalid clock info for idx %d\n", idx);
444 devm_kfree(dev, sclk);
445 continue;
446 }
447
448 sclk->id = idx;
449 sclk->ph = ph;
450 sclk->dev = dev;
451
452 /*
453 * Note that the scmi_clk_ops_db is on the stack, not global,
454 * because it cannot be shared between mulitple probe-sequences
455 * to avoid sharing the devm_ allocated clk_ops between multiple
456 * SCMI clk driver instances.
457 */
458 scmi_ops = scmi_clk_ops_select(sclk, transport_is_atomic,
459 atomic_threshold_us,
460 scmi_clk_ops_db,
461 ARRAY_SIZE(scmi_clk_ops_db));
462 if (!scmi_ops)
463 return -ENOMEM;
464
465 /* Initialize clock parent data. */
466 if (sclk->info->num_parents > 0) {
467 sclk->parent_data = devm_kcalloc(dev, sclk->info->num_parents,
468 sizeof(*sclk->parent_data), GFP_KERNEL);
469 if (!sclk->parent_data)
470 return -ENOMEM;
471
472 for (int i = 0; i < sclk->info->num_parents; i++) {
473 sclk->parent_data[i].index = sclk->info->parents[i];
474 sclk->parent_data[i].hw = hws[sclk->info->parents[i]];
475 }
476 }
477
478 err = scmi_clk_ops_init(dev, sclk, scmi_ops);
479 if (err) {
480 dev_err(dev, "failed to register clock %d\n", idx);
481 devm_kfree(dev, sclk->parent_data);
482 devm_kfree(dev, sclk);
483 hws[idx] = NULL;
484 } else {
485 dev_dbg(dev, "Registered clock:%s%s\n",
486 sclk->info->name,
487 scmi_ops->enable ? " (atomic ops)" : "");
488 hws[idx] = &sclk->hw;
489 }
490 }
491
492 return devm_of_clk_add_hw_provider(dev, of_clk_hw_onecell_get,
493 clk_data);
494}
495
496static const struct scmi_device_id scmi_id_table[] = {
497 { SCMI_PROTOCOL_CLOCK, "clocks" },
498 { },
499};
500MODULE_DEVICE_TABLE(scmi, scmi_id_table);
501
502static struct scmi_driver scmi_clocks_driver = {
503 .name = "scmi-clocks",
504 .probe = scmi_clocks_probe,
505 .id_table = scmi_id_table,
506};
507module_scmi_driver(scmi_clocks_driver);
508
509MODULE_AUTHOR("Sudeep Holla <sudeep.holla@arm.com>");
510MODULE_DESCRIPTION("ARM SCMI clock driver");
511MODULE_LICENSE("GPL v2");
1// SPDX-License-Identifier: GPL-2.0
2/*
3 * System Control and Power Interface (SCMI) Protocol based clock driver
4 *
5 * Copyright (C) 2018-2022 ARM Ltd.
6 */
7
8#include <linux/clk-provider.h>
9#include <linux/device.h>
10#include <linux/err.h>
11#include <linux/of.h>
12#include <linux/module.h>
13#include <linux/scmi_protocol.h>
14#include <asm/div64.h>
15
16#define NOT_ATOMIC false
17#define ATOMIC true
18
19static const struct scmi_clk_proto_ops *scmi_proto_clk_ops;
20
21struct scmi_clk {
22 u32 id;
23 struct device *dev;
24 struct clk_hw hw;
25 const struct scmi_clock_info *info;
26 const struct scmi_protocol_handle *ph;
27 struct clk_parent_data *parent_data;
28};
29
30#define to_scmi_clk(clk) container_of(clk, struct scmi_clk, hw)
31
32static unsigned long scmi_clk_recalc_rate(struct clk_hw *hw,
33 unsigned long parent_rate)
34{
35 int ret;
36 u64 rate;
37 struct scmi_clk *clk = to_scmi_clk(hw);
38
39 ret = scmi_proto_clk_ops->rate_get(clk->ph, clk->id, &rate);
40 if (ret)
41 return 0;
42 return rate;
43}
44
45static long scmi_clk_round_rate(struct clk_hw *hw, unsigned long rate,
46 unsigned long *parent_rate)
47{
48 u64 fmin, fmax, ftmp;
49 struct scmi_clk *clk = to_scmi_clk(hw);
50
51 /*
52 * We can't figure out what rate it will be, so just return the
53 * rate back to the caller. scmi_clk_recalc_rate() will be called
54 * after the rate is set and we'll know what rate the clock is
55 * running at then.
56 */
57 if (clk->info->rate_discrete)
58 return rate;
59
60 fmin = clk->info->range.min_rate;
61 fmax = clk->info->range.max_rate;
62 if (rate <= fmin)
63 return fmin;
64 else if (rate >= fmax)
65 return fmax;
66
67 ftmp = rate - fmin;
68 ftmp += clk->info->range.step_size - 1; /* to round up */
69 do_div(ftmp, clk->info->range.step_size);
70
71 return ftmp * clk->info->range.step_size + fmin;
72}
73
74static int scmi_clk_set_rate(struct clk_hw *hw, unsigned long rate,
75 unsigned long parent_rate)
76{
77 struct scmi_clk *clk = to_scmi_clk(hw);
78
79 return scmi_proto_clk_ops->rate_set(clk->ph, clk->id, rate);
80}
81
82static int scmi_clk_set_parent(struct clk_hw *hw, u8 parent_index)
83{
84 struct scmi_clk *clk = to_scmi_clk(hw);
85
86 return scmi_proto_clk_ops->parent_set(clk->ph, clk->id, parent_index);
87}
88
89static u8 scmi_clk_get_parent(struct clk_hw *hw)
90{
91 struct scmi_clk *clk = to_scmi_clk(hw);
92 u32 parent_id, p_idx;
93 int ret;
94
95 ret = scmi_proto_clk_ops->parent_get(clk->ph, clk->id, &parent_id);
96 if (ret)
97 return 0;
98
99 for (p_idx = 0; p_idx < clk->info->num_parents; p_idx++) {
100 if (clk->parent_data[p_idx].index == parent_id)
101 break;
102 }
103
104 if (p_idx == clk->info->num_parents)
105 return 0;
106
107 return p_idx;
108}
109
110static int scmi_clk_determine_rate(struct clk_hw *hw, struct clk_rate_request *req)
111{
112 /*
113 * Suppose all the requested rates are supported, and let firmware
114 * to handle the left work.
115 */
116 return 0;
117}
118
119static int scmi_clk_enable(struct clk_hw *hw)
120{
121 struct scmi_clk *clk = to_scmi_clk(hw);
122
123 return scmi_proto_clk_ops->enable(clk->ph, clk->id, NOT_ATOMIC);
124}
125
126static void scmi_clk_disable(struct clk_hw *hw)
127{
128 struct scmi_clk *clk = to_scmi_clk(hw);
129
130 scmi_proto_clk_ops->disable(clk->ph, clk->id, NOT_ATOMIC);
131}
132
133static int scmi_clk_atomic_enable(struct clk_hw *hw)
134{
135 struct scmi_clk *clk = to_scmi_clk(hw);
136
137 return scmi_proto_clk_ops->enable(clk->ph, clk->id, ATOMIC);
138}
139
140static void scmi_clk_atomic_disable(struct clk_hw *hw)
141{
142 struct scmi_clk *clk = to_scmi_clk(hw);
143
144 scmi_proto_clk_ops->disable(clk->ph, clk->id, ATOMIC);
145}
146
147static int scmi_clk_atomic_is_enabled(struct clk_hw *hw)
148{
149 int ret;
150 bool enabled = false;
151 struct scmi_clk *clk = to_scmi_clk(hw);
152
153 ret = scmi_proto_clk_ops->state_get(clk->ph, clk->id, &enabled, ATOMIC);
154 if (ret)
155 dev_warn(clk->dev,
156 "Failed to get state for clock ID %d\n", clk->id);
157
158 return !!enabled;
159}
160
161/*
162 * We can provide enable/disable/is_enabled atomic callbacks only if the
163 * underlying SCMI transport for an SCMI instance is configured to handle
164 * SCMI commands in an atomic manner.
165 *
166 * When no SCMI atomic transport support is available we instead provide only
167 * the prepare/unprepare API, as allowed by the clock framework when atomic
168 * calls are not available.
169 *
170 * Two distinct sets of clk_ops are provided since we could have multiple SCMI
171 * instances with different underlying transport quality, so they cannot be
172 * shared.
173 */
174static const struct clk_ops scmi_clk_ops = {
175 .recalc_rate = scmi_clk_recalc_rate,
176 .round_rate = scmi_clk_round_rate,
177 .set_rate = scmi_clk_set_rate,
178 .prepare = scmi_clk_enable,
179 .unprepare = scmi_clk_disable,
180 .set_parent = scmi_clk_set_parent,
181 .get_parent = scmi_clk_get_parent,
182 .determine_rate = scmi_clk_determine_rate,
183};
184
185static const struct clk_ops scmi_atomic_clk_ops = {
186 .recalc_rate = scmi_clk_recalc_rate,
187 .round_rate = scmi_clk_round_rate,
188 .set_rate = scmi_clk_set_rate,
189 .enable = scmi_clk_atomic_enable,
190 .disable = scmi_clk_atomic_disable,
191 .is_enabled = scmi_clk_atomic_is_enabled,
192 .set_parent = scmi_clk_set_parent,
193 .get_parent = scmi_clk_get_parent,
194 .determine_rate = scmi_clk_determine_rate,
195};
196
197static int scmi_clk_ops_init(struct device *dev, struct scmi_clk *sclk,
198 const struct clk_ops *scmi_ops)
199{
200 int ret;
201 unsigned long min_rate, max_rate;
202
203 struct clk_init_data init = {
204 .flags = CLK_GET_RATE_NOCACHE,
205 .num_parents = sclk->info->num_parents,
206 .ops = scmi_ops,
207 .name = sclk->info->name,
208 .parent_data = sclk->parent_data,
209 };
210
211 sclk->hw.init = &init;
212 ret = devm_clk_hw_register(dev, &sclk->hw);
213 if (ret)
214 return ret;
215
216 if (sclk->info->rate_discrete) {
217 int num_rates = sclk->info->list.num_rates;
218
219 if (num_rates <= 0)
220 return -EINVAL;
221
222 min_rate = sclk->info->list.rates[0];
223 max_rate = sclk->info->list.rates[num_rates - 1];
224 } else {
225 min_rate = sclk->info->range.min_rate;
226 max_rate = sclk->info->range.max_rate;
227 }
228
229 clk_hw_set_rate_range(&sclk->hw, min_rate, max_rate);
230 return ret;
231}
232
233static int scmi_clocks_probe(struct scmi_device *sdev)
234{
235 int idx, count, err;
236 unsigned int atomic_threshold;
237 bool is_atomic;
238 struct clk_hw **hws;
239 struct clk_hw_onecell_data *clk_data;
240 struct device *dev = &sdev->dev;
241 struct device_node *np = dev->of_node;
242 const struct scmi_handle *handle = sdev->handle;
243 struct scmi_protocol_handle *ph;
244
245 if (!handle)
246 return -ENODEV;
247
248 scmi_proto_clk_ops =
249 handle->devm_protocol_get(sdev, SCMI_PROTOCOL_CLOCK, &ph);
250 if (IS_ERR(scmi_proto_clk_ops))
251 return PTR_ERR(scmi_proto_clk_ops);
252
253 count = scmi_proto_clk_ops->count_get(ph);
254 if (count < 0) {
255 dev_err(dev, "%pOFn: invalid clock output count\n", np);
256 return -EINVAL;
257 }
258
259 clk_data = devm_kzalloc(dev, struct_size(clk_data, hws, count),
260 GFP_KERNEL);
261 if (!clk_data)
262 return -ENOMEM;
263
264 clk_data->num = count;
265 hws = clk_data->hws;
266
267 is_atomic = handle->is_transport_atomic(handle, &atomic_threshold);
268
269 for (idx = 0; idx < count; idx++) {
270 struct scmi_clk *sclk;
271 const struct clk_ops *scmi_ops;
272
273 sclk = devm_kzalloc(dev, sizeof(*sclk), GFP_KERNEL);
274 if (!sclk)
275 return -ENOMEM;
276
277 sclk->info = scmi_proto_clk_ops->info_get(ph, idx);
278 if (!sclk->info) {
279 dev_dbg(dev, "invalid clock info for idx %d\n", idx);
280 devm_kfree(dev, sclk);
281 continue;
282 }
283
284 sclk->id = idx;
285 sclk->ph = ph;
286 sclk->dev = dev;
287
288 /*
289 * Note that when transport is atomic but SCMI protocol did not
290 * specify (or support) an enable_latency associated with a
291 * clock, we default to use atomic operations mode.
292 */
293 if (is_atomic &&
294 sclk->info->enable_latency <= atomic_threshold)
295 scmi_ops = &scmi_atomic_clk_ops;
296 else
297 scmi_ops = &scmi_clk_ops;
298
299 /* Initialize clock parent data. */
300 if (sclk->info->num_parents > 0) {
301 sclk->parent_data = devm_kcalloc(dev, sclk->info->num_parents,
302 sizeof(*sclk->parent_data), GFP_KERNEL);
303 if (!sclk->parent_data)
304 return -ENOMEM;
305
306 for (int i = 0; i < sclk->info->num_parents; i++) {
307 sclk->parent_data[i].index = sclk->info->parents[i];
308 sclk->parent_data[i].hw = hws[sclk->info->parents[i]];
309 }
310 }
311
312 err = scmi_clk_ops_init(dev, sclk, scmi_ops);
313 if (err) {
314 dev_err(dev, "failed to register clock %d\n", idx);
315 devm_kfree(dev, sclk->parent_data);
316 devm_kfree(dev, sclk);
317 hws[idx] = NULL;
318 } else {
319 dev_dbg(dev, "Registered clock:%s%s\n",
320 sclk->info->name,
321 scmi_ops == &scmi_atomic_clk_ops ?
322 " (atomic ops)" : "");
323 hws[idx] = &sclk->hw;
324 }
325 }
326
327 return devm_of_clk_add_hw_provider(dev, of_clk_hw_onecell_get,
328 clk_data);
329}
330
331static const struct scmi_device_id scmi_id_table[] = {
332 { SCMI_PROTOCOL_CLOCK, "clocks" },
333 { },
334};
335MODULE_DEVICE_TABLE(scmi, scmi_id_table);
336
337static struct scmi_driver scmi_clocks_driver = {
338 .name = "scmi-clocks",
339 .probe = scmi_clocks_probe,
340 .id_table = scmi_id_table,
341};
342module_scmi_driver(scmi_clocks_driver);
343
344MODULE_AUTHOR("Sudeep Holla <sudeep.holla@arm.com>");
345MODULE_DESCRIPTION("ARM SCMI clock driver");
346MODULE_LICENSE("GPL v2");