blob: d5e20071859e91467d7d52f52a1dae98d1ca153b [file] [log] [blame]
Tom Rini10e47792018-05-06 17:58:06 -04001// SPDX-License-Identifier: GPL-2.0+
Simon Glass36ad2342015-06-23 15:39:15 -06002/*
3 * Copyright (C) 2015 Google, Inc
4 * Written by Simon Glass <sjg@chromium.org>
Stephen Warrena9622432016-06-17 09:44:00 -06005 * Copyright (c) 2016, NVIDIA CORPORATION.
Philipp Tomsich9cf03b02018-01-08 13:59:18 +01006 * Copyright (c) 2018, Theobroma Systems Design und Consulting GmbH
Simon Glass36ad2342015-06-23 15:39:15 -06007 */
8
Patrick Delaunay81313352021-04-27 11:02:19 +02009#define LOG_CATEGORY UCLASS_CLK
10
Simon Glass36ad2342015-06-23 15:39:15 -060011#include <common.h>
12#include <clk.h>
Stephen Warrena9622432016-06-17 09:44:00 -060013#include <clk-uclass.h>
Simon Glass36ad2342015-06-23 15:39:15 -060014#include <dm.h>
Simon Glass589d9152016-07-04 11:58:03 -060015#include <dt-structs.h>
Simon Glass36ad2342015-06-23 15:39:15 -060016#include <errno.h>
Simon Glass0f2af882020-05-10 11:40:05 -060017#include <log.h>
Simon Glass9bc15642020-02-03 07:36:16 -070018#include <malloc.h>
Patrick Delaunay283dadf2021-11-19 15:12:06 +010019#include <asm/global_data.h>
Sean Andersond7ac3732021-04-08 22:13:03 -040020#include <dm/device_compat.h>
Claudiu Bezneac8c16002020-09-07 17:46:34 +030021#include <dm/device-internal.h>
Simon Glassd66c5f72020-02-03 07:36:15 -070022#include <dm/devres.h>
23#include <dm/read.h>
Simon Glassc06c1be2020-05-10 11:40:08 -060024#include <linux/bug.h>
Lukasz Majewski9e38dc32019-06-24 15:50:42 +020025#include <linux/clk-provider.h>
Simon Glassd66c5f72020-02-03 07:36:15 -070026#include <linux/err.h>
Simon Glass36ad2342015-06-23 15:39:15 -060027
Mario Six799fe562018-01-15 11:06:51 +010028static inline const struct clk_ops *clk_dev_ops(struct udevice *dev)
Simon Glass36ad2342015-06-23 15:39:15 -060029{
Mario Six799fe562018-01-15 11:06:51 +010030 return (const struct clk_ops *)dev->driver->ops;
Simon Glass36ad2342015-06-23 15:39:15 -060031}
32
Simon Glass43033962020-07-19 10:15:56 -060033struct clk *dev_get_clk_ptr(struct udevice *dev)
34{
35 return (struct clk *)dev_get_uclass_priv(dev);
36}
37
Simon Glass3580f6d2021-08-07 07:24:03 -060038#if CONFIG_IS_ENABLED(OF_PLATDATA)
Simon Glass1257efc2021-08-07 07:24:09 -060039int clk_get_by_phandle(struct udevice *dev, const struct phandle_1_arg *cells,
40 struct clk *clk)
Simon Glass589d9152016-07-04 11:58:03 -060041{
42 int ret;
43
Simon Glass0000e0d2021-03-15 17:25:28 +130044 ret = device_get_by_ofplat_idx(cells->idx, &clk->dev);
Simon Glass589d9152016-07-04 11:58:03 -060045 if (ret)
46 return ret;
Walter Lozanodc5b4372020-06-25 01:10:13 -030047 clk->id = cells->arg[0];
Simon Glass589d9152016-07-04 11:58:03 -060048
49 return 0;
50}
Simon Glass3580f6d2021-08-07 07:24:03 -060051#endif
52
53#if CONFIG_IS_ENABLED(OF_REAL)
Stephen Warrena9622432016-06-17 09:44:00 -060054static int clk_of_xlate_default(struct clk *clk,
Simon Glassb7ae2772017-05-18 20:09:40 -060055 struct ofnode_phandle_args *args)
Simon Glass36ad2342015-06-23 15:39:15 -060056{
Stephen Warrena9622432016-06-17 09:44:00 -060057 debug("%s(clk=%p)\n", __func__, clk);
Simon Glass36ad2342015-06-23 15:39:15 -060058
Stephen Warrena9622432016-06-17 09:44:00 -060059 if (args->args_count > 1) {
Sean Andersona1b654b2021-12-01 14:26:53 -050060 debug("Invalid args_count: %d\n", args->args_count);
Stephen Warrena9622432016-06-17 09:44:00 -060061 return -EINVAL;
62 }
Simon Glass36ad2342015-06-23 15:39:15 -060063
Stephen Warrena9622432016-06-17 09:44:00 -060064 if (args->args_count)
65 clk->id = args->args[0];
66 else
67 clk->id = 0;
Simon Glass36ad2342015-06-23 15:39:15 -060068
Sekhar Nori3d23abd2019-07-11 14:30:24 +053069 clk->data = 0;
70
Stephen Warrena9622432016-06-17 09:44:00 -060071 return 0;
Simon Glass36ad2342015-06-23 15:39:15 -060072}
Simon Glass0342bd22016-01-20 19:43:02 -070073
Jagan Tekifc7c7ce2019-02-28 00:26:52 +053074static int clk_get_by_index_tail(int ret, ofnode node,
75 struct ofnode_phandle_args *args,
76 const char *list_name, int index,
77 struct clk *clk)
78{
79 struct udevice *dev_clk;
80 const struct clk_ops *ops;
81
82 assert(clk);
83 clk->dev = NULL;
84 if (ret)
85 goto err;
86
87 ret = uclass_get_device_by_ofnode(UCLASS_CLK, args->node, &dev_clk);
88 if (ret) {
89 debug("%s: uclass_get_device_by_of_offset failed: err=%d\n",
90 __func__, ret);
Simon Glassf73f5812021-01-21 13:57:11 -070091 return log_msg_ret("get", ret);
Jagan Tekifc7c7ce2019-02-28 00:26:52 +053092 }
93
94 clk->dev = dev_clk;
95
96 ops = clk_dev_ops(dev_clk);
97
98 if (ops->of_xlate)
99 ret = ops->of_xlate(clk, args);
100 else
101 ret = clk_of_xlate_default(clk, args);
102 if (ret) {
103 debug("of_xlate() failed: %d\n", ret);
Simon Glassf73f5812021-01-21 13:57:11 -0700104 return log_msg_ret("xlate", ret);
Jagan Tekifc7c7ce2019-02-28 00:26:52 +0530105 }
106
107 return clk_request(dev_clk, clk);
108err:
109 debug("%s: Node '%s', property '%s', failed to request CLK index %d: %d\n",
110 __func__, ofnode_get_name(node), list_name, index, ret);
Simon Glassf73f5812021-01-21 13:57:11 -0700111
112 return log_msg_ret("prop", ret);
Jagan Tekifc7c7ce2019-02-28 00:26:52 +0530113}
114
Philipp Tomsichf7604342018-01-08 11:18:18 +0100115static int clk_get_by_indexed_prop(struct udevice *dev, const char *prop_name,
116 int index, struct clk *clk)
Simon Glass0342bd22016-01-20 19:43:02 -0700117{
Simon Glass0342bd22016-01-20 19:43:02 -0700118 int ret;
Simon Glass2558bff2017-05-30 21:47:29 -0600119 struct ofnode_phandle_args args;
Simon Glass0342bd22016-01-20 19:43:02 -0700120
Stephen Warrena9622432016-06-17 09:44:00 -0600121 debug("%s(dev=%p, index=%d, clk=%p)\n", __func__, dev, index, clk);
122
123 assert(clk);
Patrice Chotard96fc03d2017-07-18 11:57:07 +0200124 clk->dev = NULL;
125
Philipp Tomsichf7604342018-01-08 11:18:18 +0100126 ret = dev_read_phandle_with_args(dev, prop_name, "#clock-cells", 0,
Mario Six799fe562018-01-15 11:06:51 +0100127 index, &args);
Simon Glass0342bd22016-01-20 19:43:02 -0700128 if (ret) {
129 debug("%s: fdtdec_parse_phandle_with_args failed: err=%d\n",
130 __func__, ret);
Simon Glassf73f5812021-01-21 13:57:11 -0700131 return log_ret(ret);
Simon Glass0342bd22016-01-20 19:43:02 -0700132 }
133
Stephen Warrena9622432016-06-17 09:44:00 -0600134
Jagan Tekia77add32019-02-28 00:26:53 +0530135 return clk_get_by_index_tail(ret, dev_ofnode(dev), &args, "clocks",
Sean Andersonf0d5a6b2020-06-24 06:41:08 -0400136 index, clk);
Stephen Warrena9622432016-06-17 09:44:00 -0600137}
Philipp Tomsichf7604342018-01-08 11:18:18 +0100138
139int clk_get_by_index(struct udevice *dev, int index, struct clk *clk)
140{
Sean Anderson07435de2022-02-27 14:01:13 -0500141 return clk_get_by_index_nodev(dev_ofnode(dev), index, clk);
Jagan Tekifc7c7ce2019-02-28 00:26:52 +0530142}
143
144int clk_get_by_index_nodev(ofnode node, int index, struct clk *clk)
145{
146 struct ofnode_phandle_args args;
147 int ret;
148
149 ret = ofnode_parse_phandle_with_args(node, "clocks", "#clock-cells", 0,
Sean Andersonf0d5a6b2020-06-24 06:41:08 -0400150 index, &args);
Jagan Tekifc7c7ce2019-02-28 00:26:52 +0530151
152 return clk_get_by_index_tail(ret, node, &args, "clocks",
Sean Andersonf0d5a6b2020-06-24 06:41:08 -0400153 index, clk);
Philipp Tomsichf7604342018-01-08 11:18:18 +0100154}
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100155
Neil Armstrong8a275a02018-04-03 11:44:18 +0200156int clk_get_bulk(struct udevice *dev, struct clk_bulk *bulk)
157{
158 int i, ret, err, count;
Patrick Delaunayb9c32142021-04-27 10:57:54 +0200159
Neil Armstrong8a275a02018-04-03 11:44:18 +0200160 bulk->count = 0;
161
Patrick Delaunayd776a842020-09-25 09:41:14 +0200162 count = dev_count_phandle_with_args(dev, "clocks", "#clock-cells", 0);
Neil Armstrong52b26d92018-04-17 11:30:31 +0200163 if (count < 1)
164 return count;
Neil Armstrong8a275a02018-04-03 11:44:18 +0200165
166 bulk->clks = devm_kcalloc(dev, count, sizeof(struct clk), GFP_KERNEL);
167 if (!bulk->clks)
168 return -ENOMEM;
169
170 for (i = 0; i < count; i++) {
171 ret = clk_get_by_index(dev, i, &bulk->clks[i]);
172 if (ret < 0)
173 goto bulk_get_err;
174
175 ++bulk->count;
176 }
177
178 return 0;
179
180bulk_get_err:
181 err = clk_release_all(bulk->clks, bulk->count);
182 if (err)
183 debug("%s: could release all clocks for %p\n",
184 __func__, dev);
185
186 return ret;
187}
188
Claudiu Bezneab91eee62020-09-07 17:46:36 +0300189static struct clk *clk_set_default_get_by_id(struct clk *clk)
190{
191 struct clk *c = clk;
192
193 if (CONFIG_IS_ENABLED(CLK_CCF)) {
194 int ret = clk_get_by_id(clk->id, &c);
195
196 if (ret) {
197 debug("%s(): could not get parent clock pointer, id %lu\n",
198 __func__, clk->id);
199 ERR_PTR(ret);
200 }
201 }
202
203 return c;
204}
205
Sean Anderson08d531c2021-06-11 00:16:07 -0400206static int clk_set_default_parents(struct udevice *dev,
207 enum clk_defaults_stage stage)
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100208{
Claudiu Bezneab91eee62020-09-07 17:46:36 +0300209 struct clk clk, parent_clk, *c, *p;
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100210 int index;
211 int num_parents;
212 int ret;
213
214 num_parents = dev_count_phandle_with_args(dev, "assigned-clock-parents",
Patrick Delaunayd776a842020-09-25 09:41:14 +0200215 "#clock-cells", 0);
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100216 if (num_parents < 0) {
217 debug("%s: could not read assigned-clock-parents for %p\n",
218 __func__, dev);
219 return 0;
220 }
221
222 for (index = 0; index < num_parents; index++) {
223 ret = clk_get_by_indexed_prop(dev, "assigned-clock-parents",
224 index, &parent_clk);
Neil Armstrongf3cc6312018-07-26 15:19:32 +0200225 /* If -ENOENT, this is a no-op entry */
226 if (ret == -ENOENT)
227 continue;
228
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100229 if (ret) {
230 debug("%s: could not get parent clock %d for %s\n",
231 __func__, index, dev_read_name(dev));
232 return ret;
233 }
234
Claudiu Bezneab91eee62020-09-07 17:46:36 +0300235 p = clk_set_default_get_by_id(&parent_clk);
236 if (IS_ERR(p))
237 return PTR_ERR(p);
238
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100239 ret = clk_get_by_indexed_prop(dev, "assigned-clocks",
240 index, &clk);
Tero Kristod41b2b32021-06-11 11:45:11 +0300241 /*
242 * If the clock provider is not ready yet, let it handle
243 * the re-programming later.
244 */
245 if (ret == -EPROBE_DEFER) {
246 ret = 0;
247 continue;
248 }
249
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100250 if (ret) {
251 debug("%s: could not get assigned clock %d for %s\n",
252 __func__, index, dev_read_name(dev));
253 return ret;
254 }
255
Jean-Jacques Hiblot9601f322019-10-22 14:00:06 +0200256 /* This is clk provider device trying to reparent itself
257 * It cannot be done right now but need to wait after the
258 * device is probed
259 */
Sean Anderson08d531c2021-06-11 00:16:07 -0400260 if (stage == CLK_DEFAULTS_PRE && clk.dev == dev)
Jean-Jacques Hiblot9601f322019-10-22 14:00:06 +0200261 continue;
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100262
Sean Anderson08d531c2021-06-11 00:16:07 -0400263 if (stage != CLK_DEFAULTS_PRE && clk.dev != dev)
Jean-Jacques Hiblot9601f322019-10-22 14:00:06 +0200264 /* do not setup twice the parent clocks */
265 continue;
266
Claudiu Bezneab91eee62020-09-07 17:46:36 +0300267 c = clk_set_default_get_by_id(&clk);
268 if (IS_ERR(c))
269 return PTR_ERR(c);
270
271 ret = clk_set_parent(c, p);
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100272 /*
273 * Not all drivers may support clock-reparenting (as of now).
274 * Ignore errors due to this.
275 */
276 if (ret == -ENOSYS)
277 continue;
278
Jean-Jacques Hiblotb2320812019-09-26 15:42:42 +0200279 if (ret < 0) {
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100280 debug("%s: failed to reparent clock %d for %s\n",
281 __func__, index, dev_read_name(dev));
282 return ret;
283 }
284 }
285
286 return 0;
287}
288
Sean Anderson08d531c2021-06-11 00:16:07 -0400289static int clk_set_default_rates(struct udevice *dev,
290 enum clk_defaults_stage stage)
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100291{
Claudiu Bezneab91eee62020-09-07 17:46:36 +0300292 struct clk clk, *c;
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100293 int index;
294 int num_rates;
295 int size;
296 int ret = 0;
297 u32 *rates = NULL;
298
299 size = dev_read_size(dev, "assigned-clock-rates");
300 if (size < 0)
301 return 0;
302
303 num_rates = size / sizeof(u32);
304 rates = calloc(num_rates, sizeof(u32));
305 if (!rates)
306 return -ENOMEM;
307
308 ret = dev_read_u32_array(dev, "assigned-clock-rates", rates, num_rates);
309 if (ret)
310 goto fail;
311
312 for (index = 0; index < num_rates; index++) {
Neil Armstrongf3cc6312018-07-26 15:19:32 +0200313 /* If 0 is passed, this is a no-op */
314 if (!rates[index])
315 continue;
316
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100317 ret = clk_get_by_indexed_prop(dev, "assigned-clocks",
318 index, &clk);
Tero Kristod41b2b32021-06-11 11:45:11 +0300319 /*
320 * If the clock provider is not ready yet, let it handle
321 * the re-programming later.
322 */
323 if (ret == -EPROBE_DEFER) {
324 ret = 0;
325 continue;
326 }
327
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100328 if (ret) {
Sean Andersond7ac3732021-04-08 22:13:03 -0400329 dev_dbg(dev,
330 "could not get assigned clock %d (err = %d)\n",
331 index, ret);
Ashok Reddy Soma8f03cef2023-08-30 10:31:42 +0200332 /* Skip if it is empty */
333 if (ret == -ENOENT) {
334 ret = 0;
335 continue;
336 }
337
338 return ret;
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100339 }
340
Jean-Jacques Hiblot9601f322019-10-22 14:00:06 +0200341 /* This is clk provider device trying to program itself
342 * It cannot be done right now but need to wait after the
343 * device is probed
344 */
Sean Anderson08d531c2021-06-11 00:16:07 -0400345 if (stage == CLK_DEFAULTS_PRE && clk.dev == dev)
Jean-Jacques Hiblot9601f322019-10-22 14:00:06 +0200346 continue;
347
Sean Anderson08d531c2021-06-11 00:16:07 -0400348 if (stage != CLK_DEFAULTS_PRE && clk.dev != dev)
Jean-Jacques Hiblot9601f322019-10-22 14:00:06 +0200349 /* do not setup twice the parent clocks */
350 continue;
351
Claudiu Bezneab91eee62020-09-07 17:46:36 +0300352 c = clk_set_default_get_by_id(&clk);
353 if (IS_ERR(c))
354 return PTR_ERR(c);
355
356 ret = clk_set_rate(c, rates[index]);
Jean-Jacques Hiblot9601f322019-10-22 14:00:06 +0200357
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100358 if (ret < 0) {
Sean Andersond7ac3732021-04-08 22:13:03 -0400359 dev_warn(dev,
360 "failed to set rate on clock index %d (%ld) (error = %d)\n",
361 index, clk.id, ret);
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100362 break;
363 }
364 }
365
366fail:
367 free(rates);
368 return ret;
369}
370
Sean Anderson08d531c2021-06-11 00:16:07 -0400371int clk_set_defaults(struct udevice *dev, enum clk_defaults_stage stage)
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100372{
373 int ret;
374
Simon Glassf1d50f72020-12-19 10:40:13 -0700375 if (!dev_has_ofnode(dev))
Peng Fan40ec4e42019-07-31 07:01:49 +0000376 return 0;
377
Sean Anderson08d531c2021-06-11 00:16:07 -0400378 /*
379 * To avoid setting defaults twice, don't set them before relocation.
380 * However, still set them for SPL. And still set them if explicitly
381 * asked.
382 */
Philipp Tomsiche546ec82018-11-26 20:20:19 +0100383 if (!(IS_ENABLED(CONFIG_SPL_BUILD) || (gd->flags & GD_FLG_RELOC)))
Sean Anderson08d531c2021-06-11 00:16:07 -0400384 if (stage != CLK_DEFAULTS_POST_FORCE)
385 return 0;
Philipp Tomsiche546ec82018-11-26 20:20:19 +0100386
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100387 debug("%s(%s)\n", __func__, dev_read_name(dev));
388
Jean-Jacques Hiblot9601f322019-10-22 14:00:06 +0200389 ret = clk_set_default_parents(dev, stage);
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100390 if (ret)
391 return ret;
392
Jean-Jacques Hiblot9601f322019-10-22 14:00:06 +0200393 ret = clk_set_default_rates(dev, stage);
Philipp Tomsich9cf03b02018-01-08 13:59:18 +0100394 if (ret < 0)
395 return ret;
396
397 return 0;
398}
Stephen Warrena9622432016-06-17 09:44:00 -0600399
400int clk_get_by_name(struct udevice *dev, const char *name, struct clk *clk)
401{
Sean Anderson07435de2022-02-27 14:01:13 -0500402 return clk_get_by_name_nodev(dev_ofnode(dev), name, clk);
Simon Glass0342bd22016-01-20 19:43:02 -0700403}
Simon Glass1257efc2021-08-07 07:24:09 -0600404#endif /* OF_REAL */
Patrice Chotardcafc3412017-07-25 13:24:45 +0200405
developerbdc786d2020-01-09 11:35:07 +0800406int clk_get_by_name_nodev(ofnode node, const char *name, struct clk *clk)
407{
Samuel Hollandbae0f4f2023-01-21 18:02:51 -0600408 int index = 0;
developerbdc786d2020-01-09 11:35:07 +0800409
410 debug("%s(node=%p, name=%s, clk=%p)\n", __func__,
411 ofnode_get_name(node), name, clk);
412 clk->dev = NULL;
413
Samuel Hollandbae0f4f2023-01-21 18:02:51 -0600414 if (name) {
415 index = ofnode_stringlist_search(node, "clock-names", name);
416 if (index < 0) {
417 debug("fdt_stringlist_search() failed: %d\n", index);
418 return index;
419 }
developerbdc786d2020-01-09 11:35:07 +0800420 }
421
422 return clk_get_by_index_nodev(node, index, clk);
423}
424
Eugen Hristev70e32ba2023-06-19 13:47:52 +0300425int clk_release_all(struct clk *clk, unsigned int count)
Patrice Chotardcafc3412017-07-25 13:24:45 +0200426{
Eugen Hristev70e32ba2023-06-19 13:47:52 +0300427 unsigned int i;
428 int ret;
Patrice Chotardcafc3412017-07-25 13:24:45 +0200429
430 for (i = 0; i < count; i++) {
Eugen Hristev70e32ba2023-06-19 13:47:52 +0300431 debug("%s(clk[%u]=%p)\n", __func__, i, &clk[i]);
Patrice Chotardcafc3412017-07-25 13:24:45 +0200432
433 /* check if clock has been previously requested */
434 if (!clk[i].dev)
435 continue;
436
437 ret = clk_disable(&clk[i]);
438 if (ret && ret != -ENOSYS)
439 return ret;
440
Sean Anderson9a665722022-01-15 17:25:04 -0500441 clk_free(&clk[i]);
Patrice Chotardcafc3412017-07-25 13:24:45 +0200442 }
443
444 return 0;
445}
446
Stephen Warrena9622432016-06-17 09:44:00 -0600447int clk_request(struct udevice *dev, struct clk *clk)
448{
Jean-Jacques Hiblot718039b2019-10-22 14:00:03 +0200449 const struct clk_ops *ops;
Stephen Warrena9622432016-06-17 09:44:00 -0600450
451 debug("%s(dev=%p, clk=%p)\n", __func__, dev, clk);
Jean-Jacques Hiblot718039b2019-10-22 14:00:03 +0200452 if (!clk)
453 return 0;
454 ops = clk_dev_ops(dev);
Stephen Warrena9622432016-06-17 09:44:00 -0600455
456 clk->dev = dev;
457
458 if (!ops->request)
459 return 0;
460
461 return ops->request(clk);
462}
463
Sean Anderson9a665722022-01-15 17:25:04 -0500464void clk_free(struct clk *clk)
Stephen Warrena9622432016-06-17 09:44:00 -0600465{
Jean-Jacques Hiblot718039b2019-10-22 14:00:03 +0200466 const struct clk_ops *ops;
Stephen Warrena9622432016-06-17 09:44:00 -0600467
468 debug("%s(clk=%p)\n", __func__, clk);
developerdc338d32020-01-09 11:35:06 +0800469 if (!clk_valid(clk))
Sean Anderson9a665722022-01-15 17:25:04 -0500470 return;
Jean-Jacques Hiblot718039b2019-10-22 14:00:03 +0200471 ops = clk_dev_ops(clk->dev);
Stephen Warrena9622432016-06-17 09:44:00 -0600472
Sean Anderson553935f2022-01-15 17:24:58 -0500473 if (ops->rfree)
474 ops->rfree(clk);
Sean Anderson9a665722022-01-15 17:25:04 -0500475 return;
Stephen Warrena9622432016-06-17 09:44:00 -0600476}
477
478ulong clk_get_rate(struct clk *clk)
479{
Jean-Jacques Hiblot718039b2019-10-22 14:00:03 +0200480 const struct clk_ops *ops;
Stephen Warrena9622432016-06-17 09:44:00 -0600481
482 debug("%s(clk=%p)\n", __func__, clk);
developerdc338d32020-01-09 11:35:06 +0800483 if (!clk_valid(clk))
Jean-Jacques Hiblot718039b2019-10-22 14:00:03 +0200484 return 0;
485 ops = clk_dev_ops(clk->dev);
Stephen Warrena9622432016-06-17 09:44:00 -0600486
487 if (!ops->get_rate)
488 return -ENOSYS;
489
Julien Massonb5de0b92023-12-15 15:09:43 +0100490 return ops->get_rate(clk);
Stephen Warrena9622432016-06-17 09:44:00 -0600491}
492
Lukasz Majewski9e38dc32019-06-24 15:50:42 +0200493struct clk *clk_get_parent(struct clk *clk)
494{
495 struct udevice *pdev;
496 struct clk *pclk;
497
498 debug("%s(clk=%p)\n", __func__, clk);
developerdc338d32020-01-09 11:35:06 +0800499 if (!clk_valid(clk))
Jean-Jacques Hiblot718039b2019-10-22 14:00:03 +0200500 return NULL;
Lukasz Majewski9e38dc32019-06-24 15:50:42 +0200501
502 pdev = dev_get_parent(clk->dev);
Tero Kristof04dfff2021-06-11 11:45:08 +0300503 if (!pdev)
504 return ERR_PTR(-ENODEV);
Lukasz Majewski9e38dc32019-06-24 15:50:42 +0200505 pclk = dev_get_clk_ptr(pdev);
506 if (!pclk)
507 return ERR_PTR(-ENODEV);
508
509 return pclk;
510}
511
Michal Suchanek0d4d5e42022-09-28 12:37:57 +0200512ulong clk_get_parent_rate(struct clk *clk)
Lukasz Majewski53155da2019-06-24 15:50:43 +0200513{
514 const struct clk_ops *ops;
515 struct clk *pclk;
516
517 debug("%s(clk=%p)\n", __func__, clk);
developerdc338d32020-01-09 11:35:06 +0800518 if (!clk_valid(clk))
Jean-Jacques Hiblot718039b2019-10-22 14:00:03 +0200519 return 0;
Lukasz Majewski53155da2019-06-24 15:50:43 +0200520
521 pclk = clk_get_parent(clk);
522 if (IS_ERR(pclk))
523 return -ENODEV;
524
525 ops = clk_dev_ops(pclk->dev);
526 if (!ops->get_rate)
527 return -ENOSYS;
528
Lukasz Majewski4ef32172019-06-24 15:50:46 +0200529 /* Read the 'rate' if not already set or if proper flag set*/
530 if (!pclk->rate || pclk->flags & CLK_GET_RATE_NOCACHE)
Lukasz Majewski53155da2019-06-24 15:50:43 +0200531 pclk->rate = clk_get_rate(pclk);
532
533 return pclk->rate;
534}
535
Dario Binacchib7f85892020-12-30 00:06:31 +0100536ulong clk_round_rate(struct clk *clk, ulong rate)
537{
538 const struct clk_ops *ops;
539
540 debug("%s(clk=%p, rate=%lu)\n", __func__, clk, rate);
541 if (!clk_valid(clk))
542 return 0;
543
544 ops = clk_dev_ops(clk->dev);
545 if (!ops->round_rate)
546 return -ENOSYS;
547
548 return ops->round_rate(clk, rate);
549}
550
Patrick Delaunayd867a2872022-06-20 15:37:25 +0200551static void clk_get_priv(struct clk *clk, struct clk **clkp)
552{
553 *clkp = clk;
554
555 /* get private clock struct associated to the provided clock */
556 if (CONFIG_IS_ENABLED(CLK_CCF)) {
557 /* Take id 0 as a non-valid clk, such as dummy */
558 if (clk->id)
559 clk_get_by_id(clk->id, clkp);
560 }
561}
562
563/* clean cache, called with private clock struct */
Tero Kristo9ab78c12021-06-11 11:45:12 +0300564static void clk_clean_rate_cache(struct clk *clk)
565{
566 struct udevice *child_dev;
567 struct clk *clkp;
568
569 if (!clk)
570 return;
571
572 clk->rate = 0;
573
574 list_for_each_entry(child_dev, &clk->dev->child_head, sibling_node) {
575 clkp = dev_get_clk_ptr(child_dev);
576 clk_clean_rate_cache(clkp);
577 }
578}
579
Stephen Warrena9622432016-06-17 09:44:00 -0600580ulong clk_set_rate(struct clk *clk, ulong rate)
581{
Jean-Jacques Hiblot718039b2019-10-22 14:00:03 +0200582 const struct clk_ops *ops;
Patrick Delaunayd867a2872022-06-20 15:37:25 +0200583 struct clk *clkp;
Stephen Warrena9622432016-06-17 09:44:00 -0600584
585 debug("%s(clk=%p, rate=%lu)\n", __func__, clk, rate);
developerdc338d32020-01-09 11:35:06 +0800586 if (!clk_valid(clk))
Jean-Jacques Hiblot718039b2019-10-22 14:00:03 +0200587 return 0;
588 ops = clk_dev_ops(clk->dev);
Stephen Warrena9622432016-06-17 09:44:00 -0600589
590 if (!ops->set_rate)
591 return -ENOSYS;
592
Patrick Delaunayd867a2872022-06-20 15:37:25 +0200593 /* get private clock struct used for cache */
594 clk_get_priv(clk, &clkp);
Tero Kristo9ab78c12021-06-11 11:45:12 +0300595 /* Clean up cached rates for us and all child clocks */
Patrick Delaunayd867a2872022-06-20 15:37:25 +0200596 clk_clean_rate_cache(clkp);
Tero Kristo9ab78c12021-06-11 11:45:12 +0300597
Stephen Warrena9622432016-06-17 09:44:00 -0600598 return ops->set_rate(clk, rate);
599}
600
Philipp Tomsichf8e02b22018-01-08 11:15:08 +0100601int clk_set_parent(struct clk *clk, struct clk *parent)
602{
Jean-Jacques Hiblot718039b2019-10-22 14:00:03 +0200603 const struct clk_ops *ops;
Claudiu Bezneac8c16002020-09-07 17:46:34 +0300604 int ret;
Philipp Tomsichf8e02b22018-01-08 11:15:08 +0100605
606 debug("%s(clk=%p, parent=%p)\n", __func__, clk, parent);
developerdc338d32020-01-09 11:35:06 +0800607 if (!clk_valid(clk))
Jean-Jacques Hiblot718039b2019-10-22 14:00:03 +0200608 return 0;
609 ops = clk_dev_ops(clk->dev);
Philipp Tomsichf8e02b22018-01-08 11:15:08 +0100610
611 if (!ops->set_parent)
612 return -ENOSYS;
613
Claudiu Bezneac8c16002020-09-07 17:46:34 +0300614 ret = ops->set_parent(clk, parent);
615 if (ret)
616 return ret;
617
618 if (CONFIG_IS_ENABLED(CLK_CCF))
619 ret = device_reparent(clk->dev, parent->dev);
620
621 return ret;
Philipp Tomsichf8e02b22018-01-08 11:15:08 +0100622}
623
Stephen Warrena9622432016-06-17 09:44:00 -0600624int clk_enable(struct clk *clk)
625{
Jean-Jacques Hiblot718039b2019-10-22 14:00:03 +0200626 const struct clk_ops *ops;
Peng Fan82628e22019-08-21 13:35:09 +0000627 struct clk *clkp = NULL;
628 int ret;
Stephen Warrena9622432016-06-17 09:44:00 -0600629
630 debug("%s(clk=%p)\n", __func__, clk);
developerdc338d32020-01-09 11:35:06 +0800631 if (!clk_valid(clk))
Jean-Jacques Hiblot718039b2019-10-22 14:00:03 +0200632 return 0;
633 ops = clk_dev_ops(clk->dev);
Stephen Warrena9622432016-06-17 09:44:00 -0600634
Peng Fan82628e22019-08-21 13:35:09 +0000635 if (CONFIG_IS_ENABLED(CLK_CCF)) {
636 /* Take id 0 as a non-valid clk, such as dummy */
637 if (clk->id && !clk_get_by_id(clk->id, &clkp)) {
Yang Xiwencb34b1c2023-11-19 06:10:06 +0800638 ops = clk_dev_ops(clkp->dev);
Peng Fan82628e22019-08-21 13:35:09 +0000639 if (clkp->enable_count) {
640 clkp->enable_count++;
641 return 0;
642 }
643 if (clkp->dev->parent &&
Patrick Delaunaydb9b1a12022-01-24 14:17:14 +0100644 device_get_uclass_id(clkp->dev->parent) == UCLASS_CLK) {
Peng Fan82628e22019-08-21 13:35:09 +0000645 ret = clk_enable(dev_get_clk_ptr(clkp->dev->parent));
646 if (ret) {
647 printf("Enable %s failed\n",
648 clkp->dev->parent->name);
649 return ret;
650 }
651 }
652 }
Stephen Warrena9622432016-06-17 09:44:00 -0600653
Peng Fan82628e22019-08-21 13:35:09 +0000654 if (ops->enable) {
Maksim Kiselev77e11512023-09-06 01:16:49 +0300655 ret = ops->enable(clkp ? clkp : clk);
Peng Fan82628e22019-08-21 13:35:09 +0000656 if (ret) {
657 printf("Enable %s failed\n", clk->dev->name);
658 return ret;
659 }
660 }
661 if (clkp)
662 clkp->enable_count++;
663 } else {
664 if (!ops->enable)
665 return -ENOSYS;
666 return ops->enable(clk);
667 }
668
669 return 0;
Stephen Warrena9622432016-06-17 09:44:00 -0600670}
671
Neil Armstrong8a275a02018-04-03 11:44:18 +0200672int clk_enable_bulk(struct clk_bulk *bulk)
673{
674 int i, ret;
675
676 for (i = 0; i < bulk->count; i++) {
677 ret = clk_enable(&bulk->clks[i]);
678 if (ret < 0 && ret != -ENOSYS)
679 return ret;
680 }
681
682 return 0;
683}
684
Stephen Warrena9622432016-06-17 09:44:00 -0600685int clk_disable(struct clk *clk)
686{
Jean-Jacques Hiblot718039b2019-10-22 14:00:03 +0200687 const struct clk_ops *ops;
Peng Fan82628e22019-08-21 13:35:09 +0000688 struct clk *clkp = NULL;
689 int ret;
Stephen Warrena9622432016-06-17 09:44:00 -0600690
691 debug("%s(clk=%p)\n", __func__, clk);
developerdc338d32020-01-09 11:35:06 +0800692 if (!clk_valid(clk))
Jean-Jacques Hiblot718039b2019-10-22 14:00:03 +0200693 return 0;
694 ops = clk_dev_ops(clk->dev);
Stephen Warrena9622432016-06-17 09:44:00 -0600695
Peng Fan82628e22019-08-21 13:35:09 +0000696 if (CONFIG_IS_ENABLED(CLK_CCF)) {
697 if (clk->id && !clk_get_by_id(clk->id, &clkp)) {
Yang Xiwencb34b1c2023-11-19 06:10:06 +0800698 ops = clk_dev_ops(clkp->dev);
Claudiu Bezneab02e8dd2020-09-07 17:46:35 +0300699 if (clkp->flags & CLK_IS_CRITICAL)
700 return 0;
701
Peng Fan82628e22019-08-21 13:35:09 +0000702 if (clkp->enable_count == 0) {
703 printf("clk %s already disabled\n",
704 clkp->dev->name);
705 return 0;
706 }
Stephen Warrena9622432016-06-17 09:44:00 -0600707
Peng Fan82628e22019-08-21 13:35:09 +0000708 if (--clkp->enable_count > 0)
709 return 0;
710 }
711
712 if (ops->disable) {
Maksim Kiselev77e11512023-09-06 01:16:49 +0300713 ret = ops->disable(clkp ? clkp : clk);
Peng Fan82628e22019-08-21 13:35:09 +0000714 if (ret)
715 return ret;
716 }
717
718 if (clkp && clkp->dev->parent &&
Patrick Delaunaydb9b1a12022-01-24 14:17:14 +0100719 device_get_uclass_id(clkp->dev->parent) == UCLASS_CLK) {
Peng Fan82628e22019-08-21 13:35:09 +0000720 ret = clk_disable(dev_get_clk_ptr(clkp->dev->parent));
721 if (ret) {
722 printf("Disable %s failed\n",
723 clkp->dev->parent->name);
724 return ret;
725 }
726 }
727 } else {
728 if (!ops->disable)
729 return -ENOSYS;
730
731 return ops->disable(clk);
732 }
733
734 return 0;
Stephen Warrena9622432016-06-17 09:44:00 -0600735}
Simon Glass36ad2342015-06-23 15:39:15 -0600736
Neil Armstrong8a275a02018-04-03 11:44:18 +0200737int clk_disable_bulk(struct clk_bulk *bulk)
738{
739 int i, ret;
740
741 for (i = 0; i < bulk->count; i++) {
742 ret = clk_disable(&bulk->clks[i]);
743 if (ret < 0 && ret != -ENOSYS)
744 return ret;
745 }
746
747 return 0;
748}
749
Lukasz Majewski12014be2019-06-24 15:50:44 +0200750int clk_get_by_id(ulong id, struct clk **clkp)
751{
752 struct udevice *dev;
753 struct uclass *uc;
754 int ret;
755
756 ret = uclass_get(UCLASS_CLK, &uc);
757 if (ret)
758 return ret;
759
760 uclass_foreach_dev(dev, uc) {
761 struct clk *clk = dev_get_clk_ptr(dev);
762
763 if (clk && clk->id == id) {
764 *clkp = clk;
765 return 0;
766 }
767 }
768
769 return -ENOENT;
770}
771
Sekhar Noricf3119d2019-08-01 19:12:55 +0530772bool clk_is_match(const struct clk *p, const struct clk *q)
773{
774 /* trivial case: identical struct clk's or both NULL */
775 if (p == q)
776 return true;
777
Jean-Jacques Hiblot718039b2019-10-22 14:00:03 +0200778 /* trivial case #2: on the clk pointer is NULL */
779 if (!p || !q)
780 return false;
781
Sekhar Noricf3119d2019-08-01 19:12:55 +0530782 /* same device, id and data */
783 if (p->dev == q->dev && p->id == q->id && p->data == q->data)
784 return true;
785
786 return false;
787}
788
Jean-Jacques Hiblot6e66b2d2019-10-22 14:00:04 +0200789static void devm_clk_release(struct udevice *dev, void *res)
790{
791 clk_free(res);
792}
793
794static int devm_clk_match(struct udevice *dev, void *res, void *data)
795{
796 return res == data;
797}
798
799struct clk *devm_clk_get(struct udevice *dev, const char *id)
800{
801 int rc;
802 struct clk *clk;
803
804 clk = devres_alloc(devm_clk_release, sizeof(struct clk), __GFP_ZERO);
805 if (unlikely(!clk))
806 return ERR_PTR(-ENOMEM);
807
808 rc = clk_get_by_name(dev, id, clk);
809 if (rc)
810 return ERR_PTR(rc);
811
812 devres_add(dev, clk);
813 return clk;
814}
815
Jean-Jacques Hiblot6e66b2d2019-10-22 14:00:04 +0200816void devm_clk_put(struct udevice *dev, struct clk *clk)
817{
818 int rc;
819
820 if (!clk)
821 return;
822
823 rc = devres_release(dev, devm_clk_release, devm_clk_match, clk);
824 WARN_ON(rc);
825}
826
Jean-Jacques Hiblot9601f322019-10-22 14:00:06 +0200827int clk_uclass_post_probe(struct udevice *dev)
828{
829 /*
830 * when a clock provider is probed. Call clk_set_defaults()
831 * also after the device is probed. This takes care of cases
832 * where the DT is used to setup default parents and rates
833 * using assigned-clocks
834 */
Marek Vasut05e3d8e2022-01-01 19:51:39 +0100835 clk_set_defaults(dev, CLK_DEFAULTS_POST);
Jean-Jacques Hiblot9601f322019-10-22 14:00:06 +0200836
837 return 0;
838}
839
Simon Glass36ad2342015-06-23 15:39:15 -0600840UCLASS_DRIVER(clk) = {
841 .id = UCLASS_CLK,
842 .name = "clk",
Jean-Jacques Hiblot9601f322019-10-22 14:00:06 +0200843 .post_probe = clk_uclass_post_probe,
Simon Glass36ad2342015-06-23 15:39:15 -0600844};