1 // SPDX-License-Identifier: GPL-2.0+
3 * Copyright (C) 2015 Google, Inc
4 * Written by Simon Glass <sjg@chromium.org>
5 * Copyright (c) 2016, NVIDIA CORPORATION.
6 * Copyright (c) 2018, Theobroma Systems Design und Consulting GmbH
9 #define LOG_CATEGORY UCLASS_CLK
13 #include <clk-uclass.h>
15 #include <dt-structs.h>
19 #include <asm/global_data.h>
20 #include <dm/device_compat.h>
21 #include <dm/device-internal.h>
22 #include <dm/devres.h>
24 #include <linux/bug.h>
25 #include <linux/clk-provider.h>
26 #include <linux/err.h>
28 static inline const struct clk_ops *clk_dev_ops(struct udevice *dev)
30 return (const struct clk_ops *)dev->driver->ops;
33 struct clk *dev_get_clk_ptr(struct udevice *dev)
35 return (struct clk *)dev_get_uclass_priv(dev);
38 #if CONFIG_IS_ENABLED(OF_PLATDATA)
39 int clk_get_by_phandle(struct udevice *dev, const struct phandle_1_arg *cells,
44 ret = device_get_by_ofplat_idx(cells->idx, &clk->dev);
47 clk->id = cells->arg[0];
53 #if CONFIG_IS_ENABLED(OF_REAL)
54 static int clk_of_xlate_default(struct clk *clk,
55 struct ofnode_phandle_args *args)
57 debug("%s(clk=%p)\n", __func__, clk);
59 if (args->args_count > 1) {
60 debug("Invalid args_count: %d\n", args->args_count);
65 clk->id = args->args[0];
74 static int clk_get_by_index_tail(int ret, ofnode node,
75 struct ofnode_phandle_args *args,
76 const char *list_name, int index,
79 struct udevice *dev_clk;
80 const struct clk_ops *ops;
87 ret = uclass_get_device_by_ofnode(UCLASS_CLK, args->node, &dev_clk);
89 debug("%s: uclass_get_device_by_of_offset failed: err=%d\n",
91 return log_msg_ret("get", ret);
96 ops = clk_dev_ops(dev_clk);
99 ret = ops->of_xlate(clk, args);
101 ret = clk_of_xlate_default(clk, args);
103 debug("of_xlate() failed: %d\n", ret);
104 return log_msg_ret("xlate", ret);
107 return clk_request(dev_clk, clk);
109 debug("%s: Node '%s', property '%s', failed to request CLK index %d: %d\n",
110 __func__, ofnode_get_name(node), list_name, index, ret);
112 return log_msg_ret("prop", ret);
115 static int clk_get_by_indexed_prop(struct udevice *dev, const char *prop_name,
116 int index, struct clk *clk)
119 struct ofnode_phandle_args args;
121 debug("%s(dev=%p, index=%d, clk=%p)\n", __func__, dev, index, clk);
126 ret = dev_read_phandle_with_args(dev, prop_name, "#clock-cells", 0,
129 debug("%s: fdtdec_parse_phandle_with_args failed: err=%d\n",
135 return clk_get_by_index_tail(ret, dev_ofnode(dev), &args, "clocks",
139 int clk_get_by_index(struct udevice *dev, int index, struct clk *clk)
141 return clk_get_by_index_nodev(dev_ofnode(dev), index, clk);
144 int clk_get_by_index_nodev(ofnode node, int index, struct clk *clk)
146 struct ofnode_phandle_args args;
149 ret = ofnode_parse_phandle_with_args(node, "clocks", "#clock-cells", 0,
152 return clk_get_by_index_tail(ret, node, &args, "clocks",
156 int clk_get_bulk(struct udevice *dev, struct clk_bulk *bulk)
158 int i, ret, err, count;
162 count = dev_count_phandle_with_args(dev, "clocks", "#clock-cells", 0);
166 bulk->clks = devm_kcalloc(dev, count, sizeof(struct clk), GFP_KERNEL);
170 for (i = 0; i < count; i++) {
171 ret = clk_get_by_index(dev, i, &bulk->clks[i]);
181 err = clk_release_all(bulk->clks, bulk->count);
183 debug("%s: could release all clocks for %p\n",
189 static struct clk *clk_set_default_get_by_id(struct clk *clk)
193 if (CONFIG_IS_ENABLED(CLK_CCF)) {
194 int ret = clk_get_by_id(clk->id, &c);
197 debug("%s(): could not get parent clock pointer, id %lu\n",
206 static int clk_set_default_parents(struct udevice *dev,
207 enum clk_defaults_stage stage)
209 struct clk clk, parent_clk, *c, *p;
214 num_parents = dev_count_phandle_with_args(dev, "assigned-clock-parents",
216 if (num_parents < 0) {
217 debug("%s: could not read assigned-clock-parents for %p\n",
222 for (index = 0; index < num_parents; index++) {
223 ret = clk_get_by_indexed_prop(dev, "assigned-clock-parents",
225 /* If -ENOENT, this is a no-op entry */
230 debug("%s: could not get parent clock %d for %s\n",
231 __func__, index, dev_read_name(dev));
235 p = clk_set_default_get_by_id(&parent_clk);
239 ret = clk_get_by_indexed_prop(dev, "assigned-clocks",
242 * If the clock provider is not ready yet, let it handle
243 * the re-programming later.
245 if (ret == -EPROBE_DEFER) {
251 debug("%s: could not get assigned clock %d for %s\n",
252 __func__, index, dev_read_name(dev));
256 /* This is clk provider device trying to reparent itself
257 * It cannot be done right now but need to wait after the
260 if (stage == CLK_DEFAULTS_PRE && clk.dev == dev)
263 if (stage != CLK_DEFAULTS_PRE && clk.dev != dev)
264 /* do not setup twice the parent clocks */
267 c = clk_set_default_get_by_id(&clk);
271 ret = clk_set_parent(c, p);
273 * Not all drivers may support clock-reparenting (as of now).
274 * Ignore errors due to this.
280 debug("%s: failed to reparent clock %d for %s\n",
281 __func__, index, dev_read_name(dev));
289 static int clk_set_default_rates(struct udevice *dev,
290 enum clk_defaults_stage stage)
299 size = dev_read_size(dev, "assigned-clock-rates");
303 num_rates = size / sizeof(u32);
304 rates = calloc(num_rates, sizeof(u32));
308 ret = dev_read_u32_array(dev, "assigned-clock-rates", rates, num_rates);
312 for (index = 0; index < num_rates; index++) {
313 /* If 0 is passed, this is a no-op */
317 ret = clk_get_by_indexed_prop(dev, "assigned-clocks",
320 * If the clock provider is not ready yet, let it handle
321 * the re-programming later.
323 if (ret == -EPROBE_DEFER) {
330 "could not get assigned clock %d (err = %d)\n",
335 /* This is clk provider device trying to program itself
336 * It cannot be done right now but need to wait after the
339 if (stage == CLK_DEFAULTS_PRE && clk.dev == dev)
342 if (stage != CLK_DEFAULTS_PRE && clk.dev != dev)
343 /* do not setup twice the parent clocks */
346 c = clk_set_default_get_by_id(&clk);
350 ret = clk_set_rate(c, rates[index]);
354 "failed to set rate on clock index %d (%ld) (error = %d)\n",
365 int clk_set_defaults(struct udevice *dev, enum clk_defaults_stage stage)
369 if (!dev_has_ofnode(dev))
373 * To avoid setting defaults twice, don't set them before relocation.
374 * However, still set them for SPL. And still set them if explicitly
377 if (!(IS_ENABLED(CONFIG_SPL_BUILD) || (gd->flags & GD_FLG_RELOC)))
378 if (stage != CLK_DEFAULTS_POST_FORCE)
381 debug("%s(%s)\n", __func__, dev_read_name(dev));
383 ret = clk_set_default_parents(dev, stage);
387 ret = clk_set_default_rates(dev, stage);
394 int clk_get_by_name(struct udevice *dev, const char *name, struct clk *clk)
396 return clk_get_by_name_nodev(dev_ofnode(dev), name, clk);
400 int clk_get_by_name_nodev(ofnode node, const char *name, struct clk *clk)
404 debug("%s(node=%p, name=%s, clk=%p)\n", __func__,
405 ofnode_get_name(node), name, clk);
409 index = ofnode_stringlist_search(node, "clock-names", name);
411 debug("fdt_stringlist_search() failed: %d\n", index);
416 return clk_get_by_index_nodev(node, index, clk);
419 int clk_release_all(struct clk *clk, int count)
423 for (i = 0; i < count; i++) {
424 debug("%s(clk[%d]=%p)\n", __func__, i, &clk[i]);
426 /* check if clock has been previously requested */
430 ret = clk_disable(&clk[i]);
431 if (ret && ret != -ENOSYS)
440 int clk_request(struct udevice *dev, struct clk *clk)
442 const struct clk_ops *ops;
444 debug("%s(dev=%p, clk=%p)\n", __func__, dev, clk);
447 ops = clk_dev_ops(dev);
454 return ops->request(clk);
457 void clk_free(struct clk *clk)
459 const struct clk_ops *ops;
461 debug("%s(clk=%p)\n", __func__, clk);
464 ops = clk_dev_ops(clk->dev);
471 ulong clk_get_rate(struct clk *clk)
473 const struct clk_ops *ops;
476 debug("%s(clk=%p)\n", __func__, clk);
479 ops = clk_dev_ops(clk->dev);
484 ret = ops->get_rate(clk);
491 struct clk *clk_get_parent(struct clk *clk)
493 struct udevice *pdev;
496 debug("%s(clk=%p)\n", __func__, clk);
500 pdev = dev_get_parent(clk->dev);
502 return ERR_PTR(-ENODEV);
503 pclk = dev_get_clk_ptr(pdev);
505 return ERR_PTR(-ENODEV);
510 ulong clk_get_parent_rate(struct clk *clk)
512 const struct clk_ops *ops;
515 debug("%s(clk=%p)\n", __func__, clk);
519 pclk = clk_get_parent(clk);
523 ops = clk_dev_ops(pclk->dev);
527 /* Read the 'rate' if not already set or if proper flag set*/
528 if (!pclk->rate || pclk->flags & CLK_GET_RATE_NOCACHE)
529 pclk->rate = clk_get_rate(pclk);
534 ulong clk_round_rate(struct clk *clk, ulong rate)
536 const struct clk_ops *ops;
538 debug("%s(clk=%p, rate=%lu)\n", __func__, clk, rate);
542 ops = clk_dev_ops(clk->dev);
543 if (!ops->round_rate)
546 return ops->round_rate(clk, rate);
549 static void clk_get_priv(struct clk *clk, struct clk **clkp)
553 /* get private clock struct associated to the provided clock */
554 if (CONFIG_IS_ENABLED(CLK_CCF)) {
555 /* Take id 0 as a non-valid clk, such as dummy */
557 clk_get_by_id(clk->id, clkp);
561 /* clean cache, called with private clock struct */
562 static void clk_clean_rate_cache(struct clk *clk)
564 struct udevice *child_dev;
572 list_for_each_entry(child_dev, &clk->dev->child_head, sibling_node) {
573 clkp = dev_get_clk_ptr(child_dev);
574 clk_clean_rate_cache(clkp);
578 ulong clk_set_rate(struct clk *clk, ulong rate)
580 const struct clk_ops *ops;
583 debug("%s(clk=%p, rate=%lu)\n", __func__, clk, rate);
586 ops = clk_dev_ops(clk->dev);
591 /* get private clock struct used for cache */
592 clk_get_priv(clk, &clkp);
593 /* Clean up cached rates for us and all child clocks */
594 clk_clean_rate_cache(clkp);
596 return ops->set_rate(clk, rate);
599 int clk_set_parent(struct clk *clk, struct clk *parent)
601 const struct clk_ops *ops;
604 debug("%s(clk=%p, parent=%p)\n", __func__, clk, parent);
607 ops = clk_dev_ops(clk->dev);
609 if (!ops->set_parent)
612 ret = ops->set_parent(clk, parent);
616 if (CONFIG_IS_ENABLED(CLK_CCF))
617 ret = device_reparent(clk->dev, parent->dev);
622 int clk_enable(struct clk *clk)
624 const struct clk_ops *ops;
625 struct clk *clkp = NULL;
628 debug("%s(clk=%p)\n", __func__, clk);
631 ops = clk_dev_ops(clk->dev);
633 if (CONFIG_IS_ENABLED(CLK_CCF)) {
634 /* Take id 0 as a non-valid clk, such as dummy */
635 if (clk->id && !clk_get_by_id(clk->id, &clkp)) {
636 if (clkp->enable_count) {
637 clkp->enable_count++;
640 if (clkp->dev->parent &&
641 device_get_uclass_id(clkp->dev->parent) == UCLASS_CLK) {
642 ret = clk_enable(dev_get_clk_ptr(clkp->dev->parent));
644 printf("Enable %s failed\n",
645 clkp->dev->parent->name);
652 ret = ops->enable(clk);
654 printf("Enable %s failed\n", clk->dev->name);
659 clkp->enable_count++;
663 return ops->enable(clk);
669 int clk_enable_bulk(struct clk_bulk *bulk)
673 for (i = 0; i < bulk->count; i++) {
674 ret = clk_enable(&bulk->clks[i]);
675 if (ret < 0 && ret != -ENOSYS)
682 int clk_disable(struct clk *clk)
684 const struct clk_ops *ops;
685 struct clk *clkp = NULL;
688 debug("%s(clk=%p)\n", __func__, clk);
691 ops = clk_dev_ops(clk->dev);
693 if (CONFIG_IS_ENABLED(CLK_CCF)) {
694 if (clk->id && !clk_get_by_id(clk->id, &clkp)) {
695 if (clkp->flags & CLK_IS_CRITICAL)
698 if (clkp->enable_count == 0) {
699 printf("clk %s already disabled\n",
704 if (--clkp->enable_count > 0)
709 ret = ops->disable(clk);
714 if (clkp && clkp->dev->parent &&
715 device_get_uclass_id(clkp->dev->parent) == UCLASS_CLK) {
716 ret = clk_disable(dev_get_clk_ptr(clkp->dev->parent));
718 printf("Disable %s failed\n",
719 clkp->dev->parent->name);
727 return ops->disable(clk);
733 int clk_disable_bulk(struct clk_bulk *bulk)
737 for (i = 0; i < bulk->count; i++) {
738 ret = clk_disable(&bulk->clks[i]);
739 if (ret < 0 && ret != -ENOSYS)
746 int clk_get_by_id(ulong id, struct clk **clkp)
752 ret = uclass_get(UCLASS_CLK, &uc);
756 uclass_foreach_dev(dev, uc) {
757 struct clk *clk = dev_get_clk_ptr(dev);
759 if (clk && clk->id == id) {
768 bool clk_is_match(const struct clk *p, const struct clk *q)
770 /* trivial case: identical struct clk's or both NULL */
774 /* trivial case #2: on the clk pointer is NULL */
778 /* same device, id and data */
779 if (p->dev == q->dev && p->id == q->id && p->data == q->data)
785 static void devm_clk_release(struct udevice *dev, void *res)
790 static int devm_clk_match(struct udevice *dev, void *res, void *data)
795 struct clk *devm_clk_get(struct udevice *dev, const char *id)
800 clk = devres_alloc(devm_clk_release, sizeof(struct clk), __GFP_ZERO);
802 return ERR_PTR(-ENOMEM);
804 rc = clk_get_by_name(dev, id, clk);
808 devres_add(dev, clk);
812 void devm_clk_put(struct udevice *dev, struct clk *clk)
819 rc = devres_release(dev, devm_clk_release, devm_clk_match, clk);
823 int clk_uclass_post_probe(struct udevice *dev)
826 * when a clock provider is probed. Call clk_set_defaults()
827 * also after the device is probed. This takes care of cases
828 * where the DT is used to setup default parents and rates
829 * using assigned-clocks
831 clk_set_defaults(dev, CLK_DEFAULTS_POST);
836 UCLASS_DRIVER(clk) = {
839 .post_probe = clk_uclass_post_probe,