2018-05-06 21:58:06 +00:00
|
|
|
// SPDX-License-Identifier: GPL-2.0+
|
2015-06-23 21:39:15 +00:00
|
|
|
/*
|
|
|
|
* Copyright (C) 2015 Google, Inc
|
|
|
|
* Written by Simon Glass <sjg@chromium.org>
|
2016-06-17 15:44:00 +00:00
|
|
|
* Copyright (c) 2016, NVIDIA CORPORATION.
|
2018-01-08 12:59:18 +00:00
|
|
|
* Copyright (c) 2018, Theobroma Systems Design und Consulting GmbH
|
2015-06-23 21:39:15 +00:00
|
|
|
*/
|
|
|
|
|
2021-04-27 09:02:19 +00:00
|
|
|
#define LOG_CATEGORY UCLASS_CLK
|
|
|
|
|
2015-06-23 21:39:15 +00:00
|
|
|
#include <common.h>
|
|
|
|
#include <clk.h>
|
2016-06-17 15:44:00 +00:00
|
|
|
#include <clk-uclass.h>
|
2015-06-23 21:39:15 +00:00
|
|
|
#include <dm.h>
|
2016-07-04 17:58:03 +00:00
|
|
|
#include <dt-structs.h>
|
2015-06-23 21:39:15 +00:00
|
|
|
#include <errno.h>
|
2020-05-10 17:40:05 +00:00
|
|
|
#include <log.h>
|
2020-02-03 14:36:16 +00:00
|
|
|
#include <malloc.h>
|
2021-11-19 14:12:06 +00:00
|
|
|
#include <asm/global_data.h>
|
2021-04-09 02:13:03 +00:00
|
|
|
#include <dm/device_compat.h>
|
2020-09-07 14:46:34 +00:00
|
|
|
#include <dm/device-internal.h>
|
2020-02-03 14:36:15 +00:00
|
|
|
#include <dm/devres.h>
|
|
|
|
#include <dm/read.h>
|
2020-05-10 17:40:08 +00:00
|
|
|
#include <linux/bug.h>
|
2019-06-24 13:50:42 +00:00
|
|
|
#include <linux/clk-provider.h>
|
2020-02-03 14:36:15 +00:00
|
|
|
#include <linux/err.h>
|
2015-06-23 21:39:15 +00:00
|
|
|
|
2018-01-15 10:06:51 +00:00
|
|
|
static inline const struct clk_ops *clk_dev_ops(struct udevice *dev)
|
2015-06-23 21:39:15 +00:00
|
|
|
{
|
2018-01-15 10:06:51 +00:00
|
|
|
return (const struct clk_ops *)dev->driver->ops;
|
2015-06-23 21:39:15 +00:00
|
|
|
}
|
|
|
|
|
2020-07-19 16:15:56 +00:00
|
|
|
struct clk *dev_get_clk_ptr(struct udevice *dev)
|
|
|
|
{
|
|
|
|
return (struct clk *)dev_get_uclass_priv(dev);
|
|
|
|
}
|
|
|
|
|
2021-08-07 13:24:03 +00:00
|
|
|
#if CONFIG_IS_ENABLED(OF_PLATDATA)
|
2021-08-07 13:24:09 +00:00
|
|
|
int clk_get_by_phandle(struct udevice *dev, const struct phandle_1_arg *cells,
|
|
|
|
struct clk *clk)
|
2016-07-04 17:58:03 +00:00
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
2021-03-15 04:25:28 +00:00
|
|
|
ret = device_get_by_ofplat_idx(cells->idx, &clk->dev);
|
2016-07-04 17:58:03 +00:00
|
|
|
if (ret)
|
|
|
|
return ret;
|
2020-06-25 04:10:13 +00:00
|
|
|
clk->id = cells->arg[0];
|
2016-07-04 17:58:03 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
2021-08-07 13:24:03 +00:00
|
|
|
#endif
|
|
|
|
|
|
|
|
#if CONFIG_IS_ENABLED(OF_REAL)
|
2016-06-17 15:44:00 +00:00
|
|
|
static int clk_of_xlate_default(struct clk *clk,
|
2017-05-19 02:09:40 +00:00
|
|
|
struct ofnode_phandle_args *args)
|
2015-06-23 21:39:15 +00:00
|
|
|
{
|
2016-06-17 15:44:00 +00:00
|
|
|
debug("%s(clk=%p)\n", __func__, clk);
|
2015-06-23 21:39:15 +00:00
|
|
|
|
2016-06-17 15:44:00 +00:00
|
|
|
if (args->args_count > 1) {
|
2021-12-01 19:26:53 +00:00
|
|
|
debug("Invalid args_count: %d\n", args->args_count);
|
2016-06-17 15:44:00 +00:00
|
|
|
return -EINVAL;
|
|
|
|
}
|
2015-06-23 21:39:15 +00:00
|
|
|
|
2016-06-17 15:44:00 +00:00
|
|
|
if (args->args_count)
|
|
|
|
clk->id = args->args[0];
|
|
|
|
else
|
|
|
|
clk->id = 0;
|
2015-06-23 21:39:15 +00:00
|
|
|
|
2019-07-11 09:00:24 +00:00
|
|
|
clk->data = 0;
|
|
|
|
|
2016-06-17 15:44:00 +00:00
|
|
|
return 0;
|
2015-06-23 21:39:15 +00:00
|
|
|
}
|
|
|
|
|
2019-02-27 18:56:52 +00:00
|
|
|
static int clk_get_by_index_tail(int ret, ofnode node,
|
|
|
|
struct ofnode_phandle_args *args,
|
|
|
|
const char *list_name, int index,
|
|
|
|
struct clk *clk)
|
|
|
|
{
|
|
|
|
struct udevice *dev_clk;
|
|
|
|
const struct clk_ops *ops;
|
|
|
|
|
|
|
|
assert(clk);
|
|
|
|
clk->dev = NULL;
|
|
|
|
if (ret)
|
|
|
|
goto err;
|
|
|
|
|
|
|
|
ret = uclass_get_device_by_ofnode(UCLASS_CLK, args->node, &dev_clk);
|
|
|
|
if (ret) {
|
|
|
|
debug("%s: uclass_get_device_by_of_offset failed: err=%d\n",
|
|
|
|
__func__, ret);
|
2021-01-21 20:57:11 +00:00
|
|
|
return log_msg_ret("get", ret);
|
2019-02-27 18:56:52 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
clk->dev = dev_clk;
|
|
|
|
|
|
|
|
ops = clk_dev_ops(dev_clk);
|
|
|
|
|
|
|
|
if (ops->of_xlate)
|
|
|
|
ret = ops->of_xlate(clk, args);
|
|
|
|
else
|
|
|
|
ret = clk_of_xlate_default(clk, args);
|
|
|
|
if (ret) {
|
|
|
|
debug("of_xlate() failed: %d\n", ret);
|
2021-01-21 20:57:11 +00:00
|
|
|
return log_msg_ret("xlate", ret);
|
2019-02-27 18:56:52 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return clk_request(dev_clk, clk);
|
|
|
|
err:
|
|
|
|
debug("%s: Node '%s', property '%s', failed to request CLK index %d: %d\n",
|
|
|
|
__func__, ofnode_get_name(node), list_name, index, ret);
|
2021-01-21 20:57:11 +00:00
|
|
|
|
|
|
|
return log_msg_ret("prop", ret);
|
2019-02-27 18:56:52 +00:00
|
|
|
}
|
|
|
|
|
2018-01-08 10:18:18 +00:00
|
|
|
static int clk_get_by_indexed_prop(struct udevice *dev, const char *prop_name,
|
|
|
|
int index, struct clk *clk)
|
2016-01-21 02:43:02 +00:00
|
|
|
{
|
|
|
|
int ret;
|
2017-05-31 03:47:29 +00:00
|
|
|
struct ofnode_phandle_args args;
|
2016-06-17 15:44:00 +00:00
|
|
|
|
|
|
|
debug("%s(dev=%p, index=%d, clk=%p)\n", __func__, dev, index, clk);
|
2016-01-21 02:43:02 +00:00
|
|
|
|
2016-06-17 15:44:00 +00:00
|
|
|
assert(clk);
|
2017-07-18 09:57:07 +00:00
|
|
|
clk->dev = NULL;
|
|
|
|
|
2018-01-08 10:18:18 +00:00
|
|
|
ret = dev_read_phandle_with_args(dev, prop_name, "#clock-cells", 0,
|
2018-01-15 10:06:51 +00:00
|
|
|
index, &args);
|
2016-01-21 02:43:02 +00:00
|
|
|
if (ret) {
|
|
|
|
debug("%s: fdtdec_parse_phandle_with_args failed: err=%d\n",
|
|
|
|
__func__, ret);
|
2021-01-21 20:57:11 +00:00
|
|
|
return log_ret(ret);
|
2016-01-21 02:43:02 +00:00
|
|
|
}
|
|
|
|
|
2016-06-17 15:44:00 +00:00
|
|
|
|
2019-02-27 18:56:53 +00:00
|
|
|
return clk_get_by_index_tail(ret, dev_ofnode(dev), &args, "clocks",
|
2020-06-24 10:41:08 +00:00
|
|
|
index, clk);
|
2016-06-17 15:44:00 +00:00
|
|
|
}
|
2018-01-08 10:18:18 +00:00
|
|
|
|
|
|
|
int clk_get_by_index(struct udevice *dev, int index, struct clk *clk)
|
|
|
|
{
|
2022-02-27 19:01:13 +00:00
|
|
|
return clk_get_by_index_nodev(dev_ofnode(dev), index, clk);
|
2019-02-27 18:56:52 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
int clk_get_by_index_nodev(ofnode node, int index, struct clk *clk)
|
|
|
|
{
|
|
|
|
struct ofnode_phandle_args args;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = ofnode_parse_phandle_with_args(node, "clocks", "#clock-cells", 0,
|
2020-06-24 10:41:08 +00:00
|
|
|
index, &args);
|
2019-02-27 18:56:52 +00:00
|
|
|
|
|
|
|
return clk_get_by_index_tail(ret, node, &args, "clocks",
|
2020-06-24 10:41:08 +00:00
|
|
|
index, clk);
|
2018-01-08 10:18:18 +00:00
|
|
|
}
|
2018-01-08 12:59:18 +00:00
|
|
|
|
2018-04-03 09:44:18 +00:00
|
|
|
int clk_get_bulk(struct udevice *dev, struct clk_bulk *bulk)
|
|
|
|
{
|
|
|
|
int i, ret, err, count;
|
2021-04-27 08:57:54 +00:00
|
|
|
|
2018-04-03 09:44:18 +00:00
|
|
|
bulk->count = 0;
|
|
|
|
|
2020-09-25 07:41:14 +00:00
|
|
|
count = dev_count_phandle_with_args(dev, "clocks", "#clock-cells", 0);
|
2018-04-17 09:30:31 +00:00
|
|
|
if (count < 1)
|
|
|
|
return count;
|
2018-04-03 09:44:18 +00:00
|
|
|
|
|
|
|
bulk->clks = devm_kcalloc(dev, count, sizeof(struct clk), GFP_KERNEL);
|
|
|
|
if (!bulk->clks)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
for (i = 0; i < count; i++) {
|
|
|
|
ret = clk_get_by_index(dev, i, &bulk->clks[i]);
|
|
|
|
if (ret < 0)
|
|
|
|
goto bulk_get_err;
|
|
|
|
|
|
|
|
++bulk->count;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
bulk_get_err:
|
|
|
|
err = clk_release_all(bulk->clks, bulk->count);
|
|
|
|
if (err)
|
|
|
|
debug("%s: could release all clocks for %p\n",
|
|
|
|
__func__, dev);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2020-09-07 14:46:36 +00:00
|
|
|
static struct clk *clk_set_default_get_by_id(struct clk *clk)
|
|
|
|
{
|
|
|
|
struct clk *c = clk;
|
|
|
|
|
|
|
|
if (CONFIG_IS_ENABLED(CLK_CCF)) {
|
|
|
|
int ret = clk_get_by_id(clk->id, &c);
|
|
|
|
|
|
|
|
if (ret) {
|
|
|
|
debug("%s(): could not get parent clock pointer, id %lu\n",
|
|
|
|
__func__, clk->id);
|
|
|
|
ERR_PTR(ret);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return c;
|
|
|
|
}
|
|
|
|
|
2021-06-11 04:16:07 +00:00
|
|
|
static int clk_set_default_parents(struct udevice *dev,
|
|
|
|
enum clk_defaults_stage stage)
|
2018-01-08 12:59:18 +00:00
|
|
|
{
|
2020-09-07 14:46:36 +00:00
|
|
|
struct clk clk, parent_clk, *c, *p;
|
2018-01-08 12:59:18 +00:00
|
|
|
int index;
|
|
|
|
int num_parents;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
num_parents = dev_count_phandle_with_args(dev, "assigned-clock-parents",
|
2020-09-25 07:41:14 +00:00
|
|
|
"#clock-cells", 0);
|
2018-01-08 12:59:18 +00:00
|
|
|
if (num_parents < 0) {
|
|
|
|
debug("%s: could not read assigned-clock-parents for %p\n",
|
|
|
|
__func__, dev);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (index = 0; index < num_parents; index++) {
|
|
|
|
ret = clk_get_by_indexed_prop(dev, "assigned-clock-parents",
|
|
|
|
index, &parent_clk);
|
2018-07-26 13:19:32 +00:00
|
|
|
/* If -ENOENT, this is a no-op entry */
|
|
|
|
if (ret == -ENOENT)
|
|
|
|
continue;
|
|
|
|
|
2018-01-08 12:59:18 +00:00
|
|
|
if (ret) {
|
|
|
|
debug("%s: could not get parent clock %d for %s\n",
|
|
|
|
__func__, index, dev_read_name(dev));
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2020-09-07 14:46:36 +00:00
|
|
|
p = clk_set_default_get_by_id(&parent_clk);
|
|
|
|
if (IS_ERR(p))
|
|
|
|
return PTR_ERR(p);
|
|
|
|
|
2018-01-08 12:59:18 +00:00
|
|
|
ret = clk_get_by_indexed_prop(dev, "assigned-clocks",
|
|
|
|
index, &clk);
|
2021-06-11 08:45:11 +00:00
|
|
|
/*
|
|
|
|
* If the clock provider is not ready yet, let it handle
|
|
|
|
* the re-programming later.
|
|
|
|
*/
|
|
|
|
if (ret == -EPROBE_DEFER) {
|
|
|
|
ret = 0;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
2018-01-08 12:59:18 +00:00
|
|
|
if (ret) {
|
|
|
|
debug("%s: could not get assigned clock %d for %s\n",
|
|
|
|
__func__, index, dev_read_name(dev));
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2019-10-22 12:00:06 +00:00
|
|
|
/* This is clk provider device trying to reparent itself
|
|
|
|
* It cannot be done right now but need to wait after the
|
|
|
|
* device is probed
|
|
|
|
*/
|
2021-06-11 04:16:07 +00:00
|
|
|
if (stage == CLK_DEFAULTS_PRE && clk.dev == dev)
|
2019-10-22 12:00:06 +00:00
|
|
|
continue;
|
|
|
|
|
2021-06-11 04:16:07 +00:00
|
|
|
if (stage != CLK_DEFAULTS_PRE && clk.dev != dev)
|
2019-10-22 12:00:06 +00:00
|
|
|
/* do not setup twice the parent clocks */
|
|
|
|
continue;
|
2018-01-08 12:59:18 +00:00
|
|
|
|
2020-09-07 14:46:36 +00:00
|
|
|
c = clk_set_default_get_by_id(&clk);
|
|
|
|
if (IS_ERR(c))
|
|
|
|
return PTR_ERR(c);
|
|
|
|
|
|
|
|
ret = clk_set_parent(c, p);
|
2018-01-08 12:59:18 +00:00
|
|
|
/*
|
|
|
|
* Not all drivers may support clock-reparenting (as of now).
|
|
|
|
* Ignore errors due to this.
|
|
|
|
*/
|
|
|
|
if (ret == -ENOSYS)
|
|
|
|
continue;
|
|
|
|
|
2019-09-26 13:42:42 +00:00
|
|
|
if (ret < 0) {
|
2018-01-08 12:59:18 +00:00
|
|
|
debug("%s: failed to reparent clock %d for %s\n",
|
|
|
|
__func__, index, dev_read_name(dev));
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2021-06-11 04:16:07 +00:00
|
|
|
static int clk_set_default_rates(struct udevice *dev,
|
|
|
|
enum clk_defaults_stage stage)
|
2018-01-08 12:59:18 +00:00
|
|
|
{
|
2020-09-07 14:46:36 +00:00
|
|
|
struct clk clk, *c;
|
2018-01-08 12:59:18 +00:00
|
|
|
int index;
|
|
|
|
int num_rates;
|
|
|
|
int size;
|
|
|
|
int ret = 0;
|
|
|
|
u32 *rates = NULL;
|
|
|
|
|
|
|
|
size = dev_read_size(dev, "assigned-clock-rates");
|
|
|
|
if (size < 0)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
num_rates = size / sizeof(u32);
|
|
|
|
rates = calloc(num_rates, sizeof(u32));
|
|
|
|
if (!rates)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
ret = dev_read_u32_array(dev, "assigned-clock-rates", rates, num_rates);
|
|
|
|
if (ret)
|
|
|
|
goto fail;
|
|
|
|
|
|
|
|
for (index = 0; index < num_rates; index++) {
|
2018-07-26 13:19:32 +00:00
|
|
|
/* If 0 is passed, this is a no-op */
|
|
|
|
if (!rates[index])
|
|
|
|
continue;
|
|
|
|
|
2018-01-08 12:59:18 +00:00
|
|
|
ret = clk_get_by_indexed_prop(dev, "assigned-clocks",
|
|
|
|
index, &clk);
|
2021-06-11 08:45:11 +00:00
|
|
|
/*
|
|
|
|
* If the clock provider is not ready yet, let it handle
|
|
|
|
* the re-programming later.
|
|
|
|
*/
|
|
|
|
if (ret == -EPROBE_DEFER) {
|
|
|
|
ret = 0;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
2018-01-08 12:59:18 +00:00
|
|
|
if (ret) {
|
2021-04-09 02:13:03 +00:00
|
|
|
dev_dbg(dev,
|
|
|
|
"could not get assigned clock %d (err = %d)\n",
|
|
|
|
index, ret);
|
2023-08-30 08:31:42 +00:00
|
|
|
/* Skip if it is empty */
|
|
|
|
if (ret == -ENOENT) {
|
|
|
|
ret = 0;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
2018-01-08 12:59:18 +00:00
|
|
|
}
|
|
|
|
|
2019-10-22 12:00:06 +00:00
|
|
|
/* This is clk provider device trying to program itself
|
|
|
|
* It cannot be done right now but need to wait after the
|
|
|
|
* device is probed
|
|
|
|
*/
|
2021-06-11 04:16:07 +00:00
|
|
|
if (stage == CLK_DEFAULTS_PRE && clk.dev == dev)
|
2019-10-22 12:00:06 +00:00
|
|
|
continue;
|
|
|
|
|
2021-06-11 04:16:07 +00:00
|
|
|
if (stage != CLK_DEFAULTS_PRE && clk.dev != dev)
|
2019-10-22 12:00:06 +00:00
|
|
|
/* do not setup twice the parent clocks */
|
|
|
|
continue;
|
|
|
|
|
2020-09-07 14:46:36 +00:00
|
|
|
c = clk_set_default_get_by_id(&clk);
|
|
|
|
if (IS_ERR(c))
|
|
|
|
return PTR_ERR(c);
|
|
|
|
|
|
|
|
ret = clk_set_rate(c, rates[index]);
|
2019-10-22 12:00:06 +00:00
|
|
|
|
2018-01-08 12:59:18 +00:00
|
|
|
if (ret < 0) {
|
2021-04-09 02:13:03 +00:00
|
|
|
dev_warn(dev,
|
|
|
|
"failed to set rate on clock index %d (%ld) (error = %d)\n",
|
|
|
|
index, clk.id, ret);
|
2018-01-08 12:59:18 +00:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
fail:
|
|
|
|
free(rates);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2021-06-11 04:16:07 +00:00
|
|
|
int clk_set_defaults(struct udevice *dev, enum clk_defaults_stage stage)
|
2018-01-08 12:59:18 +00:00
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
2020-12-19 17:40:13 +00:00
|
|
|
if (!dev_has_ofnode(dev))
|
2019-07-31 07:01:49 +00:00
|
|
|
return 0;
|
|
|
|
|
2021-06-11 04:16:07 +00:00
|
|
|
/*
|
|
|
|
* To avoid setting defaults twice, don't set them before relocation.
|
|
|
|
* However, still set them for SPL. And still set them if explicitly
|
|
|
|
* asked.
|
|
|
|
*/
|
2018-11-26 19:20:19 +00:00
|
|
|
if (!(IS_ENABLED(CONFIG_SPL_BUILD) || (gd->flags & GD_FLG_RELOC)))
|
2021-06-11 04:16:07 +00:00
|
|
|
if (stage != CLK_DEFAULTS_POST_FORCE)
|
|
|
|
return 0;
|
2018-11-26 19:20:19 +00:00
|
|
|
|
2018-01-08 12:59:18 +00:00
|
|
|
debug("%s(%s)\n", __func__, dev_read_name(dev));
|
|
|
|
|
2019-10-22 12:00:06 +00:00
|
|
|
ret = clk_set_default_parents(dev, stage);
|
2018-01-08 12:59:18 +00:00
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
|
2019-10-22 12:00:06 +00:00
|
|
|
ret = clk_set_default_rates(dev, stage);
|
2018-01-08 12:59:18 +00:00
|
|
|
if (ret < 0)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
2016-06-17 15:44:00 +00:00
|
|
|
|
|
|
|
int clk_get_by_name(struct udevice *dev, const char *name, struct clk *clk)
|
|
|
|
{
|
2022-02-27 19:01:13 +00:00
|
|
|
return clk_get_by_name_nodev(dev_ofnode(dev), name, clk);
|
2016-01-21 02:43:02 +00:00
|
|
|
}
|
2021-08-07 13:24:09 +00:00
|
|
|
#endif /* OF_REAL */
|
2017-07-25 11:24:45 +00:00
|
|
|
|
2020-01-09 03:35:07 +00:00
|
|
|
int clk_get_by_name_nodev(ofnode node, const char *name, struct clk *clk)
|
|
|
|
{
|
2023-01-22 00:02:51 +00:00
|
|
|
int index = 0;
|
2020-01-09 03:35:07 +00:00
|
|
|
|
|
|
|
debug("%s(node=%p, name=%s, clk=%p)\n", __func__,
|
|
|
|
ofnode_get_name(node), name, clk);
|
|
|
|
clk->dev = NULL;
|
|
|
|
|
2023-01-22 00:02:51 +00:00
|
|
|
if (name) {
|
|
|
|
index = ofnode_stringlist_search(node, "clock-names", name);
|
|
|
|
if (index < 0) {
|
|
|
|
debug("fdt_stringlist_search() failed: %d\n", index);
|
|
|
|
return index;
|
|
|
|
}
|
2020-01-09 03:35:07 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return clk_get_by_index_nodev(node, index, clk);
|
|
|
|
}
|
|
|
|
|
2023-06-19 10:47:52 +00:00
|
|
|
int clk_release_all(struct clk *clk, unsigned int count)
|
2017-07-25 11:24:45 +00:00
|
|
|
{
|
2023-06-19 10:47:52 +00:00
|
|
|
unsigned int i;
|
|
|
|
int ret;
|
2017-07-25 11:24:45 +00:00
|
|
|
|
|
|
|
for (i = 0; i < count; i++) {
|
2023-06-19 10:47:52 +00:00
|
|
|
debug("%s(clk[%u]=%p)\n", __func__, i, &clk[i]);
|
2017-07-25 11:24:45 +00:00
|
|
|
|
|
|
|
/* check if clock has been previously requested */
|
|
|
|
if (!clk[i].dev)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
ret = clk_disable(&clk[i]);
|
|
|
|
if (ret && ret != -ENOSYS)
|
|
|
|
return ret;
|
|
|
|
|
2022-01-15 22:25:04 +00:00
|
|
|
clk_free(&clk[i]);
|
2017-07-25 11:24:45 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2016-06-17 15:44:00 +00:00
|
|
|
int clk_request(struct udevice *dev, struct clk *clk)
|
|
|
|
{
|
2019-10-22 12:00:03 +00:00
|
|
|
const struct clk_ops *ops;
|
2016-06-17 15:44:00 +00:00
|
|
|
|
|
|
|
debug("%s(dev=%p, clk=%p)\n", __func__, dev, clk);
|
2019-10-22 12:00:03 +00:00
|
|
|
if (!clk)
|
|
|
|
return 0;
|
|
|
|
ops = clk_dev_ops(dev);
|
2016-06-17 15:44:00 +00:00
|
|
|
|
|
|
|
clk->dev = dev;
|
|
|
|
|
|
|
|
if (!ops->request)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
return ops->request(clk);
|
|
|
|
}
|
|
|
|
|
2022-01-15 22:25:04 +00:00
|
|
|
void clk_free(struct clk *clk)
|
2016-06-17 15:44:00 +00:00
|
|
|
{
|
2019-10-22 12:00:03 +00:00
|
|
|
const struct clk_ops *ops;
|
2016-06-17 15:44:00 +00:00
|
|
|
|
|
|
|
debug("%s(clk=%p)\n", __func__, clk);
|
2020-01-09 03:35:06 +00:00
|
|
|
if (!clk_valid(clk))
|
2022-01-15 22:25:04 +00:00
|
|
|
return;
|
2019-10-22 12:00:03 +00:00
|
|
|
ops = clk_dev_ops(clk->dev);
|
2016-06-17 15:44:00 +00:00
|
|
|
|
2022-01-15 22:24:58 +00:00
|
|
|
if (ops->rfree)
|
|
|
|
ops->rfree(clk);
|
2022-01-15 22:25:04 +00:00
|
|
|
return;
|
2016-06-17 15:44:00 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
ulong clk_get_rate(struct clk *clk)
|
|
|
|
{
|
2019-10-22 12:00:03 +00:00
|
|
|
const struct clk_ops *ops;
|
2023-05-15 19:49:58 +00:00
|
|
|
ulong ret;
|
2016-06-17 15:44:00 +00:00
|
|
|
|
|
|
|
debug("%s(clk=%p)\n", __func__, clk);
|
2020-01-09 03:35:06 +00:00
|
|
|
if (!clk_valid(clk))
|
2019-10-22 12:00:03 +00:00
|
|
|
return 0;
|
|
|
|
ops = clk_dev_ops(clk->dev);
|
2016-06-17 15:44:00 +00:00
|
|
|
|
|
|
|
if (!ops->get_rate)
|
|
|
|
return -ENOSYS;
|
|
|
|
|
2021-01-21 20:57:11 +00:00
|
|
|
ret = ops->get_rate(clk);
|
|
|
|
if (ret)
|
|
|
|
return log_ret(ret);
|
|
|
|
|
|
|
|
return 0;
|
2016-06-17 15:44:00 +00:00
|
|
|
}
|
|
|
|
|
2019-06-24 13:50:42 +00:00
|
|
|
struct clk *clk_get_parent(struct clk *clk)
|
|
|
|
{
|
|
|
|
struct udevice *pdev;
|
|
|
|
struct clk *pclk;
|
|
|
|
|
|
|
|
debug("%s(clk=%p)\n", __func__, clk);
|
2020-01-09 03:35:06 +00:00
|
|
|
if (!clk_valid(clk))
|
2019-10-22 12:00:03 +00:00
|
|
|
return NULL;
|
2019-06-24 13:50:42 +00:00
|
|
|
|
|
|
|
pdev = dev_get_parent(clk->dev);
|
2021-06-11 08:45:08 +00:00
|
|
|
if (!pdev)
|
|
|
|
return ERR_PTR(-ENODEV);
|
2019-06-24 13:50:42 +00:00
|
|
|
pclk = dev_get_clk_ptr(pdev);
|
|
|
|
if (!pclk)
|
|
|
|
return ERR_PTR(-ENODEV);
|
|
|
|
|
|
|
|
return pclk;
|
|
|
|
}
|
|
|
|
|
2022-09-28 10:37:57 +00:00
|
|
|
ulong clk_get_parent_rate(struct clk *clk)
|
2019-06-24 13:50:43 +00:00
|
|
|
{
|
|
|
|
const struct clk_ops *ops;
|
|
|
|
struct clk *pclk;
|
|
|
|
|
|
|
|
debug("%s(clk=%p)\n", __func__, clk);
|
2020-01-09 03:35:06 +00:00
|
|
|
if (!clk_valid(clk))
|
2019-10-22 12:00:03 +00:00
|
|
|
return 0;
|
2019-06-24 13:50:43 +00:00
|
|
|
|
|
|
|
pclk = clk_get_parent(clk);
|
|
|
|
if (IS_ERR(pclk))
|
|
|
|
return -ENODEV;
|
|
|
|
|
|
|
|
ops = clk_dev_ops(pclk->dev);
|
|
|
|
if (!ops->get_rate)
|
|
|
|
return -ENOSYS;
|
|
|
|
|
2019-06-24 13:50:46 +00:00
|
|
|
/* Read the 'rate' if not already set or if proper flag set*/
|
|
|
|
if (!pclk->rate || pclk->flags & CLK_GET_RATE_NOCACHE)
|
2019-06-24 13:50:43 +00:00
|
|
|
pclk->rate = clk_get_rate(pclk);
|
|
|
|
|
|
|
|
return pclk->rate;
|
|
|
|
}
|
|
|
|
|
2020-12-29 23:06:31 +00:00
|
|
|
ulong clk_round_rate(struct clk *clk, ulong rate)
|
|
|
|
{
|
|
|
|
const struct clk_ops *ops;
|
|
|
|
|
|
|
|
debug("%s(clk=%p, rate=%lu)\n", __func__, clk, rate);
|
|
|
|
if (!clk_valid(clk))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
ops = clk_dev_ops(clk->dev);
|
|
|
|
if (!ops->round_rate)
|
|
|
|
return -ENOSYS;
|
|
|
|
|
|
|
|
return ops->round_rate(clk, rate);
|
|
|
|
}
|
|
|
|
|
2022-06-20 13:37:25 +00:00
|
|
|
static void clk_get_priv(struct clk *clk, struct clk **clkp)
|
|
|
|
{
|
|
|
|
*clkp = clk;
|
|
|
|
|
|
|
|
/* get private clock struct associated to the provided clock */
|
|
|
|
if (CONFIG_IS_ENABLED(CLK_CCF)) {
|
|
|
|
/* Take id 0 as a non-valid clk, such as dummy */
|
|
|
|
if (clk->id)
|
|
|
|
clk_get_by_id(clk->id, clkp);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* clean cache, called with private clock struct */
|
2021-06-11 08:45:12 +00:00
|
|
|
static void clk_clean_rate_cache(struct clk *clk)
|
|
|
|
{
|
|
|
|
struct udevice *child_dev;
|
|
|
|
struct clk *clkp;
|
|
|
|
|
|
|
|
if (!clk)
|
|
|
|
return;
|
|
|
|
|
|
|
|
clk->rate = 0;
|
|
|
|
|
|
|
|
list_for_each_entry(child_dev, &clk->dev->child_head, sibling_node) {
|
|
|
|
clkp = dev_get_clk_ptr(child_dev);
|
|
|
|
clk_clean_rate_cache(clkp);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-06-17 15:44:00 +00:00
|
|
|
ulong clk_set_rate(struct clk *clk, ulong rate)
|
|
|
|
{
|
2019-10-22 12:00:03 +00:00
|
|
|
const struct clk_ops *ops;
|
2022-06-20 13:37:25 +00:00
|
|
|
struct clk *clkp;
|
2016-06-17 15:44:00 +00:00
|
|
|
|
|
|
|
debug("%s(clk=%p, rate=%lu)\n", __func__, clk, rate);
|
2020-01-09 03:35:06 +00:00
|
|
|
if (!clk_valid(clk))
|
2019-10-22 12:00:03 +00:00
|
|
|
return 0;
|
|
|
|
ops = clk_dev_ops(clk->dev);
|
2016-06-17 15:44:00 +00:00
|
|
|
|
|
|
|
if (!ops->set_rate)
|
|
|
|
return -ENOSYS;
|
|
|
|
|
2022-06-20 13:37:25 +00:00
|
|
|
/* get private clock struct used for cache */
|
|
|
|
clk_get_priv(clk, &clkp);
|
2021-06-11 08:45:12 +00:00
|
|
|
/* Clean up cached rates for us and all child clocks */
|
2022-06-20 13:37:25 +00:00
|
|
|
clk_clean_rate_cache(clkp);
|
2021-06-11 08:45:12 +00:00
|
|
|
|
2016-06-17 15:44:00 +00:00
|
|
|
return ops->set_rate(clk, rate);
|
|
|
|
}
|
|
|
|
|
2018-01-08 10:15:08 +00:00
|
|
|
int clk_set_parent(struct clk *clk, struct clk *parent)
|
|
|
|
{
|
2019-10-22 12:00:03 +00:00
|
|
|
const struct clk_ops *ops;
|
2020-09-07 14:46:34 +00:00
|
|
|
int ret;
|
2018-01-08 10:15:08 +00:00
|
|
|
|
|
|
|
debug("%s(clk=%p, parent=%p)\n", __func__, clk, parent);
|
2020-01-09 03:35:06 +00:00
|
|
|
if (!clk_valid(clk))
|
2019-10-22 12:00:03 +00:00
|
|
|
return 0;
|
|
|
|
ops = clk_dev_ops(clk->dev);
|
2018-01-08 10:15:08 +00:00
|
|
|
|
|
|
|
if (!ops->set_parent)
|
|
|
|
return -ENOSYS;
|
|
|
|
|
2020-09-07 14:46:34 +00:00
|
|
|
ret = ops->set_parent(clk, parent);
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
if (CONFIG_IS_ENABLED(CLK_CCF))
|
|
|
|
ret = device_reparent(clk->dev, parent->dev);
|
|
|
|
|
|
|
|
return ret;
|
2018-01-08 10:15:08 +00:00
|
|
|
}
|
|
|
|
|
2016-06-17 15:44:00 +00:00
|
|
|
int clk_enable(struct clk *clk)
|
|
|
|
{
|
2019-10-22 12:00:03 +00:00
|
|
|
const struct clk_ops *ops;
|
2019-08-21 13:35:09 +00:00
|
|
|
struct clk *clkp = NULL;
|
|
|
|
int ret;
|
2016-06-17 15:44:00 +00:00
|
|
|
|
|
|
|
debug("%s(clk=%p)\n", __func__, clk);
|
2020-01-09 03:35:06 +00:00
|
|
|
if (!clk_valid(clk))
|
2019-10-22 12:00:03 +00:00
|
|
|
return 0;
|
|
|
|
ops = clk_dev_ops(clk->dev);
|
2016-06-17 15:44:00 +00:00
|
|
|
|
2019-08-21 13:35:09 +00:00
|
|
|
if (CONFIG_IS_ENABLED(CLK_CCF)) {
|
|
|
|
/* Take id 0 as a non-valid clk, such as dummy */
|
|
|
|
if (clk->id && !clk_get_by_id(clk->id, &clkp)) {
|
2023-11-18 22:10:06 +00:00
|
|
|
ops = clk_dev_ops(clkp->dev);
|
2019-08-21 13:35:09 +00:00
|
|
|
if (clkp->enable_count) {
|
|
|
|
clkp->enable_count++;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
if (clkp->dev->parent &&
|
2022-01-24 13:17:14 +00:00
|
|
|
device_get_uclass_id(clkp->dev->parent) == UCLASS_CLK) {
|
2019-08-21 13:35:09 +00:00
|
|
|
ret = clk_enable(dev_get_clk_ptr(clkp->dev->parent));
|
|
|
|
if (ret) {
|
|
|
|
printf("Enable %s failed\n",
|
|
|
|
clkp->dev->parent->name);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2016-06-17 15:44:00 +00:00
|
|
|
|
2019-08-21 13:35:09 +00:00
|
|
|
if (ops->enable) {
|
2023-09-05 22:16:49 +00:00
|
|
|
ret = ops->enable(clkp ? clkp : clk);
|
2019-08-21 13:35:09 +00:00
|
|
|
if (ret) {
|
|
|
|
printf("Enable %s failed\n", clk->dev->name);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (clkp)
|
|
|
|
clkp->enable_count++;
|
|
|
|
} else {
|
|
|
|
if (!ops->enable)
|
|
|
|
return -ENOSYS;
|
|
|
|
return ops->enable(clk);
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
2016-06-17 15:44:00 +00:00
|
|
|
}
|
|
|
|
|
2018-04-03 09:44:18 +00:00
|
|
|
int clk_enable_bulk(struct clk_bulk *bulk)
|
|
|
|
{
|
|
|
|
int i, ret;
|
|
|
|
|
|
|
|
for (i = 0; i < bulk->count; i++) {
|
|
|
|
ret = clk_enable(&bulk->clks[i]);
|
|
|
|
if (ret < 0 && ret != -ENOSYS)
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2016-06-17 15:44:00 +00:00
|
|
|
int clk_disable(struct clk *clk)
|
|
|
|
{
|
2019-10-22 12:00:03 +00:00
|
|
|
const struct clk_ops *ops;
|
2019-08-21 13:35:09 +00:00
|
|
|
struct clk *clkp = NULL;
|
|
|
|
int ret;
|
2016-06-17 15:44:00 +00:00
|
|
|
|
|
|
|
debug("%s(clk=%p)\n", __func__, clk);
|
2020-01-09 03:35:06 +00:00
|
|
|
if (!clk_valid(clk))
|
2019-10-22 12:00:03 +00:00
|
|
|
return 0;
|
|
|
|
ops = clk_dev_ops(clk->dev);
|
2016-06-17 15:44:00 +00:00
|
|
|
|
2019-08-21 13:35:09 +00:00
|
|
|
if (CONFIG_IS_ENABLED(CLK_CCF)) {
|
|
|
|
if (clk->id && !clk_get_by_id(clk->id, &clkp)) {
|
2023-11-18 22:10:06 +00:00
|
|
|
ops = clk_dev_ops(clkp->dev);
|
2020-09-07 14:46:35 +00:00
|
|
|
if (clkp->flags & CLK_IS_CRITICAL)
|
|
|
|
return 0;
|
|
|
|
|
2019-08-21 13:35:09 +00:00
|
|
|
if (clkp->enable_count == 0) {
|
|
|
|
printf("clk %s already disabled\n",
|
|
|
|
clkp->dev->name);
|
|
|
|
return 0;
|
|
|
|
}
|
2016-06-17 15:44:00 +00:00
|
|
|
|
2019-08-21 13:35:09 +00:00
|
|
|
if (--clkp->enable_count > 0)
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ops->disable) {
|
2023-09-05 22:16:49 +00:00
|
|
|
ret = ops->disable(clkp ? clkp : clk);
|
2019-08-21 13:35:09 +00:00
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (clkp && clkp->dev->parent &&
|
2022-01-24 13:17:14 +00:00
|
|
|
device_get_uclass_id(clkp->dev->parent) == UCLASS_CLK) {
|
2019-08-21 13:35:09 +00:00
|
|
|
ret = clk_disable(dev_get_clk_ptr(clkp->dev->parent));
|
|
|
|
if (ret) {
|
|
|
|
printf("Disable %s failed\n",
|
|
|
|
clkp->dev->parent->name);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
if (!ops->disable)
|
|
|
|
return -ENOSYS;
|
|
|
|
|
|
|
|
return ops->disable(clk);
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
2016-06-17 15:44:00 +00:00
|
|
|
}
|
2016-01-21 02:43:02 +00:00
|
|
|
|
2018-04-03 09:44:18 +00:00
|
|
|
int clk_disable_bulk(struct clk_bulk *bulk)
|
|
|
|
{
|
|
|
|
int i, ret;
|
|
|
|
|
|
|
|
for (i = 0; i < bulk->count; i++) {
|
|
|
|
ret = clk_disable(&bulk->clks[i]);
|
|
|
|
if (ret < 0 && ret != -ENOSYS)
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2019-06-24 13:50:44 +00:00
|
|
|
int clk_get_by_id(ulong id, struct clk **clkp)
|
|
|
|
{
|
|
|
|
struct udevice *dev;
|
|
|
|
struct uclass *uc;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = uclass_get(UCLASS_CLK, &uc);
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
uclass_foreach_dev(dev, uc) {
|
|
|
|
struct clk *clk = dev_get_clk_ptr(dev);
|
|
|
|
|
|
|
|
if (clk && clk->id == id) {
|
|
|
|
*clkp = clk;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return -ENOENT;
|
|
|
|
}
|
|
|
|
|
2019-08-01 13:42:55 +00:00
|
|
|
bool clk_is_match(const struct clk *p, const struct clk *q)
|
|
|
|
{
|
|
|
|
/* trivial case: identical struct clk's or both NULL */
|
|
|
|
if (p == q)
|
|
|
|
return true;
|
|
|
|
|
2019-10-22 12:00:03 +00:00
|
|
|
/* trivial case #2: on the clk pointer is NULL */
|
|
|
|
if (!p || !q)
|
|
|
|
return false;
|
|
|
|
|
2019-08-01 13:42:55 +00:00
|
|
|
/* same device, id and data */
|
|
|
|
if (p->dev == q->dev && p->id == q->id && p->data == q->data)
|
|
|
|
return true;
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2019-10-22 12:00:04 +00:00
|
|
|
static void devm_clk_release(struct udevice *dev, void *res)
|
|
|
|
{
|
|
|
|
clk_free(res);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int devm_clk_match(struct udevice *dev, void *res, void *data)
|
|
|
|
{
|
|
|
|
return res == data;
|
|
|
|
}
|
|
|
|
|
|
|
|
struct clk *devm_clk_get(struct udevice *dev, const char *id)
|
|
|
|
{
|
|
|
|
int rc;
|
|
|
|
struct clk *clk;
|
|
|
|
|
|
|
|
clk = devres_alloc(devm_clk_release, sizeof(struct clk), __GFP_ZERO);
|
|
|
|
if (unlikely(!clk))
|
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
|
|
|
|
rc = clk_get_by_name(dev, id, clk);
|
|
|
|
if (rc)
|
|
|
|
return ERR_PTR(rc);
|
|
|
|
|
|
|
|
devres_add(dev, clk);
|
|
|
|
return clk;
|
|
|
|
}
|
|
|
|
|
|
|
|
void devm_clk_put(struct udevice *dev, struct clk *clk)
|
|
|
|
{
|
|
|
|
int rc;
|
|
|
|
|
|
|
|
if (!clk)
|
|
|
|
return;
|
|
|
|
|
|
|
|
rc = devres_release(dev, devm_clk_release, devm_clk_match, clk);
|
|
|
|
WARN_ON(rc);
|
|
|
|
}
|
|
|
|
|
2019-10-22 12:00:06 +00:00
|
|
|
int clk_uclass_post_probe(struct udevice *dev)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* when a clock provider is probed. Call clk_set_defaults()
|
|
|
|
* also after the device is probed. This takes care of cases
|
|
|
|
* where the DT is used to setup default parents and rates
|
|
|
|
* using assigned-clocks
|
|
|
|
*/
|
2022-01-01 18:51:39 +00:00
|
|
|
clk_set_defaults(dev, CLK_DEFAULTS_POST);
|
2019-10-22 12:00:06 +00:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2015-06-23 21:39:15 +00:00
|
|
|
UCLASS_DRIVER(clk) = {
|
|
|
|
.id = UCLASS_CLK,
|
|
|
|
.name = "clk",
|
2019-10-22 12:00:06 +00:00
|
|
|
.post_probe = clk_uclass_post_probe,
|
2015-06-23 21:39:15 +00:00
|
|
|
};
|