649 lines
16 KiB
C
649 lines
16 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Copyright (C) 2020 BAIKAL ELECTRONICS, JSC
|
|
*
|
|
* Authors:
|
|
* Serge Semin <Sergey.Semin@baikalelectronics.ru>
|
|
* Dmitry Dunaev <dmitry.dunaev@baikalelectronics.ru>
|
|
*
|
|
* Baikal-T1 CCU Dividers interface driver
|
|
*/
|
|
|
|
#define pr_fmt(fmt) "bt1-ccu-div: " fmt
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/printk.h>
|
|
#include <linux/bits.h>
|
|
#include <linux/bitfield.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/clk-provider.h>
|
|
#include <linux/of.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/regmap.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/time64.h>
|
|
#include <linux/debugfs.h>
|
|
|
|
#include "ccu-div.h"
|
|
|
|
#define CCU_DIV_CTL 0x00
|
|
#define CCU_DIV_CTL_EN BIT(0)
|
|
#define CCU_DIV_CTL_RST BIT(1)
|
|
#define CCU_DIV_CTL_SET_CLKDIV BIT(2)
|
|
#define CCU_DIV_CTL_CLKDIV_FLD 4
|
|
#define CCU_DIV_CTL_CLKDIV_MASK(_width) \
|
|
GENMASK((_width) + CCU_DIV_CTL_CLKDIV_FLD - 1, CCU_DIV_CTL_CLKDIV_FLD)
|
|
#define CCU_DIV_CTL_LOCK_SHIFTED BIT(27)
|
|
#define CCU_DIV_CTL_GATE_REF_BUF BIT(28)
|
|
#define CCU_DIV_CTL_LOCK_NORMAL BIT(31)
|
|
|
|
#define CCU_DIV_LOCK_CHECK_RETRIES 50
|
|
|
|
#define CCU_DIV_CLKDIV_MIN 0
|
|
#define CCU_DIV_CLKDIV_MAX(_mask) \
|
|
((_mask) >> CCU_DIV_CTL_CLKDIV_FLD)
|
|
|
|
/*
|
|
* Use the next two methods until there are generic field setter and
|
|
* getter available with non-constant mask support.
|
|
*/
|
|
static inline u32 ccu_div_get(u32 mask, u32 val)
|
|
{
|
|
return (val & mask) >> CCU_DIV_CTL_CLKDIV_FLD;
|
|
}
|
|
|
|
static inline u32 ccu_div_prep(u32 mask, u32 val)
|
|
{
|
|
return (val << CCU_DIV_CTL_CLKDIV_FLD) & mask;
|
|
}
|
|
|
|
static inline unsigned long ccu_div_lock_delay_ns(unsigned long ref_clk,
|
|
unsigned long div)
|
|
{
|
|
u64 ns = 4ULL * (div ?: 1) * NSEC_PER_SEC;
|
|
|
|
do_div(ns, ref_clk);
|
|
|
|
return ns;
|
|
}
|
|
|
|
static inline unsigned long ccu_div_calc_freq(unsigned long ref_clk,
|
|
unsigned long div)
|
|
{
|
|
return ref_clk / (div ?: 1);
|
|
}
|
|
|
|
static int ccu_div_var_update_clkdiv(struct ccu_div *div,
|
|
unsigned long parent_rate,
|
|
unsigned long divider)
|
|
{
|
|
unsigned long nd;
|
|
u32 val = 0;
|
|
u32 lock;
|
|
int count;
|
|
|
|
nd = ccu_div_lock_delay_ns(parent_rate, divider);
|
|
|
|
if (div->features & CCU_DIV_LOCK_SHIFTED)
|
|
lock = CCU_DIV_CTL_LOCK_SHIFTED;
|
|
else
|
|
lock = CCU_DIV_CTL_LOCK_NORMAL;
|
|
|
|
regmap_update_bits(div->sys_regs, div->reg_ctl,
|
|
CCU_DIV_CTL_SET_CLKDIV, CCU_DIV_CTL_SET_CLKDIV);
|
|
|
|
/*
|
|
* Until there is nsec-version of readl_poll_timeout() is available
|
|
* we have to implement the next polling loop.
|
|
*/
|
|
count = CCU_DIV_LOCK_CHECK_RETRIES;
|
|
do {
|
|
ndelay(nd);
|
|
regmap_read(div->sys_regs, div->reg_ctl, &val);
|
|
if (val & lock)
|
|
return 0;
|
|
} while (--count);
|
|
|
|
return -ETIMEDOUT;
|
|
}
|
|
|
|
static int ccu_div_var_enable(struct clk_hw *hw)
|
|
{
|
|
struct clk_hw *parent_hw = clk_hw_get_parent(hw);
|
|
struct ccu_div *div = to_ccu_div(hw);
|
|
unsigned long flags;
|
|
u32 val = 0;
|
|
int ret;
|
|
|
|
if (!parent_hw) {
|
|
pr_err("Can't enable '%s' with no parent", clk_hw_get_name(hw));
|
|
return -EINVAL;
|
|
}
|
|
|
|
regmap_read(div->sys_regs, div->reg_ctl, &val);
|
|
if (val & CCU_DIV_CTL_EN)
|
|
return 0;
|
|
|
|
spin_lock_irqsave(&div->lock, flags);
|
|
ret = ccu_div_var_update_clkdiv(div, clk_hw_get_rate(parent_hw),
|
|
ccu_div_get(div->mask, val));
|
|
if (!ret)
|
|
regmap_update_bits(div->sys_regs, div->reg_ctl,
|
|
CCU_DIV_CTL_EN, CCU_DIV_CTL_EN);
|
|
spin_unlock_irqrestore(&div->lock, flags);
|
|
if (ret)
|
|
pr_err("Divider '%s' lock timed out\n", clk_hw_get_name(hw));
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ccu_div_gate_enable(struct clk_hw *hw)
|
|
{
|
|
struct ccu_div *div = to_ccu_div(hw);
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&div->lock, flags);
|
|
regmap_update_bits(div->sys_regs, div->reg_ctl,
|
|
CCU_DIV_CTL_EN, CCU_DIV_CTL_EN);
|
|
spin_unlock_irqrestore(&div->lock, flags);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ccu_div_gate_disable(struct clk_hw *hw)
|
|
{
|
|
struct ccu_div *div = to_ccu_div(hw);
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&div->lock, flags);
|
|
regmap_update_bits(div->sys_regs, div->reg_ctl, CCU_DIV_CTL_EN, 0);
|
|
spin_unlock_irqrestore(&div->lock, flags);
|
|
}
|
|
|
|
static int ccu_div_gate_is_enabled(struct clk_hw *hw)
|
|
{
|
|
struct ccu_div *div = to_ccu_div(hw);
|
|
u32 val = 0;
|
|
|
|
regmap_read(div->sys_regs, div->reg_ctl, &val);
|
|
|
|
return !!(val & CCU_DIV_CTL_EN);
|
|
}
|
|
|
|
static int ccu_div_buf_enable(struct clk_hw *hw)
|
|
{
|
|
struct ccu_div *div = to_ccu_div(hw);
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&div->lock, flags);
|
|
regmap_update_bits(div->sys_regs, div->reg_ctl,
|
|
CCU_DIV_CTL_GATE_REF_BUF, 0);
|
|
spin_unlock_irqrestore(&div->lock, flags);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ccu_div_buf_disable(struct clk_hw *hw)
|
|
{
|
|
struct ccu_div *div = to_ccu_div(hw);
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&div->lock, flags);
|
|
regmap_update_bits(div->sys_regs, div->reg_ctl,
|
|
CCU_DIV_CTL_GATE_REF_BUF, CCU_DIV_CTL_GATE_REF_BUF);
|
|
spin_unlock_irqrestore(&div->lock, flags);
|
|
}
|
|
|
|
static int ccu_div_buf_is_enabled(struct clk_hw *hw)
|
|
{
|
|
struct ccu_div *div = to_ccu_div(hw);
|
|
u32 val = 0;
|
|
|
|
regmap_read(div->sys_regs, div->reg_ctl, &val);
|
|
|
|
return !(val & CCU_DIV_CTL_GATE_REF_BUF);
|
|
}
|
|
|
|
static unsigned long ccu_div_var_recalc_rate(struct clk_hw *hw,
|
|
unsigned long parent_rate)
|
|
{
|
|
struct ccu_div *div = to_ccu_div(hw);
|
|
unsigned long divider;
|
|
u32 val = 0;
|
|
|
|
regmap_read(div->sys_regs, div->reg_ctl, &val);
|
|
divider = ccu_div_get(div->mask, val);
|
|
|
|
return ccu_div_calc_freq(parent_rate, divider);
|
|
}
|
|
|
|
static inline unsigned long ccu_div_var_calc_divider(unsigned long rate,
|
|
unsigned long parent_rate,
|
|
unsigned int mask)
|
|
{
|
|
unsigned long divider;
|
|
|
|
divider = parent_rate / rate;
|
|
return clamp_t(unsigned long, divider, CCU_DIV_CLKDIV_MIN,
|
|
CCU_DIV_CLKDIV_MAX(mask));
|
|
}
|
|
|
|
static long ccu_div_var_round_rate(struct clk_hw *hw, unsigned long rate,
|
|
unsigned long *parent_rate)
|
|
{
|
|
struct ccu_div *div = to_ccu_div(hw);
|
|
unsigned long divider;
|
|
|
|
divider = ccu_div_var_calc_divider(rate, *parent_rate, div->mask);
|
|
|
|
return ccu_div_calc_freq(*parent_rate, divider);
|
|
}
|
|
|
|
/*
|
|
* This method is used for the clock divider blocks, which support the
|
|
* on-the-fly rate change. So due to lacking the EN bit functionality
|
|
* they can't be gated before the rate adjustment.
|
|
*/
|
|
static int ccu_div_var_set_rate_slow(struct clk_hw *hw, unsigned long rate,
|
|
unsigned long parent_rate)
|
|
{
|
|
struct ccu_div *div = to_ccu_div(hw);
|
|
unsigned long flags, divider;
|
|
u32 val;
|
|
int ret;
|
|
|
|
divider = ccu_div_var_calc_divider(rate, parent_rate, div->mask);
|
|
if (divider == 1 && div->features & CCU_DIV_SKIP_ONE) {
|
|
divider = 0;
|
|
} else if (div->features & CCU_DIV_SKIP_ONE_TO_THREE) {
|
|
if (divider == 1 || divider == 2)
|
|
divider = 0;
|
|
else if (divider == 3)
|
|
divider = 4;
|
|
}
|
|
|
|
val = ccu_div_prep(div->mask, divider);
|
|
|
|
spin_lock_irqsave(&div->lock, flags);
|
|
regmap_update_bits(div->sys_regs, div->reg_ctl, div->mask, val);
|
|
ret = ccu_div_var_update_clkdiv(div, parent_rate, divider);
|
|
spin_unlock_irqrestore(&div->lock, flags);
|
|
if (ret)
|
|
pr_err("Divider '%s' lock timed out\n", clk_hw_get_name(hw));
|
|
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* This method is used for the clock divider blocks, which don't support
|
|
* the on-the-fly rate change.
|
|
*/
|
|
static int ccu_div_var_set_rate_fast(struct clk_hw *hw, unsigned long rate,
|
|
unsigned long parent_rate)
|
|
{
|
|
struct ccu_div *div = to_ccu_div(hw);
|
|
unsigned long flags, divider;
|
|
u32 val;
|
|
|
|
divider = ccu_div_var_calc_divider(rate, parent_rate, div->mask);
|
|
val = ccu_div_prep(div->mask, divider);
|
|
|
|
/*
|
|
* Also disable the clock divider block if it was enabled by default
|
|
* or by the bootloader.
|
|
*/
|
|
spin_lock_irqsave(&div->lock, flags);
|
|
regmap_update_bits(div->sys_regs, div->reg_ctl,
|
|
div->mask | CCU_DIV_CTL_EN, val);
|
|
spin_unlock_irqrestore(&div->lock, flags);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static unsigned long ccu_div_fixed_recalc_rate(struct clk_hw *hw,
|
|
unsigned long parent_rate)
|
|
{
|
|
struct ccu_div *div = to_ccu_div(hw);
|
|
|
|
return ccu_div_calc_freq(parent_rate, div->divider);
|
|
}
|
|
|
|
static long ccu_div_fixed_round_rate(struct clk_hw *hw, unsigned long rate,
|
|
unsigned long *parent_rate)
|
|
{
|
|
struct ccu_div *div = to_ccu_div(hw);
|
|
|
|
return ccu_div_calc_freq(*parent_rate, div->divider);
|
|
}
|
|
|
|
static int ccu_div_fixed_set_rate(struct clk_hw *hw, unsigned long rate,
|
|
unsigned long parent_rate)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
#ifdef CONFIG_DEBUG_FS
|
|
|
|
struct ccu_div_dbgfs_bit {
|
|
struct ccu_div *div;
|
|
const char *name;
|
|
u32 mask;
|
|
};
|
|
|
|
#define CCU_DIV_DBGFS_BIT_ATTR(_name, _mask) { \
|
|
.name = _name, \
|
|
.mask = _mask \
|
|
}
|
|
|
|
static const struct ccu_div_dbgfs_bit ccu_div_bits[] = {
|
|
CCU_DIV_DBGFS_BIT_ATTR("div_en", CCU_DIV_CTL_EN),
|
|
CCU_DIV_DBGFS_BIT_ATTR("div_rst", CCU_DIV_CTL_RST),
|
|
CCU_DIV_DBGFS_BIT_ATTR("div_bypass", CCU_DIV_CTL_SET_CLKDIV),
|
|
CCU_DIV_DBGFS_BIT_ATTR("div_buf", CCU_DIV_CTL_GATE_REF_BUF),
|
|
CCU_DIV_DBGFS_BIT_ATTR("div_lock", CCU_DIV_CTL_LOCK_NORMAL)
|
|
};
|
|
|
|
#define CCU_DIV_DBGFS_BIT_NUM ARRAY_SIZE(ccu_div_bits)
|
|
|
|
/*
|
|
* It can be dangerous to change the Divider settings behind clock framework
|
|
* back, therefore we don't provide any kernel config based compile time option
|
|
* for this feature to enable.
|
|
*/
|
|
#undef CCU_DIV_ALLOW_WRITE_DEBUGFS
|
|
#ifdef CCU_DIV_ALLOW_WRITE_DEBUGFS
|
|
|
|
static int ccu_div_dbgfs_bit_set(void *priv, u64 val)
|
|
{
|
|
const struct ccu_div_dbgfs_bit *bit = priv;
|
|
struct ccu_div *div = bit->div;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&div->lock, flags);
|
|
regmap_update_bits(div->sys_regs, div->reg_ctl,
|
|
bit->mask, val ? bit->mask : 0);
|
|
spin_unlock_irqrestore(&div->lock, flags);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ccu_div_dbgfs_var_clkdiv_set(void *priv, u64 val)
|
|
{
|
|
struct ccu_div *div = priv;
|
|
unsigned long flags;
|
|
u32 data;
|
|
|
|
val = clamp_t(u64, val, CCU_DIV_CLKDIV_MIN,
|
|
CCU_DIV_CLKDIV_MAX(div->mask));
|
|
data = ccu_div_prep(div->mask, val);
|
|
|
|
spin_lock_irqsave(&div->lock, flags);
|
|
regmap_update_bits(div->sys_regs, div->reg_ctl, div->mask, data);
|
|
spin_unlock_irqrestore(&div->lock, flags);
|
|
|
|
return 0;
|
|
}
|
|
|
|
#define ccu_div_dbgfs_mode 0644
|
|
|
|
#else /* !CCU_DIV_ALLOW_WRITE_DEBUGFS */
|
|
|
|
#define ccu_div_dbgfs_bit_set NULL
|
|
#define ccu_div_dbgfs_var_clkdiv_set NULL
|
|
#define ccu_div_dbgfs_mode 0444
|
|
|
|
#endif /* !CCU_DIV_ALLOW_WRITE_DEBUGFS */
|
|
|
|
static int ccu_div_dbgfs_bit_get(void *priv, u64 *val)
|
|
{
|
|
const struct ccu_div_dbgfs_bit *bit = priv;
|
|
struct ccu_div *div = bit->div;
|
|
u32 data = 0;
|
|
|
|
regmap_read(div->sys_regs, div->reg_ctl, &data);
|
|
*val = !!(data & bit->mask);
|
|
|
|
return 0;
|
|
}
|
|
DEFINE_DEBUGFS_ATTRIBUTE(ccu_div_dbgfs_bit_fops,
|
|
ccu_div_dbgfs_bit_get, ccu_div_dbgfs_bit_set, "%llu\n");
|
|
|
|
static int ccu_div_dbgfs_var_clkdiv_get(void *priv, u64 *val)
|
|
{
|
|
struct ccu_div *div = priv;
|
|
u32 data = 0;
|
|
|
|
regmap_read(div->sys_regs, div->reg_ctl, &data);
|
|
*val = ccu_div_get(div->mask, data);
|
|
|
|
return 0;
|
|
}
|
|
DEFINE_DEBUGFS_ATTRIBUTE(ccu_div_dbgfs_var_clkdiv_fops,
|
|
ccu_div_dbgfs_var_clkdiv_get, ccu_div_dbgfs_var_clkdiv_set, "%llu\n");
|
|
|
|
static int ccu_div_dbgfs_fixed_clkdiv_get(void *priv, u64 *val)
|
|
{
|
|
struct ccu_div *div = priv;
|
|
|
|
*val = div->divider;
|
|
|
|
return 0;
|
|
}
|
|
DEFINE_DEBUGFS_ATTRIBUTE(ccu_div_dbgfs_fixed_clkdiv_fops,
|
|
ccu_div_dbgfs_fixed_clkdiv_get, NULL, "%llu\n");
|
|
|
|
static void ccu_div_var_debug_init(struct clk_hw *hw, struct dentry *dentry)
|
|
{
|
|
struct ccu_div *div = to_ccu_div(hw);
|
|
struct ccu_div_dbgfs_bit *bits;
|
|
int didx, bidx, num = 2;
|
|
const char *name;
|
|
|
|
num += !!(div->flags & CLK_SET_RATE_GATE) +
|
|
!!(div->features & CCU_DIV_RESET_DOMAIN);
|
|
|
|
bits = kcalloc(num, sizeof(*bits), GFP_KERNEL);
|
|
if (!bits)
|
|
return;
|
|
|
|
for (didx = 0, bidx = 0; bidx < CCU_DIV_DBGFS_BIT_NUM; ++bidx) {
|
|
name = ccu_div_bits[bidx].name;
|
|
if (!(div->flags & CLK_SET_RATE_GATE) &&
|
|
!strcmp("div_en", name)) {
|
|
continue;
|
|
}
|
|
|
|
if (!(div->features & CCU_DIV_RESET_DOMAIN) &&
|
|
!strcmp("div_rst", name)) {
|
|
continue;
|
|
}
|
|
|
|
if (!strcmp("div_buf", name))
|
|
continue;
|
|
|
|
bits[didx] = ccu_div_bits[bidx];
|
|
bits[didx].div = div;
|
|
|
|
if (div->features & CCU_DIV_LOCK_SHIFTED &&
|
|
!strcmp("div_lock", name)) {
|
|
bits[didx].mask = CCU_DIV_CTL_LOCK_SHIFTED;
|
|
}
|
|
|
|
debugfs_create_file_unsafe(bits[didx].name, ccu_div_dbgfs_mode,
|
|
dentry, &bits[didx],
|
|
&ccu_div_dbgfs_bit_fops);
|
|
++didx;
|
|
}
|
|
|
|
debugfs_create_file_unsafe("div_clkdiv", ccu_div_dbgfs_mode, dentry,
|
|
div, &ccu_div_dbgfs_var_clkdiv_fops);
|
|
}
|
|
|
|
static void ccu_div_gate_debug_init(struct clk_hw *hw, struct dentry *dentry)
|
|
{
|
|
struct ccu_div *div = to_ccu_div(hw);
|
|
struct ccu_div_dbgfs_bit *bit;
|
|
|
|
bit = kmalloc(sizeof(*bit), GFP_KERNEL);
|
|
if (!bit)
|
|
return;
|
|
|
|
*bit = ccu_div_bits[0];
|
|
bit->div = div;
|
|
debugfs_create_file_unsafe(bit->name, ccu_div_dbgfs_mode, dentry, bit,
|
|
&ccu_div_dbgfs_bit_fops);
|
|
|
|
debugfs_create_file_unsafe("div_clkdiv", 0400, dentry, div,
|
|
&ccu_div_dbgfs_fixed_clkdiv_fops);
|
|
}
|
|
|
|
static void ccu_div_buf_debug_init(struct clk_hw *hw, struct dentry *dentry)
|
|
{
|
|
struct ccu_div *div = to_ccu_div(hw);
|
|
struct ccu_div_dbgfs_bit *bit;
|
|
|
|
bit = kmalloc(sizeof(*bit), GFP_KERNEL);
|
|
if (!bit)
|
|
return;
|
|
|
|
*bit = ccu_div_bits[3];
|
|
bit->div = div;
|
|
debugfs_create_file_unsafe(bit->name, ccu_div_dbgfs_mode, dentry, bit,
|
|
&ccu_div_dbgfs_bit_fops);
|
|
}
|
|
|
|
static void ccu_div_fixed_debug_init(struct clk_hw *hw, struct dentry *dentry)
|
|
{
|
|
struct ccu_div *div = to_ccu_div(hw);
|
|
|
|
debugfs_create_file_unsafe("div_clkdiv", 0400, dentry, div,
|
|
&ccu_div_dbgfs_fixed_clkdiv_fops);
|
|
}
|
|
|
|
#else /* !CONFIG_DEBUG_FS */
|
|
|
|
#define ccu_div_var_debug_init NULL
|
|
#define ccu_div_gate_debug_init NULL
|
|
#define ccu_div_buf_debug_init NULL
|
|
#define ccu_div_fixed_debug_init NULL
|
|
|
|
#endif /* !CONFIG_DEBUG_FS */
|
|
|
|
static const struct clk_ops ccu_div_var_gate_to_set_ops = {
|
|
.enable = ccu_div_var_enable,
|
|
.disable = ccu_div_gate_disable,
|
|
.is_enabled = ccu_div_gate_is_enabled,
|
|
.recalc_rate = ccu_div_var_recalc_rate,
|
|
.round_rate = ccu_div_var_round_rate,
|
|
.set_rate = ccu_div_var_set_rate_fast,
|
|
.debug_init = ccu_div_var_debug_init
|
|
};
|
|
|
|
static const struct clk_ops ccu_div_var_nogate_ops = {
|
|
.recalc_rate = ccu_div_var_recalc_rate,
|
|
.round_rate = ccu_div_var_round_rate,
|
|
.set_rate = ccu_div_var_set_rate_slow,
|
|
.debug_init = ccu_div_var_debug_init
|
|
};
|
|
|
|
static const struct clk_ops ccu_div_gate_ops = {
|
|
.enable = ccu_div_gate_enable,
|
|
.disable = ccu_div_gate_disable,
|
|
.is_enabled = ccu_div_gate_is_enabled,
|
|
.recalc_rate = ccu_div_fixed_recalc_rate,
|
|
.round_rate = ccu_div_fixed_round_rate,
|
|
.set_rate = ccu_div_fixed_set_rate,
|
|
.debug_init = ccu_div_gate_debug_init
|
|
};
|
|
|
|
static const struct clk_ops ccu_div_buf_ops = {
|
|
.enable = ccu_div_buf_enable,
|
|
.disable = ccu_div_buf_disable,
|
|
.is_enabled = ccu_div_buf_is_enabled,
|
|
.debug_init = ccu_div_buf_debug_init
|
|
};
|
|
|
|
static const struct clk_ops ccu_div_fixed_ops = {
|
|
.recalc_rate = ccu_div_fixed_recalc_rate,
|
|
.round_rate = ccu_div_fixed_round_rate,
|
|
.set_rate = ccu_div_fixed_set_rate,
|
|
.debug_init = ccu_div_fixed_debug_init
|
|
};
|
|
|
|
struct ccu_div *ccu_div_hw_register(const struct ccu_div_init_data *div_init)
|
|
{
|
|
struct clk_parent_data parent_data = { };
|
|
struct clk_init_data hw_init = { };
|
|
struct ccu_div *div;
|
|
int ret;
|
|
|
|
if (!div_init)
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
div = kzalloc(sizeof(*div), GFP_KERNEL);
|
|
if (!div)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
/*
|
|
* Note since Baikal-T1 System Controller registers are MMIO-backed
|
|
* we won't check the regmap IO operations return status, because it
|
|
* must be zero anyway.
|
|
*/
|
|
div->hw.init = &hw_init;
|
|
div->id = div_init->id;
|
|
div->reg_ctl = div_init->base + CCU_DIV_CTL;
|
|
div->sys_regs = div_init->sys_regs;
|
|
div->flags = div_init->flags;
|
|
div->features = div_init->features;
|
|
spin_lock_init(&div->lock);
|
|
|
|
hw_init.name = div_init->name;
|
|
hw_init.flags = div_init->flags;
|
|
|
|
if (div_init->type == CCU_DIV_VAR) {
|
|
if (hw_init.flags & CLK_SET_RATE_GATE)
|
|
hw_init.ops = &ccu_div_var_gate_to_set_ops;
|
|
else
|
|
hw_init.ops = &ccu_div_var_nogate_ops;
|
|
div->mask = CCU_DIV_CTL_CLKDIV_MASK(div_init->width);
|
|
} else if (div_init->type == CCU_DIV_GATE) {
|
|
hw_init.ops = &ccu_div_gate_ops;
|
|
div->divider = div_init->divider;
|
|
} else if (div_init->type == CCU_DIV_BUF) {
|
|
hw_init.ops = &ccu_div_buf_ops;
|
|
} else if (div_init->type == CCU_DIV_FIXED) {
|
|
hw_init.ops = &ccu_div_fixed_ops;
|
|
div->divider = div_init->divider;
|
|
} else {
|
|
ret = -EINVAL;
|
|
goto err_free_div;
|
|
}
|
|
|
|
if (!div_init->parent_name) {
|
|
ret = -EINVAL;
|
|
goto err_free_div;
|
|
}
|
|
parent_data.fw_name = div_init->parent_name;
|
|
parent_data.name = div_init->parent_name;
|
|
hw_init.parent_data = &parent_data;
|
|
hw_init.num_parents = 1;
|
|
|
|
ret = of_clk_hw_register(div_init->np, &div->hw);
|
|
if (ret)
|
|
goto err_free_div;
|
|
|
|
return div;
|
|
|
|
err_free_div:
|
|
kfree(div);
|
|
|
|
return ERR_PTR(ret);
|
|
}
|
|
|
|
void ccu_div_hw_unregister(struct ccu_div *div)
|
|
{
|
|
clk_hw_unregister(&div->hw);
|
|
|
|
kfree(div);
|
|
}
|