// SPDX-License-Identifier: GPL-2.0-only /* * Copyright (C) 2020 BAIKAL ELECTRONICS, JSC * * Authors: * Serge Semin * Dmitry Dunaev * * Baikal-T1 CCU Dividers interface driver */ #define pr_fmt(fmt) "bt1-ccu-div: " fmt #include #include #include #include #include #include #include #include #include #include #include #include #include "ccu-div.h" #define CCU_DIV_CTL 0x00 #define CCU_DIV_CTL_EN BIT(0) #define CCU_DIV_CTL_RST BIT(1) #define CCU_DIV_CTL_SET_CLKDIV BIT(2) #define CCU_DIV_CTL_CLKDIV_FLD 4 #define CCU_DIV_CTL_CLKDIV_MASK(_width) \ GENMASK((_width) + CCU_DIV_CTL_CLKDIV_FLD - 1, CCU_DIV_CTL_CLKDIV_FLD) #define CCU_DIV_CTL_LOCK_SHIFTED BIT(27) #define CCU_DIV_CTL_GATE_REF_BUF BIT(28) #define CCU_DIV_CTL_LOCK_NORMAL BIT(31) #define CCU_DIV_LOCK_CHECK_RETRIES 50 #define CCU_DIV_CLKDIV_MIN 0 #define CCU_DIV_CLKDIV_MAX(_mask) \ ((_mask) >> CCU_DIV_CTL_CLKDIV_FLD) /* * Use the next two methods until there are generic field setter and * getter available with non-constant mask support. */ static inline u32 ccu_div_get(u32 mask, u32 val) { return (val & mask) >> CCU_DIV_CTL_CLKDIV_FLD; } static inline u32 ccu_div_prep(u32 mask, u32 val) { return (val << CCU_DIV_CTL_CLKDIV_FLD) & mask; } static inline unsigned long ccu_div_lock_delay_ns(unsigned long ref_clk, unsigned long div) { u64 ns = 4ULL * (div ?: 1) * NSEC_PER_SEC; do_div(ns, ref_clk); return ns; } static inline unsigned long ccu_div_calc_freq(unsigned long ref_clk, unsigned long div) { return ref_clk / (div ?: 1); } static int ccu_div_var_update_clkdiv(struct ccu_div *div, unsigned long parent_rate, unsigned long divider) { unsigned long nd; u32 val = 0; u32 lock; int count; nd = ccu_div_lock_delay_ns(parent_rate, divider); if (div->features & CCU_DIV_LOCK_SHIFTED) lock = CCU_DIV_CTL_LOCK_SHIFTED; else lock = CCU_DIV_CTL_LOCK_NORMAL; regmap_update_bits(div->sys_regs, div->reg_ctl, CCU_DIV_CTL_SET_CLKDIV, CCU_DIV_CTL_SET_CLKDIV); /* * Until there is nsec-version of readl_poll_timeout() is available * we have to implement the next polling loop. */ count = CCU_DIV_LOCK_CHECK_RETRIES; do { ndelay(nd); regmap_read(div->sys_regs, div->reg_ctl, &val); if (val & lock) return 0; } while (--count); return -ETIMEDOUT; } static int ccu_div_var_enable(struct clk_hw *hw) { struct clk_hw *parent_hw = clk_hw_get_parent(hw); struct ccu_div *div = to_ccu_div(hw); unsigned long flags; u32 val = 0; int ret; if (!parent_hw) { pr_err("Can't enable '%s' with no parent", clk_hw_get_name(hw)); return -EINVAL; } regmap_read(div->sys_regs, div->reg_ctl, &val); if (val & CCU_DIV_CTL_EN) return 0; spin_lock_irqsave(&div->lock, flags); ret = ccu_div_var_update_clkdiv(div, clk_hw_get_rate(parent_hw), ccu_div_get(div->mask, val)); if (!ret) regmap_update_bits(div->sys_regs, div->reg_ctl, CCU_DIV_CTL_EN, CCU_DIV_CTL_EN); spin_unlock_irqrestore(&div->lock, flags); if (ret) pr_err("Divider '%s' lock timed out\n", clk_hw_get_name(hw)); return ret; } static int ccu_div_gate_enable(struct clk_hw *hw) { struct ccu_div *div = to_ccu_div(hw); unsigned long flags; spin_lock_irqsave(&div->lock, flags); regmap_update_bits(div->sys_regs, div->reg_ctl, CCU_DIV_CTL_EN, CCU_DIV_CTL_EN); spin_unlock_irqrestore(&div->lock, flags); return 0; } static void ccu_div_gate_disable(struct clk_hw *hw) { struct ccu_div *div = to_ccu_div(hw); unsigned long flags; spin_lock_irqsave(&div->lock, flags); regmap_update_bits(div->sys_regs, div->reg_ctl, CCU_DIV_CTL_EN, 0); spin_unlock_irqrestore(&div->lock, flags); } static int ccu_div_gate_is_enabled(struct clk_hw *hw) { struct ccu_div *div = to_ccu_div(hw); u32 val = 0; regmap_read(div->sys_regs, div->reg_ctl, &val); return !!(val & CCU_DIV_CTL_EN); } static int ccu_div_buf_enable(struct clk_hw *hw) { struct ccu_div *div = to_ccu_div(hw); unsigned long flags; spin_lock_irqsave(&div->lock, flags); regmap_update_bits(div->sys_regs, div->reg_ctl, CCU_DIV_CTL_GATE_REF_BUF, 0); spin_unlock_irqrestore(&div->lock, flags); return 0; } static void ccu_div_buf_disable(struct clk_hw *hw) { struct ccu_div *div = to_ccu_div(hw); unsigned long flags; spin_lock_irqsave(&div->lock, flags); regmap_update_bits(div->sys_regs, div->reg_ctl, CCU_DIV_CTL_GATE_REF_BUF, CCU_DIV_CTL_GATE_REF_BUF); spin_unlock_irqrestore(&div->lock, flags); } static int ccu_div_buf_is_enabled(struct clk_hw *hw) { struct ccu_div *div = to_ccu_div(hw); u32 val = 0; regmap_read(div->sys_regs, div->reg_ctl, &val); return !(val & CCU_DIV_CTL_GATE_REF_BUF); } static unsigned long ccu_div_var_recalc_rate(struct clk_hw *hw, unsigned long parent_rate) { struct ccu_div *div = to_ccu_div(hw); unsigned long divider; u32 val = 0; regmap_read(div->sys_regs, div->reg_ctl, &val); divider = ccu_div_get(div->mask, val); return ccu_div_calc_freq(parent_rate, divider); } static inline unsigned long ccu_div_var_calc_divider(unsigned long rate, unsigned long parent_rate, unsigned int mask) { unsigned long divider; divider = parent_rate / rate; return clamp_t(unsigned long, divider, CCU_DIV_CLKDIV_MIN, CCU_DIV_CLKDIV_MAX(mask)); } static long ccu_div_var_round_rate(struct clk_hw *hw, unsigned long rate, unsigned long *parent_rate) { struct ccu_div *div = to_ccu_div(hw); unsigned long divider; divider = ccu_div_var_calc_divider(rate, *parent_rate, div->mask); return ccu_div_calc_freq(*parent_rate, divider); } /* * This method is used for the clock divider blocks, which support the * on-the-fly rate change. So due to lacking the EN bit functionality * they can't be gated before the rate adjustment. */ static int ccu_div_var_set_rate_slow(struct clk_hw *hw, unsigned long rate, unsigned long parent_rate) { struct ccu_div *div = to_ccu_div(hw); unsigned long flags, divider; u32 val; int ret; divider = ccu_div_var_calc_divider(rate, parent_rate, div->mask); if (divider == 1 && div->features & CCU_DIV_SKIP_ONE) { divider = 0; } else if (div->features & CCU_DIV_SKIP_ONE_TO_THREE) { if (divider == 1 || divider == 2) divider = 0; else if (divider == 3) divider = 4; } val = ccu_div_prep(div->mask, divider); spin_lock_irqsave(&div->lock, flags); regmap_update_bits(div->sys_regs, div->reg_ctl, div->mask, val); ret = ccu_div_var_update_clkdiv(div, parent_rate, divider); spin_unlock_irqrestore(&div->lock, flags); if (ret) pr_err("Divider '%s' lock timed out\n", clk_hw_get_name(hw)); return ret; } /* * This method is used for the clock divider blocks, which don't support * the on-the-fly rate change. */ static int ccu_div_var_set_rate_fast(struct clk_hw *hw, unsigned long rate, unsigned long parent_rate) { struct ccu_div *div = to_ccu_div(hw); unsigned long flags, divider; u32 val; divider = ccu_div_var_calc_divider(rate, parent_rate, div->mask); val = ccu_div_prep(div->mask, divider); /* * Also disable the clock divider block if it was enabled by default * or by the bootloader. */ spin_lock_irqsave(&div->lock, flags); regmap_update_bits(div->sys_regs, div->reg_ctl, div->mask | CCU_DIV_CTL_EN, val); spin_unlock_irqrestore(&div->lock, flags); return 0; } static unsigned long ccu_div_fixed_recalc_rate(struct clk_hw *hw, unsigned long parent_rate) { struct ccu_div *div = to_ccu_div(hw); return ccu_div_calc_freq(parent_rate, div->divider); } static long ccu_div_fixed_round_rate(struct clk_hw *hw, unsigned long rate, unsigned long *parent_rate) { struct ccu_div *div = to_ccu_div(hw); return ccu_div_calc_freq(*parent_rate, div->divider); } static int ccu_div_fixed_set_rate(struct clk_hw *hw, unsigned long rate, unsigned long parent_rate) { return 0; } #ifdef CONFIG_DEBUG_FS struct ccu_div_dbgfs_bit { struct ccu_div *div; const char *name; u32 mask; }; #define CCU_DIV_DBGFS_BIT_ATTR(_name, _mask) { \ .name = _name, \ .mask = _mask \ } static const struct ccu_div_dbgfs_bit ccu_div_bits[] = { CCU_DIV_DBGFS_BIT_ATTR("div_en", CCU_DIV_CTL_EN), CCU_DIV_DBGFS_BIT_ATTR("div_rst", CCU_DIV_CTL_RST), CCU_DIV_DBGFS_BIT_ATTR("div_bypass", CCU_DIV_CTL_SET_CLKDIV), CCU_DIV_DBGFS_BIT_ATTR("div_buf", CCU_DIV_CTL_GATE_REF_BUF), CCU_DIV_DBGFS_BIT_ATTR("div_lock", CCU_DIV_CTL_LOCK_NORMAL) }; #define CCU_DIV_DBGFS_BIT_NUM ARRAY_SIZE(ccu_div_bits) /* * It can be dangerous to change the Divider settings behind clock framework * back, therefore we don't provide any kernel config based compile time option * for this feature to enable. */ #undef CCU_DIV_ALLOW_WRITE_DEBUGFS #ifdef CCU_DIV_ALLOW_WRITE_DEBUGFS static int ccu_div_dbgfs_bit_set(void *priv, u64 val) { const struct ccu_div_dbgfs_bit *bit = priv; struct ccu_div *div = bit->div; unsigned long flags; spin_lock_irqsave(&div->lock, flags); regmap_update_bits(div->sys_regs, div->reg_ctl, bit->mask, val ? bit->mask : 0); spin_unlock_irqrestore(&div->lock, flags); return 0; } static int ccu_div_dbgfs_var_clkdiv_set(void *priv, u64 val) { struct ccu_div *div = priv; unsigned long flags; u32 data; val = clamp_t(u64, val, CCU_DIV_CLKDIV_MIN, CCU_DIV_CLKDIV_MAX(div->mask)); data = ccu_div_prep(div->mask, val); spin_lock_irqsave(&div->lock, flags); regmap_update_bits(div->sys_regs, div->reg_ctl, div->mask, data); spin_unlock_irqrestore(&div->lock, flags); return 0; } #define ccu_div_dbgfs_mode 0644 #else /* !CCU_DIV_ALLOW_WRITE_DEBUGFS */ #define ccu_div_dbgfs_bit_set NULL #define ccu_div_dbgfs_var_clkdiv_set NULL #define ccu_div_dbgfs_mode 0444 #endif /* !CCU_DIV_ALLOW_WRITE_DEBUGFS */ static int ccu_div_dbgfs_bit_get(void *priv, u64 *val) { const struct ccu_div_dbgfs_bit *bit = priv; struct ccu_div *div = bit->div; u32 data = 0; regmap_read(div->sys_regs, div->reg_ctl, &data); *val = !!(data & bit->mask); return 0; } DEFINE_DEBUGFS_ATTRIBUTE(ccu_div_dbgfs_bit_fops, ccu_div_dbgfs_bit_get, ccu_div_dbgfs_bit_set, "%llu\n"); static int ccu_div_dbgfs_var_clkdiv_get(void *priv, u64 *val) { struct ccu_div *div = priv; u32 data = 0; regmap_read(div->sys_regs, div->reg_ctl, &data); *val = ccu_div_get(div->mask, data); return 0; } DEFINE_DEBUGFS_ATTRIBUTE(ccu_div_dbgfs_var_clkdiv_fops, ccu_div_dbgfs_var_clkdiv_get, ccu_div_dbgfs_var_clkdiv_set, "%llu\n"); static int ccu_div_dbgfs_fixed_clkdiv_get(void *priv, u64 *val) { struct ccu_div *div = priv; *val = div->divider; return 0; } DEFINE_DEBUGFS_ATTRIBUTE(ccu_div_dbgfs_fixed_clkdiv_fops, ccu_div_dbgfs_fixed_clkdiv_get, NULL, "%llu\n"); static void ccu_div_var_debug_init(struct clk_hw *hw, struct dentry *dentry) { struct ccu_div *div = to_ccu_div(hw); struct ccu_div_dbgfs_bit *bits; int didx, bidx, num = 2; const char *name; num += !!(div->flags & CLK_SET_RATE_GATE) + !!(div->features & CCU_DIV_RESET_DOMAIN); bits = kcalloc(num, sizeof(*bits), GFP_KERNEL); if (!bits) return; for (didx = 0, bidx = 0; bidx < CCU_DIV_DBGFS_BIT_NUM; ++bidx) { name = ccu_div_bits[bidx].name; if (!(div->flags & CLK_SET_RATE_GATE) && !strcmp("div_en", name)) { continue; } if (!(div->features & CCU_DIV_RESET_DOMAIN) && !strcmp("div_rst", name)) { continue; } if (!strcmp("div_buf", name)) continue; bits[didx] = ccu_div_bits[bidx]; bits[didx].div = div; if (div->features & CCU_DIV_LOCK_SHIFTED && !strcmp("div_lock", name)) { bits[didx].mask = CCU_DIV_CTL_LOCK_SHIFTED; } debugfs_create_file_unsafe(bits[didx].name, ccu_div_dbgfs_mode, dentry, &bits[didx], &ccu_div_dbgfs_bit_fops); ++didx; } debugfs_create_file_unsafe("div_clkdiv", ccu_div_dbgfs_mode, dentry, div, &ccu_div_dbgfs_var_clkdiv_fops); } static void ccu_div_gate_debug_init(struct clk_hw *hw, struct dentry *dentry) { struct ccu_div *div = to_ccu_div(hw); struct ccu_div_dbgfs_bit *bit; bit = kmalloc(sizeof(*bit), GFP_KERNEL); if (!bit) return; *bit = ccu_div_bits[0]; bit->div = div; debugfs_create_file_unsafe(bit->name, ccu_div_dbgfs_mode, dentry, bit, &ccu_div_dbgfs_bit_fops); debugfs_create_file_unsafe("div_clkdiv", 0400, dentry, div, &ccu_div_dbgfs_fixed_clkdiv_fops); } static void ccu_div_buf_debug_init(struct clk_hw *hw, struct dentry *dentry) { struct ccu_div *div = to_ccu_div(hw); struct ccu_div_dbgfs_bit *bit; bit = kmalloc(sizeof(*bit), GFP_KERNEL); if (!bit) return; *bit = ccu_div_bits[3]; bit->div = div; debugfs_create_file_unsafe(bit->name, ccu_div_dbgfs_mode, dentry, bit, &ccu_div_dbgfs_bit_fops); } static void ccu_div_fixed_debug_init(struct clk_hw *hw, struct dentry *dentry) { struct ccu_div *div = to_ccu_div(hw); debugfs_create_file_unsafe("div_clkdiv", 0400, dentry, div, &ccu_div_dbgfs_fixed_clkdiv_fops); } #else /* !CONFIG_DEBUG_FS */ #define ccu_div_var_debug_init NULL #define ccu_div_gate_debug_init NULL #define ccu_div_buf_debug_init NULL #define ccu_div_fixed_debug_init NULL #endif /* !CONFIG_DEBUG_FS */ static const struct clk_ops ccu_div_var_gate_to_set_ops = { .enable = ccu_div_var_enable, .disable = ccu_div_gate_disable, .is_enabled = ccu_div_gate_is_enabled, .recalc_rate = ccu_div_var_recalc_rate, .round_rate = ccu_div_var_round_rate, .set_rate = ccu_div_var_set_rate_fast, .debug_init = ccu_div_var_debug_init }; static const struct clk_ops ccu_div_var_nogate_ops = { .recalc_rate = ccu_div_var_recalc_rate, .round_rate = ccu_div_var_round_rate, .set_rate = ccu_div_var_set_rate_slow, .debug_init = ccu_div_var_debug_init }; static const struct clk_ops ccu_div_gate_ops = { .enable = ccu_div_gate_enable, .disable = ccu_div_gate_disable, .is_enabled = ccu_div_gate_is_enabled, .recalc_rate = ccu_div_fixed_recalc_rate, .round_rate = ccu_div_fixed_round_rate, .set_rate = ccu_div_fixed_set_rate, .debug_init = ccu_div_gate_debug_init }; static const struct clk_ops ccu_div_buf_ops = { .enable = ccu_div_buf_enable, .disable = ccu_div_buf_disable, .is_enabled = ccu_div_buf_is_enabled, .debug_init = ccu_div_buf_debug_init }; static const struct clk_ops ccu_div_fixed_ops = { .recalc_rate = ccu_div_fixed_recalc_rate, .round_rate = ccu_div_fixed_round_rate, .set_rate = ccu_div_fixed_set_rate, .debug_init = ccu_div_fixed_debug_init }; struct ccu_div *ccu_div_hw_register(const struct ccu_div_init_data *div_init) { struct clk_parent_data parent_data = { }; struct clk_init_data hw_init = { }; struct ccu_div *div; int ret; if (!div_init) return ERR_PTR(-EINVAL); div = kzalloc(sizeof(*div), GFP_KERNEL); if (!div) return ERR_PTR(-ENOMEM); /* * Note since Baikal-T1 System Controller registers are MMIO-backed * we won't check the regmap IO operations return status, because it * must be zero anyway. */ div->hw.init = &hw_init; div->id = div_init->id; div->reg_ctl = div_init->base + CCU_DIV_CTL; div->sys_regs = div_init->sys_regs; div->flags = div_init->flags; div->features = div_init->features; spin_lock_init(&div->lock); hw_init.name = div_init->name; hw_init.flags = div_init->flags; if (div_init->type == CCU_DIV_VAR) { if (hw_init.flags & CLK_SET_RATE_GATE) hw_init.ops = &ccu_div_var_gate_to_set_ops; else hw_init.ops = &ccu_div_var_nogate_ops; div->mask = CCU_DIV_CTL_CLKDIV_MASK(div_init->width); } else if (div_init->type == CCU_DIV_GATE) { hw_init.ops = &ccu_div_gate_ops; div->divider = div_init->divider; } else if (div_init->type == CCU_DIV_BUF) { hw_init.ops = &ccu_div_buf_ops; } else if (div_init->type == CCU_DIV_FIXED) { hw_init.ops = &ccu_div_fixed_ops; div->divider = div_init->divider; } else { ret = -EINVAL; goto err_free_div; } if (!div_init->parent_name) { ret = -EINVAL; goto err_free_div; } parent_data.fw_name = div_init->parent_name; parent_data.name = div_init->parent_name; hw_init.parent_data = &parent_data; hw_init.num_parents = 1; ret = of_clk_hw_register(div_init->np, &div->hw); if (ret) goto err_free_div; return div; err_free_div: kfree(div); return ERR_PTR(ret); } void ccu_div_hw_unregister(struct ccu_div *div) { clk_hw_unregister(&div->hw); kfree(div); }