852 lines
21 KiB
C
852 lines
21 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
/*
|
|
* Ingenic SoC CGU driver
|
|
*
|
|
* Copyright (c) 2013-2015 Imagination Technologies
|
|
* Author: Paul Burton <paul.burton@mips.com>
|
|
*/
|
|
|
|
#include <linux/bitops.h>
|
|
#include <linux/clk.h>
|
|
#include <linux/clk-provider.h>
|
|
#include <linux/clkdev.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/io.h>
|
|
#include <linux/iopoll.h>
|
|
#include <linux/math64.h>
|
|
#include <linux/of.h>
|
|
#include <linux/of_address.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/time.h>
|
|
|
|
#include "cgu.h"
|
|
|
|
#define MHZ (1000 * 1000)
|
|
|
|
static inline const struct ingenic_cgu_clk_info *
|
|
to_clk_info(struct ingenic_clk *clk)
|
|
{
|
|
return &clk->cgu->clock_info[clk->idx];
|
|
}
|
|
|
|
/**
|
|
* ingenic_cgu_gate_get() - get the value of clock gate register bit
|
|
* @cgu: reference to the CGU whose registers should be read
|
|
* @info: info struct describing the gate bit
|
|
*
|
|
* Retrieves the state of the clock gate bit described by info. The
|
|
* caller must hold cgu->lock.
|
|
*
|
|
* Return: true if the gate bit is set, else false.
|
|
*/
|
|
static inline bool
|
|
ingenic_cgu_gate_get(struct ingenic_cgu *cgu,
|
|
const struct ingenic_cgu_gate_info *info)
|
|
{
|
|
return !!(readl(cgu->base + info->reg) & BIT(info->bit))
|
|
^ info->clear_to_gate;
|
|
}
|
|
|
|
/**
|
|
* ingenic_cgu_gate_set() - set the value of clock gate register bit
|
|
* @cgu: reference to the CGU whose registers should be modified
|
|
* @info: info struct describing the gate bit
|
|
* @val: non-zero to gate a clock, otherwise zero
|
|
*
|
|
* Sets the given gate bit in order to gate or ungate a clock.
|
|
*
|
|
* The caller must hold cgu->lock.
|
|
*/
|
|
static inline void
|
|
ingenic_cgu_gate_set(struct ingenic_cgu *cgu,
|
|
const struct ingenic_cgu_gate_info *info, bool val)
|
|
{
|
|
u32 clkgr = readl(cgu->base + info->reg);
|
|
|
|
if (val ^ info->clear_to_gate)
|
|
clkgr |= BIT(info->bit);
|
|
else
|
|
clkgr &= ~BIT(info->bit);
|
|
|
|
writel(clkgr, cgu->base + info->reg);
|
|
}
|
|
|
|
/*
|
|
* PLL operations
|
|
*/
|
|
|
|
static unsigned long
|
|
ingenic_pll_recalc_rate(struct clk_hw *hw, unsigned long parent_rate)
|
|
{
|
|
struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
|
|
const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk);
|
|
struct ingenic_cgu *cgu = ingenic_clk->cgu;
|
|
const struct ingenic_cgu_pll_info *pll_info;
|
|
unsigned m, n, od, od_enc = 0;
|
|
bool bypass;
|
|
u32 ctl;
|
|
|
|
BUG_ON(clk_info->type != CGU_CLK_PLL);
|
|
pll_info = &clk_info->pll;
|
|
|
|
ctl = readl(cgu->base + pll_info->reg);
|
|
|
|
m = (ctl >> pll_info->m_shift) & GENMASK(pll_info->m_bits - 1, 0);
|
|
m += pll_info->m_offset;
|
|
n = (ctl >> pll_info->n_shift) & GENMASK(pll_info->n_bits - 1, 0);
|
|
n += pll_info->n_offset;
|
|
|
|
if (pll_info->od_bits > 0) {
|
|
od_enc = ctl >> pll_info->od_shift;
|
|
od_enc &= GENMASK(pll_info->od_bits - 1, 0);
|
|
}
|
|
|
|
if (pll_info->bypass_bit >= 0) {
|
|
ctl = readl(cgu->base + pll_info->bypass_reg);
|
|
|
|
bypass = !!(ctl & BIT(pll_info->bypass_bit));
|
|
|
|
if (bypass)
|
|
return parent_rate;
|
|
}
|
|
|
|
for (od = 0; od < pll_info->od_max; od++)
|
|
if (pll_info->od_encoding[od] == od_enc)
|
|
break;
|
|
|
|
/* if od_max = 0, od_bits should be 0 and od is fixed to 1. */
|
|
if (pll_info->od_max == 0)
|
|
BUG_ON(pll_info->od_bits != 0);
|
|
else
|
|
BUG_ON(od == pll_info->od_max);
|
|
od++;
|
|
|
|
return div_u64((u64)parent_rate * m * pll_info->rate_multiplier,
|
|
n * od);
|
|
}
|
|
|
|
static void
|
|
ingenic_pll_calc_m_n_od(const struct ingenic_cgu_pll_info *pll_info,
|
|
unsigned long rate, unsigned long parent_rate,
|
|
unsigned int *pm, unsigned int *pn, unsigned int *pod)
|
|
{
|
|
unsigned int m, n, od = 1;
|
|
|
|
/*
|
|
* The frequency after the input divider must be between 10 and 50 MHz.
|
|
* The highest divider yields the best resolution.
|
|
*/
|
|
n = parent_rate / (10 * MHZ);
|
|
n = min_t(unsigned int, n, 1 << pll_info->n_bits);
|
|
n = max_t(unsigned int, n, pll_info->n_offset);
|
|
|
|
m = (rate / MHZ) * od * n / (parent_rate / MHZ);
|
|
m = min_t(unsigned int, m, 1 << pll_info->m_bits);
|
|
m = max_t(unsigned int, m, pll_info->m_offset);
|
|
|
|
*pm = m;
|
|
*pn = n;
|
|
*pod = od;
|
|
}
|
|
|
|
static unsigned long
|
|
ingenic_pll_calc(const struct ingenic_cgu_clk_info *clk_info,
|
|
unsigned long rate, unsigned long parent_rate,
|
|
unsigned int *pm, unsigned int *pn, unsigned int *pod)
|
|
{
|
|
const struct ingenic_cgu_pll_info *pll_info = &clk_info->pll;
|
|
unsigned int m, n, od;
|
|
|
|
if (pll_info->calc_m_n_od)
|
|
(*pll_info->calc_m_n_od)(pll_info, rate, parent_rate, &m, &n, &od);
|
|
else
|
|
ingenic_pll_calc_m_n_od(pll_info, rate, parent_rate, &m, &n, &od);
|
|
|
|
if (pm)
|
|
*pm = m;
|
|
if (pn)
|
|
*pn = n;
|
|
if (pod)
|
|
*pod = od;
|
|
|
|
return div_u64((u64)parent_rate * m * pll_info->rate_multiplier,
|
|
n * od);
|
|
}
|
|
|
|
static long
|
|
ingenic_pll_round_rate(struct clk_hw *hw, unsigned long req_rate,
|
|
unsigned long *prate)
|
|
{
|
|
struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
|
|
const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk);
|
|
|
|
return ingenic_pll_calc(clk_info, req_rate, *prate, NULL, NULL, NULL);
|
|
}
|
|
|
|
static inline int ingenic_pll_check_stable(struct ingenic_cgu *cgu,
|
|
const struct ingenic_cgu_pll_info *pll_info)
|
|
{
|
|
u32 ctl;
|
|
|
|
if (pll_info->stable_bit < 0)
|
|
return 0;
|
|
|
|
return readl_poll_timeout(cgu->base + pll_info->reg, ctl,
|
|
ctl & BIT(pll_info->stable_bit),
|
|
0, 100 * USEC_PER_MSEC);
|
|
}
|
|
|
|
static int
|
|
ingenic_pll_set_rate(struct clk_hw *hw, unsigned long req_rate,
|
|
unsigned long parent_rate)
|
|
{
|
|
struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
|
|
struct ingenic_cgu *cgu = ingenic_clk->cgu;
|
|
const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk);
|
|
const struct ingenic_cgu_pll_info *pll_info = &clk_info->pll;
|
|
unsigned long rate, flags;
|
|
unsigned int m, n, od;
|
|
int ret = 0;
|
|
u32 ctl;
|
|
|
|
rate = ingenic_pll_calc(clk_info, req_rate, parent_rate,
|
|
&m, &n, &od);
|
|
if (rate != req_rate)
|
|
pr_info("ingenic-cgu: request '%s' rate %luHz, actual %luHz\n",
|
|
clk_info->name, req_rate, rate);
|
|
|
|
spin_lock_irqsave(&cgu->lock, flags);
|
|
ctl = readl(cgu->base + pll_info->reg);
|
|
|
|
ctl &= ~(GENMASK(pll_info->m_bits - 1, 0) << pll_info->m_shift);
|
|
ctl |= (m - pll_info->m_offset) << pll_info->m_shift;
|
|
|
|
ctl &= ~(GENMASK(pll_info->n_bits - 1, 0) << pll_info->n_shift);
|
|
ctl |= (n - pll_info->n_offset) << pll_info->n_shift;
|
|
|
|
if (pll_info->od_bits > 0) {
|
|
ctl &= ~(GENMASK(pll_info->od_bits - 1, 0) << pll_info->od_shift);
|
|
ctl |= pll_info->od_encoding[od - 1] << pll_info->od_shift;
|
|
}
|
|
|
|
writel(ctl, cgu->base + pll_info->reg);
|
|
|
|
if (pll_info->set_rate_hook)
|
|
pll_info->set_rate_hook(pll_info, rate, parent_rate);
|
|
|
|
/* If the PLL is enabled, verify that it's stable */
|
|
if (pll_info->enable_bit >= 0 && (ctl & BIT(pll_info->enable_bit)))
|
|
ret = ingenic_pll_check_stable(cgu, pll_info);
|
|
|
|
spin_unlock_irqrestore(&cgu->lock, flags);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ingenic_pll_enable(struct clk_hw *hw)
|
|
{
|
|
struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
|
|
struct ingenic_cgu *cgu = ingenic_clk->cgu;
|
|
const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk);
|
|
const struct ingenic_cgu_pll_info *pll_info = &clk_info->pll;
|
|
unsigned long flags;
|
|
int ret;
|
|
u32 ctl;
|
|
|
|
if (pll_info->enable_bit < 0)
|
|
return 0;
|
|
|
|
spin_lock_irqsave(&cgu->lock, flags);
|
|
if (pll_info->bypass_bit >= 0) {
|
|
ctl = readl(cgu->base + pll_info->bypass_reg);
|
|
|
|
ctl &= ~BIT(pll_info->bypass_bit);
|
|
|
|
writel(ctl, cgu->base + pll_info->bypass_reg);
|
|
}
|
|
|
|
ctl = readl(cgu->base + pll_info->reg);
|
|
|
|
ctl |= BIT(pll_info->enable_bit);
|
|
|
|
writel(ctl, cgu->base + pll_info->reg);
|
|
|
|
ret = ingenic_pll_check_stable(cgu, pll_info);
|
|
spin_unlock_irqrestore(&cgu->lock, flags);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void ingenic_pll_disable(struct clk_hw *hw)
|
|
{
|
|
struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
|
|
struct ingenic_cgu *cgu = ingenic_clk->cgu;
|
|
const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk);
|
|
const struct ingenic_cgu_pll_info *pll_info = &clk_info->pll;
|
|
unsigned long flags;
|
|
u32 ctl;
|
|
|
|
if (pll_info->enable_bit < 0)
|
|
return;
|
|
|
|
spin_lock_irqsave(&cgu->lock, flags);
|
|
ctl = readl(cgu->base + pll_info->reg);
|
|
|
|
ctl &= ~BIT(pll_info->enable_bit);
|
|
|
|
writel(ctl, cgu->base + pll_info->reg);
|
|
spin_unlock_irqrestore(&cgu->lock, flags);
|
|
}
|
|
|
|
static int ingenic_pll_is_enabled(struct clk_hw *hw)
|
|
{
|
|
struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
|
|
struct ingenic_cgu *cgu = ingenic_clk->cgu;
|
|
const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk);
|
|
const struct ingenic_cgu_pll_info *pll_info = &clk_info->pll;
|
|
u32 ctl;
|
|
|
|
if (pll_info->enable_bit < 0)
|
|
return true;
|
|
|
|
ctl = readl(cgu->base + pll_info->reg);
|
|
|
|
return !!(ctl & BIT(pll_info->enable_bit));
|
|
}
|
|
|
|
static const struct clk_ops ingenic_pll_ops = {
|
|
.recalc_rate = ingenic_pll_recalc_rate,
|
|
.round_rate = ingenic_pll_round_rate,
|
|
.set_rate = ingenic_pll_set_rate,
|
|
|
|
.enable = ingenic_pll_enable,
|
|
.disable = ingenic_pll_disable,
|
|
.is_enabled = ingenic_pll_is_enabled,
|
|
};
|
|
|
|
/*
|
|
* Operations for all non-PLL clocks
|
|
*/
|
|
|
|
static u8 ingenic_clk_get_parent(struct clk_hw *hw)
|
|
{
|
|
struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
|
|
const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk);
|
|
struct ingenic_cgu *cgu = ingenic_clk->cgu;
|
|
u32 reg;
|
|
u8 i, hw_idx, idx = 0;
|
|
|
|
if (clk_info->type & CGU_CLK_MUX) {
|
|
reg = readl(cgu->base + clk_info->mux.reg);
|
|
hw_idx = (reg >> clk_info->mux.shift) &
|
|
GENMASK(clk_info->mux.bits - 1, 0);
|
|
|
|
/*
|
|
* Convert the hardware index to the parent index by skipping
|
|
* over any -1's in the parents array.
|
|
*/
|
|
for (i = 0; i < hw_idx; i++) {
|
|
if (clk_info->parents[i] != -1)
|
|
idx++;
|
|
}
|
|
}
|
|
|
|
return idx;
|
|
}
|
|
|
|
static int ingenic_clk_set_parent(struct clk_hw *hw, u8 idx)
|
|
{
|
|
struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
|
|
const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk);
|
|
struct ingenic_cgu *cgu = ingenic_clk->cgu;
|
|
unsigned long flags;
|
|
u8 curr_idx, hw_idx, num_poss;
|
|
u32 reg, mask;
|
|
|
|
if (clk_info->type & CGU_CLK_MUX) {
|
|
/*
|
|
* Convert the parent index to the hardware index by adding
|
|
* 1 for any -1 in the parents array preceding the given
|
|
* index. That is, we want the index of idx'th entry in
|
|
* clk_info->parents which does not equal -1.
|
|
*/
|
|
hw_idx = curr_idx = 0;
|
|
num_poss = 1 << clk_info->mux.bits;
|
|
for (; hw_idx < num_poss; hw_idx++) {
|
|
if (clk_info->parents[hw_idx] == -1)
|
|
continue;
|
|
if (curr_idx == idx)
|
|
break;
|
|
curr_idx++;
|
|
}
|
|
|
|
/* idx should always be a valid parent */
|
|
BUG_ON(curr_idx != idx);
|
|
|
|
mask = GENMASK(clk_info->mux.bits - 1, 0);
|
|
mask <<= clk_info->mux.shift;
|
|
|
|
spin_lock_irqsave(&cgu->lock, flags);
|
|
|
|
/* write the register */
|
|
reg = readl(cgu->base + clk_info->mux.reg);
|
|
reg &= ~mask;
|
|
reg |= hw_idx << clk_info->mux.shift;
|
|
writel(reg, cgu->base + clk_info->mux.reg);
|
|
|
|
spin_unlock_irqrestore(&cgu->lock, flags);
|
|
return 0;
|
|
}
|
|
|
|
return idx ? -EINVAL : 0;
|
|
}
|
|
|
|
static unsigned long
|
|
ingenic_clk_recalc_rate(struct clk_hw *hw, unsigned long parent_rate)
|
|
{
|
|
struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
|
|
const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk);
|
|
struct ingenic_cgu *cgu = ingenic_clk->cgu;
|
|
unsigned long rate = parent_rate;
|
|
u32 div_reg, div;
|
|
u8 parent;
|
|
|
|
if (clk_info->type & CGU_CLK_DIV) {
|
|
parent = ingenic_clk_get_parent(hw);
|
|
|
|
if (!(clk_info->div.bypass_mask & BIT(parent))) {
|
|
div_reg = readl(cgu->base + clk_info->div.reg);
|
|
div = (div_reg >> clk_info->div.shift) &
|
|
GENMASK(clk_info->div.bits - 1, 0);
|
|
|
|
if (clk_info->div.div_table)
|
|
div = clk_info->div.div_table[div];
|
|
else
|
|
div = (div + 1) * clk_info->div.div;
|
|
|
|
rate /= div;
|
|
}
|
|
} else if (clk_info->type & CGU_CLK_FIXDIV) {
|
|
rate /= clk_info->fixdiv.div;
|
|
}
|
|
|
|
return rate;
|
|
}
|
|
|
|
static unsigned int
|
|
ingenic_clk_calc_hw_div(const struct ingenic_cgu_clk_info *clk_info,
|
|
unsigned int div)
|
|
{
|
|
unsigned int i, best_i = 0, best = (unsigned int)-1;
|
|
|
|
for (i = 0; i < (1 << clk_info->div.bits)
|
|
&& clk_info->div.div_table[i]; i++) {
|
|
if (clk_info->div.div_table[i] >= div &&
|
|
clk_info->div.div_table[i] < best) {
|
|
best = clk_info->div.div_table[i];
|
|
best_i = i;
|
|
|
|
if (div == best)
|
|
break;
|
|
}
|
|
}
|
|
|
|
return best_i;
|
|
}
|
|
|
|
static unsigned
|
|
ingenic_clk_calc_div(struct clk_hw *hw,
|
|
const struct ingenic_cgu_clk_info *clk_info,
|
|
unsigned long parent_rate, unsigned long req_rate)
|
|
{
|
|
unsigned int div, hw_div;
|
|
u8 parent;
|
|
|
|
parent = ingenic_clk_get_parent(hw);
|
|
if (clk_info->div.bypass_mask & BIT(parent))
|
|
return 1;
|
|
|
|
/* calculate the divide */
|
|
div = DIV_ROUND_UP(parent_rate, req_rate);
|
|
|
|
if (clk_info->div.div_table) {
|
|
hw_div = ingenic_clk_calc_hw_div(clk_info, div);
|
|
|
|
return clk_info->div.div_table[hw_div];
|
|
}
|
|
|
|
/* Impose hardware constraints */
|
|
div = clamp_t(unsigned int, div, clk_info->div.div,
|
|
clk_info->div.div << clk_info->div.bits);
|
|
|
|
/*
|
|
* If the divider value itself must be divided before being written to
|
|
* the divider register, we must ensure we don't have any bits set that
|
|
* would be lost as a result of doing so.
|
|
*/
|
|
div = DIV_ROUND_UP(div, clk_info->div.div);
|
|
div *= clk_info->div.div;
|
|
|
|
return div;
|
|
}
|
|
|
|
static long
|
|
ingenic_clk_round_rate(struct clk_hw *hw, unsigned long req_rate,
|
|
unsigned long *parent_rate)
|
|
{
|
|
struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
|
|
const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk);
|
|
unsigned int div = 1;
|
|
|
|
if (clk_info->type & CGU_CLK_DIV)
|
|
div = ingenic_clk_calc_div(hw, clk_info, *parent_rate, req_rate);
|
|
else if (clk_info->type & CGU_CLK_FIXDIV)
|
|
div = clk_info->fixdiv.div;
|
|
else if (clk_hw_can_set_rate_parent(hw))
|
|
*parent_rate = req_rate;
|
|
|
|
return DIV_ROUND_UP(*parent_rate, div);
|
|
}
|
|
|
|
static inline int ingenic_clk_check_stable(struct ingenic_cgu *cgu,
|
|
const struct ingenic_cgu_clk_info *clk_info)
|
|
{
|
|
u32 reg;
|
|
|
|
return readl_poll_timeout(cgu->base + clk_info->div.reg, reg,
|
|
!(reg & BIT(clk_info->div.busy_bit)),
|
|
0, 100 * USEC_PER_MSEC);
|
|
}
|
|
|
|
static int
|
|
ingenic_clk_set_rate(struct clk_hw *hw, unsigned long req_rate,
|
|
unsigned long parent_rate)
|
|
{
|
|
struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
|
|
const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk);
|
|
struct ingenic_cgu *cgu = ingenic_clk->cgu;
|
|
unsigned long rate, flags;
|
|
unsigned int hw_div, div;
|
|
u32 reg, mask;
|
|
int ret = 0;
|
|
|
|
if (clk_info->type & CGU_CLK_DIV) {
|
|
div = ingenic_clk_calc_div(hw, clk_info, parent_rate, req_rate);
|
|
rate = DIV_ROUND_UP(parent_rate, div);
|
|
|
|
if (rate != req_rate)
|
|
return -EINVAL;
|
|
|
|
if (clk_info->div.div_table)
|
|
hw_div = ingenic_clk_calc_hw_div(clk_info, div);
|
|
else
|
|
hw_div = ((div / clk_info->div.div) - 1);
|
|
|
|
spin_lock_irqsave(&cgu->lock, flags);
|
|
reg = readl(cgu->base + clk_info->div.reg);
|
|
|
|
/* update the divide */
|
|
mask = GENMASK(clk_info->div.bits - 1, 0);
|
|
reg &= ~(mask << clk_info->div.shift);
|
|
reg |= hw_div << clk_info->div.shift;
|
|
|
|
/* clear the stop bit */
|
|
if (clk_info->div.stop_bit != -1)
|
|
reg &= ~BIT(clk_info->div.stop_bit);
|
|
|
|
/* set the change enable bit */
|
|
if (clk_info->div.ce_bit != -1)
|
|
reg |= BIT(clk_info->div.ce_bit);
|
|
|
|
/* update the hardware */
|
|
writel(reg, cgu->base + clk_info->div.reg);
|
|
|
|
/* wait for the change to take effect */
|
|
if (clk_info->div.busy_bit != -1)
|
|
ret = ingenic_clk_check_stable(cgu, clk_info);
|
|
|
|
spin_unlock_irqrestore(&cgu->lock, flags);
|
|
return ret;
|
|
}
|
|
|
|
return -EINVAL;
|
|
}
|
|
|
|
static int ingenic_clk_enable(struct clk_hw *hw)
|
|
{
|
|
struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
|
|
const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk);
|
|
struct ingenic_cgu *cgu = ingenic_clk->cgu;
|
|
unsigned long flags;
|
|
|
|
if (clk_info->type & CGU_CLK_GATE) {
|
|
/* ungate the clock */
|
|
spin_lock_irqsave(&cgu->lock, flags);
|
|
ingenic_cgu_gate_set(cgu, &clk_info->gate, false);
|
|
spin_unlock_irqrestore(&cgu->lock, flags);
|
|
|
|
if (clk_info->gate.delay_us)
|
|
udelay(clk_info->gate.delay_us);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ingenic_clk_disable(struct clk_hw *hw)
|
|
{
|
|
struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
|
|
const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk);
|
|
struct ingenic_cgu *cgu = ingenic_clk->cgu;
|
|
unsigned long flags;
|
|
|
|
if (clk_info->type & CGU_CLK_GATE) {
|
|
/* gate the clock */
|
|
spin_lock_irqsave(&cgu->lock, flags);
|
|
ingenic_cgu_gate_set(cgu, &clk_info->gate, true);
|
|
spin_unlock_irqrestore(&cgu->lock, flags);
|
|
}
|
|
}
|
|
|
|
static int ingenic_clk_is_enabled(struct clk_hw *hw)
|
|
{
|
|
struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
|
|
const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk);
|
|
struct ingenic_cgu *cgu = ingenic_clk->cgu;
|
|
int enabled = 1;
|
|
|
|
if (clk_info->type & CGU_CLK_GATE)
|
|
enabled = !ingenic_cgu_gate_get(cgu, &clk_info->gate);
|
|
|
|
return enabled;
|
|
}
|
|
|
|
static const struct clk_ops ingenic_clk_ops = {
|
|
.get_parent = ingenic_clk_get_parent,
|
|
.set_parent = ingenic_clk_set_parent,
|
|
|
|
.recalc_rate = ingenic_clk_recalc_rate,
|
|
.round_rate = ingenic_clk_round_rate,
|
|
.set_rate = ingenic_clk_set_rate,
|
|
|
|
.enable = ingenic_clk_enable,
|
|
.disable = ingenic_clk_disable,
|
|
.is_enabled = ingenic_clk_is_enabled,
|
|
};
|
|
|
|
/*
|
|
* Setup functions.
|
|
*/
|
|
|
|
static int ingenic_register_clock(struct ingenic_cgu *cgu, unsigned idx)
|
|
{
|
|
const struct ingenic_cgu_clk_info *clk_info = &cgu->clock_info[idx];
|
|
struct clk_init_data clk_init;
|
|
struct ingenic_clk *ingenic_clk = NULL;
|
|
struct clk *clk, *parent;
|
|
const char *parent_names[4];
|
|
unsigned caps, i, num_possible;
|
|
int err = -EINVAL;
|
|
|
|
BUILD_BUG_ON(ARRAY_SIZE(clk_info->parents) > ARRAY_SIZE(parent_names));
|
|
|
|
if (clk_info->type == CGU_CLK_EXT) {
|
|
clk = of_clk_get_by_name(cgu->np, clk_info->name);
|
|
if (IS_ERR(clk)) {
|
|
pr_err("%s: no external clock '%s' provided\n",
|
|
__func__, clk_info->name);
|
|
err = -ENODEV;
|
|
goto out;
|
|
}
|
|
err = clk_register_clkdev(clk, clk_info->name, NULL);
|
|
if (err) {
|
|
clk_put(clk);
|
|
goto out;
|
|
}
|
|
cgu->clocks.clks[idx] = clk;
|
|
return 0;
|
|
}
|
|
|
|
if (!clk_info->type) {
|
|
pr_err("%s: no clock type specified for '%s'\n", __func__,
|
|
clk_info->name);
|
|
goto out;
|
|
}
|
|
|
|
ingenic_clk = kzalloc(sizeof(*ingenic_clk), GFP_KERNEL);
|
|
if (!ingenic_clk) {
|
|
err = -ENOMEM;
|
|
goto out;
|
|
}
|
|
|
|
ingenic_clk->hw.init = &clk_init;
|
|
ingenic_clk->cgu = cgu;
|
|
ingenic_clk->idx = idx;
|
|
|
|
clk_init.name = clk_info->name;
|
|
clk_init.flags = clk_info->flags;
|
|
clk_init.parent_names = parent_names;
|
|
|
|
caps = clk_info->type;
|
|
|
|
if (caps & CGU_CLK_DIV) {
|
|
caps &= ~CGU_CLK_DIV;
|
|
} else if (!(caps & CGU_CLK_CUSTOM)) {
|
|
/* pass rate changes to the parent clock */
|
|
clk_init.flags |= CLK_SET_RATE_PARENT;
|
|
}
|
|
|
|
if (caps & (CGU_CLK_MUX | CGU_CLK_CUSTOM)) {
|
|
clk_init.num_parents = 0;
|
|
|
|
if (caps & CGU_CLK_MUX)
|
|
num_possible = 1 << clk_info->mux.bits;
|
|
else
|
|
num_possible = ARRAY_SIZE(clk_info->parents);
|
|
|
|
for (i = 0; i < num_possible; i++) {
|
|
if (clk_info->parents[i] == -1)
|
|
continue;
|
|
|
|
parent = cgu->clocks.clks[clk_info->parents[i]];
|
|
parent_names[clk_init.num_parents] =
|
|
__clk_get_name(parent);
|
|
clk_init.num_parents++;
|
|
}
|
|
|
|
BUG_ON(!clk_init.num_parents);
|
|
BUG_ON(clk_init.num_parents > ARRAY_SIZE(parent_names));
|
|
} else {
|
|
BUG_ON(clk_info->parents[0] == -1);
|
|
clk_init.num_parents = 1;
|
|
parent = cgu->clocks.clks[clk_info->parents[0]];
|
|
parent_names[0] = __clk_get_name(parent);
|
|
}
|
|
|
|
if (caps & CGU_CLK_CUSTOM) {
|
|
clk_init.ops = clk_info->custom.clk_ops;
|
|
|
|
caps &= ~CGU_CLK_CUSTOM;
|
|
|
|
if (caps) {
|
|
pr_err("%s: custom clock may not be combined with type 0x%x\n",
|
|
__func__, caps);
|
|
goto out;
|
|
}
|
|
} else if (caps & CGU_CLK_PLL) {
|
|
clk_init.ops = &ingenic_pll_ops;
|
|
|
|
caps &= ~CGU_CLK_PLL;
|
|
|
|
if (caps) {
|
|
pr_err("%s: PLL may not be combined with type 0x%x\n",
|
|
__func__, caps);
|
|
goto out;
|
|
}
|
|
} else {
|
|
clk_init.ops = &ingenic_clk_ops;
|
|
}
|
|
|
|
/* nothing to do for gates or fixed dividers */
|
|
caps &= ~(CGU_CLK_GATE | CGU_CLK_FIXDIV);
|
|
|
|
if (caps & CGU_CLK_MUX) {
|
|
if (!(caps & CGU_CLK_MUX_GLITCHFREE))
|
|
clk_init.flags |= CLK_SET_PARENT_GATE;
|
|
|
|
caps &= ~(CGU_CLK_MUX | CGU_CLK_MUX_GLITCHFREE);
|
|
}
|
|
|
|
if (caps) {
|
|
pr_err("%s: unknown clock type 0x%x\n", __func__, caps);
|
|
goto out;
|
|
}
|
|
|
|
clk = clk_register(NULL, &ingenic_clk->hw);
|
|
if (IS_ERR(clk)) {
|
|
pr_err("%s: failed to register clock '%s'\n", __func__,
|
|
clk_info->name);
|
|
err = PTR_ERR(clk);
|
|
goto out;
|
|
}
|
|
|
|
err = clk_register_clkdev(clk, clk_info->name, NULL);
|
|
if (err)
|
|
goto out;
|
|
|
|
cgu->clocks.clks[idx] = clk;
|
|
out:
|
|
if (err)
|
|
kfree(ingenic_clk);
|
|
return err;
|
|
}
|
|
|
|
struct ingenic_cgu *
|
|
ingenic_cgu_new(const struct ingenic_cgu_clk_info *clock_info,
|
|
unsigned num_clocks, struct device_node *np)
|
|
{
|
|
struct ingenic_cgu *cgu;
|
|
|
|
cgu = kzalloc(sizeof(*cgu), GFP_KERNEL);
|
|
if (!cgu)
|
|
goto err_out;
|
|
|
|
cgu->base = of_iomap(np, 0);
|
|
if (!cgu->base) {
|
|
pr_err("%s: failed to map CGU registers\n", __func__);
|
|
goto err_out_free;
|
|
}
|
|
|
|
cgu->np = np;
|
|
cgu->clock_info = clock_info;
|
|
cgu->clocks.clk_num = num_clocks;
|
|
|
|
spin_lock_init(&cgu->lock);
|
|
|
|
return cgu;
|
|
|
|
err_out_free:
|
|
kfree(cgu);
|
|
err_out:
|
|
return NULL;
|
|
}
|
|
|
|
int ingenic_cgu_register_clocks(struct ingenic_cgu *cgu)
|
|
{
|
|
unsigned i;
|
|
int err;
|
|
|
|
cgu->clocks.clks = kcalloc(cgu->clocks.clk_num, sizeof(struct clk *),
|
|
GFP_KERNEL);
|
|
if (!cgu->clocks.clks) {
|
|
err = -ENOMEM;
|
|
goto err_out;
|
|
}
|
|
|
|
for (i = 0; i < cgu->clocks.clk_num; i++) {
|
|
err = ingenic_register_clock(cgu, i);
|
|
if (err)
|
|
goto err_out_unregister;
|
|
}
|
|
|
|
err = of_clk_add_provider(cgu->np, of_clk_src_onecell_get,
|
|
&cgu->clocks);
|
|
if (err)
|
|
goto err_out_unregister;
|
|
|
|
return 0;
|
|
|
|
err_out_unregister:
|
|
for (i = 0; i < cgu->clocks.clk_num; i++) {
|
|
if (!cgu->clocks.clks[i])
|
|
continue;
|
|
if (cgu->clock_info[i].type & CGU_CLK_EXT)
|
|
clk_put(cgu->clocks.clks[i]);
|
|
else
|
|
clk_unregister(cgu->clocks.clks[i]);
|
|
}
|
|
kfree(cgu->clocks.clks);
|
|
err_out:
|
|
return err;
|
|
}
|