Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Paul Burton | 3343 | 77.17% | 1 | 3.57% |
Paul Cercueil | 798 | 18.42% | 17 | 60.71% |
Aidan MacDonald | 113 | 2.61% | 4 | 14.29% |
Maxime Ripard | 29 | 0.67% | 1 | 3.57% |
Harvey Hunt | 22 | 0.51% | 1 | 3.57% |
Zhou Yanjie | 19 | 0.44% | 1 | 3.57% |
Michael Turquette | 3 | 0.07% | 1 | 3.57% |
Stephen Boyd | 3 | 0.07% | 1 | 3.57% |
Thomas Gleixner | 2 | 0.05% | 1 | 3.57% |
Total | 4332 | 28 |
// SPDX-License-Identifier: GPL-2.0-or-later /* * Ingenic SoC CGU driver * * Copyright (c) 2013-2015 Imagination Technologies * Author: Paul Burton <paul.burton@mips.com> */ #include <linux/bitops.h> #include <linux/clk.h> #include <linux/clk-provider.h> #include <linux/clkdev.h> #include <linux/delay.h> #include <linux/io.h> #include <linux/iopoll.h> #include <linux/math64.h> #include <linux/of.h> #include <linux/of_address.h> #include <linux/slab.h> #include <linux/spinlock.h> #include <linux/time.h> #include "cgu.h" #define MHZ (1000 * 1000) static inline const struct ingenic_cgu_clk_info * to_clk_info(struct ingenic_clk *clk) { return &clk->cgu->clock_info[clk->idx]; } /** * ingenic_cgu_gate_get() - get the value of clock gate register bit * @cgu: reference to the CGU whose registers should be read * @info: info struct describing the gate bit * * Retrieves the state of the clock gate bit described by info. The * caller must hold cgu->lock. * * Return: true if the gate bit is set, else false. */ static inline bool ingenic_cgu_gate_get(struct ingenic_cgu *cgu, const struct ingenic_cgu_gate_info *info) { return !!(readl(cgu->base + info->reg) & BIT(info->bit)) ^ info->clear_to_gate; } /** * ingenic_cgu_gate_set() - set the value of clock gate register bit * @cgu: reference to the CGU whose registers should be modified * @info: info struct describing the gate bit * @val: non-zero to gate a clock, otherwise zero * * Sets the given gate bit in order to gate or ungate a clock. * * The caller must hold cgu->lock. */ static inline void ingenic_cgu_gate_set(struct ingenic_cgu *cgu, const struct ingenic_cgu_gate_info *info, bool val) { u32 clkgr = readl(cgu->base + info->reg); if (val ^ info->clear_to_gate) clkgr |= BIT(info->bit); else clkgr &= ~BIT(info->bit); writel(clkgr, cgu->base + info->reg); } /* * PLL operations */ static unsigned long ingenic_pll_recalc_rate(struct clk_hw *hw, unsigned long parent_rate) { struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw); const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk); struct ingenic_cgu *cgu = ingenic_clk->cgu; const struct ingenic_cgu_pll_info *pll_info; unsigned m, n, od, od_enc = 0; bool bypass; u32 ctl; BUG_ON(clk_info->type != CGU_CLK_PLL); pll_info = &clk_info->pll; ctl = readl(cgu->base + pll_info->reg); m = (ctl >> pll_info->m_shift) & GENMASK(pll_info->m_bits - 1, 0); m += pll_info->m_offset; n = (ctl >> pll_info->n_shift) & GENMASK(pll_info->n_bits - 1, 0); n += pll_info->n_offset; if (pll_info->od_bits > 0) { od_enc = ctl >> pll_info->od_shift; od_enc &= GENMASK(pll_info->od_bits - 1, 0); } if (pll_info->bypass_bit >= 0) { ctl = readl(cgu->base + pll_info->bypass_reg); bypass = !!(ctl & BIT(pll_info->bypass_bit)); if (bypass) return parent_rate; } for (od = 0; od < pll_info->od_max; od++) if (pll_info->od_encoding[od] == od_enc) break; /* if od_max = 0, od_bits should be 0 and od is fixed to 1. */ if (pll_info->od_max == 0) BUG_ON(pll_info->od_bits != 0); else BUG_ON(od == pll_info->od_max); od++; return div_u64((u64)parent_rate * m * pll_info->rate_multiplier, n * od); } static void ingenic_pll_calc_m_n_od(const struct ingenic_cgu_pll_info *pll_info, unsigned long rate, unsigned long parent_rate, unsigned int *pm, unsigned int *pn, unsigned int *pod) { unsigned int m, n, od = 1; /* * The frequency after the input divider must be between 10 and 50 MHz. * The highest divider yields the best resolution. */ n = parent_rate / (10 * MHZ); n = min_t(unsigned int, n, 1 << pll_info->n_bits); n = max_t(unsigned int, n, pll_info->n_offset); m = (rate / MHZ) * od * n / (parent_rate / MHZ); m = min_t(unsigned int, m, 1 << pll_info->m_bits); m = max_t(unsigned int, m, pll_info->m_offset); *pm = m; *pn = n; *pod = od; } static unsigned long ingenic_pll_calc(const struct ingenic_cgu_clk_info *clk_info, unsigned long rate, unsigned long parent_rate, unsigned int *pm, unsigned int *pn, unsigned int *pod) { const struct ingenic_cgu_pll_info *pll_info = &clk_info->pll; unsigned int m, n, od; if (pll_info->calc_m_n_od) (*pll_info->calc_m_n_od)(pll_info, rate, parent_rate, &m, &n, &od); else ingenic_pll_calc_m_n_od(pll_info, rate, parent_rate, &m, &n, &od); if (pm) *pm = m; if (pn) *pn = n; if (pod) *pod = od; return div_u64((u64)parent_rate * m * pll_info->rate_multiplier, n * od); } static long ingenic_pll_round_rate(struct clk_hw *hw, unsigned long req_rate, unsigned long *prate) { struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw); const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk); return ingenic_pll_calc(clk_info, req_rate, *prate, NULL, NULL, NULL); } static inline int ingenic_pll_check_stable(struct ingenic_cgu *cgu, const struct ingenic_cgu_pll_info *pll_info) { u32 ctl; if (pll_info->stable_bit < 0) return 0; return readl_poll_timeout(cgu->base + pll_info->reg, ctl, ctl & BIT(pll_info->stable_bit), 0, 100 * USEC_PER_MSEC); } static int ingenic_pll_set_rate(struct clk_hw *hw, unsigned long req_rate, unsigned long parent_rate) { struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw); struct ingenic_cgu *cgu = ingenic_clk->cgu; const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk); const struct ingenic_cgu_pll_info *pll_info = &clk_info->pll; unsigned long rate, flags; unsigned int m, n, od; int ret = 0; u32 ctl; rate = ingenic_pll_calc(clk_info, req_rate, parent_rate, &m, &n, &od); if (rate != req_rate) pr_info("ingenic-cgu: request '%s' rate %luHz, actual %luHz\n", clk_info->name, req_rate, rate); spin_lock_irqsave(&cgu->lock, flags); ctl = readl(cgu->base + pll_info->reg); ctl &= ~(GENMASK(pll_info->m_bits - 1, 0) << pll_info->m_shift); ctl |= (m - pll_info->m_offset) << pll_info->m_shift; ctl &= ~(GENMASK(pll_info->n_bits - 1, 0) << pll_info->n_shift); ctl |= (n - pll_info->n_offset) << pll_info->n_shift; if (pll_info->od_bits > 0) { ctl &= ~(GENMASK(pll_info->od_bits - 1, 0) << pll_info->od_shift); ctl |= pll_info->od_encoding[od - 1] << pll_info->od_shift; } writel(ctl, cgu->base + pll_info->reg); if (pll_info->set_rate_hook) pll_info->set_rate_hook(pll_info, rate, parent_rate); /* If the PLL is enabled, verify that it's stable */ if (pll_info->enable_bit >= 0 && (ctl & BIT(pll_info->enable_bit))) ret = ingenic_pll_check_stable(cgu, pll_info); spin_unlock_irqrestore(&cgu->lock, flags); return ret; } static int ingenic_pll_enable(struct clk_hw *hw) { struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw); struct ingenic_cgu *cgu = ingenic_clk->cgu; const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk); const struct ingenic_cgu_pll_info *pll_info = &clk_info->pll; unsigned long flags; int ret; u32 ctl; if (pll_info->enable_bit < 0) return 0; spin_lock_irqsave(&cgu->lock, flags); if (pll_info->bypass_bit >= 0) { ctl = readl(cgu->base + pll_info->bypass_reg); ctl &= ~BIT(pll_info->bypass_bit); writel(ctl, cgu->base + pll_info->bypass_reg); } ctl = readl(cgu->base + pll_info->reg); ctl |= BIT(pll_info->enable_bit); writel(ctl, cgu->base + pll_info->reg); ret = ingenic_pll_check_stable(cgu, pll_info); spin_unlock_irqrestore(&cgu->lock, flags); return ret; } static void ingenic_pll_disable(struct clk_hw *hw) { struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw); struct ingenic_cgu *cgu = ingenic_clk->cgu; const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk); const struct ingenic_cgu_pll_info *pll_info = &clk_info->pll; unsigned long flags; u32 ctl; if (pll_info->enable_bit < 0) return; spin_lock_irqsave(&cgu->lock, flags); ctl = readl(cgu->base + pll_info->reg); ctl &= ~BIT(pll_info->enable_bit); writel(ctl, cgu->base + pll_info->reg); spin_unlock_irqrestore(&cgu->lock, flags); } static int ingenic_pll_is_enabled(struct clk_hw *hw) { struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw); struct ingenic_cgu *cgu = ingenic_clk->cgu; const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk); const struct ingenic_cgu_pll_info *pll_info = &clk_info->pll; u32 ctl; if (pll_info->enable_bit < 0) return true; ctl = readl(cgu->base + pll_info->reg); return !!(ctl & BIT(pll_info->enable_bit)); } static const struct clk_ops ingenic_pll_ops = { .recalc_rate = ingenic_pll_recalc_rate, .round_rate = ingenic_pll_round_rate, .set_rate = ingenic_pll_set_rate, .enable = ingenic_pll_enable, .disable = ingenic_pll_disable, .is_enabled = ingenic_pll_is_enabled, }; /* * Operations for all non-PLL clocks */ static u8 ingenic_clk_get_parent(struct clk_hw *hw) { struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw); const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk); struct ingenic_cgu *cgu = ingenic_clk->cgu; u32 reg; u8 i, hw_idx, idx = 0; if (clk_info->type & CGU_CLK_MUX) { reg = readl(cgu->base + clk_info->mux.reg); hw_idx = (reg >> clk_info->mux.shift) & GENMASK(clk_info->mux.bits - 1, 0); /* * Convert the hardware index to the parent index by skipping * over any -1's in the parents array. */ for (i = 0; i < hw_idx; i++) { if (clk_info->parents[i] != -1) idx++; } } return idx; } static int ingenic_clk_set_parent(struct clk_hw *hw, u8 idx) { struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw); const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk); struct ingenic_cgu *cgu = ingenic_clk->cgu; unsigned long flags; u8 curr_idx, hw_idx, num_poss; u32 reg, mask; if (clk_info->type & CGU_CLK_MUX) { /* * Convert the parent index to the hardware index by adding * 1 for any -1 in the parents array preceding the given * index. That is, we want the index of idx'th entry in * clk_info->parents which does not equal -1. */ hw_idx = curr_idx = 0; num_poss = 1 << clk_info->mux.bits; for (; hw_idx < num_poss; hw_idx++) { if (clk_info->parents[hw_idx] == -1) continue; if (curr_idx == idx) break; curr_idx++; } /* idx should always be a valid parent */ BUG_ON(curr_idx != idx); mask = GENMASK(clk_info->mux.bits - 1, 0); mask <<= clk_info->mux.shift; spin_lock_irqsave(&cgu->lock, flags); /* write the register */ reg = readl(cgu->base + clk_info->mux.reg); reg &= ~mask; reg |= hw_idx << clk_info->mux.shift; writel(reg, cgu->base + clk_info->mux.reg); spin_unlock_irqrestore(&cgu->lock, flags); return 0; } return idx ? -EINVAL : 0; } static unsigned long ingenic_clk_recalc_rate(struct clk_hw *hw, unsigned long parent_rate) { struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw); const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk); struct ingenic_cgu *cgu = ingenic_clk->cgu; unsigned long rate = parent_rate; u32 div_reg, div; u8 parent; if (clk_info->type & CGU_CLK_DIV) { parent = ingenic_clk_get_parent(hw); if (!(clk_info->div.bypass_mask & BIT(parent))) { div_reg = readl(cgu->base + clk_info->div.reg); div = (div_reg >> clk_info->div.shift) & GENMASK(clk_info->div.bits - 1, 0); if (clk_info->div.div_table) div = clk_info->div.div_table[div]; else div = (div + 1) * clk_info->div.div; rate /= div; } } else if (clk_info->type & CGU_CLK_FIXDIV) { rate /= clk_info->fixdiv.div; } return rate; } static unsigned int ingenic_clk_calc_hw_div(const struct ingenic_cgu_clk_info *clk_info, unsigned int div) { unsigned int i, best_i = 0, best = (unsigned int)-1; for (i = 0; i < (1 << clk_info->div.bits) && clk_info->div.div_table[i]; i++) { if (clk_info->div.div_table[i] >= div && clk_info->div.div_table[i] < best) { best = clk_info->div.div_table[i]; best_i = i; if (div == best) break; } } return best_i; } static unsigned ingenic_clk_calc_div(struct clk_hw *hw, const struct ingenic_cgu_clk_info *clk_info, unsigned long parent_rate, unsigned long req_rate) { unsigned int div, hw_div; u8 parent; parent = ingenic_clk_get_parent(hw); if (clk_info->div.bypass_mask & BIT(parent)) return 1; /* calculate the divide */ div = DIV_ROUND_UP(parent_rate, req_rate); if (clk_info->div.div_table) { hw_div = ingenic_clk_calc_hw_div(clk_info, div); return clk_info->div.div_table[hw_div]; } /* Impose hardware constraints */ div = clamp_t(unsigned int, div, clk_info->div.div, clk_info->div.div << clk_info->div.bits); /* * If the divider value itself must be divided before being written to * the divider register, we must ensure we don't have any bits set that * would be lost as a result of doing so. */ div = DIV_ROUND_UP(div, clk_info->div.div); div *= clk_info->div.div; return div; } static int ingenic_clk_determine_rate(struct clk_hw *hw, struct clk_rate_request *req) { struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw); const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk); unsigned int div = 1; if (clk_info->type & CGU_CLK_DIV) div = ingenic_clk_calc_div(hw, clk_info, req->best_parent_rate, req->rate); else if (clk_info->type & CGU_CLK_FIXDIV) div = clk_info->fixdiv.div; else if (clk_hw_can_set_rate_parent(hw)) req->best_parent_rate = req->rate; req->rate = DIV_ROUND_UP(req->best_parent_rate, div); return 0; } static inline int ingenic_clk_check_stable(struct ingenic_cgu *cgu, const struct ingenic_cgu_clk_info *clk_info) { u32 reg; return readl_poll_timeout(cgu->base + clk_info->div.reg, reg, !(reg & BIT(clk_info->div.busy_bit)), 0, 100 * USEC_PER_MSEC); } static int ingenic_clk_set_rate(struct clk_hw *hw, unsigned long req_rate, unsigned long parent_rate) { struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw); const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk); struct ingenic_cgu *cgu = ingenic_clk->cgu; unsigned long rate, flags; unsigned int hw_div, div; u32 reg, mask; int ret = 0; if (clk_info->type & CGU_CLK_DIV) { div = ingenic_clk_calc_div(hw, clk_info, parent_rate, req_rate); rate = DIV_ROUND_UP(parent_rate, div); if (rate != req_rate) return -EINVAL; if (clk_info->div.div_table) hw_div = ingenic_clk_calc_hw_div(clk_info, div); else hw_div = ((div / clk_info->div.div) - 1); spin_lock_irqsave(&cgu->lock, flags); reg = readl(cgu->base + clk_info->div.reg); /* update the divide */ mask = GENMASK(clk_info->div.bits - 1, 0); reg &= ~(mask << clk_info->div.shift); reg |= hw_div << clk_info->div.shift; /* clear the stop bit */ if (clk_info->div.stop_bit != -1) reg &= ~BIT(clk_info->div.stop_bit); /* set the change enable bit */ if (clk_info->div.ce_bit != -1) reg |= BIT(clk_info->div.ce_bit); /* update the hardware */ writel(reg, cgu->base + clk_info->div.reg); /* wait for the change to take effect */ if (clk_info->div.busy_bit != -1) ret = ingenic_clk_check_stable(cgu, clk_info); spin_unlock_irqrestore(&cgu->lock, flags); return ret; } return -EINVAL; } static int ingenic_clk_enable(struct clk_hw *hw) { struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw); const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk); struct ingenic_cgu *cgu = ingenic_clk->cgu; unsigned long flags; if (clk_info->type & CGU_CLK_GATE) { /* ungate the clock */ spin_lock_irqsave(&cgu->lock, flags); ingenic_cgu_gate_set(cgu, &clk_info->gate, false); spin_unlock_irqrestore(&cgu->lock, flags); if (clk_info->gate.delay_us) udelay(clk_info->gate.delay_us); } return 0; } static void ingenic_clk_disable(struct clk_hw *hw) { struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw); const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk); struct ingenic_cgu *cgu = ingenic_clk->cgu; unsigned long flags; if (clk_info->type & CGU_CLK_GATE) { /* gate the clock */ spin_lock_irqsave(&cgu->lock, flags); ingenic_cgu_gate_set(cgu, &clk_info->gate, true); spin_unlock_irqrestore(&cgu->lock, flags); } } static int ingenic_clk_is_enabled(struct clk_hw *hw) { struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw); const struct ingenic_cgu_clk_info *clk_info = to_clk_info(ingenic_clk); struct ingenic_cgu *cgu = ingenic_clk->cgu; int enabled = 1; if (clk_info->type & CGU_CLK_GATE) enabled = !ingenic_cgu_gate_get(cgu, &clk_info->gate); return enabled; } static const struct clk_ops ingenic_clk_ops = { .get_parent = ingenic_clk_get_parent, .set_parent = ingenic_clk_set_parent, .recalc_rate = ingenic_clk_recalc_rate, .determine_rate = ingenic_clk_determine_rate, .set_rate = ingenic_clk_set_rate, .enable = ingenic_clk_enable, .disable = ingenic_clk_disable, .is_enabled = ingenic_clk_is_enabled, }; /* * Setup functions. */ static int ingenic_register_clock(struct ingenic_cgu *cgu, unsigned idx) { const struct ingenic_cgu_clk_info *clk_info = &cgu->clock_info[idx]; struct clk_init_data clk_init; struct ingenic_clk *ingenic_clk = NULL; struct clk *clk, *parent; const char *parent_names[4]; unsigned caps, i, num_possible; int err = -EINVAL; BUILD_BUG_ON(ARRAY_SIZE(clk_info->parents) > ARRAY_SIZE(parent_names)); if (clk_info->type == CGU_CLK_EXT) { clk = of_clk_get_by_name(cgu->np, clk_info->name); if (IS_ERR(clk)) { pr_err("%s: no external clock '%s' provided\n", __func__, clk_info->name); err = -ENODEV; goto out; } err = clk_register_clkdev(clk, clk_info->name, NULL); if (err) { clk_put(clk); goto out; } cgu->clocks.clks[idx] = clk; return 0; } if (!clk_info->type) { pr_err("%s: no clock type specified for '%s'\n", __func__, clk_info->name); goto out; } ingenic_clk = kzalloc(sizeof(*ingenic_clk), GFP_KERNEL); if (!ingenic_clk) { err = -ENOMEM; goto out; } ingenic_clk->hw.init = &clk_init; ingenic_clk->cgu = cgu; ingenic_clk->idx = idx; clk_init.name = clk_info->name; clk_init.flags = clk_info->flags; clk_init.parent_names = parent_names; caps = clk_info->type; if (caps & CGU_CLK_DIV) { caps &= ~CGU_CLK_DIV; } else if (!(caps & CGU_CLK_CUSTOM)) { /* pass rate changes to the parent clock */ clk_init.flags |= CLK_SET_RATE_PARENT; } if (caps & (CGU_CLK_MUX | CGU_CLK_CUSTOM)) { clk_init.num_parents = 0; if (caps & CGU_CLK_MUX) num_possible = 1 << clk_info->mux.bits; else num_possible = ARRAY_SIZE(clk_info->parents); for (i = 0; i < num_possible; i++) { if (clk_info->parents[i] == -1) continue; parent = cgu->clocks.clks[clk_info->parents[i]]; parent_names[clk_init.num_parents] = __clk_get_name(parent); clk_init.num_parents++; } BUG_ON(!clk_init.num_parents); BUG_ON(clk_init.num_parents > ARRAY_SIZE(parent_names)); } else { BUG_ON(clk_info->parents[0] == -1); clk_init.num_parents = 1; parent = cgu->clocks.clks[clk_info->parents[0]]; parent_names[0] = __clk_get_name(parent); } if (caps & CGU_CLK_CUSTOM) { clk_init.ops = clk_info->custom.clk_ops; caps &= ~CGU_CLK_CUSTOM; if (caps) { pr_err("%s: custom clock may not be combined with type 0x%x\n", __func__, caps); goto out; } } else if (caps & CGU_CLK_PLL) { clk_init.ops = &ingenic_pll_ops; caps &= ~CGU_CLK_PLL; if (caps) { pr_err("%s: PLL may not be combined with type 0x%x\n", __func__, caps); goto out; } } else { clk_init.ops = &ingenic_clk_ops; } /* nothing to do for gates or fixed dividers */ caps &= ~(CGU_CLK_GATE | CGU_CLK_FIXDIV); if (caps & CGU_CLK_MUX) { if (!(caps & CGU_CLK_MUX_GLITCHFREE)) clk_init.flags |= CLK_SET_PARENT_GATE; caps &= ~(CGU_CLK_MUX | CGU_CLK_MUX_GLITCHFREE); } if (caps) { pr_err("%s: unknown clock type 0x%x\n", __func__, caps); goto out; } clk = clk_register(NULL, &ingenic_clk->hw); if (IS_ERR(clk)) { pr_err("%s: failed to register clock '%s'\n", __func__, clk_info->name); err = PTR_ERR(clk); goto out; } err = clk_register_clkdev(clk, clk_info->name, NULL); if (err) goto out; cgu->clocks.clks[idx] = clk; out: if (err) kfree(ingenic_clk); return err; } struct ingenic_cgu * ingenic_cgu_new(const struct ingenic_cgu_clk_info *clock_info, unsigned num_clocks, struct device_node *np) { struct ingenic_cgu *cgu; cgu = kzalloc(sizeof(*cgu), GFP_KERNEL); if (!cgu) goto err_out; cgu->base = of_iomap(np, 0); if (!cgu->base) { pr_err("%s: failed to map CGU registers\n", __func__); goto err_out_free; } cgu->np = np; cgu->clock_info = clock_info; cgu->clocks.clk_num = num_clocks; spin_lock_init(&cgu->lock); return cgu; err_out_free: kfree(cgu); err_out: return NULL; } int ingenic_cgu_register_clocks(struct ingenic_cgu *cgu) { unsigned i; int err; cgu->clocks.clks = kcalloc(cgu->clocks.clk_num, sizeof(struct clk *), GFP_KERNEL); if (!cgu->clocks.clks) { err = -ENOMEM; goto err_out; } for (i = 0; i < cgu->clocks.clk_num; i++) { err = ingenic_register_clock(cgu, i); if (err) goto err_out_unregister; } err = of_clk_add_provider(cgu->np, of_clk_src_onecell_get, &cgu->clocks); if (err) goto err_out_unregister; return 0; err_out_unregister: for (i = 0; i < cgu->clocks.clk_num; i++) { if (!cgu->clocks.clks[i]) continue; if (cgu->clock_info[i].type & CGU_CLK_EXT) clk_put(cgu->clocks.clks[i]); else clk_unregister(cgu->clocks.clks[i]); } kfree(cgu->clocks.clks); err_out: return err; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1