blob: a2e73a6d60fddc232fe3d4974707eacad6983835 [file] [log] [blame]
Paul Burtonb0663032015-05-24 16:11:35 +01001/*
2 * Ingenic SoC CGU driver
3 *
4 * Copyright (c) 2013-2015 Imagination Technologies
Paul Burtonfb615d62017-10-25 17:04:33 -07005 * Author: Paul Burton <paul.burton@mips.com>
Paul Burtonb0663032015-05-24 16:11:35 +01006 *
7 * This program is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU General Public License as
9 * published by the Free Software Foundation; either version 2 of
10 * the License, or (at your option) any later version.
11 *
12 * This program is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 * GNU General Public License for more details.
16 */
17
18#include <linux/bitops.h>
Stephen Boyde2a65702015-05-01 16:09:33 -070019#include <linux/clk.h>
Paul Burtonb0663032015-05-24 16:11:35 +010020#include <linux/clk-provider.h>
21#include <linux/clkdev.h>
22#include <linux/delay.h>
23#include <linux/math64.h>
24#include <linux/of.h>
25#include <linux/of_address.h>
26#include <linux/slab.h>
27#include <linux/spinlock.h>
28#include "cgu.h"
29
30#define MHZ (1000 * 1000)
31
32/**
33 * ingenic_cgu_gate_get() - get the value of clock gate register bit
34 * @cgu: reference to the CGU whose registers should be read
35 * @info: info struct describing the gate bit
36 *
37 * Retrieves the state of the clock gate bit described by info. The
38 * caller must hold cgu->lock.
39 *
40 * Return: true if the gate bit is set, else false.
41 */
42static inline bool
43ingenic_cgu_gate_get(struct ingenic_cgu *cgu,
44 const struct ingenic_cgu_gate_info *info)
45{
46 return readl(cgu->base + info->reg) & BIT(info->bit);
47}
48
49/**
50 * ingenic_cgu_gate_set() - set the value of clock gate register bit
51 * @cgu: reference to the CGU whose registers should be modified
52 * @info: info struct describing the gate bit
53 * @val: non-zero to gate a clock, otherwise zero
54 *
55 * Sets the given gate bit in order to gate or ungate a clock.
56 *
57 * The caller must hold cgu->lock.
58 */
59static inline void
60ingenic_cgu_gate_set(struct ingenic_cgu *cgu,
61 const struct ingenic_cgu_gate_info *info, bool val)
62{
63 u32 clkgr = readl(cgu->base + info->reg);
64
65 if (val)
66 clkgr |= BIT(info->bit);
67 else
68 clkgr &= ~BIT(info->bit);
69
70 writel(clkgr, cgu->base + info->reg);
71}
72
73/*
74 * PLL operations
75 */
76
77static unsigned long
78ingenic_pll_recalc_rate(struct clk_hw *hw, unsigned long parent_rate)
79{
80 struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
81 struct ingenic_cgu *cgu = ingenic_clk->cgu;
82 const struct ingenic_cgu_clk_info *clk_info;
83 const struct ingenic_cgu_pll_info *pll_info;
84 unsigned m, n, od_enc, od;
85 bool bypass, enable;
86 unsigned long flags;
87 u32 ctl;
88
89 clk_info = &cgu->clock_info[ingenic_clk->idx];
90 BUG_ON(clk_info->type != CGU_CLK_PLL);
91 pll_info = &clk_info->pll;
92
93 spin_lock_irqsave(&cgu->lock, flags);
94 ctl = readl(cgu->base + pll_info->reg);
95 spin_unlock_irqrestore(&cgu->lock, flags);
96
97 m = (ctl >> pll_info->m_shift) & GENMASK(pll_info->m_bits - 1, 0);
98 m += pll_info->m_offset;
99 n = (ctl >> pll_info->n_shift) & GENMASK(pll_info->n_bits - 1, 0);
100 n += pll_info->n_offset;
101 od_enc = ctl >> pll_info->od_shift;
102 od_enc &= GENMASK(pll_info->od_bits - 1, 0);
103 bypass = !!(ctl & BIT(pll_info->bypass_bit));
104 enable = !!(ctl & BIT(pll_info->enable_bit));
105
106 if (bypass)
107 return parent_rate;
108
109 if (!enable)
110 return 0;
111
112 for (od = 0; od < pll_info->od_max; od++) {
113 if (pll_info->od_encoding[od] == od_enc)
114 break;
115 }
116 BUG_ON(od == pll_info->od_max);
117 od++;
118
119 return div_u64((u64)parent_rate * m, n * od);
120}
121
122static unsigned long
123ingenic_pll_calc(const struct ingenic_cgu_clk_info *clk_info,
124 unsigned long rate, unsigned long parent_rate,
125 unsigned *pm, unsigned *pn, unsigned *pod)
126{
127 const struct ingenic_cgu_pll_info *pll_info;
128 unsigned m, n, od;
129
130 pll_info = &clk_info->pll;
131 od = 1;
132
133 /*
134 * The frequency after the input divider must be between 10 and 50 MHz.
135 * The highest divider yields the best resolution.
136 */
137 n = parent_rate / (10 * MHZ);
138 n = min_t(unsigned, n, 1 << clk_info->pll.n_bits);
139 n = max_t(unsigned, n, pll_info->n_offset);
140
141 m = (rate / MHZ) * od * n / (parent_rate / MHZ);
142 m = min_t(unsigned, m, 1 << clk_info->pll.m_bits);
143 m = max_t(unsigned, m, pll_info->m_offset);
144
145 if (pm)
146 *pm = m;
147 if (pn)
148 *pn = n;
149 if (pod)
150 *pod = od;
151
152 return div_u64((u64)parent_rate * m, n * od);
153}
154
155static long
156ingenic_pll_round_rate(struct clk_hw *hw, unsigned long req_rate,
157 unsigned long *prate)
158{
159 struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
160 struct ingenic_cgu *cgu = ingenic_clk->cgu;
161 const struct ingenic_cgu_clk_info *clk_info;
162
163 clk_info = &cgu->clock_info[ingenic_clk->idx];
164 BUG_ON(clk_info->type != CGU_CLK_PLL);
165
166 return ingenic_pll_calc(clk_info, req_rate, *prate, NULL, NULL, NULL);
167}
168
169static int
170ingenic_pll_set_rate(struct clk_hw *hw, unsigned long req_rate,
171 unsigned long parent_rate)
172{
173 const unsigned timeout = 100;
174 struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
175 struct ingenic_cgu *cgu = ingenic_clk->cgu;
176 const struct ingenic_cgu_clk_info *clk_info;
177 const struct ingenic_cgu_pll_info *pll_info;
178 unsigned long rate, flags;
179 unsigned m, n, od, i;
180 u32 ctl;
181
182 clk_info = &cgu->clock_info[ingenic_clk->idx];
183 BUG_ON(clk_info->type != CGU_CLK_PLL);
184 pll_info = &clk_info->pll;
185
186 rate = ingenic_pll_calc(clk_info, req_rate, parent_rate,
187 &m, &n, &od);
188 if (rate != req_rate)
189 pr_info("ingenic-cgu: request '%s' rate %luHz, actual %luHz\n",
190 clk_info->name, req_rate, rate);
191
192 spin_lock_irqsave(&cgu->lock, flags);
193 ctl = readl(cgu->base + pll_info->reg);
194
195 ctl &= ~(GENMASK(pll_info->m_bits - 1, 0) << pll_info->m_shift);
196 ctl |= (m - pll_info->m_offset) << pll_info->m_shift;
197
198 ctl &= ~(GENMASK(pll_info->n_bits - 1, 0) << pll_info->n_shift);
199 ctl |= (n - pll_info->n_offset) << pll_info->n_shift;
200
201 ctl &= ~(GENMASK(pll_info->od_bits - 1, 0) << pll_info->od_shift);
202 ctl |= pll_info->od_encoding[od - 1] << pll_info->od_shift;
203
204 ctl &= ~BIT(pll_info->bypass_bit);
205 ctl |= BIT(pll_info->enable_bit);
206
207 writel(ctl, cgu->base + pll_info->reg);
208
209 /* wait for the PLL to stabilise */
210 for (i = 0; i < timeout; i++) {
211 ctl = readl(cgu->base + pll_info->reg);
212 if (ctl & BIT(pll_info->stable_bit))
213 break;
214 mdelay(1);
215 }
216
217 spin_unlock_irqrestore(&cgu->lock, flags);
218
219 if (i == timeout)
220 return -EBUSY;
221
222 return 0;
223}
224
225static const struct clk_ops ingenic_pll_ops = {
226 .recalc_rate = ingenic_pll_recalc_rate,
227 .round_rate = ingenic_pll_round_rate,
228 .set_rate = ingenic_pll_set_rate,
229};
230
231/*
232 * Operations for all non-PLL clocks
233 */
234
235static u8 ingenic_clk_get_parent(struct clk_hw *hw)
236{
237 struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
238 struct ingenic_cgu *cgu = ingenic_clk->cgu;
239 const struct ingenic_cgu_clk_info *clk_info;
240 u32 reg;
241 u8 i, hw_idx, idx = 0;
242
243 clk_info = &cgu->clock_info[ingenic_clk->idx];
244
245 if (clk_info->type & CGU_CLK_MUX) {
246 reg = readl(cgu->base + clk_info->mux.reg);
247 hw_idx = (reg >> clk_info->mux.shift) &
248 GENMASK(clk_info->mux.bits - 1, 0);
249
250 /*
251 * Convert the hardware index to the parent index by skipping
252 * over any -1's in the parents array.
253 */
254 for (i = 0; i < hw_idx; i++) {
255 if (clk_info->parents[i] != -1)
256 idx++;
257 }
258 }
259
260 return idx;
261}
262
263static int ingenic_clk_set_parent(struct clk_hw *hw, u8 idx)
264{
265 struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
266 struct ingenic_cgu *cgu = ingenic_clk->cgu;
267 const struct ingenic_cgu_clk_info *clk_info;
268 unsigned long flags;
269 u8 curr_idx, hw_idx, num_poss;
270 u32 reg, mask;
271
272 clk_info = &cgu->clock_info[ingenic_clk->idx];
273
274 if (clk_info->type & CGU_CLK_MUX) {
275 /*
276 * Convert the parent index to the hardware index by adding
277 * 1 for any -1 in the parents array preceding the given
278 * index. That is, we want the index of idx'th entry in
279 * clk_info->parents which does not equal -1.
280 */
281 hw_idx = curr_idx = 0;
282 num_poss = 1 << clk_info->mux.bits;
283 for (; hw_idx < num_poss; hw_idx++) {
284 if (clk_info->parents[hw_idx] == -1)
285 continue;
286 if (curr_idx == idx)
287 break;
288 curr_idx++;
289 }
290
291 /* idx should always be a valid parent */
292 BUG_ON(curr_idx != idx);
293
294 mask = GENMASK(clk_info->mux.bits - 1, 0);
295 mask <<= clk_info->mux.shift;
296
297 spin_lock_irqsave(&cgu->lock, flags);
298
299 /* write the register */
300 reg = readl(cgu->base + clk_info->mux.reg);
301 reg &= ~mask;
302 reg |= hw_idx << clk_info->mux.shift;
303 writel(reg, cgu->base + clk_info->mux.reg);
304
305 spin_unlock_irqrestore(&cgu->lock, flags);
306 return 0;
307 }
308
309 return idx ? -EINVAL : 0;
310}
311
312static unsigned long
313ingenic_clk_recalc_rate(struct clk_hw *hw, unsigned long parent_rate)
314{
315 struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
316 struct ingenic_cgu *cgu = ingenic_clk->cgu;
317 const struct ingenic_cgu_clk_info *clk_info;
318 unsigned long rate = parent_rate;
319 u32 div_reg, div;
320
321 clk_info = &cgu->clock_info[ingenic_clk->idx];
322
323 if (clk_info->type & CGU_CLK_DIV) {
324 div_reg = readl(cgu->base + clk_info->div.reg);
325 div = (div_reg >> clk_info->div.shift) &
326 GENMASK(clk_info->div.bits - 1, 0);
327 div += 1;
Harvey Hunt4afe2d12016-05-09 17:29:52 +0100328 div *= clk_info->div.div;
Paul Burtonb0663032015-05-24 16:11:35 +0100329
330 rate /= div;
Paul Cercueile6cfa6432018-01-16 16:47:52 +0100331 } else if (clk_info->type & CGU_CLK_FIXDIV) {
332 rate /= clk_info->fixdiv.div;
Paul Burtonb0663032015-05-24 16:11:35 +0100333 }
334
335 return rate;
336}
337
338static unsigned
339ingenic_clk_calc_div(const struct ingenic_cgu_clk_info *clk_info,
340 unsigned long parent_rate, unsigned long req_rate)
341{
342 unsigned div;
343
344 /* calculate the divide */
345 div = DIV_ROUND_UP(parent_rate, req_rate);
346
347 /* and impose hardware constraints */
348 div = min_t(unsigned, div, 1 << clk_info->div.bits);
349 div = max_t(unsigned, div, 1);
350
Harvey Hunt4afe2d12016-05-09 17:29:52 +0100351 /*
352 * If the divider value itself must be divided before being written to
353 * the divider register, we must ensure we don't have any bits set that
354 * would be lost as a result of doing so.
355 */
356 div /= clk_info->div.div;
357 div *= clk_info->div.div;
358
Paul Burtonb0663032015-05-24 16:11:35 +0100359 return div;
360}
361
362static long
363ingenic_clk_round_rate(struct clk_hw *hw, unsigned long req_rate,
364 unsigned long *parent_rate)
365{
366 struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
367 struct ingenic_cgu *cgu = ingenic_clk->cgu;
368 const struct ingenic_cgu_clk_info *clk_info;
369 long rate = *parent_rate;
370
371 clk_info = &cgu->clock_info[ingenic_clk->idx];
372
373 if (clk_info->type & CGU_CLK_DIV)
374 rate /= ingenic_clk_calc_div(clk_info, *parent_rate, req_rate);
375 else if (clk_info->type & CGU_CLK_FIXDIV)
376 rate /= clk_info->fixdiv.div;
377
378 return rate;
379}
380
381static int
382ingenic_clk_set_rate(struct clk_hw *hw, unsigned long req_rate,
383 unsigned long parent_rate)
384{
385 struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
386 struct ingenic_cgu *cgu = ingenic_clk->cgu;
387 const struct ingenic_cgu_clk_info *clk_info;
388 const unsigned timeout = 100;
389 unsigned long rate, flags;
390 unsigned div, i;
391 u32 reg, mask;
392 int ret = 0;
393
394 clk_info = &cgu->clock_info[ingenic_clk->idx];
395
396 if (clk_info->type & CGU_CLK_DIV) {
397 div = ingenic_clk_calc_div(clk_info, parent_rate, req_rate);
398 rate = parent_rate / div;
399
400 if (rate != req_rate)
401 return -EINVAL;
402
403 spin_lock_irqsave(&cgu->lock, flags);
404 reg = readl(cgu->base + clk_info->div.reg);
405
406 /* update the divide */
407 mask = GENMASK(clk_info->div.bits - 1, 0);
408 reg &= ~(mask << clk_info->div.shift);
Harvey Hunt4afe2d12016-05-09 17:29:52 +0100409 reg |= ((div / clk_info->div.div) - 1) << clk_info->div.shift;
Paul Burtonb0663032015-05-24 16:11:35 +0100410
411 /* clear the stop bit */
412 if (clk_info->div.stop_bit != -1)
413 reg &= ~BIT(clk_info->div.stop_bit);
414
415 /* set the change enable bit */
416 if (clk_info->div.ce_bit != -1)
417 reg |= BIT(clk_info->div.ce_bit);
418
419 /* update the hardware */
420 writel(reg, cgu->base + clk_info->div.reg);
421
422 /* wait for the change to take effect */
423 if (clk_info->div.busy_bit != -1) {
424 for (i = 0; i < timeout; i++) {
425 reg = readl(cgu->base + clk_info->div.reg);
426 if (!(reg & BIT(clk_info->div.busy_bit)))
427 break;
428 mdelay(1);
429 }
430 if (i == timeout)
431 ret = -EBUSY;
432 }
433
434 spin_unlock_irqrestore(&cgu->lock, flags);
435 return ret;
436 }
437
438 return -EINVAL;
439}
440
441static int ingenic_clk_enable(struct clk_hw *hw)
442{
443 struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
444 struct ingenic_cgu *cgu = ingenic_clk->cgu;
445 const struct ingenic_cgu_clk_info *clk_info;
446 unsigned long flags;
447
448 clk_info = &cgu->clock_info[ingenic_clk->idx];
449
450 if (clk_info->type & CGU_CLK_GATE) {
451 /* ungate the clock */
452 spin_lock_irqsave(&cgu->lock, flags);
453 ingenic_cgu_gate_set(cgu, &clk_info->gate, false);
454 spin_unlock_irqrestore(&cgu->lock, flags);
455 }
456
457 return 0;
458}
459
460static void ingenic_clk_disable(struct clk_hw *hw)
461{
462 struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
463 struct ingenic_cgu *cgu = ingenic_clk->cgu;
464 const struct ingenic_cgu_clk_info *clk_info;
465 unsigned long flags;
466
467 clk_info = &cgu->clock_info[ingenic_clk->idx];
468
469 if (clk_info->type & CGU_CLK_GATE) {
470 /* gate the clock */
471 spin_lock_irqsave(&cgu->lock, flags);
472 ingenic_cgu_gate_set(cgu, &clk_info->gate, true);
473 spin_unlock_irqrestore(&cgu->lock, flags);
474 }
475}
476
477static int ingenic_clk_is_enabled(struct clk_hw *hw)
478{
479 struct ingenic_clk *ingenic_clk = to_ingenic_clk(hw);
480 struct ingenic_cgu *cgu = ingenic_clk->cgu;
481 const struct ingenic_cgu_clk_info *clk_info;
482 unsigned long flags;
483 int enabled = 1;
484
485 clk_info = &cgu->clock_info[ingenic_clk->idx];
486
487 if (clk_info->type & CGU_CLK_GATE) {
488 spin_lock_irqsave(&cgu->lock, flags);
489 enabled = !ingenic_cgu_gate_get(cgu, &clk_info->gate);
490 spin_unlock_irqrestore(&cgu->lock, flags);
491 }
492
493 return enabled;
494}
495
496static const struct clk_ops ingenic_clk_ops = {
497 .get_parent = ingenic_clk_get_parent,
498 .set_parent = ingenic_clk_set_parent,
499
500 .recalc_rate = ingenic_clk_recalc_rate,
501 .round_rate = ingenic_clk_round_rate,
502 .set_rate = ingenic_clk_set_rate,
503
504 .enable = ingenic_clk_enable,
505 .disable = ingenic_clk_disable,
506 .is_enabled = ingenic_clk_is_enabled,
507};
508
509/*
510 * Setup functions.
511 */
512
513static int ingenic_register_clock(struct ingenic_cgu *cgu, unsigned idx)
514{
515 const struct ingenic_cgu_clk_info *clk_info = &cgu->clock_info[idx];
516 struct clk_init_data clk_init;
517 struct ingenic_clk *ingenic_clk = NULL;
518 struct clk *clk, *parent;
519 const char *parent_names[4];
520 unsigned caps, i, num_possible;
521 int err = -EINVAL;
522
523 BUILD_BUG_ON(ARRAY_SIZE(clk_info->parents) > ARRAY_SIZE(parent_names));
524
525 if (clk_info->type == CGU_CLK_EXT) {
526 clk = of_clk_get_by_name(cgu->np, clk_info->name);
527 if (IS_ERR(clk)) {
528 pr_err("%s: no external clock '%s' provided\n",
529 __func__, clk_info->name);
530 err = -ENODEV;
531 goto out;
532 }
533 err = clk_register_clkdev(clk, clk_info->name, NULL);
534 if (err) {
535 clk_put(clk);
536 goto out;
537 }
538 cgu->clocks.clks[idx] = clk;
539 return 0;
540 }
541
542 if (!clk_info->type) {
543 pr_err("%s: no clock type specified for '%s'\n", __func__,
544 clk_info->name);
545 goto out;
546 }
547
548 ingenic_clk = kzalloc(sizeof(*ingenic_clk), GFP_KERNEL);
549 if (!ingenic_clk) {
550 err = -ENOMEM;
551 goto out;
552 }
553
554 ingenic_clk->hw.init = &clk_init;
555 ingenic_clk->cgu = cgu;
556 ingenic_clk->idx = idx;
557
558 clk_init.name = clk_info->name;
559 clk_init.flags = 0;
560 clk_init.parent_names = parent_names;
561
562 caps = clk_info->type;
563
564 if (caps & (CGU_CLK_MUX | CGU_CLK_CUSTOM)) {
565 clk_init.num_parents = 0;
566
567 if (caps & CGU_CLK_MUX)
568 num_possible = 1 << clk_info->mux.bits;
569 else
570 num_possible = ARRAY_SIZE(clk_info->parents);
571
572 for (i = 0; i < num_possible; i++) {
573 if (clk_info->parents[i] == -1)
574 continue;
575
576 parent = cgu->clocks.clks[clk_info->parents[i]];
577 parent_names[clk_init.num_parents] =
578 __clk_get_name(parent);
579 clk_init.num_parents++;
580 }
581
582 BUG_ON(!clk_init.num_parents);
583 BUG_ON(clk_init.num_parents > ARRAY_SIZE(parent_names));
584 } else {
585 BUG_ON(clk_info->parents[0] == -1);
586 clk_init.num_parents = 1;
587 parent = cgu->clocks.clks[clk_info->parents[0]];
588 parent_names[0] = __clk_get_name(parent);
589 }
590
591 if (caps & CGU_CLK_CUSTOM) {
592 clk_init.ops = clk_info->custom.clk_ops;
593
594 caps &= ~CGU_CLK_CUSTOM;
595
596 if (caps) {
597 pr_err("%s: custom clock may not be combined with type 0x%x\n",
598 __func__, caps);
599 goto out;
600 }
601 } else if (caps & CGU_CLK_PLL) {
602 clk_init.ops = &ingenic_pll_ops;
603
604 caps &= ~CGU_CLK_PLL;
605
606 if (caps) {
607 pr_err("%s: PLL may not be combined with type 0x%x\n",
608 __func__, caps);
609 goto out;
610 }
611 } else {
612 clk_init.ops = &ingenic_clk_ops;
613 }
614
615 /* nothing to do for gates or fixed dividers */
616 caps &= ~(CGU_CLK_GATE | CGU_CLK_FIXDIV);
617
618 if (caps & CGU_CLK_MUX) {
619 if (!(caps & CGU_CLK_MUX_GLITCHFREE))
620 clk_init.flags |= CLK_SET_PARENT_GATE;
621
622 caps &= ~(CGU_CLK_MUX | CGU_CLK_MUX_GLITCHFREE);
623 }
624
625 if (caps & CGU_CLK_DIV) {
626 caps &= ~CGU_CLK_DIV;
627 } else {
628 /* pass rate changes to the parent clock */
629 clk_init.flags |= CLK_SET_RATE_PARENT;
630 }
631
632 if (caps) {
633 pr_err("%s: unknown clock type 0x%x\n", __func__, caps);
634 goto out;
635 }
636
637 clk = clk_register(NULL, &ingenic_clk->hw);
638 if (IS_ERR(clk)) {
639 pr_err("%s: failed to register clock '%s'\n", __func__,
640 clk_info->name);
641 err = PTR_ERR(clk);
642 goto out;
643 }
644
645 err = clk_register_clkdev(clk, clk_info->name, NULL);
646 if (err)
647 goto out;
648
649 cgu->clocks.clks[idx] = clk;
650out:
651 if (err)
652 kfree(ingenic_clk);
653 return err;
654}
655
656struct ingenic_cgu *
657ingenic_cgu_new(const struct ingenic_cgu_clk_info *clock_info,
658 unsigned num_clocks, struct device_node *np)
659{
660 struct ingenic_cgu *cgu;
661
662 cgu = kzalloc(sizeof(*cgu), GFP_KERNEL);
663 if (!cgu)
664 goto err_out;
665
666 cgu->base = of_iomap(np, 0);
667 if (!cgu->base) {
668 pr_err("%s: failed to map CGU registers\n", __func__);
669 goto err_out_free;
670 }
671
672 cgu->np = np;
673 cgu->clock_info = clock_info;
674 cgu->clocks.clk_num = num_clocks;
675
676 spin_lock_init(&cgu->lock);
677
678 return cgu;
679
680err_out_free:
681 kfree(cgu);
682err_out:
683 return NULL;
684}
685
686int ingenic_cgu_register_clocks(struct ingenic_cgu *cgu)
687{
688 unsigned i;
689 int err;
690
691 cgu->clocks.clks = kcalloc(cgu->clocks.clk_num, sizeof(struct clk *),
692 GFP_KERNEL);
693 if (!cgu->clocks.clks) {
694 err = -ENOMEM;
695 goto err_out;
696 }
697
698 for (i = 0; i < cgu->clocks.clk_num; i++) {
699 err = ingenic_register_clock(cgu, i);
700 if (err)
701 goto err_out_unregister;
702 }
703
704 err = of_clk_add_provider(cgu->np, of_clk_src_onecell_get,
705 &cgu->clocks);
706 if (err)
707 goto err_out_unregister;
708
709 return 0;
710
711err_out_unregister:
712 for (i = 0; i < cgu->clocks.clk_num; i++) {
713 if (!cgu->clocks.clks[i])
714 continue;
715 if (cgu->clock_info[i].type & CGU_CLK_EXT)
716 clk_put(cgu->clocks.clks[i]);
717 else
718 clk_unregister(cgu->clocks.clks[i]);
719 }
720 kfree(cgu->clocks.clks);
721err_out:
722 return err;
723}