blob: 7a1d56af9e470c439fd73407194a02c6f31c8f6d [file] [log] [blame]
Paul Walmsley543d9372008-03-18 10:22:06 +02001/*
2 * linux/arch/arm/mach-omap2/clock.c
3 *
Tony Lindgrena16e9702008-03-18 11:56:39 +02004 * Copyright (C) 2005-2008 Texas Instruments, Inc.
5 * Copyright (C) 2004-2008 Nokia Corporation
6 *
7 * Contacts:
Paul Walmsley543d9372008-03-18 10:22:06 +02008 * Richard Woodruff <r-woodruff2@ti.com>
Paul Walmsley543d9372008-03-18 10:22:06 +02009 * Paul Walmsley
10 *
Paul Walmsley543d9372008-03-18 10:22:06 +020011 * This program is free software; you can redistribute it and/or modify
12 * it under the terms of the GNU General Public License version 2 as
13 * published by the Free Software Foundation.
14 */
15#undef DEBUG
16
17#include <linux/module.h>
18#include <linux/kernel.h>
19#include <linux/device.h>
20#include <linux/list.h>
21#include <linux/errno.h>
22#include <linux/delay.h>
23#include <linux/clk.h>
Russell Kingfced80c2008-09-06 12:10:45 +010024#include <linux/io.h>
Russell Kingfbd3bdb2008-09-06 12:13:59 +010025#include <linux/bitops.h>
Paul Walmsley543d9372008-03-18 10:22:06 +020026
Russell Kinga09e64f2008-08-05 16:14:15 +010027#include <mach/clock.h>
Paul Walmsley333943b2008-08-19 11:08:45 +030028#include <mach/clockdomain.h>
Russell Kinga09e64f2008-08-05 16:14:15 +010029#include <mach/sram.h>
30#include <mach/cpu.h>
Paul Walmsley543d9372008-03-18 10:22:06 +020031#include <asm/div64.h>
32
33#include "memory.h"
34#include "sdrc.h"
35#include "clock.h"
36#include "prm.h"
37#include "prm-regbits-24xx.h"
38#include "cm.h"
39#include "cm-regbits-24xx.h"
40#include "cm-regbits-34xx.h"
41
42#define MAX_CLOCK_ENABLE_WAIT 100000
43
Paul Walmsley88b8ba92008-07-03 12:24:46 +030044/* DPLL rate rounding: minimum DPLL multiplier, divider values */
45#define DPLL_MIN_MULTIPLIER 1
46#define DPLL_MIN_DIVIDER 1
47
48/* Possible error results from _dpll_test_mult */
49#define DPLL_MULT_UNDERFLOW (1 << 0)
50
51/*
52 * Scale factor to mitigate roundoff errors in DPLL rate rounding.
53 * The higher the scale factor, the greater the risk of arithmetic overflow,
54 * but the closer the rounded rate to the target rate. DPLL_SCALE_FACTOR
55 * must be a power of DPLL_SCALE_BASE.
56 */
57#define DPLL_SCALE_FACTOR 64
58#define DPLL_SCALE_BASE 2
59#define DPLL_ROUNDING_VAL ((DPLL_SCALE_BASE / 2) * \
60 (DPLL_SCALE_FACTOR / DPLL_SCALE_BASE))
61
Paul Walmsley543d9372008-03-18 10:22:06 +020062u8 cpu_mask;
63
64/*-------------------------------------------------------------------------
Paul Walmsley333943b2008-08-19 11:08:45 +030065 * OMAP2/3 specific clock functions
Paul Walmsley543d9372008-03-18 10:22:06 +020066 *-------------------------------------------------------------------------*/
67
68/**
Paul Walmsley333943b2008-08-19 11:08:45 +030069 * omap2_init_clk_clkdm - look up a clockdomain name, store pointer in clk
70 * @clk: OMAP clock struct ptr to use
71 *
72 * Convert a clockdomain name stored in a struct clk 'clk' into a
73 * clockdomain pointer, and save it into the struct clk. Intended to be
74 * called during clk_register(). No return value.
75 */
76void omap2_init_clk_clkdm(struct clk *clk)
77{
78 struct clockdomain *clkdm;
79
80 if (!clk->clkdm_name)
81 return;
82
83 clkdm = clkdm_lookup(clk->clkdm_name);
84 if (clkdm) {
85 pr_debug("clock: associated clk %s to clkdm %s\n",
86 clk->name, clk->clkdm_name);
87 clk->clkdm = clkdm;
88 } else {
89 pr_debug("clock: could not associate clk %s to "
90 "clkdm %s\n", clk->name, clk->clkdm_name);
91 }
92}
93
94/**
Paul Walmsley543d9372008-03-18 10:22:06 +020095 * omap2_init_clksel_parent - set a clksel clk's parent field from the hardware
96 * @clk: OMAP clock struct ptr to use
97 *
98 * Given a pointer to a source-selectable struct clk, read the hardware
99 * register and determine what its parent is currently set to. Update the
100 * clk->parent field with the appropriate clk ptr.
101 */
102void omap2_init_clksel_parent(struct clk *clk)
103{
104 const struct clksel *clks;
105 const struct clksel_rate *clkr;
106 u32 r, found = 0;
107
108 if (!clk->clksel)
109 return;
110
111 r = __raw_readl(clk->clksel_reg) & clk->clksel_mask;
112 r >>= __ffs(clk->clksel_mask);
113
114 for (clks = clk->clksel; clks->parent && !found; clks++) {
115 for (clkr = clks->rates; clkr->div && !found; clkr++) {
116 if ((clkr->flags & cpu_mask) && (clkr->val == r)) {
117 if (clk->parent != clks->parent) {
118 pr_debug("clock: inited %s parent "
119 "to %s (was %s)\n",
120 clk->name, clks->parent->name,
121 ((clk->parent) ?
122 clk->parent->name : "NULL"));
123 clk->parent = clks->parent;
124 };
125 found = 1;
126 }
127 }
128 }
129
130 if (!found)
131 printk(KERN_ERR "clock: init parent: could not find "
132 "regval %0x for clock %s\n", r, clk->name);
133
134 return;
135}
136
137/* Returns the DPLL rate */
138u32 omap2_get_dpll_rate(struct clk *clk)
139{
140 long long dpll_clk;
141 u32 dpll_mult, dpll_div, dpll;
Paul Walmsley88b8ba92008-07-03 12:24:46 +0300142 struct dpll_data *dd;
Paul Walmsley543d9372008-03-18 10:22:06 +0200143
144 dd = clk->dpll_data;
145 /* REVISIT: What do we return on error? */
146 if (!dd)
147 return 0;
148
149 dpll = __raw_readl(dd->mult_div1_reg);
150 dpll_mult = dpll & dd->mult_mask;
151 dpll_mult >>= __ffs(dd->mult_mask);
152 dpll_div = dpll & dd->div1_mask;
153 dpll_div >>= __ffs(dd->div1_mask);
154
155 dpll_clk = (long long)clk->parent->rate * dpll_mult;
156 do_div(dpll_clk, dpll_div + 1);
157
Paul Walmsley543d9372008-03-18 10:22:06 +0200158 return dpll_clk;
159}
160
161/*
162 * Used for clocks that have the same value as the parent clock,
163 * divided by some factor
164 */
165void omap2_fixed_divisor_recalc(struct clk *clk)
166{
167 WARN_ON(!clk->fixed_div);
168
169 clk->rate = clk->parent->rate / clk->fixed_div;
170
171 if (clk->flags & RATE_PROPAGATES)
172 propagate_rate(clk);
173}
174
175/**
176 * omap2_wait_clock_ready - wait for clock to enable
177 * @reg: physical address of clock IDLEST register
178 * @mask: value to mask against to determine if the clock is active
179 * @name: name of the clock (for printk)
180 *
181 * Returns 1 if the clock enabled in time, or 0 if it failed to enable
182 * in roughly MAX_CLOCK_ENABLE_WAIT microseconds.
183 */
184int omap2_wait_clock_ready(void __iomem *reg, u32 mask, const char *name)
185{
186 int i = 0;
187 int ena = 0;
188
189 /*
190 * 24xx uses 0 to indicate not ready, and 1 to indicate ready.
191 * 34xx reverses this, just to keep us on our toes
192 */
193 if (cpu_mask & (RATE_IN_242X | RATE_IN_243X)) {
194 ena = mask;
195 } else if (cpu_mask & RATE_IN_343X) {
196 ena = 0;
197 }
198
199 /* Wait for lock */
200 while (((__raw_readl(reg) & mask) != ena) &&
201 (i++ < MAX_CLOCK_ENABLE_WAIT)) {
202 udelay(1);
203 }
204
205 if (i < MAX_CLOCK_ENABLE_WAIT)
206 pr_debug("Clock %s stable after %d loops\n", name, i);
207 else
208 printk(KERN_ERR "Clock %s didn't enable in %d tries\n",
209 name, MAX_CLOCK_ENABLE_WAIT);
210
211
212 return (i < MAX_CLOCK_ENABLE_WAIT) ? 1 : 0;
213};
214
215
216/*
217 * Note: We don't need special code here for INVERT_ENABLE
218 * for the time being since INVERT_ENABLE only applies to clocks enabled by
219 * CM_CLKEN_PLL
220 */
221static void omap2_clk_wait_ready(struct clk *clk)
222{
223 void __iomem *reg, *other_reg, *st_reg;
224 u32 bit;
225
226 /*
227 * REVISIT: This code is pretty ugly. It would be nice to generalize
228 * it and pull it into struct clk itself somehow.
229 */
230 reg = clk->enable_reg;
Russell Kingc1168dc2008-11-04 21:24:00 +0000231
232 /*
233 * Convert CM_ICLKEN* <-> CM_FCLKEN*. This conversion assumes
234 * it's just a matter of XORing the bits.
235 */
236 other_reg = (void __iomem *)((u32)reg ^ (CM_FCLKEN ^ CM_ICLKEN));
Paul Walmsley543d9372008-03-18 10:22:06 +0200237
Paul Walmsley543d9372008-03-18 10:22:06 +0200238 /* Check if both functional and interface clocks
239 * are running. */
240 bit = 1 << clk->enable_bit;
241 if (!(__raw_readl(other_reg) & bit))
242 return;
243 st_reg = (void __iomem *)(((u32)other_reg & ~0xf0) | 0x20); /* CM_IDLEST* */
244
245 omap2_wait_clock_ready(st_reg, bit, clk->name);
246}
247
Russell Kingbc51da42008-11-04 18:59:32 +0000248static int omap2_dflt_clk_enable(struct clk *clk)
Paul Walmsley543d9372008-03-18 10:22:06 +0200249{
250 u32 regval32;
251
Russell Kingc0fc18c2008-09-05 15:10:27 +0100252 if (unlikely(clk->enable_reg == NULL)) {
Paul Walmsley543d9372008-03-18 10:22:06 +0200253 printk(KERN_ERR "clock.c: Enable for %s without enable code\n",
254 clk->name);
255 return 0; /* REVISIT: -EINVAL */
256 }
257
258 regval32 = __raw_readl(clk->enable_reg);
259 if (clk->flags & INVERT_ENABLE)
260 regval32 &= ~(1 << clk->enable_bit);
261 else
262 regval32 |= (1 << clk->enable_bit);
263 __raw_writel(regval32, clk->enable_reg);
264 wmb();
265
Paul Walmsley543d9372008-03-18 10:22:06 +0200266 return 0;
267}
268
Russell Kingbc51da42008-11-04 18:59:32 +0000269static int omap2_dflt_clk_enable_wait(struct clk *clk)
270{
271 int ret;
272
273 if (unlikely(clk->enable_reg == NULL)) {
274 printk(KERN_ERR "clock.c: Enable for %s without enable code\n",
275 clk->name);
276 return 0; /* REVISIT: -EINVAL */
277 }
278
279 ret = omap2_dflt_clk_enable(clk);
280 if (ret == 0)
281 omap2_clk_wait_ready(clk);
282 return ret;
283}
284
Russell Kingb36ee722008-11-04 17:59:52 +0000285static void omap2_dflt_clk_disable(struct clk *clk)
Paul Walmsley543d9372008-03-18 10:22:06 +0200286{
287 u32 regval32;
288
Russell Kingc0fc18c2008-09-05 15:10:27 +0100289 if (clk->enable_reg == NULL) {
Paul Walmsley543d9372008-03-18 10:22:06 +0200290 /*
291 * 'Independent' here refers to a clock which is not
292 * controlled by its parent.
293 */
294 printk(KERN_ERR "clock: clk_disable called on independent "
295 "clock %s which has no enable_reg\n", clk->name);
296 return;
297 }
298
299 regval32 = __raw_readl(clk->enable_reg);
300 if (clk->flags & INVERT_ENABLE)
301 regval32 |= (1 << clk->enable_bit);
302 else
303 regval32 &= ~(1 << clk->enable_bit);
304 __raw_writel(regval32, clk->enable_reg);
305 wmb();
306}
307
Russell Kingb36ee722008-11-04 17:59:52 +0000308const struct clkops clkops_omap2_dflt_wait = {
309 .enable = omap2_dflt_clk_enable_wait,
310 .disable = omap2_dflt_clk_disable,
311};
312
Russell Kingbc51da42008-11-04 18:59:32 +0000313const struct clkops clkops_omap2_dflt = {
314 .enable = omap2_dflt_clk_enable,
315 .disable = omap2_dflt_clk_disable,
316};
317
Russell Kingb36ee722008-11-04 17:59:52 +0000318/* Enables clock without considering parent dependencies or use count
319 * REVISIT: Maybe change this to use clk->enable like on omap1?
320 */
321static int _omap2_clk_enable(struct clk *clk)
322{
323 return clk->ops->enable(clk);
324}
325
326/* Disables clock without considering parent dependencies or use count */
327static void _omap2_clk_disable(struct clk *clk)
328{
329 clk->ops->disable(clk);
330}
331
Paul Walmsley543d9372008-03-18 10:22:06 +0200332void omap2_clk_disable(struct clk *clk)
333{
334 if (clk->usecount > 0 && !(--clk->usecount)) {
335 _omap2_clk_disable(clk);
336 if (likely((u32)clk->parent))
337 omap2_clk_disable(clk->parent);
Paul Walmsley333943b2008-08-19 11:08:45 +0300338 if (clk->clkdm)
339 omap2_clkdm_clk_disable(clk->clkdm, clk);
340
Paul Walmsley543d9372008-03-18 10:22:06 +0200341 }
342}
343
344int omap2_clk_enable(struct clk *clk)
345{
346 int ret = 0;
347
348 if (clk->usecount++ == 0) {
349 if (likely((u32)clk->parent))
350 ret = omap2_clk_enable(clk->parent);
351
352 if (unlikely(ret != 0)) {
353 clk->usecount--;
354 return ret;
355 }
356
Paul Walmsley333943b2008-08-19 11:08:45 +0300357 if (clk->clkdm)
358 omap2_clkdm_clk_enable(clk->clkdm, clk);
359
Paul Walmsley543d9372008-03-18 10:22:06 +0200360 ret = _omap2_clk_enable(clk);
361
Paul Walmsley333943b2008-08-19 11:08:45 +0300362 if (unlikely(ret != 0)) {
363 if (clk->clkdm)
364 omap2_clkdm_clk_disable(clk->clkdm, clk);
365
366 if (clk->parent) {
367 omap2_clk_disable(clk->parent);
368 clk->usecount--;
369 }
Paul Walmsley543d9372008-03-18 10:22:06 +0200370 }
371 }
372
373 return ret;
374}
375
376/*
377 * Used for clocks that are part of CLKSEL_xyz governed clocks.
378 * REVISIT: Maybe change to use clk->enable() functions like on omap1?
379 */
380void omap2_clksel_recalc(struct clk *clk)
381{
382 u32 div = 0;
383
384 pr_debug("clock: recalc'ing clksel clk %s\n", clk->name);
385
386 div = omap2_clksel_get_divisor(clk);
387 if (div == 0)
388 return;
389
390 if (unlikely(clk->rate == clk->parent->rate / div))
391 return;
392 clk->rate = clk->parent->rate / div;
393
394 pr_debug("clock: new clock rate is %ld (div %d)\n", clk->rate, div);
395
396 if (unlikely(clk->flags & RATE_PROPAGATES))
397 propagate_rate(clk);
398}
399
400/**
401 * omap2_get_clksel_by_parent - return clksel struct for a given clk & parent
402 * @clk: OMAP struct clk ptr to inspect
403 * @src_clk: OMAP struct clk ptr of the parent clk to search for
404 *
405 * Scan the struct clksel array associated with the clock to find
406 * the element associated with the supplied parent clock address.
407 * Returns a pointer to the struct clksel on success or NULL on error.
408 */
409const struct clksel *omap2_get_clksel_by_parent(struct clk *clk,
410 struct clk *src_clk)
411{
412 const struct clksel *clks;
413
414 if (!clk->clksel)
415 return NULL;
416
417 for (clks = clk->clksel; clks->parent; clks++) {
418 if (clks->parent == src_clk)
419 break; /* Found the requested parent */
420 }
421
422 if (!clks->parent) {
423 printk(KERN_ERR "clock: Could not find parent clock %s in "
424 "clksel array of clock %s\n", src_clk->name,
425 clk->name);
426 return NULL;
427 }
428
429 return clks;
430}
431
432/**
433 * omap2_clksel_round_rate_div - find divisor for the given clock and rate
434 * @clk: OMAP struct clk to use
435 * @target_rate: desired clock rate
436 * @new_div: ptr to where we should store the divisor
437 *
438 * Finds 'best' divider value in an array based on the source and target
439 * rates. The divider array must be sorted with smallest divider first.
440 * Note that this will not work for clocks which are part of CONFIG_PARTICIPANT,
441 * they are only settable as part of virtual_prcm set.
442 *
443 * Returns the rounded clock rate or returns 0xffffffff on error.
444 */
445u32 omap2_clksel_round_rate_div(struct clk *clk, unsigned long target_rate,
446 u32 *new_div)
447{
448 unsigned long test_rate;
449 const struct clksel *clks;
450 const struct clksel_rate *clkr;
451 u32 last_div = 0;
452
453 printk(KERN_INFO "clock: clksel_round_rate_div: %s target_rate %ld\n",
454 clk->name, target_rate);
455
456 *new_div = 1;
457
458 clks = omap2_get_clksel_by_parent(clk, clk->parent);
459 if (clks == NULL)
460 return ~0;
461
462 for (clkr = clks->rates; clkr->div; clkr++) {
463 if (!(clkr->flags & cpu_mask))
464 continue;
465
466 /* Sanity check */
467 if (clkr->div <= last_div)
468 printk(KERN_ERR "clock: clksel_rate table not sorted "
469 "for clock %s", clk->name);
470
471 last_div = clkr->div;
472
473 test_rate = clk->parent->rate / clkr->div;
474
475 if (test_rate <= target_rate)
476 break; /* found it */
477 }
478
479 if (!clkr->div) {
480 printk(KERN_ERR "clock: Could not find divisor for target "
481 "rate %ld for clock %s parent %s\n", target_rate,
482 clk->name, clk->parent->name);
483 return ~0;
484 }
485
486 *new_div = clkr->div;
487
488 printk(KERN_INFO "clock: new_div = %d, new_rate = %ld\n", *new_div,
489 (clk->parent->rate / clkr->div));
490
491 return (clk->parent->rate / clkr->div);
492}
493
494/**
495 * omap2_clksel_round_rate - find rounded rate for the given clock and rate
496 * @clk: OMAP struct clk to use
497 * @target_rate: desired clock rate
498 *
499 * Compatibility wrapper for OMAP clock framework
500 * Finds best target rate based on the source clock and possible dividers.
501 * rates. The divider array must be sorted with smallest divider first.
502 * Note that this will not work for clocks which are part of CONFIG_PARTICIPANT,
503 * they are only settable as part of virtual_prcm set.
504 *
505 * Returns the rounded clock rate or returns 0xffffffff on error.
506 */
507long omap2_clksel_round_rate(struct clk *clk, unsigned long target_rate)
508{
509 u32 new_div;
510
511 return omap2_clksel_round_rate_div(clk, target_rate, &new_div);
512}
513
514
515/* Given a clock and a rate apply a clock specific rounding function */
516long omap2_clk_round_rate(struct clk *clk, unsigned long rate)
517{
Russell Kingc0fc18c2008-09-05 15:10:27 +0100518 if (clk->round_rate != NULL)
Paul Walmsley543d9372008-03-18 10:22:06 +0200519 return clk->round_rate(clk, rate);
520
521 if (clk->flags & RATE_FIXED)
522 printk(KERN_ERR "clock: generic omap2_clk_round_rate called "
523 "on fixed-rate clock %s\n", clk->name);
524
525 return clk->rate;
526}
527
528/**
529 * omap2_clksel_to_divisor() - turn clksel field value into integer divider
530 * @clk: OMAP struct clk to use
531 * @field_val: register field value to find
532 *
533 * Given a struct clk of a rate-selectable clksel clock, and a register field
534 * value to search for, find the corresponding clock divisor. The register
535 * field value should be pre-masked and shifted down so the LSB is at bit 0
536 * before calling. Returns 0 on error
537 */
538u32 omap2_clksel_to_divisor(struct clk *clk, u32 field_val)
539{
540 const struct clksel *clks;
541 const struct clksel_rate *clkr;
542
543 clks = omap2_get_clksel_by_parent(clk, clk->parent);
544 if (clks == NULL)
545 return 0;
546
547 for (clkr = clks->rates; clkr->div; clkr++) {
548 if ((clkr->flags & cpu_mask) && (clkr->val == field_val))
549 break;
550 }
551
552 if (!clkr->div) {
553 printk(KERN_ERR "clock: Could not find fieldval %d for "
554 "clock %s parent %s\n", field_val, clk->name,
555 clk->parent->name);
556 return 0;
557 }
558
559 return clkr->div;
560}
561
562/**
563 * omap2_divisor_to_clksel() - turn clksel integer divisor into a field value
564 * @clk: OMAP struct clk to use
565 * @div: integer divisor to search for
566 *
567 * Given a struct clk of a rate-selectable clksel clock, and a clock divisor,
568 * find the corresponding register field value. The return register value is
569 * the value before left-shifting. Returns 0xffffffff on error
570 */
571u32 omap2_divisor_to_clksel(struct clk *clk, u32 div)
572{
573 const struct clksel *clks;
574 const struct clksel_rate *clkr;
575
576 /* should never happen */
577 WARN_ON(div == 0);
578
579 clks = omap2_get_clksel_by_parent(clk, clk->parent);
580 if (clks == NULL)
581 return 0;
582
583 for (clkr = clks->rates; clkr->div; clkr++) {
584 if ((clkr->flags & cpu_mask) && (clkr->div == div))
585 break;
586 }
587
588 if (!clkr->div) {
589 printk(KERN_ERR "clock: Could not find divisor %d for "
590 "clock %s parent %s\n", div, clk->name,
591 clk->parent->name);
592 return 0;
593 }
594
595 return clkr->val;
596}
597
598/**
599 * omap2_get_clksel - find clksel register addr & field mask for a clk
600 * @clk: struct clk to use
601 * @field_mask: ptr to u32 to store the register field mask
602 *
603 * Returns the address of the clksel register upon success or NULL on error.
604 */
605void __iomem *omap2_get_clksel(struct clk *clk, u32 *field_mask)
606{
Russell Kingc0fc18c2008-09-05 15:10:27 +0100607 if (unlikely((clk->clksel_reg == NULL) || (clk->clksel_mask == NULL)))
Paul Walmsley543d9372008-03-18 10:22:06 +0200608 return NULL;
609
610 *field_mask = clk->clksel_mask;
611
612 return clk->clksel_reg;
613}
614
615/**
616 * omap2_clksel_get_divisor - get current divider applied to parent clock.
617 * @clk: OMAP struct clk to use.
618 *
619 * Returns the integer divisor upon success or 0 on error.
620 */
621u32 omap2_clksel_get_divisor(struct clk *clk)
622{
623 u32 field_mask, field_val;
624 void __iomem *div_addr;
625
626 div_addr = omap2_get_clksel(clk, &field_mask);
Russell Kingc0fc18c2008-09-05 15:10:27 +0100627 if (div_addr == NULL)
Paul Walmsley543d9372008-03-18 10:22:06 +0200628 return 0;
629
630 field_val = __raw_readl(div_addr) & field_mask;
631 field_val >>= __ffs(field_mask);
632
633 return omap2_clksel_to_divisor(clk, field_val);
634}
635
636int omap2_clksel_set_rate(struct clk *clk, unsigned long rate)
637{
638 u32 field_mask, field_val, reg_val, validrate, new_div = 0;
639 void __iomem *div_addr;
640
641 validrate = omap2_clksel_round_rate_div(clk, rate, &new_div);
642 if (validrate != rate)
643 return -EINVAL;
644
645 div_addr = omap2_get_clksel(clk, &field_mask);
Russell Kingc0fc18c2008-09-05 15:10:27 +0100646 if (div_addr == NULL)
Paul Walmsley543d9372008-03-18 10:22:06 +0200647 return -EINVAL;
648
649 field_val = omap2_divisor_to_clksel(clk, new_div);
650 if (field_val == ~0)
651 return -EINVAL;
652
653 reg_val = __raw_readl(div_addr);
654 reg_val &= ~field_mask;
655 reg_val |= (field_val << __ffs(field_mask));
656 __raw_writel(reg_val, div_addr);
657 wmb();
658
659 clk->rate = clk->parent->rate / new_div;
660
661 if (clk->flags & DELAYED_APP && cpu_is_omap24xx()) {
Tony Lindgrenc2d43e32008-07-03 12:24:38 +0300662 prm_write_mod_reg(OMAP24XX_VALID_CONFIG,
663 OMAP24XX_GR_MOD, OMAP24XX_PRCM_CLKCFG_CTRL_OFFSET);
Paul Walmsley543d9372008-03-18 10:22:06 +0200664 wmb();
665 }
666
667 return 0;
668}
669
670
671/* Set the clock rate for a clock source */
672int omap2_clk_set_rate(struct clk *clk, unsigned long rate)
673{
674 int ret = -EINVAL;
675
676 pr_debug("clock: set_rate for clock %s to rate %ld\n", clk->name, rate);
677
678 /* CONFIG_PARTICIPANT clocks are changed only in sets via the
679 rate table mechanism, driven by mpu_speed */
680 if (clk->flags & CONFIG_PARTICIPANT)
681 return -EINVAL;
682
683 /* dpll_ck, core_ck, virt_prcm_set; plus all clksel clocks */
Russell Kingc0fc18c2008-09-05 15:10:27 +0100684 if (clk->set_rate != NULL)
Paul Walmsley543d9372008-03-18 10:22:06 +0200685 ret = clk->set_rate(clk, rate);
686
Paul Walmsley543d9372008-03-18 10:22:06 +0200687 return ret;
688}
689
690/*
691 * Converts encoded control register address into a full address
692 * On error, *src_addr will be returned as 0.
693 */
694static u32 omap2_clksel_get_src_field(void __iomem **src_addr,
695 struct clk *src_clk, u32 *field_mask,
696 struct clk *clk, u32 *parent_div)
697{
698 const struct clksel *clks;
699 const struct clksel_rate *clkr;
700
701 *parent_div = 0;
Russell Kingc0fc18c2008-09-05 15:10:27 +0100702 *src_addr = NULL;
Paul Walmsley543d9372008-03-18 10:22:06 +0200703
704 clks = omap2_get_clksel_by_parent(clk, src_clk);
705 if (clks == NULL)
706 return 0;
707
708 for (clkr = clks->rates; clkr->div; clkr++) {
709 if (clkr->flags & (cpu_mask | DEFAULT_RATE))
710 break; /* Found the default rate for this platform */
711 }
712
713 if (!clkr->div) {
714 printk(KERN_ERR "clock: Could not find default rate for "
715 "clock %s parent %s\n", clk->name,
716 src_clk->parent->name);
717 return 0;
718 }
719
720 /* Should never happen. Add a clksel mask to the struct clk. */
721 WARN_ON(clk->clksel_mask == 0);
722
723 *field_mask = clk->clksel_mask;
724 *src_addr = clk->clksel_reg;
725 *parent_div = clkr->div;
726
727 return clkr->val;
728}
729
730int omap2_clk_set_parent(struct clk *clk, struct clk *new_parent)
731{
732 void __iomem *src_addr;
733 u32 field_val, field_mask, reg_val, parent_div;
734
735 if (unlikely(clk->flags & CONFIG_PARTICIPANT))
736 return -EINVAL;
737
738 if (!clk->clksel)
739 return -EINVAL;
740
741 field_val = omap2_clksel_get_src_field(&src_addr, new_parent,
742 &field_mask, clk, &parent_div);
Russell Kingc0fc18c2008-09-05 15:10:27 +0100743 if (src_addr == NULL)
Paul Walmsley543d9372008-03-18 10:22:06 +0200744 return -EINVAL;
745
746 if (clk->usecount > 0)
747 _omap2_clk_disable(clk);
748
749 /* Set new source value (previous dividers if any in effect) */
750 reg_val = __raw_readl(src_addr) & ~field_mask;
751 reg_val |= (field_val << __ffs(field_mask));
752 __raw_writel(reg_val, src_addr);
753 wmb();
754
755 if (clk->flags & DELAYED_APP && cpu_is_omap24xx()) {
756 __raw_writel(OMAP24XX_VALID_CONFIG, OMAP24XX_PRCM_CLKCFG_CTRL);
757 wmb();
758 }
759
760 if (clk->usecount > 0)
761 _omap2_clk_enable(clk);
762
763 clk->parent = new_parent;
764
765 /* CLKSEL clocks follow their parents' rates, divided by a divisor */
766 clk->rate = new_parent->rate;
767
768 if (parent_div > 0)
769 clk->rate /= parent_div;
770
771 pr_debug("clock: set parent of %s to %s (new rate %ld)\n",
772 clk->name, clk->parent->name, clk->rate);
773
Paul Walmsley543d9372008-03-18 10:22:06 +0200774 return 0;
775}
776
Paul Walmsley88b8ba92008-07-03 12:24:46 +0300777/* DPLL rate rounding code */
778
779/**
780 * omap2_dpll_set_rate_tolerance: set the error tolerance during rate rounding
781 * @clk: struct clk * of the DPLL
782 * @tolerance: maximum rate error tolerance
783 *
784 * Set the maximum DPLL rate error tolerance for the rate rounding
785 * algorithm. The rate tolerance is an attempt to balance DPLL power
786 * saving (the least divider value "n") vs. rate fidelity (the least
787 * difference between the desired DPLL target rate and the rounded
788 * rate out of the algorithm). So, increasing the tolerance is likely
789 * to decrease DPLL power consumption and increase DPLL rate error.
790 * Returns -EINVAL if provided a null clock ptr or a clk that is not a
791 * DPLL; or 0 upon success.
792 */
793int omap2_dpll_set_rate_tolerance(struct clk *clk, unsigned int tolerance)
794{
795 if (!clk || !clk->dpll_data)
796 return -EINVAL;
797
798 clk->dpll_data->rate_tolerance = tolerance;
799
800 return 0;
801}
802
803static unsigned long _dpll_compute_new_rate(unsigned long parent_rate, unsigned int m, unsigned int n)
804{
805 unsigned long long num;
806
807 num = (unsigned long long)parent_rate * m;
808 do_div(num, n);
809 return num;
810}
811
812/*
813 * _dpll_test_mult - test a DPLL multiplier value
814 * @m: pointer to the DPLL m (multiplier) value under test
815 * @n: current DPLL n (divider) value under test
816 * @new_rate: pointer to storage for the resulting rounded rate
817 * @target_rate: the desired DPLL rate
818 * @parent_rate: the DPLL's parent clock rate
819 *
820 * This code tests a DPLL multiplier value, ensuring that the
821 * resulting rate will not be higher than the target_rate, and that
822 * the multiplier value itself is valid for the DPLL. Initially, the
823 * integer pointed to by the m argument should be prescaled by
824 * multiplying by DPLL_SCALE_FACTOR. The code will replace this with
825 * a non-scaled m upon return. This non-scaled m will result in a
826 * new_rate as close as possible to target_rate (but not greater than
827 * target_rate) given the current (parent_rate, n, prescaled m)
828 * triple. Returns DPLL_MULT_UNDERFLOW in the event that the
829 * non-scaled m attempted to underflow, which can allow the calling
830 * function to bail out early; or 0 upon success.
831 */
832static int _dpll_test_mult(int *m, int n, unsigned long *new_rate,
833 unsigned long target_rate,
834 unsigned long parent_rate)
835{
836 int flags = 0, carry = 0;
837
838 /* Unscale m and round if necessary */
839 if (*m % DPLL_SCALE_FACTOR >= DPLL_ROUNDING_VAL)
840 carry = 1;
841 *m = (*m / DPLL_SCALE_FACTOR) + carry;
842
843 /*
844 * The new rate must be <= the target rate to avoid programming
845 * a rate that is impossible for the hardware to handle
846 */
847 *new_rate = _dpll_compute_new_rate(parent_rate, *m, n);
848 if (*new_rate > target_rate) {
849 (*m)--;
850 *new_rate = 0;
851 }
852
853 /* Guard against m underflow */
854 if (*m < DPLL_MIN_MULTIPLIER) {
855 *m = DPLL_MIN_MULTIPLIER;
856 *new_rate = 0;
857 flags = DPLL_MULT_UNDERFLOW;
858 }
859
860 if (*new_rate == 0)
861 *new_rate = _dpll_compute_new_rate(parent_rate, *m, n);
862
863 return flags;
864}
865
866/**
867 * omap2_dpll_round_rate - round a target rate for an OMAP DPLL
868 * @clk: struct clk * for a DPLL
869 * @target_rate: desired DPLL clock rate
870 *
871 * Given a DPLL, a desired target rate, and a rate tolerance, round
872 * the target rate to a possible, programmable rate for this DPLL.
873 * Rate tolerance is assumed to be set by the caller before this
874 * function is called. Attempts to select the minimum possible n
875 * within the tolerance to reduce power consumption. Stores the
876 * computed (m, n) in the DPLL's dpll_data structure so set_rate()
877 * will not need to call this (expensive) function again. Returns ~0
878 * if the target rate cannot be rounded, either because the rate is
879 * too low or because the rate tolerance is set too tightly; or the
880 * rounded rate upon success.
881 */
882long omap2_dpll_round_rate(struct clk *clk, unsigned long target_rate)
883{
884 int m, n, r, e, scaled_max_m;
885 unsigned long scaled_rt_rp, new_rate;
886 int min_e = -1, min_e_m = -1, min_e_n = -1;
887
888 if (!clk || !clk->dpll_data)
889 return ~0;
890
891 pr_debug("clock: starting DPLL round_rate for clock %s, target rate "
892 "%ld\n", clk->name, target_rate);
893
894 scaled_rt_rp = target_rate / (clk->parent->rate / DPLL_SCALE_FACTOR);
895 scaled_max_m = clk->dpll_data->max_multiplier * DPLL_SCALE_FACTOR;
896
897 clk->dpll_data->last_rounded_rate = 0;
898
899 for (n = clk->dpll_data->max_divider; n >= DPLL_MIN_DIVIDER; n--) {
900
901 /* Compute the scaled DPLL multiplier, based on the divider */
902 m = scaled_rt_rp * n;
903
904 /*
905 * Since we're counting n down, a m overflow means we can
906 * can immediately skip to the next n
907 */
908 if (m > scaled_max_m)
909 continue;
910
911 r = _dpll_test_mult(&m, n, &new_rate, target_rate,
912 clk->parent->rate);
913
914 e = target_rate - new_rate;
915 pr_debug("clock: n = %d: m = %d: rate error is %d "
916 "(new_rate = %ld)\n", n, m, e, new_rate);
917
918 if (min_e == -1 ||
919 min_e >= (int)(abs(e) - clk->dpll_data->rate_tolerance)) {
920 min_e = e;
921 min_e_m = m;
922 min_e_n = n;
923
924 pr_debug("clock: found new least error %d\n", min_e);
925 }
926
927 /*
928 * Since we're counting n down, a m underflow means we
929 * can bail out completely (since as n decreases in
930 * the next iteration, there's no way that m can
931 * increase beyond the current m)
932 */
933 if (r & DPLL_MULT_UNDERFLOW)
934 break;
935 }
936
937 if (min_e < 0) {
938 pr_debug("clock: error: target rate or tolerance too low\n");
939 return ~0;
940 }
941
942 clk->dpll_data->last_rounded_m = min_e_m;
943 clk->dpll_data->last_rounded_n = min_e_n;
944 clk->dpll_data->last_rounded_rate =
945 _dpll_compute_new_rate(clk->parent->rate, min_e_m, min_e_n);
946
947 pr_debug("clock: final least error: e = %d, m = %d, n = %d\n",
948 min_e, min_e_m, min_e_n);
949 pr_debug("clock: final rate: %ld (target rate: %ld)\n",
950 clk->dpll_data->last_rounded_rate, target_rate);
951
952 return clk->dpll_data->last_rounded_rate;
953}
954
Paul Walmsley543d9372008-03-18 10:22:06 +0200955/*-------------------------------------------------------------------------
956 * Omap2 clock reset and init functions
957 *-------------------------------------------------------------------------*/
958
959#ifdef CONFIG_OMAP_RESET_CLOCKS
960void omap2_clk_disable_unused(struct clk *clk)
961{
962 u32 regval32, v;
963
964 v = (clk->flags & INVERT_ENABLE) ? (1 << clk->enable_bit) : 0;
965
966 regval32 = __raw_readl(clk->enable_reg);
967 if ((regval32 & (1 << clk->enable_bit)) == v)
968 return;
969
970 printk(KERN_INFO "Disabling unused clock \"%s\"\n", clk->name);
971 _omap2_clk_disable(clk);
972}
973#endif