blob: 8935570008d2eb192d9d6698c2b4580ea3c0882c [file] [log] [blame]
Paul Mundtaa016662006-01-16 22:14:18 -08001/*
2 * arch/sh/kernel/timers/timer-tmu.c - TMU Timer Support
3 *
Paul Mundt57be2b42007-05-09 17:33:24 +09004 * Copyright (C) 2005 - 2007 Paul Mundt
Paul Mundtaa016662006-01-16 22:14:18 -08005 *
6 * TMU handling code hacked out of arch/sh/kernel/time.c
7 *
8 * Copyright (C) 1999 Tetsuya Okada & Niibe Yutaka
9 * Copyright (C) 2000 Philipp Rumpf <prumpf@tux.org>
10 * Copyright (C) 2002, 2003, 2004 Paul Mundt
11 * Copyright (C) 2002 M. R. Brown <mrbrown@linux-sh.org>
12 *
13 * This file is subject to the terms and conditions of the GNU General Public
14 * License. See the file "COPYING" in the main directory of this archive
15 * for more details.
16 */
17#include <linux/init.h>
18#include <linux/kernel.h>
19#include <linux/interrupt.h>
Paul Mundtaa016662006-01-16 22:14:18 -080020#include <linux/seqlock.h>
Paul Mundt57be2b42007-05-09 17:33:24 +090021#include <linux/clockchips.h>
Paul Mundtaa016662006-01-16 22:14:18 -080022#include <asm/timer.h>
23#include <asm/rtc.h>
24#include <asm/io.h>
25#include <asm/irq.h>
26#include <asm/clock.h>
27
28#define TMU_TOCR_INIT 0x00
Paul Mundt57be2b42007-05-09 17:33:24 +090029#define TMU_TCR_INIT 0x0020
Paul Mundtaa016662006-01-16 22:14:18 -080030
Paul Mundt57be2b42007-05-09 17:33:24 +090031static int tmu_timer_start(void)
Paul Mundtaa016662006-01-16 22:14:18 -080032{
Paul Mundt2b1bd1a2007-06-20 18:27:10 +090033 ctrl_outb(ctrl_inb(TMU_012_TSTR) | 0x3, TMU_012_TSTR);
Paul Mundt57be2b42007-05-09 17:33:24 +090034 return 0;
Paul Mundtaa016662006-01-16 22:14:18 -080035}
36
Paul Mundt57be2b42007-05-09 17:33:24 +090037static void tmu0_timer_set_interval(unsigned long interval, unsigned int reload)
38{
39 ctrl_outl(interval, TMU0_TCNT);
40
41 /*
42 * TCNT reloads from TCOR on underflow, clear it if we don't
43 * intend to auto-reload
44 */
45 if (reload)
46 ctrl_outl(interval, TMU0_TCOR);
47 else
48 ctrl_outl(0, TMU0_TCOR);
49
50 tmu_timer_start();
51}
52
53static int tmu_timer_stop(void)
54{
Paul Mundt2b1bd1a2007-06-20 18:27:10 +090055 ctrl_outb(ctrl_inb(TMU_012_TSTR) & ~0x3, TMU_012_TSTR);
Paul Mundt57be2b42007-05-09 17:33:24 +090056 return 0;
57}
58
59static cycle_t tmu_timer_read(void)
60{
61 return ~ctrl_inl(TMU1_TCNT);
62}
63
64static int tmu_set_next_event(unsigned long cycles,
65 struct clock_event_device *evt)
66{
67 tmu0_timer_set_interval(cycles, 1);
68 return 0;
69}
70
71static void tmu_set_mode(enum clock_event_mode mode,
72 struct clock_event_device *evt)
73{
74 switch (mode) {
75 case CLOCK_EVT_MODE_PERIODIC:
76 ctrl_outl(ctrl_inl(TMU0_TCNT), TMU0_TCOR);
77 break;
78 case CLOCK_EVT_MODE_ONESHOT:
79 ctrl_outl(0, TMU0_TCOR);
80 break;
81 case CLOCK_EVT_MODE_UNUSED:
82 case CLOCK_EVT_MODE_SHUTDOWN:
Thomas Gleixner18de5bc2007-07-21 04:37:34 -070083 case CLOCK_EVT_MODE_RESUME:
Paul Mundt57be2b42007-05-09 17:33:24 +090084 break;
85 }
86}
87
88static struct clock_event_device tmu0_clockevent = {
89 .name = "tmu0",
90 .shift = 32,
91 .features = CLOCK_EVT_FEAT_PERIODIC | CLOCK_EVT_FEAT_ONESHOT,
92 .set_mode = tmu_set_mode,
93 .set_next_event = tmu_set_next_event,
94};
95
Paul Mundt35f3c512006-10-06 15:31:16 +090096static irqreturn_t tmu_timer_interrupt(int irq, void *dummy)
Paul Mundtaa016662006-01-16 22:14:18 -080097{
Paul Mundt57be2b42007-05-09 17:33:24 +090098 struct clock_event_device *evt = &tmu0_clockevent;
Paul Mundtaa016662006-01-16 22:14:18 -080099 unsigned long timer_status;
100
101 /* Clear UNF bit */
102 timer_status = ctrl_inw(TMU0_TCR);
103 timer_status &= ~0x100;
104 ctrl_outw(timer_status, TMU0_TCR);
105
Paul Mundt57be2b42007-05-09 17:33:24 +0900106 evt->event_handler(evt);
Paul Mundtaa016662006-01-16 22:14:18 -0800107
108 return IRQ_HANDLED;
109}
110
Paul Mundt57be2b42007-05-09 17:33:24 +0900111static struct irqaction tmu0_irq = {
112 .name = "periodic timer",
Paul Mundtaa016662006-01-16 22:14:18 -0800113 .handler = tmu_timer_interrupt,
Bernhard Wallee9485ba2007-05-08 00:35:34 -0700114 .flags = IRQF_DISABLED | IRQF_TIMER | IRQF_IRQPOLL,
Paul Mundtaa016662006-01-16 22:14:18 -0800115 .mask = CPU_MASK_NONE,
116};
117
Paul Mundt57be2b42007-05-09 17:33:24 +0900118static void tmu0_clk_init(struct clk *clk)
Paul Mundtaa016662006-01-16 22:14:18 -0800119{
Paul Mundt57be2b42007-05-09 17:33:24 +0900120 u8 divisor = TMU_TCR_INIT & 0x7;
121 ctrl_outw(TMU_TCR_INIT, TMU0_TCR);
Paul Mundtaa016662006-01-16 22:14:18 -0800122 clk->rate = clk->parent->rate / (4 << (divisor << 1));
123}
124
Paul Mundt57be2b42007-05-09 17:33:24 +0900125static void tmu0_clk_recalc(struct clk *clk)
Paul Mundtaa016662006-01-16 22:14:18 -0800126{
127 u8 divisor = ctrl_inw(TMU0_TCR) & 0x7;
128 clk->rate = clk->parent->rate / (4 << (divisor << 1));
129}
130
Paul Mundt57be2b42007-05-09 17:33:24 +0900131static struct clk_ops tmu0_clk_ops = {
132 .init = tmu0_clk_init,
133 .recalc = tmu0_clk_recalc,
Paul Mundtaa016662006-01-16 22:14:18 -0800134};
135
136static struct clk tmu0_clk = {
137 .name = "tmu0_clk",
Paul Mundt57be2b42007-05-09 17:33:24 +0900138 .ops = &tmu0_clk_ops,
Paul Mundtaa016662006-01-16 22:14:18 -0800139};
140
Paul Mundt57be2b42007-05-09 17:33:24 +0900141static void tmu1_clk_init(struct clk *clk)
Andriy Skulysh3aa770e2006-09-27 16:20:22 +0900142{
Paul Mundt57be2b42007-05-09 17:33:24 +0900143 u8 divisor = TMU_TCR_INIT & 0x7;
144 ctrl_outw(divisor, TMU1_TCR);
145 clk->rate = clk->parent->rate / (4 << (divisor << 1));
Andriy Skulysh3aa770e2006-09-27 16:20:22 +0900146}
147
Paul Mundt57be2b42007-05-09 17:33:24 +0900148static void tmu1_clk_recalc(struct clk *clk)
Andriy Skulysh3aa770e2006-09-27 16:20:22 +0900149{
Paul Mundt57be2b42007-05-09 17:33:24 +0900150 u8 divisor = ctrl_inw(TMU1_TCR) & 0x7;
151 clk->rate = clk->parent->rate / (4 << (divisor << 1));
Andriy Skulysh3aa770e2006-09-27 16:20:22 +0900152}
153
Paul Mundt57be2b42007-05-09 17:33:24 +0900154static struct clk_ops tmu1_clk_ops = {
155 .init = tmu1_clk_init,
156 .recalc = tmu1_clk_recalc,
157};
158
159static struct clk tmu1_clk = {
160 .name = "tmu1_clk",
161 .ops = &tmu1_clk_ops,
162};
163
Paul Mundtaa016662006-01-16 22:14:18 -0800164static int tmu_timer_init(void)
165{
166 unsigned long interval;
Paul Mundt57be2b42007-05-09 17:33:24 +0900167 unsigned long frequency;
Paul Mundtaa016662006-01-16 22:14:18 -0800168
Paul Mundt57be2b42007-05-09 17:33:24 +0900169 setup_irq(CONFIG_SH_TIMER_IRQ, &tmu0_irq);
Paul Mundtaa016662006-01-16 22:14:18 -0800170
Paul Mundt1d118562006-12-01 13:15:14 +0900171 tmu0_clk.parent = clk_get(NULL, "module_clk");
Paul Mundt57be2b42007-05-09 17:33:24 +0900172 tmu1_clk.parent = clk_get(NULL, "module_clk");
Paul Mundtaa016662006-01-16 22:14:18 -0800173
Andriy Skulysh3aa770e2006-09-27 16:20:22 +0900174 tmu_timer_stop();
Paul Mundt57be2b42007-05-09 17:33:24 +0900175
Markus Brunner3ea6bc32007-08-20 08:59:33 +0900176#if !defined(CONFIG_CPU_SUBTYPE_SH7720) && \
Yoshihiro Shimoda31a49c42007-12-26 11:45:06 +0900177 !defined(CONFIG_CPU_SUBTYPE_SH7721) && \
Markus Brunner3ea6bc32007-08-20 08:59:33 +0900178 !defined(CONFIG_CPU_SUBTYPE_SH7760) && \
Paul Mundt2b1bd1a2007-06-20 18:27:10 +0900179 !defined(CONFIG_CPU_SUBTYPE_SH7785) && \
180 !defined(CONFIG_CPU_SUBTYPE_SHX3)
Paul Mundtaa016662006-01-16 22:14:18 -0800181 ctrl_outb(TMU_TOCR_INIT, TMU_TOCR);
182#endif
183
184 clk_register(&tmu0_clk);
Paul Mundt57be2b42007-05-09 17:33:24 +0900185 clk_register(&tmu1_clk);
Paul Mundtaa016662006-01-16 22:14:18 -0800186 clk_enable(&tmu0_clk);
Paul Mundt57be2b42007-05-09 17:33:24 +0900187 clk_enable(&tmu1_clk);
Paul Mundtaa016662006-01-16 22:14:18 -0800188
Paul Mundt57be2b42007-05-09 17:33:24 +0900189 frequency = clk_get_rate(&tmu0_clk);
190 interval = (frequency + HZ / 2) / HZ;
Paul Mundtaa016662006-01-16 22:14:18 -0800191
Paul Mundt57be2b42007-05-09 17:33:24 +0900192 sh_hpt_frequency = clk_get_rate(&tmu1_clk);
193 ctrl_outl(~0, TMU1_TCNT);
194 ctrl_outl(~0, TMU1_TCOR);
Paul Mundtaa016662006-01-16 22:14:18 -0800195
Paul Mundt57be2b42007-05-09 17:33:24 +0900196 tmu0_timer_set_interval(interval, 1);
197
198 tmu0_clockevent.mult = div_sc(frequency, NSEC_PER_SEC,
199 tmu0_clockevent.shift);
200 tmu0_clockevent.max_delta_ns =
201 clockevent_delta2ns(-1, &tmu0_clockevent);
202 tmu0_clockevent.min_delta_ns =
203 clockevent_delta2ns(1, &tmu0_clockevent);
204
205 tmu0_clockevent.cpumask = cpumask_of_cpu(0);
206
207 clockevents_register_device(&tmu0_clockevent);
Paul Mundtaa016662006-01-16 22:14:18 -0800208
209 return 0;
210}
211
212struct sys_timer_ops tmu_timer_ops = {
213 .init = tmu_timer_init,
Andriy Skulysh3aa770e2006-09-27 16:20:22 +0900214 .start = tmu_timer_start,
215 .stop = tmu_timer_stop,
Paul Mundt57be2b42007-05-09 17:33:24 +0900216 .read = tmu_timer_read,
Paul Mundtaa016662006-01-16 22:14:18 -0800217};
218
219struct sys_timer tmu_timer = {
220 .name = "tmu",
221 .ops = &tmu_timer_ops,
222};