blob: 73c33429f0e075c2dbebf98c71eaba0a4ace3556 [file] [log] [blame]
john stultz8d016ef2006-06-26 00:25:09 -07001/*
2 * i8253.c 8253/PIT functions
3 *
4 */
Thomas Gleixnere9e2cdb2007-02-16 01:28:04 -08005#include <linux/clockchips.h>
john stultz8d016ef2006-06-26 00:25:09 -07006#include <linux/init.h>
Thomas Gleixner18de5bc2007-07-21 04:37:34 -07007#include <linux/interrupt.h>
8#include <linux/jiffies.h>
9#include <linux/module.h>
10#include <linux/spinlock.h>
john stultz8d016ef2006-06-26 00:25:09 -070011
12#include <asm/smp.h>
13#include <asm/delay.h>
14#include <asm/i8253.h>
15#include <asm/io.h>
Adrian Bunk3f9c8d12007-07-21 17:10:28 +020016#include <asm/timer.h>
john stultz8d016ef2006-06-26 00:25:09 -070017
18#include "io_ports.h"
19
20DEFINE_SPINLOCK(i8253_lock);
21EXPORT_SYMBOL(i8253_lock);
22
Thomas Gleixnere9e2cdb2007-02-16 01:28:04 -080023/*
24 * HPET replaces the PIT, when enabled. So we need to know, which of
25 * the two timers is used
26 */
27struct clock_event_device *global_clock_event;
28
29/*
30 * Initialize the PIT timer.
31 *
32 * This is also called after resume to bring the PIT into operation again.
33 */
34static void init_pit_timer(enum clock_event_mode mode,
35 struct clock_event_device *evt)
john stultz8d016ef2006-06-26 00:25:09 -070036{
37 unsigned long flags;
38
39 spin_lock_irqsave(&i8253_lock, flags);
Thomas Gleixnere9e2cdb2007-02-16 01:28:04 -080040
41 switch(mode) {
42 case CLOCK_EVT_MODE_PERIODIC:
43 /* binary, mode 2, LSB/MSB, ch 0 */
44 outb_p(0x34, PIT_MODE);
Thomas Gleixnere9e2cdb2007-02-16 01:28:04 -080045 outb_p(LATCH & 0xff , PIT_CH0); /* LSB */
Thomas Gleixnere9e2cdb2007-02-16 01:28:04 -080046 outb(LATCH >> 8 , PIT_CH0); /* MSB */
47 break;
48
Thomas Gleixnere9e2cdb2007-02-16 01:28:04 -080049 case CLOCK_EVT_MODE_SHUTDOWN:
50 case CLOCK_EVT_MODE_UNUSED:
Thomas Gleixner76719882007-07-21 04:37:38 -070051 if (evt->mode == CLOCK_EVT_MODE_PERIODIC ||
52 evt->mode == CLOCK_EVT_MODE_ONESHOT) {
53 outb_p(0x30, PIT_MODE);
54 outb_p(0, PIT_CH0);
55 outb_p(0, PIT_CH0);
56 }
Thomas Gleixner18de5bc2007-07-21 04:37:34 -070057 break;
58
Thomas Gleixner6b3964c2007-03-22 22:46:18 +010059 case CLOCK_EVT_MODE_ONESHOT:
Thomas Gleixnere9e2cdb2007-02-16 01:28:04 -080060 /* One shot setup */
61 outb_p(0x38, PIT_MODE);
Thomas Gleixner18de5bc2007-07-21 04:37:34 -070062 break;
63
64 case CLOCK_EVT_MODE_RESUME:
65 /* Nothing to do here */
Thomas Gleixnere9e2cdb2007-02-16 01:28:04 -080066 break;
67 }
john stultz8d016ef2006-06-26 00:25:09 -070068 spin_unlock_irqrestore(&i8253_lock, flags);
69}
john stultz5d0cf412006-06-26 00:25:12 -070070
71/*
Thomas Gleixnere9e2cdb2007-02-16 01:28:04 -080072 * Program the next event in oneshot mode
73 *
74 * Delta is given in PIT ticks
75 */
76static int pit_next_event(unsigned long delta, struct clock_event_device *evt)
77{
78 unsigned long flags;
79
80 spin_lock_irqsave(&i8253_lock, flags);
81 outb_p(delta & 0xff , PIT_CH0); /* LSB */
82 outb(delta >> 8 , PIT_CH0); /* MSB */
83 spin_unlock_irqrestore(&i8253_lock, flags);
84
85 return 0;
86}
87
88/*
89 * On UP the PIT can serve all of the possible timer functions. On SMP systems
90 * it can be solely used for the global tick.
91 *
92 * The profiling and update capabilites are switched off once the local apic is
93 * registered. This mechanism replaces the previous #ifdef LOCAL_APIC -
94 * !using_apic_timer decisions in do_timer_interrupt_hook()
95 */
96struct clock_event_device pit_clockevent = {
97 .name = "pit",
98 .features = CLOCK_EVT_FEAT_PERIODIC | CLOCK_EVT_FEAT_ONESHOT,
99 .set_mode = init_pit_timer,
100 .set_next_event = pit_next_event,
101 .shift = 32,
102 .irq = 0,
103};
104
105/*
106 * Initialize the conversion factor and the min/max deltas of the clock event
107 * structure and register the clock event source with the framework.
108 */
109void __init setup_pit_timer(void)
110{
111 /*
112 * Start pit with the boot cpu mask and make it global after the
113 * IO_APIC has been initialized.
114 */
James Bottomley2feae212007-04-30 11:27:25 -0500115 pit_clockevent.cpumask = cpumask_of_cpu(smp_processor_id());
Thomas Gleixnere9e2cdb2007-02-16 01:28:04 -0800116 pit_clockevent.mult = div_sc(CLOCK_TICK_RATE, NSEC_PER_SEC, 32);
117 pit_clockevent.max_delta_ns =
118 clockevent_delta2ns(0x7FFF, &pit_clockevent);
119 pit_clockevent.min_delta_ns =
120 clockevent_delta2ns(0xF, &pit_clockevent);
121 clockevents_register_device(&pit_clockevent);
122 global_clock_event = &pit_clockevent;
123}
124
125/*
john stultz5d0cf412006-06-26 00:25:12 -0700126 * Since the PIT overflows every tick, its not very useful
127 * to just read by itself. So use jiffies to emulate a free
128 * running counter:
129 */
130static cycle_t pit_read(void)
131{
132 unsigned long flags;
133 int count;
john stultz6415ce92006-06-26 00:25:16 -0700134 u32 jifs;
135 static int old_count;
136 static u32 old_jifs;
john stultz5d0cf412006-06-26 00:25:12 -0700137
138 spin_lock_irqsave(&i8253_lock, flags);
Thomas Gleixnere9e2cdb2007-02-16 01:28:04 -0800139 /*
john stultz6415ce92006-06-26 00:25:16 -0700140 * Although our caller may have the read side of xtime_lock,
141 * this is now a seqlock, and we are cheating in this routine
142 * by having side effects on state that we cannot undo if
143 * there is a collision on the seqlock and our caller has to
144 * retry. (Namely, old_jifs and old_count.) So we must treat
145 * jiffies as volatile despite the lock. We read jiffies
146 * before latching the timer count to guarantee that although
147 * the jiffies value might be older than the count (that is,
148 * the counter may underflow between the last point where
149 * jiffies was incremented and the point where we latch the
150 * count), it cannot be newer.
151 */
152 jifs = jiffies;
john stultz5d0cf412006-06-26 00:25:12 -0700153 outb_p(0x00, PIT_MODE); /* latch the count ASAP */
154 count = inb_p(PIT_CH0); /* read the latched count */
155 count |= inb_p(PIT_CH0) << 8;
156
157 /* VIA686a test code... reset the latch if count > max + 1 */
158 if (count > LATCH) {
159 outb_p(0x34, PIT_MODE);
160 outb_p(LATCH & 0xff, PIT_CH0);
161 outb(LATCH >> 8, PIT_CH0);
162 count = LATCH - 1;
163 }
john stultz6415ce92006-06-26 00:25:16 -0700164
165 /*
166 * It's possible for count to appear to go the wrong way for a
167 * couple of reasons:
168 *
169 * 1. The timer counter underflows, but we haven't handled the
170 * resulting interrupt and incremented jiffies yet.
171 * 2. Hardware problem with the timer, not giving us continuous time,
172 * the counter does small "jumps" upwards on some Pentium systems,
173 * (see c't 95/10 page 335 for Neptun bug.)
174 *
175 * Previous attempts to handle these cases intelligently were
176 * buggy, so we just do the simple thing now.
177 */
178 if (count > old_count && jifs == old_jifs) {
179 count = old_count;
180 }
181 old_count = count;
182 old_jifs = jifs;
183
john stultz5d0cf412006-06-26 00:25:12 -0700184 spin_unlock_irqrestore(&i8253_lock, flags);
185
john stultz6415ce92006-06-26 00:25:16 -0700186 count = (LATCH - 1) - count;
john stultz5d0cf412006-06-26 00:25:12 -0700187
188 return (cycle_t)(jifs * LATCH) + count;
189}
190
191static struct clocksource clocksource_pit = {
192 .name = "pit",
193 .rating = 110,
194 .read = pit_read,
john stultz6415ce92006-06-26 00:25:16 -0700195 .mask = CLOCKSOURCE_MASK(32),
john stultz5d0cf412006-06-26 00:25:12 -0700196 .mult = 0,
197 .shift = 20,
198};
199
200static int __init init_pit_clocksource(void)
201{
john stultz3f4a0b92006-10-17 00:09:32 -0700202 if (num_possible_cpus() > 1) /* PIT does not scale! */
john stultz5d0cf412006-06-26 00:25:12 -0700203 return 0;
204
205 clocksource_pit.mult = clocksource_hz2mult(CLOCK_TICK_RATE, 20);
john stultza2752542006-06-26 00:25:14 -0700206 return clocksource_register(&clocksource_pit);
john stultz5d0cf412006-06-26 00:25:12 -0700207}
john stultz6bb74df2007-03-05 00:30:50 -0800208arch_initcall(init_pit_clocksource);