Partial revert "Basic kernel memory functionality for the Memory Controller"
[linux-2.6/libata-dev.git] / include / linux / clocksource.h
blobc86c940d1de3a58b73fd4027b80524daca997478
1 /* linux/include/linux/clocksource.h
3 * This file contains the structure definitions for clocksources.
5 * If you are not a clocksource, or timekeeping code, you should
6 * not be including this file!
7 */
8 #ifndef _LINUX_CLOCKSOURCE_H
9 #define _LINUX_CLOCKSOURCE_H
11 #include <linux/types.h>
12 #include <linux/timex.h>
13 #include <linux/time.h>
14 #include <linux/list.h>
15 #include <linux/cache.h>
16 #include <linux/timer.h>
17 #include <linux/init.h>
18 #include <asm/div64.h>
19 #include <asm/io.h>
21 /* clocksource cycle base type */
22 typedef u64 cycle_t;
23 struct clocksource;
25 #ifdef CONFIG_ARCH_CLOCKSOURCE_DATA
26 #include <asm/clocksource.h>
27 #endif
29 /**
30 * struct cyclecounter - hardware abstraction for a free running counter
31 * Provides completely state-free accessors to the underlying hardware.
32 * Depending on which hardware it reads, the cycle counter may wrap
33 * around quickly. Locking rules (if necessary) have to be defined
34 * by the implementor and user of specific instances of this API.
36 * @read: returns the current cycle value
37 * @mask: bitmask for two's complement
38 * subtraction of non 64 bit counters,
39 * see CLOCKSOURCE_MASK() helper macro
40 * @mult: cycle to nanosecond multiplier
41 * @shift: cycle to nanosecond divisor (power of two)
43 struct cyclecounter {
44 cycle_t (*read)(const struct cyclecounter *cc);
45 cycle_t mask;
46 u32 mult;
47 u32 shift;
50 /**
51 * struct timecounter - layer above a %struct cyclecounter which counts nanoseconds
52 * Contains the state needed by timecounter_read() to detect
53 * cycle counter wrap around. Initialize with
54 * timecounter_init(). Also used to convert cycle counts into the
55 * corresponding nanosecond counts with timecounter_cyc2time(). Users
56 * of this code are responsible for initializing the underlying
57 * cycle counter hardware, locking issues and reading the time
58 * more often than the cycle counter wraps around. The nanosecond
59 * counter will only wrap around after ~585 years.
61 * @cc: the cycle counter used by this instance
62 * @cycle_last: most recent cycle counter value seen by
63 * timecounter_read()
64 * @nsec: continuously increasing count
66 struct timecounter {
67 const struct cyclecounter *cc;
68 cycle_t cycle_last;
69 u64 nsec;
72 /**
73 * cyclecounter_cyc2ns - converts cycle counter cycles to nanoseconds
74 * @tc: Pointer to cycle counter.
75 * @cycles: Cycles
77 * XXX - This could use some mult_lxl_ll() asm optimization. Same code
78 * as in cyc2ns, but with unsigned result.
80 static inline u64 cyclecounter_cyc2ns(const struct cyclecounter *cc,
81 cycle_t cycles)
83 u64 ret = (u64)cycles;
84 ret = (ret * cc->mult) >> cc->shift;
85 return ret;
88 /**
89 * timecounter_init - initialize a time counter
90 * @tc: Pointer to time counter which is to be initialized/reset
91 * @cc: A cycle counter, ready to be used.
92 * @start_tstamp: Arbitrary initial time stamp.
94 * After this call the current cycle register (roughly) corresponds to
95 * the initial time stamp. Every call to timecounter_read() increments
96 * the time stamp counter by the number of elapsed nanoseconds.
98 extern void timecounter_init(struct timecounter *tc,
99 const struct cyclecounter *cc,
100 u64 start_tstamp);
103 * timecounter_read - return nanoseconds elapsed since timecounter_init()
104 * plus the initial time stamp
105 * @tc: Pointer to time counter.
107 * In other words, keeps track of time since the same epoch as
108 * the function which generated the initial time stamp.
110 extern u64 timecounter_read(struct timecounter *tc);
113 * timecounter_cyc2time - convert a cycle counter to same
114 * time base as values returned by
115 * timecounter_read()
116 * @tc: Pointer to time counter.
117 * @cycle: a value returned by tc->cc->read()
119 * Cycle counts that are converted correctly as long as they
120 * fall into the interval [-1/2 max cycle count, +1/2 max cycle count],
121 * with "max cycle count" == cs->mask+1.
123 * This allows conversion of cycle counter values which were generated
124 * in the past.
126 extern u64 timecounter_cyc2time(struct timecounter *tc,
127 cycle_t cycle_tstamp);
130 * struct clocksource - hardware abstraction for a free running counter
131 * Provides mostly state-free accessors to the underlying hardware.
132 * This is the structure used for system time.
134 * @name: ptr to clocksource name
135 * @list: list head for registration
136 * @rating: rating value for selection (higher is better)
137 * To avoid rating inflation the following
138 * list should give you a guide as to how
139 * to assign your clocksource a rating
140 * 1-99: Unfit for real use
141 * Only available for bootup and testing purposes.
142 * 100-199: Base level usability.
143 * Functional for real use, but not desired.
144 * 200-299: Good.
145 * A correct and usable clocksource.
146 * 300-399: Desired.
147 * A reasonably fast and accurate clocksource.
148 * 400-499: Perfect
149 * The ideal clocksource. A must-use where
150 * available.
151 * @read: returns a cycle value, passes clocksource as argument
152 * @enable: optional function to enable the clocksource
153 * @disable: optional function to disable the clocksource
154 * @mask: bitmask for two's complement
155 * subtraction of non 64 bit counters
156 * @mult: cycle to nanosecond multiplier
157 * @shift: cycle to nanosecond divisor (power of two)
158 * @max_idle_ns: max idle time permitted by the clocksource (nsecs)
159 * @maxadj maximum adjustment value to mult (~11%)
160 * @flags: flags describing special properties
161 * @archdata: arch-specific data
162 * @suspend: suspend function for the clocksource, if necessary
163 * @resume: resume function for the clocksource, if necessary
165 struct clocksource {
167 * Hotpath data, fits in a single cache line when the
168 * clocksource itself is cacheline aligned.
170 cycle_t (*read)(struct clocksource *cs);
171 cycle_t cycle_last;
172 cycle_t mask;
173 u32 mult;
174 u32 shift;
175 u64 max_idle_ns;
176 u32 maxadj;
177 #ifdef CONFIG_ARCH_CLOCKSOURCE_DATA
178 struct arch_clocksource_data archdata;
179 #endif
181 const char *name;
182 struct list_head list;
183 int rating;
184 int (*enable)(struct clocksource *cs);
185 void (*disable)(struct clocksource *cs);
186 unsigned long flags;
187 void (*suspend)(struct clocksource *cs);
188 void (*resume)(struct clocksource *cs);
190 #ifdef CONFIG_CLOCKSOURCE_WATCHDOG
191 /* Watchdog related data, used by the framework */
192 struct list_head wd_list;
193 cycle_t cs_last;
194 cycle_t wd_last;
195 #endif
196 } ____cacheline_aligned;
199 * Clock source flags bits::
201 #define CLOCK_SOURCE_IS_CONTINUOUS 0x01
202 #define CLOCK_SOURCE_MUST_VERIFY 0x02
204 #define CLOCK_SOURCE_WATCHDOG 0x10
205 #define CLOCK_SOURCE_VALID_FOR_HRES 0x20
206 #define CLOCK_SOURCE_UNSTABLE 0x40
208 /* simplify initialization of mask field */
209 #define CLOCKSOURCE_MASK(bits) (cycle_t)((bits) < 64 ? ((1ULL<<(bits))-1) : -1)
212 * clocksource_khz2mult - calculates mult from khz and shift
213 * @khz: Clocksource frequency in KHz
214 * @shift_constant: Clocksource shift factor
216 * Helper functions that converts a khz counter frequency to a timsource
217 * multiplier, given the clocksource shift value
219 static inline u32 clocksource_khz2mult(u32 khz, u32 shift_constant)
221 /* khz = cyc/(Million ns)
222 * mult/2^shift = ns/cyc
223 * mult = ns/cyc * 2^shift
224 * mult = 1Million/khz * 2^shift
225 * mult = 1000000 * 2^shift / khz
226 * mult = (1000000<<shift) / khz
228 u64 tmp = ((u64)1000000) << shift_constant;
230 tmp += khz/2; /* round for do_div */
231 do_div(tmp, khz);
233 return (u32)tmp;
237 * clocksource_hz2mult - calculates mult from hz and shift
238 * @hz: Clocksource frequency in Hz
239 * @shift_constant: Clocksource shift factor
241 * Helper functions that converts a hz counter
242 * frequency to a timsource multiplier, given the
243 * clocksource shift value
245 static inline u32 clocksource_hz2mult(u32 hz, u32 shift_constant)
247 /* hz = cyc/(Billion ns)
248 * mult/2^shift = ns/cyc
249 * mult = ns/cyc * 2^shift
250 * mult = 1Billion/hz * 2^shift
251 * mult = 1000000000 * 2^shift / hz
252 * mult = (1000000000<<shift) / hz
254 u64 tmp = ((u64)1000000000) << shift_constant;
256 tmp += hz/2; /* round for do_div */
257 do_div(tmp, hz);
259 return (u32)tmp;
263 * clocksource_cyc2ns - converts clocksource cycles to nanoseconds
265 * Converts cycles to nanoseconds, using the given mult and shift.
267 * XXX - This could use some mult_lxl_ll() asm optimization
269 static inline s64 clocksource_cyc2ns(cycle_t cycles, u32 mult, u32 shift)
271 return ((u64) cycles * mult) >> shift;
275 extern int clocksource_register(struct clocksource*);
276 extern void clocksource_unregister(struct clocksource*);
277 extern void clocksource_touch_watchdog(void);
278 extern struct clocksource* clocksource_get_next(void);
279 extern void clocksource_change_rating(struct clocksource *cs, int rating);
280 extern void clocksource_suspend(void);
281 extern void clocksource_resume(void);
282 extern struct clocksource * __init __weak clocksource_default_clock(void);
283 extern void clocksource_mark_unstable(struct clocksource *cs);
285 extern void
286 clocks_calc_mult_shift(u32 *mult, u32 *shift, u32 from, u32 to, u32 minsec);
289 * Don't call __clocksource_register_scale directly, use
290 * clocksource_register_hz/khz
292 extern int
293 __clocksource_register_scale(struct clocksource *cs, u32 scale, u32 freq);
294 extern void
295 __clocksource_updatefreq_scale(struct clocksource *cs, u32 scale, u32 freq);
297 static inline int clocksource_register_hz(struct clocksource *cs, u32 hz)
299 return __clocksource_register_scale(cs, 1, hz);
302 static inline int clocksource_register_khz(struct clocksource *cs, u32 khz)
304 return __clocksource_register_scale(cs, 1000, khz);
307 static inline void __clocksource_updatefreq_hz(struct clocksource *cs, u32 hz)
309 __clocksource_updatefreq_scale(cs, 1, hz);
312 static inline void __clocksource_updatefreq_khz(struct clocksource *cs, u32 khz)
314 __clocksource_updatefreq_scale(cs, 1000, khz);
317 static inline void
318 clocksource_calc_mult_shift(struct clocksource *cs, u32 freq, u32 minsec)
320 return clocks_calc_mult_shift(&cs->mult, &cs->shift, freq,
321 NSEC_PER_SEC, minsec);
324 #ifdef CONFIG_GENERIC_TIME_VSYSCALL
325 extern void
326 update_vsyscall(struct timespec *ts, struct timespec *wtm,
327 struct clocksource *c, u32 mult);
328 extern void update_vsyscall_tz(void);
329 #else
330 static inline void
331 update_vsyscall(struct timespec *ts, struct timespec *wtm,
332 struct clocksource *c, u32 mult)
336 static inline void update_vsyscall_tz(void)
339 #endif
341 extern void timekeeping_notify(struct clocksource *clock);
343 extern cycle_t clocksource_mmio_readl_up(struct clocksource *);
344 extern cycle_t clocksource_mmio_readl_down(struct clocksource *);
345 extern cycle_t clocksource_mmio_readw_up(struct clocksource *);
346 extern cycle_t clocksource_mmio_readw_down(struct clocksource *);
348 extern int clocksource_mmio_init(void __iomem *, const char *,
349 unsigned long, int, unsigned, cycle_t (*)(struct clocksource *));
351 extern int clocksource_i8253_init(void);
353 #endif /* _LINUX_CLOCKSOURCE_H */