split dev_queue
[cor.git] / kernel / time / time.c
blob704ccd9451b0946145643467f60272c926a1c8b5
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * Copyright (C) 1991, 1992 Linus Torvalds
5 * This file contains the interface functions for the various time related
6 * system calls: time, stime, gettimeofday, settimeofday, adjtime
8 * Modification history:
10 * 1993-09-02 Philip Gladstone
11 * Created file with time related functions from sched/core.c and adjtimex()
12 * 1993-10-08 Torsten Duwe
13 * adjtime interface update and CMOS clock write code
14 * 1995-08-13 Torsten Duwe
15 * kernel PLL updated to 1994-12-13 specs (rfc-1589)
16 * 1999-01-16 Ulrich Windl
17 * Introduced error checking for many cases in adjtimex().
18 * Updated NTP code according to technical memorandum Jan '96
19 * "A Kernel Model for Precision Timekeeping" by Dave Mills
20 * Allow time_constant larger than MAXTC(6) for NTP v4 (MAXTC == 10)
21 * (Even though the technical memorandum forbids it)
22 * 2004-07-14 Christoph Lameter
23 * Added getnstimeofday to allow the posix timer functions to return
24 * with nanosecond accuracy
27 #include <linux/export.h>
28 #include <linux/kernel.h>
29 #include <linux/timex.h>
30 #include <linux/capability.h>
31 #include <linux/timekeeper_internal.h>
32 #include <linux/errno.h>
33 #include <linux/syscalls.h>
34 #include <linux/security.h>
35 #include <linux/fs.h>
36 #include <linux/math64.h>
37 #include <linux/ptrace.h>
39 #include <linux/uaccess.h>
40 #include <linux/compat.h>
41 #include <asm/unistd.h>
43 #include <generated/timeconst.h>
44 #include "timekeeping.h"
47 * The timezone where the local system is located. Used as a default by some
48 * programs who obtain this value by using gettimeofday.
50 struct timezone sys_tz;
52 EXPORT_SYMBOL(sys_tz);
54 #ifdef __ARCH_WANT_SYS_TIME
57 * sys_time() can be implemented in user-level using
58 * sys_gettimeofday(). Is this for backwards compatibility? If so,
59 * why not move it into the appropriate arch directory (for those
60 * architectures that need it).
62 SYSCALL_DEFINE1(time, __kernel_old_time_t __user *, tloc)
64 __kernel_old_time_t i = (__kernel_old_time_t)ktime_get_real_seconds();
66 if (tloc) {
67 if (put_user(i,tloc))
68 return -EFAULT;
70 force_successful_syscall_return();
71 return i;
75 * sys_stime() can be implemented in user-level using
76 * sys_settimeofday(). Is this for backwards compatibility? If so,
77 * why not move it into the appropriate arch directory (for those
78 * architectures that need it).
81 SYSCALL_DEFINE1(stime, __kernel_old_time_t __user *, tptr)
83 struct timespec64 tv;
84 int err;
86 if (get_user(tv.tv_sec, tptr))
87 return -EFAULT;
89 tv.tv_nsec = 0;
91 err = security_settime64(&tv, NULL);
92 if (err)
93 return err;
95 do_settimeofday64(&tv);
96 return 0;
99 #endif /* __ARCH_WANT_SYS_TIME */
101 #ifdef CONFIG_COMPAT_32BIT_TIME
102 #ifdef __ARCH_WANT_SYS_TIME32
104 /* old_time32_t is a 32 bit "long" and needs to get converted. */
105 SYSCALL_DEFINE1(time32, old_time32_t __user *, tloc)
107 old_time32_t i;
109 i = (old_time32_t)ktime_get_real_seconds();
111 if (tloc) {
112 if (put_user(i,tloc))
113 return -EFAULT;
115 force_successful_syscall_return();
116 return i;
119 SYSCALL_DEFINE1(stime32, old_time32_t __user *, tptr)
121 struct timespec64 tv;
122 int err;
124 if (get_user(tv.tv_sec, tptr))
125 return -EFAULT;
127 tv.tv_nsec = 0;
129 err = security_settime64(&tv, NULL);
130 if (err)
131 return err;
133 do_settimeofday64(&tv);
134 return 0;
137 #endif /* __ARCH_WANT_SYS_TIME32 */
138 #endif
140 SYSCALL_DEFINE2(gettimeofday, struct __kernel_old_timeval __user *, tv,
141 struct timezone __user *, tz)
143 if (likely(tv != NULL)) {
144 struct timespec64 ts;
146 ktime_get_real_ts64(&ts);
147 if (put_user(ts.tv_sec, &tv->tv_sec) ||
148 put_user(ts.tv_nsec / 1000, &tv->tv_usec))
149 return -EFAULT;
151 if (unlikely(tz != NULL)) {
152 if (copy_to_user(tz, &sys_tz, sizeof(sys_tz)))
153 return -EFAULT;
155 return 0;
159 * In case for some reason the CMOS clock has not already been running
160 * in UTC, but in some local time: The first time we set the timezone,
161 * we will warp the clock so that it is ticking UTC time instead of
162 * local time. Presumably, if someone is setting the timezone then we
163 * are running in an environment where the programs understand about
164 * timezones. This should be done at boot time in the /etc/rc script,
165 * as soon as possible, so that the clock can be set right. Otherwise,
166 * various programs will get confused when the clock gets warped.
169 int do_sys_settimeofday64(const struct timespec64 *tv, const struct timezone *tz)
171 static int firsttime = 1;
172 int error = 0;
174 if (tv && !timespec64_valid_settod(tv))
175 return -EINVAL;
177 error = security_settime64(tv, tz);
178 if (error)
179 return error;
181 if (tz) {
182 /* Verify we're within the +-15 hrs range */
183 if (tz->tz_minuteswest > 15*60 || tz->tz_minuteswest < -15*60)
184 return -EINVAL;
186 sys_tz = *tz;
187 update_vsyscall_tz();
188 if (firsttime) {
189 firsttime = 0;
190 if (!tv)
191 timekeeping_warp_clock();
194 if (tv)
195 return do_settimeofday64(tv);
196 return 0;
199 SYSCALL_DEFINE2(settimeofday, struct __kernel_old_timeval __user *, tv,
200 struct timezone __user *, tz)
202 struct timespec64 new_ts;
203 struct timezone new_tz;
205 if (tv) {
206 if (get_user(new_ts.tv_sec, &tv->tv_sec) ||
207 get_user(new_ts.tv_nsec, &tv->tv_usec))
208 return -EFAULT;
210 if (new_ts.tv_nsec > USEC_PER_SEC || new_ts.tv_nsec < 0)
211 return -EINVAL;
213 new_ts.tv_nsec *= NSEC_PER_USEC;
215 if (tz) {
216 if (copy_from_user(&new_tz, tz, sizeof(*tz)))
217 return -EFAULT;
220 return do_sys_settimeofday64(tv ? &new_ts : NULL, tz ? &new_tz : NULL);
223 #ifdef CONFIG_COMPAT
224 COMPAT_SYSCALL_DEFINE2(gettimeofday, struct old_timeval32 __user *, tv,
225 struct timezone __user *, tz)
227 if (tv) {
228 struct timespec64 ts;
230 ktime_get_real_ts64(&ts);
231 if (put_user(ts.tv_sec, &tv->tv_sec) ||
232 put_user(ts.tv_nsec / 1000, &tv->tv_usec))
233 return -EFAULT;
235 if (tz) {
236 if (copy_to_user(tz, &sys_tz, sizeof(sys_tz)))
237 return -EFAULT;
240 return 0;
243 COMPAT_SYSCALL_DEFINE2(settimeofday, struct old_timeval32 __user *, tv,
244 struct timezone __user *, tz)
246 struct timespec64 new_ts;
247 struct timezone new_tz;
249 if (tv) {
250 if (get_user(new_ts.tv_sec, &tv->tv_sec) ||
251 get_user(new_ts.tv_nsec, &tv->tv_usec))
252 return -EFAULT;
254 if (new_ts.tv_nsec > USEC_PER_SEC || new_ts.tv_nsec < 0)
255 return -EINVAL;
257 new_ts.tv_nsec *= NSEC_PER_USEC;
259 if (tz) {
260 if (copy_from_user(&new_tz, tz, sizeof(*tz)))
261 return -EFAULT;
264 return do_sys_settimeofday64(tv ? &new_ts : NULL, tz ? &new_tz : NULL);
266 #endif
268 #ifdef CONFIG_64BIT
269 SYSCALL_DEFINE1(adjtimex, struct __kernel_timex __user *, txc_p)
271 struct __kernel_timex txc; /* Local copy of parameter */
272 int ret;
274 /* Copy the user data space into the kernel copy
275 * structure. But bear in mind that the structures
276 * may change
278 if (copy_from_user(&txc, txc_p, sizeof(struct __kernel_timex)))
279 return -EFAULT;
280 ret = do_adjtimex(&txc);
281 return copy_to_user(txc_p, &txc, sizeof(struct __kernel_timex)) ? -EFAULT : ret;
283 #endif
285 #ifdef CONFIG_COMPAT_32BIT_TIME
286 int get_old_timex32(struct __kernel_timex *txc, const struct old_timex32 __user *utp)
288 struct old_timex32 tx32;
290 memset(txc, 0, sizeof(struct __kernel_timex));
291 if (copy_from_user(&tx32, utp, sizeof(struct old_timex32)))
292 return -EFAULT;
294 txc->modes = tx32.modes;
295 txc->offset = tx32.offset;
296 txc->freq = tx32.freq;
297 txc->maxerror = tx32.maxerror;
298 txc->esterror = tx32.esterror;
299 txc->status = tx32.status;
300 txc->constant = tx32.constant;
301 txc->precision = tx32.precision;
302 txc->tolerance = tx32.tolerance;
303 txc->time.tv_sec = tx32.time.tv_sec;
304 txc->time.tv_usec = tx32.time.tv_usec;
305 txc->tick = tx32.tick;
306 txc->ppsfreq = tx32.ppsfreq;
307 txc->jitter = tx32.jitter;
308 txc->shift = tx32.shift;
309 txc->stabil = tx32.stabil;
310 txc->jitcnt = tx32.jitcnt;
311 txc->calcnt = tx32.calcnt;
312 txc->errcnt = tx32.errcnt;
313 txc->stbcnt = tx32.stbcnt;
315 return 0;
318 int put_old_timex32(struct old_timex32 __user *utp, const struct __kernel_timex *txc)
320 struct old_timex32 tx32;
322 memset(&tx32, 0, sizeof(struct old_timex32));
323 tx32.modes = txc->modes;
324 tx32.offset = txc->offset;
325 tx32.freq = txc->freq;
326 tx32.maxerror = txc->maxerror;
327 tx32.esterror = txc->esterror;
328 tx32.status = txc->status;
329 tx32.constant = txc->constant;
330 tx32.precision = txc->precision;
331 tx32.tolerance = txc->tolerance;
332 tx32.time.tv_sec = txc->time.tv_sec;
333 tx32.time.tv_usec = txc->time.tv_usec;
334 tx32.tick = txc->tick;
335 tx32.ppsfreq = txc->ppsfreq;
336 tx32.jitter = txc->jitter;
337 tx32.shift = txc->shift;
338 tx32.stabil = txc->stabil;
339 tx32.jitcnt = txc->jitcnt;
340 tx32.calcnt = txc->calcnt;
341 tx32.errcnt = txc->errcnt;
342 tx32.stbcnt = txc->stbcnt;
343 tx32.tai = txc->tai;
344 if (copy_to_user(utp, &tx32, sizeof(struct old_timex32)))
345 return -EFAULT;
346 return 0;
349 SYSCALL_DEFINE1(adjtimex_time32, struct old_timex32 __user *, utp)
351 struct __kernel_timex txc;
352 int err, ret;
354 err = get_old_timex32(&txc, utp);
355 if (err)
356 return err;
358 ret = do_adjtimex(&txc);
360 err = put_old_timex32(utp, &txc);
361 if (err)
362 return err;
364 return ret;
366 #endif
369 * Convert jiffies to milliseconds and back.
371 * Avoid unnecessary multiplications/divisions in the
372 * two most common HZ cases:
374 unsigned int jiffies_to_msecs(const unsigned long j)
376 #if HZ <= MSEC_PER_SEC && !(MSEC_PER_SEC % HZ)
377 return (MSEC_PER_SEC / HZ) * j;
378 #elif HZ > MSEC_PER_SEC && !(HZ % MSEC_PER_SEC)
379 return (j + (HZ / MSEC_PER_SEC) - 1)/(HZ / MSEC_PER_SEC);
380 #else
381 # if BITS_PER_LONG == 32
382 return (HZ_TO_MSEC_MUL32 * j + (1ULL << HZ_TO_MSEC_SHR32) - 1) >>
383 HZ_TO_MSEC_SHR32;
384 # else
385 return DIV_ROUND_UP(j * HZ_TO_MSEC_NUM, HZ_TO_MSEC_DEN);
386 # endif
387 #endif
389 EXPORT_SYMBOL(jiffies_to_msecs);
391 unsigned int jiffies_to_usecs(const unsigned long j)
394 * Hz usually doesn't go much further MSEC_PER_SEC.
395 * jiffies_to_usecs() and usecs_to_jiffies() depend on that.
397 BUILD_BUG_ON(HZ > USEC_PER_SEC);
399 #if !(USEC_PER_SEC % HZ)
400 return (USEC_PER_SEC / HZ) * j;
401 #else
402 # if BITS_PER_LONG == 32
403 return (HZ_TO_USEC_MUL32 * j) >> HZ_TO_USEC_SHR32;
404 # else
405 return (j * HZ_TO_USEC_NUM) / HZ_TO_USEC_DEN;
406 # endif
407 #endif
409 EXPORT_SYMBOL(jiffies_to_usecs);
412 * mktime64 - Converts date to seconds.
413 * Converts Gregorian date to seconds since 1970-01-01 00:00:00.
414 * Assumes input in normal date format, i.e. 1980-12-31 23:59:59
415 * => year=1980, mon=12, day=31, hour=23, min=59, sec=59.
417 * [For the Julian calendar (which was used in Russia before 1917,
418 * Britain & colonies before 1752, anywhere else before 1582,
419 * and is still in use by some communities) leave out the
420 * -year/100+year/400 terms, and add 10.]
422 * This algorithm was first published by Gauss (I think).
424 * A leap second can be indicated by calling this function with sec as
425 * 60 (allowable under ISO 8601). The leap second is treated the same
426 * as the following second since they don't exist in UNIX time.
428 * An encoding of midnight at the end of the day as 24:00:00 - ie. midnight
429 * tomorrow - (allowable under ISO 8601) is supported.
431 time64_t mktime64(const unsigned int year0, const unsigned int mon0,
432 const unsigned int day, const unsigned int hour,
433 const unsigned int min, const unsigned int sec)
435 unsigned int mon = mon0, year = year0;
437 /* 1..12 -> 11,12,1..10 */
438 if (0 >= (int) (mon -= 2)) {
439 mon += 12; /* Puts Feb last since it has leap day */
440 year -= 1;
443 return ((((time64_t)
444 (year/4 - year/100 + year/400 + 367*mon/12 + day) +
445 year*365 - 719499
446 )*24 + hour /* now have hours - midnight tomorrow handled here */
447 )*60 + min /* now have minutes */
448 )*60 + sec; /* finally seconds */
450 EXPORT_SYMBOL(mktime64);
453 * ns_to_timespec - Convert nanoseconds to timespec
454 * @nsec: the nanoseconds value to be converted
456 * Returns the timespec representation of the nsec parameter.
458 struct timespec ns_to_timespec(const s64 nsec)
460 struct timespec ts;
461 s32 rem;
463 if (!nsec)
464 return (struct timespec) {0, 0};
466 ts.tv_sec = div_s64_rem(nsec, NSEC_PER_SEC, &rem);
467 if (unlikely(rem < 0)) {
468 ts.tv_sec--;
469 rem += NSEC_PER_SEC;
471 ts.tv_nsec = rem;
473 return ts;
475 EXPORT_SYMBOL(ns_to_timespec);
478 * ns_to_timeval - Convert nanoseconds to timeval
479 * @nsec: the nanoseconds value to be converted
481 * Returns the timeval representation of the nsec parameter.
483 struct timeval ns_to_timeval(const s64 nsec)
485 struct timespec ts = ns_to_timespec(nsec);
486 struct timeval tv;
488 tv.tv_sec = ts.tv_sec;
489 tv.tv_usec = (suseconds_t) ts.tv_nsec / 1000;
491 return tv;
493 EXPORT_SYMBOL(ns_to_timeval);
495 struct __kernel_old_timeval ns_to_kernel_old_timeval(const s64 nsec)
497 struct timespec64 ts = ns_to_timespec64(nsec);
498 struct __kernel_old_timeval tv;
500 tv.tv_sec = ts.tv_sec;
501 tv.tv_usec = (suseconds_t)ts.tv_nsec / 1000;
503 return tv;
505 EXPORT_SYMBOL(ns_to_kernel_old_timeval);
508 * set_normalized_timespec - set timespec sec and nsec parts and normalize
510 * @ts: pointer to timespec variable to be set
511 * @sec: seconds to set
512 * @nsec: nanoseconds to set
514 * Set seconds and nanoseconds field of a timespec variable and
515 * normalize to the timespec storage format
517 * Note: The tv_nsec part is always in the range of
518 * 0 <= tv_nsec < NSEC_PER_SEC
519 * For negative values only the tv_sec field is negative !
521 void set_normalized_timespec64(struct timespec64 *ts, time64_t sec, s64 nsec)
523 while (nsec >= NSEC_PER_SEC) {
525 * The following asm() prevents the compiler from
526 * optimising this loop into a modulo operation. See
527 * also __iter_div_u64_rem() in include/linux/time.h
529 asm("" : "+rm"(nsec));
530 nsec -= NSEC_PER_SEC;
531 ++sec;
533 while (nsec < 0) {
534 asm("" : "+rm"(nsec));
535 nsec += NSEC_PER_SEC;
536 --sec;
538 ts->tv_sec = sec;
539 ts->tv_nsec = nsec;
541 EXPORT_SYMBOL(set_normalized_timespec64);
544 * ns_to_timespec64 - Convert nanoseconds to timespec64
545 * @nsec: the nanoseconds value to be converted
547 * Returns the timespec64 representation of the nsec parameter.
549 struct timespec64 ns_to_timespec64(const s64 nsec)
551 struct timespec64 ts = { 0, 0 };
552 s32 rem;
554 if (likely(nsec > 0)) {
555 ts.tv_sec = div_u64_rem(nsec, NSEC_PER_SEC, &rem);
556 ts.tv_nsec = rem;
557 } else if (nsec < 0) {
559 * With negative times, tv_sec points to the earlier
560 * second, and tv_nsec counts the nanoseconds since
561 * then, so tv_nsec is always a positive number.
563 ts.tv_sec = -div_u64_rem(-nsec - 1, NSEC_PER_SEC, &rem) - 1;
564 ts.tv_nsec = NSEC_PER_SEC - rem - 1;
567 return ts;
569 EXPORT_SYMBOL(ns_to_timespec64);
572 * msecs_to_jiffies: - convert milliseconds to jiffies
573 * @m: time in milliseconds
575 * conversion is done as follows:
577 * - negative values mean 'infinite timeout' (MAX_JIFFY_OFFSET)
579 * - 'too large' values [that would result in larger than
580 * MAX_JIFFY_OFFSET values] mean 'infinite timeout' too.
582 * - all other values are converted to jiffies by either multiplying
583 * the input value by a factor or dividing it with a factor and
584 * handling any 32-bit overflows.
585 * for the details see __msecs_to_jiffies()
587 * msecs_to_jiffies() checks for the passed in value being a constant
588 * via __builtin_constant_p() allowing gcc to eliminate most of the
589 * code, __msecs_to_jiffies() is called if the value passed does not
590 * allow constant folding and the actual conversion must be done at
591 * runtime.
592 * the _msecs_to_jiffies helpers are the HZ dependent conversion
593 * routines found in include/linux/jiffies.h
595 unsigned long __msecs_to_jiffies(const unsigned int m)
598 * Negative value, means infinite timeout:
600 if ((int)m < 0)
601 return MAX_JIFFY_OFFSET;
602 return _msecs_to_jiffies(m);
604 EXPORT_SYMBOL(__msecs_to_jiffies);
606 unsigned long __usecs_to_jiffies(const unsigned int u)
608 if (u > jiffies_to_usecs(MAX_JIFFY_OFFSET))
609 return MAX_JIFFY_OFFSET;
610 return _usecs_to_jiffies(u);
612 EXPORT_SYMBOL(__usecs_to_jiffies);
615 * The TICK_NSEC - 1 rounds up the value to the next resolution. Note
616 * that a remainder subtract here would not do the right thing as the
617 * resolution values don't fall on second boundries. I.e. the line:
618 * nsec -= nsec % TICK_NSEC; is NOT a correct resolution rounding.
619 * Note that due to the small error in the multiplier here, this
620 * rounding is incorrect for sufficiently large values of tv_nsec, but
621 * well formed timespecs should have tv_nsec < NSEC_PER_SEC, so we're
622 * OK.
624 * Rather, we just shift the bits off the right.
626 * The >> (NSEC_JIFFIE_SC - SEC_JIFFIE_SC) converts the scaled nsec
627 * value to a scaled second value.
629 static unsigned long
630 __timespec64_to_jiffies(u64 sec, long nsec)
632 nsec = nsec + TICK_NSEC - 1;
634 if (sec >= MAX_SEC_IN_JIFFIES){
635 sec = MAX_SEC_IN_JIFFIES;
636 nsec = 0;
638 return ((sec * SEC_CONVERSION) +
639 (((u64)nsec * NSEC_CONVERSION) >>
640 (NSEC_JIFFIE_SC - SEC_JIFFIE_SC))) >> SEC_JIFFIE_SC;
644 static unsigned long
645 __timespec_to_jiffies(unsigned long sec, long nsec)
647 return __timespec64_to_jiffies((u64)sec, nsec);
650 unsigned long
651 timespec64_to_jiffies(const struct timespec64 *value)
653 return __timespec64_to_jiffies(value->tv_sec, value->tv_nsec);
655 EXPORT_SYMBOL(timespec64_to_jiffies);
657 void
658 jiffies_to_timespec64(const unsigned long jiffies, struct timespec64 *value)
661 * Convert jiffies to nanoseconds and separate with
662 * one divide.
664 u32 rem;
665 value->tv_sec = div_u64_rem((u64)jiffies * TICK_NSEC,
666 NSEC_PER_SEC, &rem);
667 value->tv_nsec = rem;
669 EXPORT_SYMBOL(jiffies_to_timespec64);
672 * We could use a similar algorithm to timespec_to_jiffies (with a
673 * different multiplier for usec instead of nsec). But this has a
674 * problem with rounding: we can't exactly add TICK_NSEC - 1 to the
675 * usec value, since it's not necessarily integral.
677 * We could instead round in the intermediate scaled representation
678 * (i.e. in units of 1/2^(large scale) jiffies) but that's also
679 * perilous: the scaling introduces a small positive error, which
680 * combined with a division-rounding-upward (i.e. adding 2^(scale) - 1
681 * units to the intermediate before shifting) leads to accidental
682 * overflow and overestimates.
684 * At the cost of one additional multiplication by a constant, just
685 * use the timespec implementation.
687 unsigned long
688 timeval_to_jiffies(const struct timeval *value)
690 return __timespec_to_jiffies(value->tv_sec,
691 value->tv_usec * NSEC_PER_USEC);
693 EXPORT_SYMBOL(timeval_to_jiffies);
695 void jiffies_to_timeval(const unsigned long jiffies, struct timeval *value)
698 * Convert jiffies to nanoseconds and separate with
699 * one divide.
701 u32 rem;
703 value->tv_sec = div_u64_rem((u64)jiffies * TICK_NSEC,
704 NSEC_PER_SEC, &rem);
705 value->tv_usec = rem / NSEC_PER_USEC;
707 EXPORT_SYMBOL(jiffies_to_timeval);
710 * Convert jiffies/jiffies_64 to clock_t and back.
712 clock_t jiffies_to_clock_t(unsigned long x)
714 #if (TICK_NSEC % (NSEC_PER_SEC / USER_HZ)) == 0
715 # if HZ < USER_HZ
716 return x * (USER_HZ / HZ);
717 # else
718 return x / (HZ / USER_HZ);
719 # endif
720 #else
721 return div_u64((u64)x * TICK_NSEC, NSEC_PER_SEC / USER_HZ);
722 #endif
724 EXPORT_SYMBOL(jiffies_to_clock_t);
726 unsigned long clock_t_to_jiffies(unsigned long x)
728 #if (HZ % USER_HZ)==0
729 if (x >= ~0UL / (HZ / USER_HZ))
730 return ~0UL;
731 return x * (HZ / USER_HZ);
732 #else
733 /* Don't worry about loss of precision here .. */
734 if (x >= ~0UL / HZ * USER_HZ)
735 return ~0UL;
737 /* .. but do try to contain it here */
738 return div_u64((u64)x * HZ, USER_HZ);
739 #endif
741 EXPORT_SYMBOL(clock_t_to_jiffies);
743 u64 jiffies_64_to_clock_t(u64 x)
745 #if (TICK_NSEC % (NSEC_PER_SEC / USER_HZ)) == 0
746 # if HZ < USER_HZ
747 x = div_u64(x * USER_HZ, HZ);
748 # elif HZ > USER_HZ
749 x = div_u64(x, HZ / USER_HZ);
750 # else
751 /* Nothing to do */
752 # endif
753 #else
755 * There are better ways that don't overflow early,
756 * but even this doesn't overflow in hundreds of years
757 * in 64 bits, so..
759 x = div_u64(x * TICK_NSEC, (NSEC_PER_SEC / USER_HZ));
760 #endif
761 return x;
763 EXPORT_SYMBOL(jiffies_64_to_clock_t);
765 u64 nsec_to_clock_t(u64 x)
767 #if (NSEC_PER_SEC % USER_HZ) == 0
768 return div_u64(x, NSEC_PER_SEC / USER_HZ);
769 #elif (USER_HZ % 512) == 0
770 return div_u64(x * USER_HZ / 512, NSEC_PER_SEC / 512);
771 #else
773 * max relative error 5.7e-8 (1.8s per year) for USER_HZ <= 1024,
774 * overflow after 64.99 years.
775 * exact for HZ=60, 72, 90, 120, 144, 180, 300, 600, 900, ...
777 return div_u64(x * 9, (9ull * NSEC_PER_SEC + (USER_HZ / 2)) / USER_HZ);
778 #endif
781 u64 jiffies64_to_nsecs(u64 j)
783 #if !(NSEC_PER_SEC % HZ)
784 return (NSEC_PER_SEC / HZ) * j;
785 # else
786 return div_u64(j * HZ_TO_NSEC_NUM, HZ_TO_NSEC_DEN);
787 #endif
789 EXPORT_SYMBOL(jiffies64_to_nsecs);
791 u64 jiffies64_to_msecs(const u64 j)
793 #if HZ <= MSEC_PER_SEC && !(MSEC_PER_SEC % HZ)
794 return (MSEC_PER_SEC / HZ) * j;
795 #else
796 return div_u64(j * HZ_TO_MSEC_NUM, HZ_TO_MSEC_DEN);
797 #endif
799 EXPORT_SYMBOL(jiffies64_to_msecs);
802 * nsecs_to_jiffies64 - Convert nsecs in u64 to jiffies64
804 * @n: nsecs in u64
806 * Unlike {m,u}secs_to_jiffies, type of input is not unsigned int but u64.
807 * And this doesn't return MAX_JIFFY_OFFSET since this function is designed
808 * for scheduler, not for use in device drivers to calculate timeout value.
810 * note:
811 * NSEC_PER_SEC = 10^9 = (5^9 * 2^9) = (1953125 * 512)
812 * ULLONG_MAX ns = 18446744073.709551615 secs = about 584 years
814 u64 nsecs_to_jiffies64(u64 n)
816 #if (NSEC_PER_SEC % HZ) == 0
817 /* Common case, HZ = 100, 128, 200, 250, 256, 500, 512, 1000 etc. */
818 return div_u64(n, NSEC_PER_SEC / HZ);
819 #elif (HZ % 512) == 0
820 /* overflow after 292 years if HZ = 1024 */
821 return div_u64(n * HZ / 512, NSEC_PER_SEC / 512);
822 #else
824 * Generic case - optimized for cases where HZ is a multiple of 3.
825 * overflow after 64.99 years, exact for HZ = 60, 72, 90, 120 etc.
827 return div_u64(n * 9, (9ull * NSEC_PER_SEC + HZ / 2) / HZ);
828 #endif
830 EXPORT_SYMBOL(nsecs_to_jiffies64);
833 * nsecs_to_jiffies - Convert nsecs in u64 to jiffies
835 * @n: nsecs in u64
837 * Unlike {m,u}secs_to_jiffies, type of input is not unsigned int but u64.
838 * And this doesn't return MAX_JIFFY_OFFSET since this function is designed
839 * for scheduler, not for use in device drivers to calculate timeout value.
841 * note:
842 * NSEC_PER_SEC = 10^9 = (5^9 * 2^9) = (1953125 * 512)
843 * ULLONG_MAX ns = 18446744073.709551615 secs = about 584 years
845 unsigned long nsecs_to_jiffies(u64 n)
847 return (unsigned long)nsecs_to_jiffies64(n);
849 EXPORT_SYMBOL_GPL(nsecs_to_jiffies);
852 * Add two timespec64 values and do a safety check for overflow.
853 * It's assumed that both values are valid (>= 0).
854 * And, each timespec64 is in normalized form.
856 struct timespec64 timespec64_add_safe(const struct timespec64 lhs,
857 const struct timespec64 rhs)
859 struct timespec64 res;
861 set_normalized_timespec64(&res, (timeu64_t) lhs.tv_sec + rhs.tv_sec,
862 lhs.tv_nsec + rhs.tv_nsec);
864 if (unlikely(res.tv_sec < lhs.tv_sec || res.tv_sec < rhs.tv_sec)) {
865 res.tv_sec = TIME64_MAX;
866 res.tv_nsec = 0;
869 return res;
872 int get_timespec64(struct timespec64 *ts,
873 const struct __kernel_timespec __user *uts)
875 struct __kernel_timespec kts;
876 int ret;
878 ret = copy_from_user(&kts, uts, sizeof(kts));
879 if (ret)
880 return -EFAULT;
882 ts->tv_sec = kts.tv_sec;
884 /* Zero out the padding in compat mode */
885 if (in_compat_syscall())
886 kts.tv_nsec &= 0xFFFFFFFFUL;
888 /* In 32-bit mode, this drops the padding */
889 ts->tv_nsec = kts.tv_nsec;
891 return 0;
893 EXPORT_SYMBOL_GPL(get_timespec64);
895 int put_timespec64(const struct timespec64 *ts,
896 struct __kernel_timespec __user *uts)
898 struct __kernel_timespec kts = {
899 .tv_sec = ts->tv_sec,
900 .tv_nsec = ts->tv_nsec
903 return copy_to_user(uts, &kts, sizeof(kts)) ? -EFAULT : 0;
905 EXPORT_SYMBOL_GPL(put_timespec64);
907 static int __get_old_timespec32(struct timespec64 *ts64,
908 const struct old_timespec32 __user *cts)
910 struct old_timespec32 ts;
911 int ret;
913 ret = copy_from_user(&ts, cts, sizeof(ts));
914 if (ret)
915 return -EFAULT;
917 ts64->tv_sec = ts.tv_sec;
918 ts64->tv_nsec = ts.tv_nsec;
920 return 0;
923 static int __put_old_timespec32(const struct timespec64 *ts64,
924 struct old_timespec32 __user *cts)
926 struct old_timespec32 ts = {
927 .tv_sec = ts64->tv_sec,
928 .tv_nsec = ts64->tv_nsec
930 return copy_to_user(cts, &ts, sizeof(ts)) ? -EFAULT : 0;
933 int get_old_timespec32(struct timespec64 *ts, const void __user *uts)
935 if (COMPAT_USE_64BIT_TIME)
936 return copy_from_user(ts, uts, sizeof(*ts)) ? -EFAULT : 0;
937 else
938 return __get_old_timespec32(ts, uts);
940 EXPORT_SYMBOL_GPL(get_old_timespec32);
942 int put_old_timespec32(const struct timespec64 *ts, void __user *uts)
944 if (COMPAT_USE_64BIT_TIME)
945 return copy_to_user(uts, ts, sizeof(*ts)) ? -EFAULT : 0;
946 else
947 return __put_old_timespec32(ts, uts);
949 EXPORT_SYMBOL_GPL(put_old_timespec32);
951 int get_itimerspec64(struct itimerspec64 *it,
952 const struct __kernel_itimerspec __user *uit)
954 int ret;
956 ret = get_timespec64(&it->it_interval, &uit->it_interval);
957 if (ret)
958 return ret;
960 ret = get_timespec64(&it->it_value, &uit->it_value);
962 return ret;
964 EXPORT_SYMBOL_GPL(get_itimerspec64);
966 int put_itimerspec64(const struct itimerspec64 *it,
967 struct __kernel_itimerspec __user *uit)
969 int ret;
971 ret = put_timespec64(&it->it_interval, &uit->it_interval);
972 if (ret)
973 return ret;
975 ret = put_timespec64(&it->it_value, &uit->it_value);
977 return ret;
979 EXPORT_SYMBOL_GPL(put_itimerspec64);
981 int get_old_itimerspec32(struct itimerspec64 *its,
982 const struct old_itimerspec32 __user *uits)
985 if (__get_old_timespec32(&its->it_interval, &uits->it_interval) ||
986 __get_old_timespec32(&its->it_value, &uits->it_value))
987 return -EFAULT;
988 return 0;
990 EXPORT_SYMBOL_GPL(get_old_itimerspec32);
992 int put_old_itimerspec32(const struct itimerspec64 *its,
993 struct old_itimerspec32 __user *uits)
995 if (__put_old_timespec32(&its->it_interval, &uits->it_interval) ||
996 __put_old_timespec32(&its->it_value, &uits->it_value))
997 return -EFAULT;
998 return 0;
1000 EXPORT_SYMBOL_GPL(put_old_itimerspec32);