1 /* TCC runtime library.
2 Parts of this code are (c) 2002 Fabrice Bellard
4 Copyright (C) 1987, 1988, 1992, 1994, 1995 Free Software Foundation, Inc.
6 This file is free software; you can redistribute it and/or modify it
7 under the terms of the GNU General Public License as published by the
8 Free Software Foundation; either version 2, or (at your option) any
11 In addition to the permissions in the GNU General Public License, the
12 Free Software Foundation gives you unlimited permission to link the
13 compiled version of this file into combinations with other programs,
14 and to distribute those combinations without any restriction coming
15 from the use of this file. (The General Public License restrictions
16 do apply in other respects; for example, they cover modification of
17 the file, and distribution when not linked into a combine
20 This file is distributed in the hope that it will be useful, but
21 WITHOUT ANY WARRANTY; without even the implied warranty of
22 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
23 General Public License for more details.
25 You should have received a copy of the GNU General Public License
26 along with this program; see the file COPYING. If not, write to
27 the Free Software Foundation, 59 Temple Place - Suite 330,
28 Boston, MA 02111-1307, USA.
33 #define W_TYPE_SIZE 32
34 #define BITS_PER_UNIT 8
37 typedef unsigned int UWtype
;
38 typedef unsigned int USItype
;
39 typedef long long DWtype
;
40 typedef unsigned long long UDWtype
;
52 typedef long double XFtype
;
53 #define WORD_SIZE (sizeof (Wtype) * BITS_PER_UNIT)
54 #define HIGH_WORD_COEFF (((UDWtype) 1) << WORD_SIZE)
56 /* the following deal with IEEE single-precision numbers */
58 #define SIGNBIT 0x80000000
59 #define HIDDEN (1 << 23)
60 #define SIGN(fp) ((fp) & SIGNBIT)
61 #define EXP(fp) (((fp) >> 23) & 0xFF)
62 #define MANT(fp) (((fp) & 0x7FFFFF) | HIDDEN)
63 #define PACK(s,e,m) ((s) | ((e) << 23) | (m))
65 /* the following deal with IEEE double-precision numbers */
67 #define HIDDEND (1 << 20)
68 #define EXPD(fp) (((fp.l.upper) >> 20) & 0x7FF)
69 #define SIGND(fp) ((fp.l.upper) & SIGNBIT)
70 #define MANTD(fp) (((((fp.l.upper) & 0xFFFFF) | HIDDEND) << 10) | \
72 #define HIDDEND_LL ((long long)1 << 52)
73 #define MANTD_LL(fp) ((fp.ll & (HIDDEND_LL-1)) | HIDDEND_LL)
74 #define PACKD_LL(s,e,m) (((long long)((s)+((e)<<20))<<32)|(m))
76 /* the following deal with x86 long double-precision numbers */
77 #define EXCESSLD 16382
78 #define EXPLD(fp) (fp.l.upper & 0x7fff)
79 #define SIGNLD(fp) ((fp.l.upper) & 0x8000)
85 unsigned long long lower
;
111 /* XXX: we don't support several builtin supports for now */
112 #if !defined(TCC_TARGET_X86_64) && !defined(TCC_TARGET_ARM)
114 /* XXX: use gcc/tcc intrinsic ? */
115 #if defined(TCC_TARGET_I386)
116 #define sub_ddmmss(sh, sl, ah, al, bh, bl) \
117 __asm__ ("subl %5,%1\n\tsbbl %3,%0" \
118 : "=r" ((USItype) (sh)), \
119 "=&r" ((USItype) (sl)) \
120 : "0" ((USItype) (ah)), \
121 "g" ((USItype) (bh)), \
122 "1" ((USItype) (al)), \
123 "g" ((USItype) (bl)))
124 #define umul_ppmm(w1, w0, u, v) \
126 : "=a" ((USItype) (w0)), \
127 "=d" ((USItype) (w1)) \
128 : "%0" ((USItype) (u)), \
129 "rm" ((USItype) (v)))
130 #define udiv_qrnnd(q, r, n1, n0, dv) \
132 : "=a" ((USItype) (q)), \
133 "=d" ((USItype) (r)) \
134 : "0" ((USItype) (n0)), \
135 "1" ((USItype) (n1)), \
136 "rm" ((USItype) (dv)))
137 #define count_leading_zeros(count, x) \
140 __asm__ ("bsrl %1,%0" \
141 : "=r" (__cbtmp) : "rm" ((USItype) (x))); \
142 (count) = __cbtmp ^ 31; \
145 #error unsupported CPU type
148 /* most of this code is taken from libgcc2.c from gcc */
150 static UDWtype
__udivmoddi4 (UDWtype n
, UDWtype d
, UDWtype
*rp
)
155 UWtype d0
, d1
, n0
, n1
, n2
;
167 #if !defined(UDIV_NEEDS_NORMALIZATION)
174 udiv_qrnnd (q0
, n0
, n1
, n0
, d0
);
177 /* Remainder in n0. */
184 d0
= 1 / d0
; /* Divide intentionally by zero. */
186 udiv_qrnnd (q1
, n1
, 0, n1
, d0
);
187 udiv_qrnnd (q0
, n0
, n1
, n0
, d0
);
189 /* Remainder in n0. */
200 #else /* UDIV_NEEDS_NORMALIZATION */
208 count_leading_zeros (bm
, d0
);
212 /* Normalize, i.e. make the most significant bit of the
216 n1
= (n1
<< bm
) | (n0
>> (W_TYPE_SIZE
- bm
));
220 udiv_qrnnd (q0
, n0
, n1
, n0
, d0
);
223 /* Remainder in n0 >> bm. */
230 d0
= 1 / d0
; /* Divide intentionally by zero. */
232 count_leading_zeros (bm
, d0
);
236 /* From (n1 >= d0) /\ (the most significant bit of d0 is set),
237 conclude (the most significant bit of n1 is set) /\ (the
238 leading quotient digit q1 = 1).
240 This special case is necessary, not an optimization.
241 (Shifts counts of W_TYPE_SIZE are undefined.) */
250 b
= W_TYPE_SIZE
- bm
;
254 n1
= (n1
<< bm
) | (n0
>> b
);
257 udiv_qrnnd (q1
, n1
, n2
, n1
, d0
);
262 udiv_qrnnd (q0
, n0
, n1
, n0
, d0
);
264 /* Remainder in n0 >> bm. */
274 #endif /* UDIV_NEEDS_NORMALIZATION */
285 /* Remainder in n1n0. */
297 count_leading_zeros (bm
, d1
);
300 /* From (n1 >= d1) /\ (the most significant bit of d1 is set),
301 conclude (the most significant bit of n1 is set) /\ (the
302 quotient digit q0 = 0 or 1).
304 This special case is necessary, not an optimization. */
306 /* The condition on the next line takes advantage of that
307 n1 >= d1 (true due to program flow). */
308 if (n1
> d1
|| n0
>= d0
)
311 sub_ddmmss (n1
, n0
, n1
, n0
, d1
, d0
);
330 b
= W_TYPE_SIZE
- bm
;
332 d1
= (d1
<< bm
) | (d0
>> b
);
335 n1
= (n1
<< bm
) | (n0
>> b
);
338 udiv_qrnnd (q0
, n1
, n2
, n1
, d1
);
339 umul_ppmm (m1
, m0
, q0
, d0
);
341 if (m1
> n1
|| (m1
== n1
&& m0
> n0
))
344 sub_ddmmss (m1
, m0
, m1
, m0
, d1
, d0
);
349 /* Remainder in (n1n0 - m1m0) >> bm. */
352 sub_ddmmss (n1
, n0
, n1
, n0
, m1
, m0
);
353 rr
.s
.low
= (n1
<< b
) | (n0
>> bm
);
354 rr
.s
.high
= n1
>> bm
;
366 #define __negdi2(a) (-(a))
368 long long __divdi3(long long u
, long long v
)
379 uu
.ll
= __negdi2 (uu
.ll
);
383 vv
.ll
= __negdi2 (vv
.ll
);
385 w
= __udivmoddi4 (uu
.ll
, vv
.ll
, (UDWtype
*) 0);
391 long long __moddi3(long long u
, long long v
)
402 uu
.ll
= __negdi2 (uu
.ll
);
405 vv
.ll
= __negdi2 (vv
.ll
);
407 __udivmoddi4 (uu
.ll
, vv
.ll
, (UDWtype
*) &w
);
413 unsigned long long __udivdi3(unsigned long long u
, unsigned long long v
)
415 return __udivmoddi4 (u
, v
, (UDWtype
*) 0);
418 unsigned long long __umoddi3(unsigned long long u
, unsigned long long v
)
422 __udivmoddi4 (u
, v
, &w
);
426 /* XXX: fix tcc's code generator to do this instead */
427 long long __ashrdi3(long long a
, int b
)
433 u
.s
.low
= u
.s
.high
>> (b
- 32);
434 u
.s
.high
= u
.s
.high
>> 31;
436 u
.s
.low
= ((unsigned)u
.s
.low
>> b
) | (u
.s
.high
<< (32 - b
));
437 u
.s
.high
= u
.s
.high
>> b
;
445 /* XXX: fix tcc's code generator to do this instead */
446 unsigned long long __lshrdi3(unsigned long long a
, int b
)
452 u
.s
.low
= (unsigned)u
.s
.high
>> (b
- 32);
455 u
.s
.low
= ((unsigned)u
.s
.low
>> b
) | (u
.s
.high
<< (32 - b
));
456 u
.s
.high
= (unsigned)u
.s
.high
>> b
;
464 /* XXX: fix tcc's code generator to do this instead */
465 long long __ashldi3(long long a
, int b
)
471 u
.s
.high
= (unsigned)u
.s
.low
<< (b
- 32);
474 u
.s
.high
= ((unsigned)u
.s
.high
<< b
) | ((unsigned)u
.s
.low
>> (32 - b
));
475 u
.s
.low
= (unsigned)u
.s
.low
<< b
;
483 #ifndef COMMIT_4ad186c5ef61_IS_FIXED
484 long long __tcc_cvt_ftol(long double x
)
488 __asm__
__volatile__ ("fnstcw %0" : "=m" (c0
));
490 __asm__
__volatile__ ("fldcw %0" : : "m" (c1
));
491 __asm__
__volatile__ ("fistpll %0" : "=m" (ret
));
492 __asm__
__volatile__ ("fldcw %0" : : "m" (c0
));
497 #endif /* !__x86_64__ */
499 /* XXX: fix tcc's code generator to do this instead */
500 float __floatundisf(unsigned long long a
)
506 if (uu
.s
.high
>= 0) {
510 r
+= 18446744073709551616.0;
515 double __floatundidf(unsigned long long a
)
521 if (uu
.s
.high
>= 0) {
522 return (double)uu
.ll
;
525 r
+= 18446744073709551616.0;
530 long double __floatundixf(unsigned long long a
)
536 if (uu
.s
.high
>= 0) {
537 return (long double)uu
.ll
;
540 r
+= 18446744073709551616.0;
541 return (long double)r
;
545 unsigned long long __fixunssfdi (float a1
)
547 register union float_long fl1
;
549 register unsigned long l
;
556 exp
= EXP (fl1
.l
) - EXCESS
- 24;
560 return (unsigned long long)-1;
562 return (unsigned long long)l
<< exp
;
569 unsigned long long __fixunsdfdi (double a1
)
571 register union double_long dl1
;
573 register unsigned long long l
;
580 exp
= EXPD (dl1
) - EXCESSD
- 53;
585 return (unsigned long long)-1;
594 unsigned long long __fixunsxfdi (long double a1
)
596 register union ldouble_long dl1
;
598 register unsigned long long l
;
602 if (dl1
.l
.lower
== 0 && dl1
.l
.upper
== 0)
605 exp
= EXPLD (dl1
) - EXCESSLD
- 64;
610 return (unsigned long long)-1;
617 long long __fixsfdi (float a1
)
619 long long ret
; int s
;
620 ret
= __fixunssfdi((s
= a1
>= 0) ? a1
: -a1
);
621 return s
? ret
: -ret
;
624 long long __fixdfdi (double a1
)
626 long long ret
; int s
;
627 ret
= __fixunsdfdi((s
= a1
>= 0) ? a1
: -a1
);
628 return s
? ret
: -ret
;
631 long long __fixxfdi (long double a1
)
633 long long ret
; int s
;
634 ret
= __fixunsxfdi((s
= a1
>= 0) ? a1
: -a1
);
635 return s
? ret
: -ret
;
638 #if defined(TCC_TARGET_X86_64) && !defined(_WIN64)
645 /* Avoid including stdlib.h because it is not easily available when
647 #include <stddef.h> /* size_t definition is needed for a x86_64-tcc to parse memset() */
648 extern void *malloc(unsigned long long);
649 extern void *memset(void *s
, int c
, size_t n
);
650 extern void free(void*);
651 extern void abort(void);
655 __va_gen_reg
, __va_float_reg
, __va_stack
658 //This should be in sync with the declaration on our include/stdarg.h
659 /* GCC compatible definition of va_list. */
661 unsigned int gp_offset
;
662 unsigned int fp_offset
;
664 unsigned int overflow_offset
;
665 char *overflow_arg_area
;
675 void __va_start(__va_list_struct
*ap
, void *fp
)
677 memset(ap
, 0, sizeof(__va_list_struct
));
678 *ap
= *(__va_list_struct
*)((char *)fp
- 16);
679 ap
->overflow_arg_area
= (char *)fp
+ ap
->overflow_offset
;
680 ap
->reg_save_area
= (char *)fp
- 176 - 16;
683 void *__va_arg(__va_list_struct
*ap
,
684 enum __va_arg_type arg_type
,
687 size
= (size
+ 7) & ~7;
688 align
= (align
+ 7) & ~7;
691 if (ap
->gp_offset
+ size
<= 48) {
692 ap
->gp_offset
+= size
;
693 return ap
->reg_save_area
+ ap
->gp_offset
- size
;
695 goto use_overflow_area
;
698 if (ap
->fp_offset
< 128 + 48) {
700 return ap
->reg_save_area
+ ap
->fp_offset
- 16;
703 goto use_overflow_area
;
707 ap
->overflow_arg_area
+= size
;
708 ap
->overflow_arg_area
= (char*)((intptr_t)(ap
->overflow_arg_area
+ align
- 1) & -(intptr_t)align
);
709 return ap
->overflow_arg_area
- size
;
713 fprintf(stderr
, "unknown ABI type for __va_arg\n");
719 #endif /* __x86_64__ */
721 /* Flushing for tccrun */
722 #if defined(TCC_TARGET_X86_64) || defined(TCC_TARGET_I386)
724 void __clear_cache(char *beginning
, char *end
)
728 #elif defined(TCC_TARGET_ARM)
732 #include <sys/syscall.h>
735 void __clear_cache(char *beginning
, char *end
)
737 /* __ARM_NR_cacheflush is kernel private and should not be used in user space.
738 * However, there is no ARM asm parser in tcc so we use it for now */
740 syscall(__ARM_NR_cacheflush
, beginning
, end
, 0);
742 __asm__ ("push {r7}\n\t"
743 "mov r7, #0xf0002\n\t"
752 #warning __clear_cache not defined for this architecture, avoid using tcc -run