2 * Copyright 1988, 1989 Hans-J. Boehm, Alan J. Demers
3 * Copyright (c) 1991-1995 by Xerox Corporation. All rights reserved.
4 * Copyright (c) 1996-1999 by Silicon Graphics. All rights reserved.
5 * Copyright (c) 1999 by Hewlett-Packard Company. All rights reserved.
7 * THIS MATERIAL IS PROVIDED AS IS, WITH ABSOLUTELY NO WARRANTY EXPRESSED
8 * OR IMPLIED. ANY USE IS AT YOUR OWN RISK.
10 * Permission is hereby granted to use or copy this program
11 * for any purpose, provided the above notices are retained on all copies.
12 * Permission to modify the code and to distribute modified code is granted,
13 * provided the above notices are retained, and a notice that the code was
14 * modified is included with the above copyright notice.
17 # include "private/gc_priv.h"
19 # if defined(LINUX) && !defined(POWERPC)
20 # include <linux/version.h>
21 # if (LINUX_VERSION_CODE <= 0x10400)
22 /* Ugly hack to get struct sigcontext_struct definition. Required */
23 /* for some early 1.3.X releases. Will hopefully go away soon. */
24 /* in some later Linux releases, asm/sigcontext.h may have to */
25 /* be included instead. */
27 # include <asm/signal.h>
30 /* Kernels prior to 2.1.1 defined struct sigcontext_struct instead of */
31 /* struct sigcontext. libc6 (glibc2) uses "struct sigcontext" in */
32 /* prototypes, so we have to include the top-level sigcontext.h to */
33 /* make sure the former gets defined to be the latter if appropriate. */
34 # include <features.h>
36 # if 2 == __GLIBC__ && 0 == __GLIBC_MINOR__
37 /* glibc 2.1 no longer has sigcontext.h. But signal.h */
38 /* has the right declaration for glibc 2.1. */
39 # include <sigcontext.h>
40 # endif /* 0 == __GLIBC_MINOR__ */
41 # else /* not 2 <= __GLIBC__ */
42 /* libc5 doesn't have <sigcontext.h>: go directly with the kernel */
43 /* one. Check LINUX_VERSION_CODE to see which we should reference. */
44 # include <asm/sigcontext.h>
45 # endif /* 2 <= __GLIBC__ */
48 # if !defined(OS2) && !defined(PCR) && !defined(AMIGA) && !defined(MACOS) \
50 # include <sys/types.h>
51 # if !defined(MSWIN32) && !defined(SUNOS4)
58 # define SIGSEGV 0 /* value is irrelevant */
63 #if defined(LINUX) || defined(LINUX_STACKBOTTOM)
67 /* Blatantly OS dependent routines, except for those that are related */
68 /* to dynamic loading. */
70 # if defined(HEURISTIC2) || defined(SEARCH_FOR_DATA_START)
71 # define NEED_FIND_LIMIT
74 # if !defined(STACKBOTTOM) && defined(HEURISTIC2)
75 # define NEED_FIND_LIMIT
78 # if (defined(SUNOS4) && defined(DYNAMIC_LOADING)) && !defined(PCR)
79 # define NEED_FIND_LIMIT
82 # if (defined(SVR4) || defined(AUX) || defined(DGUX) \
83 || (defined(LINUX) && defined(SPARC))) && !defined(PCR)
84 # define NEED_FIND_LIMIT
87 #if defined(FREEBSD) && (defined(I386) || defined(X86_64) || defined(powerpc) || defined(__powerpc__))
88 # include <machine/trap.h>
90 # define NEED_FIND_LIMIT
94 #if (defined(NETBSD) || defined(OPENBSD)) && defined(__ELF__) \
95 && !defined(NEED_FIND_LIMIT)
96 /* Used by GC_init_netbsd_elf() below. */
97 # define NEED_FIND_LIMIT
100 #ifdef NEED_FIND_LIMIT
105 # define GC_AMIGA_DEF
106 # include "AmigaOS.c"
110 #if defined(MSWIN32) || defined(MSWINCE) || defined(CYGWIN32)
111 # define WIN32_LEAN_AND_MEAN
113 # include <windows.h>
117 # include <Processes.h>
121 # include <sys/uio.h>
122 # include <malloc.h> /* for locking */
124 #if defined(USE_MMAP) || defined(USE_MUNMAP)
126 --> USE_MUNMAP requires USE_MMAP
128 # include <sys/types.h>
129 # include <sys/mman.h>
130 # include <sys/stat.h>
136 # if defined(SUNOS5SIGS) && !defined(FREEBSD)
137 # include <sys/siginfo.h>
139 /* Define SETJMP and friends to be the version that restores */
140 /* the signal mask. */
141 # define SETJMP(env) sigsetjmp(env, 1)
142 # define LONGJMP(env, val) siglongjmp(env, val)
143 # define JMP_BUF sigjmp_buf
145 # define SETJMP(env) setjmp(env)
146 # define LONGJMP(env, val) longjmp(env, val)
147 # define JMP_BUF jmp_buf
151 /* for get_etext and friends */
152 #include <mach-o/getsect.h>
156 /* Apparently necessary for djgpp 2.01. May cause problems with */
157 /* other versions. */
158 typedef long unsigned int caddr_t
;
162 # include "il/PCR_IL.h"
163 # include "th/PCR_ThCtl.h"
164 # include "mm/PCR_MM.h"
167 #if !defined(NO_EXECUTE_PERMISSION)
168 # define OPT_PROT_EXEC PROT_EXEC
170 # define OPT_PROT_EXEC 0
173 #if defined(LINUX) && \
174 (defined(USE_PROC_FOR_LIBRARIES) || defined(IA64) || !defined(SMALL_CONFIG))
176 /* We need to parse /proc/self/maps, either to find dynamic libraries, */
177 /* and/or to find the register backing store base (IA64). Do it once */
182 /* Repeatedly perform a read call until the buffer is filled or */
183 /* we encounter EOF. */
184 ssize_t
GC_repeat_read(int fd
, char *buf
, size_t count
)
186 ssize_t num_read
= 0;
189 while (num_read
< count
) {
190 result
= READ(fd
, buf
+ num_read
, count
- num_read
);
191 if (result
< 0) return result
;
192 if (result
== 0) break;
199 * Apply fn to a buffer containing the contents of /proc/self/maps.
200 * Return the result of fn or, if we failed, 0.
201 * We currently do nothing to /proc/self/maps other than simply read
202 * it. This code could be simplified if we could determine its size
206 word
GC_apply_to_maps(word (*fn
)(char *))
210 size_t maps_size
= 4000; /* Initial guess. */
211 static char init_buf
[1];
212 static char *maps_buf
= init_buf
;
213 static size_t maps_buf_sz
= 1;
215 /* Read /proc/self/maps, growing maps_buf as necessary. */
216 /* Note that we may not allocate conventionally, and */
217 /* thus can't use stdio. */
219 if (maps_size
>= maps_buf_sz
) {
220 /* Grow only by powers of 2, since we leak "too small" buffers. */
221 while (maps_size
>= maps_buf_sz
) maps_buf_sz
*= 2;
222 maps_buf
= GC_scratch_alloc(maps_buf_sz
);
223 if (maps_buf
== 0) return 0;
225 f
= open("/proc/self/maps", O_RDONLY
);
226 if (-1 == f
) return 0;
229 result
= GC_repeat_read(f
, maps_buf
, maps_buf_sz
-1);
230 if (result
<= 0) return 0;
232 } while (result
== maps_buf_sz
-1);
234 } while (maps_size
>= maps_buf_sz
);
235 maps_buf
[maps_size
] = '\0';
237 /* Apply fn to result. */
241 #endif /* Need GC_apply_to_maps */
243 #if defined(LINUX) && (defined(USE_PROC_FOR_LIBRARIES) || defined(IA64))
245 // GC_parse_map_entry parses an entry from /proc/self/maps so we can
246 // locate all writable data segments that belong to shared libraries.
247 // The format of one of these entries and the fields we care about
249 // XXXXXXXX-XXXXXXXX r-xp 00000000 30:05 260537 name of mapping...\n
250 // ^^^^^^^^ ^^^^^^^^ ^^^^ ^^
251 // start end prot maj_dev
253 // Note that since about auguat 2003 kernels, the columns no longer have
254 // fixed offsets on 64-bit kernels. Hence we no longer rely on fixed offsets
255 // anywhere, which is safer anyway.
259 * Assign various fields of the first line in buf_ptr to *start, *end,
260 * *prot_buf and *maj_dev. Only *prot_buf may be set for unwritable maps.
262 char *GC_parse_map_entry(char *buf_ptr
, word
*start
, word
*end
,
263 char *prot_buf
, unsigned int *maj_dev
)
265 char *start_start
, *end_start
, *prot_start
, *maj_dev_start
;
269 if (buf_ptr
== NULL
|| *buf_ptr
== '\0') {
274 while (isspace(*p
)) ++p
;
276 GC_ASSERT(isxdigit(*start_start
));
277 *start
= strtoul(start_start
, &endp
, 16); p
= endp
;
282 GC_ASSERT(isxdigit(*end_start
));
283 *end
= strtoul(end_start
, &endp
, 16); p
= endp
;
284 GC_ASSERT(isspace(*p
));
286 while (isspace(*p
)) ++p
;
288 GC_ASSERT(*prot_start
== 'r' || *prot_start
== '-');
289 memcpy(prot_buf
, prot_start
, 4);
291 if (prot_buf
[1] == 'w') {/* we can skip the rest if it's not writable. */
292 /* Skip past protection field to offset field */
293 while (!isspace(*p
)) ++p
; while (isspace(*p
)) ++p
;
294 GC_ASSERT(isxdigit(*p
));
295 /* Skip past offset field, which we ignore */
296 while (!isspace(*p
)) ++p
; while (isspace(*p
)) ++p
;
298 GC_ASSERT(isxdigit(*maj_dev_start
));
299 *maj_dev
= strtoul(maj_dev_start
, NULL
, 16);
302 while (*p
&& *p
++ != '\n');
307 #endif /* Need to parse /proc/self/maps. */
309 #if defined(SEARCH_FOR_DATA_START)
310 /* The I386 case can be handled without a search. The Alpha case */
311 /* used to be handled differently as well, but the rules changed */
312 /* for recent Linux versions. This seems to be the easiest way to */
313 /* cover all versions. */
316 /* Some Linux distributions arrange to define __data_start. Some */
317 /* define data_start as a weak symbol. The latter is technically */
318 /* broken, since the user program may define data_start, in which */
319 /* case we lose. Nonetheless, we try both, prefering __data_start. */
320 /* We assume gcc-compatible pragmas. */
321 # pragma weak __data_start
322 extern int __data_start
[];
323 # pragma weak data_start
324 extern int data_start
[];
330 void GC_init_linux_data_start()
332 extern ptr_t
GC_find_limit();
335 /* Try the easy approaches first: */
336 if ((ptr_t
)__data_start
!= 0) {
337 GC_data_start
= (ptr_t
)(__data_start
);
340 if ((ptr_t
)data_start
!= 0) {
341 GC_data_start
= (ptr_t
)(data_start
);
345 GC_data_start
= GC_find_limit((ptr_t
)(_end
), FALSE
);
351 # ifndef ECOS_GC_MEMORY_SIZE
352 # define ECOS_GC_MEMORY_SIZE (448 * 1024)
353 # endif /* ECOS_GC_MEMORY_SIZE */
355 // setjmp() function, as described in ANSI para 7.6.1.1
357 #define SETJMP( __env__ ) hal_setjmp( __env__ )
359 // FIXME: This is a simple way of allocating memory which is
360 // compatible with ECOS early releases. Later releases use a more
361 // sophisticated means of allocating memory than this simple static
362 // allocator, but this method is at least bound to work.
363 static char memory
[ECOS_GC_MEMORY_SIZE
];
364 static char *brk
= memory
;
366 static void *tiny_sbrk(ptrdiff_t increment
)
372 if (brk
> memory
+ sizeof memory
)
380 #define sbrk tiny_sbrk
383 #if (defined(NETBSD) || defined(OPENBSD)) && defined(__ELF__)
386 void GC_init_netbsd_elf()
388 extern ptr_t
GC_find_limit();
389 extern char **environ
;
390 /* This may need to be environ, without the underscore, for */
392 GC_data_start
= GC_find_limit((ptr_t
)&environ
, FALSE
);
400 # if !defined(__IBMC__) && !defined(__WATCOMC__) /* e.g. EMX */
403 unsigned short magic_number
;
404 unsigned short padding
[29];
408 #define E_MAGIC(x) (x).magic_number
409 #define EMAGIC 0x5A4D
410 #define E_LFANEW(x) (x).new_exe_offset
413 unsigned char magic_number
[2];
414 unsigned char byte_order
;
415 unsigned char word_order
;
416 unsigned long exe_format_level
;
419 unsigned long padding1
[13];
420 unsigned long object_table_offset
;
421 unsigned long object_count
;
422 unsigned long padding2
[31];
425 #define E32_MAGIC1(x) (x).magic_number[0]
426 #define E32MAGIC1 'L'
427 #define E32_MAGIC2(x) (x).magic_number[1]
428 #define E32MAGIC2 'X'
429 #define E32_BORDER(x) (x).byte_order
431 #define E32_WORDER(x) (x).word_order
433 #define E32_CPU(x) (x).cpu
435 #define E32_OBJTAB(x) (x).object_table_offset
436 #define E32_OBJCNT(x) (x).object_count
442 unsigned long pagemap
;
443 unsigned long mapsize
;
444 unsigned long reserved
;
447 #define O32_FLAGS(x) (x).flags
448 #define OBJREAD 0x0001L
449 #define OBJWRITE 0x0002L
450 #define OBJINVALID 0x0080L
451 #define O32_SIZE(x) (x).size
452 #define O32_BASE(x) (x).base
454 # else /* IBM's compiler */
456 /* A kludge to get around what appears to be a header file bug */
458 # define WORD unsigned short
461 # define DWORD unsigned long
468 # endif /* __IBMC__ */
470 # define INCL_DOSEXCEPTIONS
471 # define INCL_DOSPROCESS
472 # define INCL_DOSERRORS
473 # define INCL_DOSMODULEMGR
474 # define INCL_DOSMEMMGR
478 /* Disable and enable signals during nontrivial allocations */
480 void GC_disable_signals(void)
484 DosEnterMustComplete(&nest
);
485 if (nest
!= 1) ABORT("nested GC_disable_signals");
488 void GC_enable_signals(void)
492 DosExitMustComplete(&nest
);
493 if (nest
!= 0) ABORT("GC_enable_signals");
499 # if !defined(PCR) && !defined(AMIGA) && !defined(MSWIN32) \
500 && !defined(MSWINCE) \
501 && !defined(MACOS) && !defined(DJGPP) && !defined(DOS4GW) \
502 && !defined(NOSYS) && !defined(ECOS)
504 # if defined(SIG_BLOCK)
505 /* Use POSIX/SYSV interface */
506 # define SIGSET_T sigset_t
507 # define SIG_DEL(set, signal) sigdelset(&(set), (signal))
508 # define SIG_FILL(set) sigfillset(&set)
509 # define SIGSETMASK(old, new) sigprocmask(SIG_SETMASK, &(new), &(old))
510 # elif defined(sigmask) && !defined(UTS4) && !defined(HURD)
511 /* Use the traditional BSD interface */
512 # define SIGSET_T int
513 # define SIG_DEL(set, signal) (set) &= ~(sigmask(signal))
514 # define SIG_FILL(set) (set) = 0x7fffffff
515 /* Setting the leading bit appears to provoke a bug in some */
516 /* longjmp implementations. Most systems appear not to have */
518 # define SIGSETMASK(old, new) (old) = sigsetmask(new)
520 # error undetectable signal API
523 static GC_bool mask_initialized
= FALSE
;
525 static SIGSET_T new_mask
;
527 static SIGSET_T old_mask
;
529 static SIGSET_T dummy
;
531 #if defined(PRINTSTATS) && !defined(THREADS)
532 # define CHECK_SIGNALS
533 int GC_sig_disabled
= 0;
536 void GC_disable_signals()
538 if (!mask_initialized
) {
541 SIG_DEL(new_mask
, SIGSEGV
);
542 SIG_DEL(new_mask
, SIGILL
);
543 SIG_DEL(new_mask
, SIGQUIT
);
545 SIG_DEL(new_mask
, SIGBUS
);
548 SIG_DEL(new_mask
, SIGIOT
);
551 SIG_DEL(new_mask
, SIGEMT
);
554 SIG_DEL(new_mask
, SIGTRAP
);
556 mask_initialized
= TRUE
;
558 # ifdef CHECK_SIGNALS
559 if (GC_sig_disabled
!= 0) ABORT("Nested disables");
562 SIGSETMASK(old_mask
,new_mask
);
565 void GC_enable_signals()
567 # ifdef CHECK_SIGNALS
568 if (GC_sig_disabled
!= 1) ABORT("Unmatched enable");
571 SIGSETMASK(dummy
,old_mask
);
578 /* Ivan Demakov: simplest way (to me) */
580 void GC_disable_signals() { }
581 void GC_enable_signals() { }
584 /* Find the page size */
587 # if defined(MSWIN32) || defined(MSWINCE) || defined (CYGWIN32)
588 void GC_setpagesize()
590 GetSystemInfo(&GC_sysinfo
);
591 GC_page_size
= GC_sysinfo
.dwPageSize
;
595 # if defined(MPROTECT_VDB) || defined(PROC_VDB) || defined(USE_MMAP) \
596 || defined(USE_MUNMAP)
597 void GC_setpagesize()
599 GC_page_size
= GETPAGESIZE();
602 /* It's acceptable to fake it. */
603 void GC_setpagesize()
605 GC_page_size
= HBLKSIZE
;
611 * Find the base of the stack.
612 * Used only in single-threaded environment.
613 * With threads, GC_mark_roots needs to know how to do this.
614 * Called with allocator lock held.
616 # if defined(MSWIN32) || defined(MSWINCE)
617 # define is_writable(prot) ((prot) == PAGE_READWRITE \
618 || (prot) == PAGE_WRITECOPY \
619 || (prot) == PAGE_EXECUTE_READWRITE \
620 || (prot) == PAGE_EXECUTE_WRITECOPY)
621 /* Return the number of bytes that are writable starting at p. */
622 /* The pointer p is assumed to be page aligned. */
623 /* If base is not 0, *base becomes the beginning of the */
624 /* allocation region containing p. */
625 word
GC_get_writable_length(ptr_t p
, ptr_t
*base
)
627 MEMORY_BASIC_INFORMATION buf
;
631 result
= VirtualQuery(p
, &buf
, sizeof(buf
));
632 if (result
!= sizeof(buf
)) ABORT("Weird VirtualQuery result");
633 if (base
!= 0) *base
= (ptr_t
)(buf
.AllocationBase
);
634 protect
= (buf
.Protect
& ~(PAGE_GUARD
| PAGE_NOCACHE
));
635 if (!is_writable(protect
)) {
638 if (buf
.State
!= MEM_COMMIT
) return(0);
639 return(buf
.RegionSize
);
642 ptr_t
GC_get_stack_base()
645 ptr_t sp
= (ptr_t
)(&dummy
);
646 ptr_t trunc_sp
= (ptr_t
)((word
)sp
& ~(GC_page_size
- 1));
647 word size
= GC_get_writable_length(trunc_sp
, 0);
649 return(trunc_sp
+ size
);
653 # endif /* MS Windows */
656 # include <kernel/OS.h>
657 ptr_t
GC_get_stack_base(){
659 get_thread_info(find_thread(NULL
),&th
);
667 ptr_t
GC_get_stack_base()
672 if (DosGetInfoBlocks(&ptib
, &ppib
) != NO_ERROR
) {
673 GC_err_printf0("DosGetInfoBlocks failed\n");
674 ABORT("DosGetInfoBlocks failed\n");
676 return((ptr_t
)(ptib
-> tib_pstacklimit
));
683 # include "AmigaOS.c"
687 # if defined(NEED_FIND_LIMIT) || defined(UNIX_LIKE)
690 typedef void (*handler
)(int);
692 typedef void (*handler
)();
695 # if defined(SUNOS5SIGS) || defined(IRIX5) || defined(OSF1) \
696 || defined(HURD) || defined(NETBSD)
697 static struct sigaction old_segv_act
;
698 # if defined(IRIX5) || defined(HPUX) \
699 || defined(HURD) || defined(NETBSD)
700 static struct sigaction old_bus_act
;
703 static handler old_segv_handler
, old_bus_handler
;
707 void GC_set_and_save_fault_handler(handler h
)
709 void GC_set_and_save_fault_handler(h
)
713 # if defined(SUNOS5SIGS) || defined(IRIX5) \
714 || defined(OSF1) || defined(HURD) || defined(NETBSD)
715 struct sigaction act
;
718 # if 0 /* Was necessary for Solaris 2.3 and very temporary */
720 act
.sa_flags
= SA_RESTART
| SA_NODEFER
;
722 act
.sa_flags
= SA_RESTART
;
725 (void) sigemptyset(&act
.sa_mask
);
726 # ifdef GC_IRIX_THREADS
727 /* Older versions have a bug related to retrieving and */
728 /* and setting a handler at the same time. */
729 (void) sigaction(SIGSEGV
, 0, &old_segv_act
);
730 (void) sigaction(SIGSEGV
, &act
, 0);
731 (void) sigaction(SIGBUS
, 0, &old_bus_act
);
732 (void) sigaction(SIGBUS
, &act
, 0);
734 (void) sigaction(SIGSEGV
, &act
, &old_segv_act
);
735 # if defined(IRIX5) \
736 || defined(HPUX) || defined(HURD) || defined(NETBSD)
737 /* Under Irix 5.x or HP/UX, we may get SIGBUS. */
738 /* Pthreads doesn't exist under Irix 5.x, so we */
739 /* don't have to worry in the threads case. */
740 (void) sigaction(SIGBUS
, &act
, &old_bus_act
);
742 # endif /* GC_IRIX_THREADS */
744 old_segv_handler
= signal(SIGSEGV
, h
);
746 old_bus_handler
= signal(SIGBUS
, h
);
750 # endif /* NEED_FIND_LIMIT || UNIX_LIKE */
752 # ifdef NEED_FIND_LIMIT
753 /* Some tools to implement HEURISTIC2 */
754 # define MIN_PAGE_SIZE 256 /* Smallest conceivable page size, bytes */
755 /* static */ JMP_BUF GC_jmp_buf
;
758 void GC_fault_handler(sig
)
761 LONGJMP(GC_jmp_buf
, 1);
764 void GC_setup_temporary_fault_handler()
766 GC_set_and_save_fault_handler(GC_fault_handler
);
769 void GC_reset_fault_handler()
771 # if defined(SUNOS5SIGS) || defined(IRIX5) \
772 || defined(OSF1) || defined(HURD) || defined(NETBSD)
773 (void) sigaction(SIGSEGV
, &old_segv_act
, 0);
774 # if defined(IRIX5) \
775 || defined(HPUX) || defined(HURD) || defined(NETBSD)
776 (void) sigaction(SIGBUS
, &old_bus_act
, 0);
779 (void) signal(SIGSEGV
, old_segv_handler
);
781 (void) signal(SIGBUS
, old_bus_handler
);
786 /* Return the first nonaddressible location > p (up) or */
787 /* the smallest location q s.t. [q,p) is addressable (!up). */
788 /* We assume that p (up) or p-1 (!up) is addressable. */
789 ptr_t
GC_find_limit(p
, up
)
793 static VOLATILE ptr_t result
;
794 /* Needs to be static, since otherwise it may not be */
795 /* preserved across the longjmp. Can safely be */
796 /* static since it's only called once, with the */
797 /* allocation lock held. */
800 GC_setup_temporary_fault_handler();
801 if (SETJMP(GC_jmp_buf
) == 0) {
802 result
= (ptr_t
)(((word
)(p
))
803 & ~(MIN_PAGE_SIZE
-1));
806 result
+= MIN_PAGE_SIZE
;
808 result
-= MIN_PAGE_SIZE
;
810 GC_noop1((word
)(*result
));
813 GC_reset_fault_handler();
815 result
+= MIN_PAGE_SIZE
;
821 #if defined(ECOS) || defined(NOSYS)
822 ptr_t
GC_get_stack_base()
828 #ifdef HPUX_STACKBOTTOM
830 #include <sys/param.h>
831 #include <sys/pstat.h>
833 ptr_t
GC_get_register_stack_base(void)
835 struct pst_vm_status vm_status
;
838 while (pstat_getprocvm(&vm_status
, sizeof(vm_status
), 0, i
++) == 1) {
839 if (vm_status
.pst_type
== PS_RSESTACK
) {
840 return (ptr_t
) vm_status
.pst_vaddr
;
844 /* old way to get the register stackbottom */
845 return (ptr_t
)(((word
)GC_stackbottom
- BACKING_STORE_DISPLACEMENT
- 1)
846 & ~(BACKING_STORE_ALIGNMENT
- 1));
849 #endif /* HPUX_STACK_BOTTOM */
851 #ifdef LINUX_STACKBOTTOM
853 #include <sys/types.h>
854 #include <sys/stat.h>
856 # define STAT_SKIP 27 /* Number of fields preceding startstack */
857 /* field in /proc/self/stat */
859 #ifdef USE_LIBC_PRIVATES
860 # pragma weak __libc_stack_end
861 extern ptr_t __libc_stack_end
;
865 /* Try to read the backing store base from /proc/self/maps. */
866 /* We look for the writable mapping with a 0 major device, */
867 /* which is as close to our frame as possible, but below it.*/
868 static word
backing_store_base_from_maps(char *maps
)
871 char *buf_ptr
= maps
;
873 unsigned int maj_dev
;
874 word current_best
= 0;
878 buf_ptr
= GC_parse_map_entry(buf_ptr
, &start
, &end
, prot_buf
, &maj_dev
);
879 if (buf_ptr
== NULL
) return current_best
;
880 if (prot_buf
[1] == 'w' && maj_dev
== 0) {
881 if (end
< (word
)(&dummy
) && start
> current_best
) current_best
= start
;
887 static word
backing_store_base_from_proc(void)
889 return GC_apply_to_maps(backing_store_base_from_maps
);
892 # ifdef USE_LIBC_PRIVATES
893 # pragma weak __libc_ia64_register_backing_store_base
894 extern ptr_t __libc_ia64_register_backing_store_base
;
897 ptr_t
GC_get_register_stack_base(void)
899 # ifdef USE_LIBC_PRIVATES
900 if (0 != &__libc_ia64_register_backing_store_base
901 && 0 != __libc_ia64_register_backing_store_base
) {
902 /* Glibc 2.2.4 has a bug such that for dynamically linked */
903 /* executables __libc_ia64_register_backing_store_base is */
904 /* defined but uninitialized during constructor calls. */
905 /* Hence we check for both nonzero address and value. */
906 return __libc_ia64_register_backing_store_base
;
909 word result
= backing_store_base_from_proc();
911 /* Use dumb heuristics. Works only for default configuration. */
912 result
= (word
)GC_stackbottom
- BACKING_STORE_DISPLACEMENT
;
913 result
+= BACKING_STORE_ALIGNMENT
- 1;
914 result
&= ~(BACKING_STORE_ALIGNMENT
- 1);
915 /* Verify that it's at least readable. If not, we goofed. */
916 GC_noop1(*(word
*)result
);
918 return (ptr_t
)result
;
922 ptr_t
GC_linux_stack_base(void)
924 /* We read the stack base value from /proc/self/stat. We do this */
925 /* using direct I/O system calls in order to avoid calling malloc */
926 /* in case REDIRECT_MALLOC is defined. */
927 # define STAT_BUF_SIZE 4096
928 # define STAT_READ read
929 /* Should probably call the real read, if read is wrapped. */
930 char stat_buf
[STAT_BUF_SIZE
];
934 size_t i
, buf_offset
= 0;
936 /* First try the easy way. This should work for glibc 2.2 */
937 /* This fails in a prelinked ("prelink" command) executable */
938 /* since the correct value of __libc_stack_end never */
939 /* becomes visible to us. The second test works around */
941 # ifdef USE_LIBC_PRIVATES
942 if (0 != &__libc_stack_end
&& 0 != __libc_stack_end
) {
944 /* Some versions of glibc set the address 16 bytes too */
945 /* low while the initialization code is running. */
946 if (((word
)__libc_stack_end
& 0xfff) + 0x10 < 0x1000) {
947 return __libc_stack_end
+ 0x10;
948 } /* Otherwise it's not safe to add 16 bytes and we fall */
949 /* back to using /proc. */
952 /* Older versions of glibc for 64-bit Sparc do not set
953 * this variable correctly, it gets set to either zero
956 if (__libc_stack_end
!= (ptr_t
) (unsigned long)0x1)
957 return __libc_stack_end
;
959 return __libc_stack_end
;
964 f
= open("/proc/self/stat", O_RDONLY
);
965 if (f
< 0 || STAT_READ(f
, stat_buf
, STAT_BUF_SIZE
) < 2 * STAT_SKIP
) {
966 ABORT("Couldn't read /proc/self/stat");
968 c
= stat_buf
[buf_offset
++];
969 /* Skip the required number of fields. This number is hopefully */
970 /* constant across all Linux implementations. */
971 for (i
= 0; i
< STAT_SKIP
; ++i
) {
972 while (isspace(c
)) c
= stat_buf
[buf_offset
++];
973 while (!isspace(c
)) c
= stat_buf
[buf_offset
++];
975 while (isspace(c
)) c
= stat_buf
[buf_offset
++];
979 c
= stat_buf
[buf_offset
++];
982 if (result
< 0x10000000) ABORT("Absurd stack bottom value");
983 return (ptr_t
)result
;
986 #endif /* LINUX_STACKBOTTOM */
988 #ifdef FREEBSD_STACKBOTTOM
990 /* This uses an undocumented sysctl call, but at least one expert */
991 /* believes it will stay. */
994 #include <sys/types.h>
995 #include <sys/sysctl.h>
997 ptr_t
GC_freebsd_stack_base(void)
999 int nm
[2] = {CTL_KERN
, KERN_USRSTACK
};
1001 size_t len
= sizeof(ptr_t
);
1002 int r
= sysctl(nm
, 2, &base
, &len
, NULL
, 0);
1004 if (r
) ABORT("Error getting stack base");
1009 #endif /* FREEBSD_STACKBOTTOM */
1011 #if !defined(BEOS) && !defined(AMIGA) && !defined(MSWIN32) \
1012 && !defined(MSWINCE) && !defined(OS2) && !defined(NOSYS) && !defined(ECOS)
1014 ptr_t
GC_get_stack_base()
1016 # if defined(HEURISTIC1) || defined(HEURISTIC2) || \
1017 defined(LINUX_STACKBOTTOM) || defined(FREEBSD_STACKBOTTOM)
1022 # define STACKBOTTOM_ALIGNMENT_M1 ((word)STACK_GRAN - 1)
1025 return(STACKBOTTOM
);
1028 # ifdef STACK_GROWS_DOWN
1029 result
= (ptr_t
)((((word
)(&dummy
))
1030 + STACKBOTTOM_ALIGNMENT_M1
)
1031 & ~STACKBOTTOM_ALIGNMENT_M1
);
1033 result
= (ptr_t
)(((word
)(&dummy
))
1034 & ~STACKBOTTOM_ALIGNMENT_M1
);
1036 # endif /* HEURISTIC1 */
1037 # ifdef LINUX_STACKBOTTOM
1038 result
= GC_linux_stack_base();
1040 # ifdef FREEBSD_STACKBOTTOM
1041 result
= GC_freebsd_stack_base();
1044 # ifdef STACK_GROWS_DOWN
1045 result
= GC_find_limit((ptr_t
)(&dummy
), TRUE
);
1046 # ifdef HEURISTIC2_LIMIT
1047 if (result
> HEURISTIC2_LIMIT
1048 && (ptr_t
)(&dummy
) < HEURISTIC2_LIMIT
) {
1049 result
= HEURISTIC2_LIMIT
;
1053 result
= GC_find_limit((ptr_t
)(&dummy
), FALSE
);
1054 # ifdef HEURISTIC2_LIMIT
1055 if (result
< HEURISTIC2_LIMIT
1056 && (ptr_t
)(&dummy
) > HEURISTIC2_LIMIT
) {
1057 result
= HEURISTIC2_LIMIT
;
1062 # endif /* HEURISTIC2 */
1063 # ifdef STACK_GROWS_DOWN
1064 if (result
== 0) result
= (ptr_t
)(signed_word
)(-sizeof(ptr_t
));
1067 # endif /* STACKBOTTOM */
1070 # endif /* ! AMIGA, !OS 2, ! MS Windows, !BEOS, !NOSYS, !ECOS */
1073 * Register static data segment(s) as roots.
1074 * If more data segments are added later then they need to be registered
1075 * add that point (as we do with SunOS dynamic loading),
1076 * or GC_mark_roots needs to check for them (as we do with PCR).
1077 * Called with allocator lock held.
1082 void GC_register_data_segments()
1086 HMODULE module_handle
;
1087 # define PBUFSIZ 512
1088 UCHAR path
[PBUFSIZ
];
1090 struct exe_hdr hdrdos
; /* MSDOS header. */
1091 struct e32_exe hdr386
; /* Real header for my executable */
1092 struct o32_obj seg
; /* Currrent segment */
1096 if (DosGetInfoBlocks(&ptib
, &ppib
) != NO_ERROR
) {
1097 GC_err_printf0("DosGetInfoBlocks failed\n");
1098 ABORT("DosGetInfoBlocks failed\n");
1100 module_handle
= ppib
-> pib_hmte
;
1101 if (DosQueryModuleName(module_handle
, PBUFSIZ
, path
) != NO_ERROR
) {
1102 GC_err_printf0("DosQueryModuleName failed\n");
1103 ABORT("DosGetInfoBlocks failed\n");
1105 myexefile
= fopen(path
, "rb");
1106 if (myexefile
== 0) {
1107 GC_err_puts("Couldn't open executable ");
1108 GC_err_puts(path
); GC_err_puts("\n");
1109 ABORT("Failed to open executable\n");
1111 if (fread((char *)(&hdrdos
), 1, sizeof hdrdos
, myexefile
) < sizeof hdrdos
) {
1112 GC_err_puts("Couldn't read MSDOS header from ");
1113 GC_err_puts(path
); GC_err_puts("\n");
1114 ABORT("Couldn't read MSDOS header");
1116 if (E_MAGIC(hdrdos
) != EMAGIC
) {
1117 GC_err_puts("Executable has wrong DOS magic number: ");
1118 GC_err_puts(path
); GC_err_puts("\n");
1119 ABORT("Bad DOS magic number");
1121 if (fseek(myexefile
, E_LFANEW(hdrdos
), SEEK_SET
) != 0) {
1122 GC_err_puts("Seek to new header failed in ");
1123 GC_err_puts(path
); GC_err_puts("\n");
1124 ABORT("Bad DOS magic number");
1126 if (fread((char *)(&hdr386
), 1, sizeof hdr386
, myexefile
) < sizeof hdr386
) {
1127 GC_err_puts("Couldn't read MSDOS header from ");
1128 GC_err_puts(path
); GC_err_puts("\n");
1129 ABORT("Couldn't read OS/2 header");
1131 if (E32_MAGIC1(hdr386
) != E32MAGIC1
|| E32_MAGIC2(hdr386
) != E32MAGIC2
) {
1132 GC_err_puts("Executable has wrong OS/2 magic number:");
1133 GC_err_puts(path
); GC_err_puts("\n");
1134 ABORT("Bad OS/2 magic number");
1136 if ( E32_BORDER(hdr386
) != E32LEBO
|| E32_WORDER(hdr386
) != E32LEWO
) {
1137 GC_err_puts("Executable %s has wrong byte order: ");
1138 GC_err_puts(path
); GC_err_puts("\n");
1139 ABORT("Bad byte order");
1141 if ( E32_CPU(hdr386
) == E32CPU286
) {
1142 GC_err_puts("GC can't handle 80286 executables: ");
1143 GC_err_puts(path
); GC_err_puts("\n");
1146 if (fseek(myexefile
, E_LFANEW(hdrdos
) + E32_OBJTAB(hdr386
),
1148 GC_err_puts("Seek to object table failed: ");
1149 GC_err_puts(path
); GC_err_puts("\n");
1150 ABORT("Seek to object table failed");
1152 for (nsegs
= E32_OBJCNT(hdr386
); nsegs
> 0; nsegs
--) {
1154 if (fread((char *)(&seg
), 1, sizeof seg
, myexefile
) < sizeof seg
) {
1155 GC_err_puts("Couldn't read obj table entry from ");
1156 GC_err_puts(path
); GC_err_puts("\n");
1157 ABORT("Couldn't read obj table entry");
1159 flags
= O32_FLAGS(seg
);
1160 if (!(flags
& OBJWRITE
)) continue;
1161 if (!(flags
& OBJREAD
)) continue;
1162 if (flags
& OBJINVALID
) {
1163 GC_err_printf0("Object with invalid pages?\n");
1166 GC_add_roots_inner(O32_BASE(seg
), O32_BASE(seg
)+O32_SIZE(seg
), FALSE
);
1172 # if defined(MSWIN32) || defined(MSWINCE) || defined (CYGWIN32)
1175 # define GC_no_win32_dlls (FALSE)
1179 /* Unfortunately, we have to handle win32s very differently from NT, */
1180 /* Since VirtualQuery has very different semantics. In particular, */
1181 /* under win32s a VirtualQuery call on an unmapped page returns an */
1182 /* invalid result. Under NT, GC_register_data_segments is a noop and */
1183 /* all real work is done by GC_register_dynamic_libraries. Under */
1184 /* win32s, we cannot find the data segments associated with dll's. */
1185 /* We register the main data segment here. */
1186 GC_bool GC_no_win32_dlls
= FALSE
;
1187 /* This used to be set for gcc, to avoid dealing with */
1188 /* the structured exception handling issues. But we now have */
1189 /* assembly code to do that right. */
1190 GC_bool GC_wnt
= FALSE
;
1191 /* This is a Windows NT derivative, i.e. NT, W2K, XP or later. */
1193 void GC_init_win32()
1195 /* if we're running under win32s, assume that no DLLs will be loaded */
1196 DWORD v
= GetVersion();
1197 GC_wnt
= !(v
& 0x80000000);
1198 GC_no_win32_dlls
|= ((!GC_wnt
) && (v
& 0xff) <= 3);
1201 /* Return the smallest address a such that VirtualQuery */
1202 /* returns correct results for all addresses between a and start. */
1203 /* Assumes VirtualQuery returns correct information for start. */
1204 ptr_t
GC_least_described_address(ptr_t start
)
1206 MEMORY_BASIC_INFORMATION buf
;
1212 limit
= GC_sysinfo
.lpMinimumApplicationAddress
;
1213 p
= (ptr_t
)((word
)start
& ~(GC_page_size
- 1));
1215 q
= (LPVOID
)(p
- GC_page_size
);
1216 if ((ptr_t
)q
> (ptr_t
)p
/* underflow */ || q
< limit
) break;
1217 result
= VirtualQuery(q
, &buf
, sizeof(buf
));
1218 if (result
!= sizeof(buf
) || buf
.AllocationBase
== 0) break;
1219 p
= (ptr_t
)(buf
.AllocationBase
);
1225 # ifndef REDIRECT_MALLOC
1226 /* We maintain a linked list of AllocationBase values that we know */
1227 /* correspond to malloc heap sections. Currently this is only called */
1228 /* during a GC. But there is some hope that for long running */
1229 /* programs we will eventually see most heap sections. */
1231 /* In the long run, it would be more reliable to occasionally walk */
1232 /* the malloc heap with HeapWalk on the default heap. But that */
1233 /* apparently works only for NT-based Windows. */
1235 /* In the long run, a better data structure would also be nice ... */
1236 struct GC_malloc_heap_list
{
1237 void * allocation_base
;
1238 struct GC_malloc_heap_list
*next
;
1239 } *GC_malloc_heap_l
= 0;
1241 /* Is p the base of one of the malloc heap sections we already know */
1243 GC_bool
GC_is_malloc_heap_base(ptr_t p
)
1245 struct GC_malloc_heap_list
*q
= GC_malloc_heap_l
;
1248 if (q
-> allocation_base
== p
) return TRUE
;
1254 void *GC_get_allocation_base(void *p
)
1256 MEMORY_BASIC_INFORMATION buf
;
1257 DWORD result
= VirtualQuery(p
, &buf
, sizeof(buf
));
1258 if (result
!= sizeof(buf
)) {
1259 ABORT("Weird VirtualQuery result");
1261 return buf
.AllocationBase
;
1264 size_t GC_max_root_size
= 100000; /* Appr. largest root size. */
1266 void GC_add_current_malloc_heap()
1268 struct GC_malloc_heap_list
*new_l
=
1269 malloc(sizeof(struct GC_malloc_heap_list
));
1270 void * candidate
= GC_get_allocation_base(new_l
);
1272 if (new_l
== 0) return;
1273 if (GC_is_malloc_heap_base(candidate
)) {
1274 /* Try a little harder to find malloc heap. */
1275 size_t req_size
= 10000;
1277 void *p
= malloc(req_size
);
1278 if (0 == p
) { free(new_l
); return; }
1279 candidate
= GC_get_allocation_base(p
);
1282 } while (GC_is_malloc_heap_base(candidate
)
1283 && req_size
< GC_max_root_size
/10 && req_size
< 500000);
1284 if (GC_is_malloc_heap_base(candidate
)) {
1285 free(new_l
); return;
1290 GC_printf1("Found new system malloc AllocationBase at 0x%lx\n",
1293 new_l
-> allocation_base
= candidate
;
1294 new_l
-> next
= GC_malloc_heap_l
;
1295 GC_malloc_heap_l
= new_l
;
1297 # endif /* REDIRECT_MALLOC */
1299 /* Is p the start of either the malloc heap, or of one of our */
1300 /* heap sections? */
1301 GC_bool
GC_is_heap_base (ptr_t p
)
1306 # ifndef REDIRECT_MALLOC
1307 static word last_gc_no
= -1;
1309 if (last_gc_no
!= GC_gc_no
) {
1310 GC_add_current_malloc_heap();
1311 last_gc_no
= GC_gc_no
;
1313 if (GC_root_size
> GC_max_root_size
) GC_max_root_size
= GC_root_size
;
1314 if (GC_is_malloc_heap_base(p
)) return TRUE
;
1316 for (i
= 0; i
< GC_n_heap_bases
; i
++) {
1317 if (GC_heap_bases
[i
] == p
) return TRUE
;
1323 void GC_register_root_section(ptr_t static_root
)
1325 MEMORY_BASIC_INFORMATION buf
;
1330 char * limit
, * new_limit
;
1332 if (!GC_no_win32_dlls
) return;
1333 p
= base
= limit
= GC_least_described_address(static_root
);
1334 while (p
< GC_sysinfo
.lpMaximumApplicationAddress
) {
1335 result
= VirtualQuery(p
, &buf
, sizeof(buf
));
1336 if (result
!= sizeof(buf
) || buf
.AllocationBase
== 0
1337 || GC_is_heap_base(buf
.AllocationBase
)) break;
1338 new_limit
= (char *)p
+ buf
.RegionSize
;
1339 protect
= buf
.Protect
;
1340 if (buf
.State
== MEM_COMMIT
1341 && is_writable(protect
)) {
1342 if ((char *)p
== limit
) {
1345 if (base
!= limit
) GC_add_roots_inner(base
, limit
, FALSE
);
1350 if (p
> (LPVOID
)new_limit
/* overflow */) break;
1351 p
= (LPVOID
)new_limit
;
1353 if (base
!= limit
) GC_add_roots_inner(base
, limit
, FALSE
);
1357 void GC_register_data_segments()
1361 GC_register_root_section((ptr_t
)(&dummy
));
1365 # else /* !OS2 && !Windows */
1367 # if (defined(SVR4) || defined(AUX) || defined(DGUX) \
1368 || (defined(LINUX) && defined(SPARC))) && !defined(PCR)
1369 ptr_t
GC_SysVGetDataStart(max_page_size
, etext_addr
)
1373 word text_end
= ((word
)(etext_addr
) + sizeof(word
) - 1)
1374 & ~(sizeof(word
) - 1);
1375 /* etext rounded to word boundary */
1376 word next_page
= ((text_end
+ (word
)max_page_size
- 1)
1377 & ~((word
)max_page_size
- 1));
1378 word page_offset
= (text_end
& ((word
)max_page_size
- 1));
1379 VOLATILE
char * result
= (char *)(next_page
+ page_offset
);
1380 /* Note that this isnt equivalent to just adding */
1381 /* max_page_size to &etext if &etext is at a page boundary */
1383 GC_setup_temporary_fault_handler();
1384 if (SETJMP(GC_jmp_buf
) == 0) {
1385 /* Try writing to the address. */
1387 GC_reset_fault_handler();
1389 GC_reset_fault_handler();
1390 /* We got here via a longjmp. The address is not readable. */
1391 /* This is known to happen under Solaris 2.4 + gcc, which place */
1392 /* string constants in the text segment, but after etext. */
1393 /* Use plan B. Note that we now know there is a gap between */
1394 /* text and data segments, so plan A bought us something. */
1395 result
= (char *)GC_find_limit((ptr_t
)(DATAEND
), FALSE
);
1397 return((ptr_t
)result
);
1401 # if defined(FREEBSD) && (defined(I386) || defined(X86_64) || defined(powerpc) || defined(__powerpc__)) && !defined(PCR)
1402 /* Its unclear whether this should be identical to the above, or */
1403 /* whether it should apply to non-X86 architectures. */
1404 /* For now we don't assume that there is always an empty page after */
1405 /* etext. But in some cases there actually seems to be slightly more. */
1406 /* This also deals with holes between read-only data and writable data. */
1407 ptr_t
GC_FreeBSDGetDataStart(max_page_size
, etext_addr
)
1411 word text_end
= ((word
)(etext_addr
) + sizeof(word
) - 1)
1412 & ~(sizeof(word
) - 1);
1413 /* etext rounded to word boundary */
1414 VOLATILE word next_page
= (text_end
+ (word
)max_page_size
- 1)
1415 & ~((word
)max_page_size
- 1);
1416 VOLATILE ptr_t result
= (ptr_t
)text_end
;
1417 GC_setup_temporary_fault_handler();
1418 if (SETJMP(GC_jmp_buf
) == 0) {
1419 /* Try reading at the address. */
1420 /* This should happen before there is another thread. */
1421 for (; next_page
< (word
)(DATAEND
); next_page
+= (word
)max_page_size
)
1422 *(VOLATILE
char *)next_page
;
1423 GC_reset_fault_handler();
1425 GC_reset_fault_handler();
1426 /* As above, we go to plan B */
1427 result
= GC_find_limit((ptr_t
)(DATAEND
), FALSE
);
1437 # define GC_AMIGA_DS
1438 # include "AmigaOS.c"
1441 #else /* !OS2 && !Windows && !AMIGA */
1443 void GC_register_data_segments()
1445 # if !defined(PCR) && !defined(SRC_M3) && !defined(MACOS)
1446 # if defined(REDIRECT_MALLOC) && defined(GC_SOLARIS_THREADS)
1447 /* As of Solaris 2.3, the Solaris threads implementation */
1448 /* allocates the data structure for the initial thread with */
1449 /* sbrk at process startup. It needs to be scanned, so that */
1450 /* we don't lose some malloc allocated data structures */
1451 /* hanging from it. We're on thin ice here ... */
1452 extern caddr_t
sbrk();
1454 GC_add_roots_inner(DATASTART
, (char *)sbrk(0), FALSE
);
1456 GC_add_roots_inner(DATASTART
, (char *)(DATAEND
), FALSE
);
1457 # if defined(DATASTART2)
1458 GC_add_roots_inner(DATASTART2
, (char *)(DATAEND2
), FALSE
);
1464 # if defined(THINK_C)
1465 extern void* GC_MacGetDataStart(void);
1466 /* globals begin above stack and end at a5. */
1467 GC_add_roots_inner((ptr_t
)GC_MacGetDataStart(),
1468 (ptr_t
)LMGetCurrentA5(), FALSE
);
1470 # if defined(__MWERKS__)
1472 extern void* GC_MacGetDataStart(void);
1473 /* MATTHEW: Function to handle Far Globals (CW Pro 3) */
1474 # if __option(far_data)
1475 extern void* GC_MacGetDataEnd(void);
1477 /* globals begin above stack and end at a5. */
1478 GC_add_roots_inner((ptr_t
)GC_MacGetDataStart(),
1479 (ptr_t
)LMGetCurrentA5(), FALSE
);
1480 /* MATTHEW: Handle Far Globals */
1481 # if __option(far_data)
1482 /* Far globals follow he QD globals: */
1483 GC_add_roots_inner((ptr_t
)LMGetCurrentA5(),
1484 (ptr_t
)GC_MacGetDataEnd(), FALSE
);
1487 extern char __data_start__
[], __data_end__
[];
1488 GC_add_roots_inner((ptr_t
)&__data_start__
,
1489 (ptr_t
)&__data_end__
, FALSE
);
1490 # endif /* __POWERPC__ */
1491 # endif /* __MWERKS__ */
1492 # endif /* !THINK_C */
1496 /* Dynamic libraries are added at every collection, since they may */
1500 # endif /* ! AMIGA */
1501 # endif /* ! MSWIN32 && ! MSWINCE*/
1505 * Auxiliary routines for obtaining memory from OS.
1508 # if !defined(OS2) && !defined(PCR) && !defined(AMIGA) \
1509 && !defined(MSWIN32) && !defined(MSWINCE) \
1510 && !defined(MACOS) && !defined(DOS4GW)
1513 extern caddr_t
sbrk();
1516 # define SBRK_ARG_T ptrdiff_t
1518 # define SBRK_ARG_T int
1522 # if 0 && defined(RS6000) /* We now use mmap */
1523 /* The compiler seems to generate speculative reads one past the end of */
1524 /* an allocated object. Hence we need to make sure that the page */
1525 /* following the last heap page is also mapped. */
1526 ptr_t
GC_unix_get_mem(bytes
)
1529 caddr_t cur_brk
= (caddr_t
)sbrk(0);
1531 SBRK_ARG_T lsbs
= (word
)cur_brk
& (GC_page_size
-1);
1532 static caddr_t my_brk_val
= 0;
1534 if ((SBRK_ARG_T
)bytes
< 0) return(0); /* too big */
1536 if((caddr_t
)(sbrk(GC_page_size
- lsbs
)) == (caddr_t
)(-1)) return(0);
1538 if (cur_brk
== my_brk_val
) {
1539 /* Use the extra block we allocated last time. */
1540 result
= (ptr_t
)sbrk((SBRK_ARG_T
)bytes
);
1541 if (result
== (caddr_t
)(-1)) return(0);
1542 result
-= GC_page_size
;
1544 result
= (ptr_t
)sbrk(GC_page_size
+ (SBRK_ARG_T
)bytes
);
1545 if (result
== (caddr_t
)(-1)) return(0);
1547 my_brk_val
= result
+ bytes
+ GC_page_size
; /* Always page aligned */
1548 return((ptr_t
)result
);
1551 #else /* Not RS6000 */
1553 #if defined(USE_MMAP) || defined(USE_MUNMAP)
1555 #ifdef USE_MMAP_FIXED
1556 # define GC_MMAP_FLAGS MAP_FIXED | MAP_PRIVATE
1557 /* Seems to yield better performance on Solaris 2, but can */
1558 /* be unreliable if something is already mapped at the address. */
1560 # define GC_MMAP_FLAGS MAP_PRIVATE
1563 #ifdef USE_MMAP_ANON
1565 # if defined(MAP_ANONYMOUS)
1566 # define OPT_MAP_ANON MAP_ANONYMOUS
1568 # define OPT_MAP_ANON MAP_ANON
1572 # define OPT_MAP_ANON 0
1575 #endif /* defined(USE_MMAP) || defined(USE_MUNMAP) */
1577 #if defined(USE_MMAP)
1578 /* Tested only under Linux, IRIX5 and Solaris 2 */
1581 # define HEAP_START 0
1584 ptr_t
GC_unix_get_mem(bytes
)
1588 static ptr_t last_addr
= HEAP_START
;
1590 # ifndef USE_MMAP_ANON
1591 static GC_bool initialized
= FALSE
;
1594 zero_fd
= open("/dev/zero", O_RDONLY
);
1595 fcntl(zero_fd
, F_SETFD
, FD_CLOEXEC
);
1600 if (bytes
& (GC_page_size
-1)) ABORT("Bad GET_MEM arg");
1601 result
= mmap(last_addr
, bytes
, PROT_READ
| PROT_WRITE
| OPT_PROT_EXEC
,
1602 GC_MMAP_FLAGS
| OPT_MAP_ANON
, zero_fd
, 0/* offset */);
1603 if (result
== MAP_FAILED
) return(0);
1604 last_addr
= (ptr_t
)result
+ bytes
+ GC_page_size
- 1;
1605 last_addr
= (ptr_t
)((word
)last_addr
& ~(GC_page_size
- 1));
1606 # if !defined(LINUX)
1607 if (last_addr
== 0) {
1608 /* Oops. We got the end of the address space. This isn't */
1609 /* usable by arbitrary C code, since one-past-end pointers */
1610 /* don't work, so we discard it and try again. */
1611 munmap(result
, (size_t)(-GC_page_size
) - (size_t)result
);
1612 /* Leave last page mapped, so we can't repeat. */
1613 return GC_unix_get_mem(bytes
);
1616 GC_ASSERT(last_addr
!= 0);
1618 return((ptr_t
)result
);
1621 #else /* Not RS6000, not USE_MMAP */
1622 ptr_t
GC_unix_get_mem(bytes
)
1627 /* Bare sbrk isn't thread safe. Play by malloc rules. */
1628 /* The equivalent may be needed on other systems as well. */
1632 ptr_t cur_brk
= (ptr_t
)sbrk(0);
1633 SBRK_ARG_T lsbs
= (word
)cur_brk
& (GC_page_size
-1);
1635 if ((SBRK_ARG_T
)bytes
< 0) return(0); /* too big */
1637 if((ptr_t
)sbrk(GC_page_size
- lsbs
) == (ptr_t
)(-1)) return(0);
1639 result
= (ptr_t
)sbrk((SBRK_ARG_T
)bytes
);
1640 if (result
== (ptr_t
)(-1)) result
= 0;
1648 #endif /* Not USE_MMAP */
1649 #endif /* Not RS6000 */
1655 void * os2_alloc(size_t bytes
)
1659 if (DosAllocMem(&result
, bytes
, PAG_EXECUTE
| PAG_READ
|
1660 PAG_WRITE
| PAG_COMMIT
)
1664 if (result
== 0) return(os2_alloc(bytes
));
1671 # if defined(MSWIN32) || defined(MSWINCE) || defined(CYGWIN32)
1672 SYSTEM_INFO GC_sysinfo
;
1675 # if defined(MSWIN32) || defined(CYGWIN32)
1677 word GC_n_heap_bases
= 0;
1679 # ifdef USE_GLOBAL_ALLOC
1680 # define GLOBAL_ALLOC_TEST 1
1682 # define GLOBAL_ALLOC_TEST GC_no_win32_dlls
1685 ptr_t
GC_win32_get_mem(bytes
)
1691 result
= GC_unix_get_mem (bytes
);
1693 if (GLOBAL_ALLOC_TEST
) {
1694 /* VirtualAlloc doesn't like PAGE_EXECUTE_READWRITE. */
1695 /* There are also unconfirmed rumors of other */
1696 /* problems, so we dodge the issue. */
1697 result
= (ptr_t
) GlobalAlloc(0, bytes
+ HBLKSIZE
);
1698 result
= (ptr_t
)(((word
)result
+ HBLKSIZE
) & ~(HBLKSIZE
-1));
1700 /* VirtualProtect only works on regions returned by a */
1701 /* single VirtualAlloc call. Thus we allocate one */
1702 /* extra page, which will prevent merging of blocks */
1703 /* in separate regions, and eliminate any temptation */
1704 /* to call VirtualProtect on a range spanning regions. */
1705 /* This wastes a small amount of memory, and risks */
1706 /* increased fragmentation. But better alternatives */
1707 /* would require effort. */
1708 result
= (ptr_t
) VirtualAlloc(NULL
, bytes
+ 1,
1709 MEM_COMMIT
| MEM_RESERVE
,
1710 PAGE_EXECUTE_READWRITE
);
1713 if (HBLKDISPL(result
) != 0) ABORT("Bad VirtualAlloc result");
1714 /* If I read the documentation correctly, this can */
1715 /* only happen if HBLKSIZE > 64k or not a power of 2. */
1716 if (GC_n_heap_bases
>= MAX_HEAP_SECTS
) ABORT("Too many heap sections");
1717 GC_heap_bases
[GC_n_heap_bases
++] = result
;
1721 void GC_win32_free_heap ()
1723 if (GC_no_win32_dlls
) {
1724 while (GC_n_heap_bases
> 0) {
1726 free (GC_heap_bases
[--GC_n_heap_bases
]);
1728 GlobalFree (GC_heap_bases
[--GC_n_heap_bases
]);
1730 GC_heap_bases
[GC_n_heap_bases
] = 0;
1737 # define GC_AMIGA_AM
1738 # include "AmigaOS.c"
1744 word GC_n_heap_bases
= 0;
1746 ptr_t
GC_wince_get_mem(bytes
)
1752 /* Round up allocation size to multiple of page size */
1753 bytes
= (bytes
+ GC_page_size
-1) & ~(GC_page_size
-1);
1755 /* Try to find reserved, uncommitted pages */
1756 for (i
= 0; i
< GC_n_heap_bases
; i
++) {
1757 if (((word
)(-(signed_word
)GC_heap_lengths
[i
])
1758 & (GC_sysinfo
.dwAllocationGranularity
-1))
1760 result
= GC_heap_bases
[i
] + GC_heap_lengths
[i
];
1765 if (i
== GC_n_heap_bases
) {
1766 /* Reserve more pages */
1767 word res_bytes
= (bytes
+ GC_sysinfo
.dwAllocationGranularity
-1)
1768 & ~(GC_sysinfo
.dwAllocationGranularity
-1);
1769 /* If we ever support MPROTECT_VDB here, we will probably need to */
1770 /* ensure that res_bytes is strictly > bytes, so that VirtualProtect */
1771 /* never spans regions. It seems to be OK for a VirtualFree argument */
1772 /* to span regions, so we should be OK for now. */
1773 result
= (ptr_t
) VirtualAlloc(NULL
, res_bytes
,
1774 MEM_RESERVE
| MEM_TOP_DOWN
,
1775 PAGE_EXECUTE_READWRITE
);
1776 if (HBLKDISPL(result
) != 0) ABORT("Bad VirtualAlloc result");
1777 /* If I read the documentation correctly, this can */
1778 /* only happen if HBLKSIZE > 64k or not a power of 2. */
1779 if (GC_n_heap_bases
>= MAX_HEAP_SECTS
) ABORT("Too many heap sections");
1780 GC_heap_bases
[GC_n_heap_bases
] = result
;
1781 GC_heap_lengths
[GC_n_heap_bases
] = 0;
1786 result
= (ptr_t
) VirtualAlloc(result
, bytes
,
1788 PAGE_EXECUTE_READWRITE
);
1789 if (result
!= NULL
) {
1790 if (HBLKDISPL(result
) != 0) ABORT("Bad VirtualAlloc result");
1791 GC_heap_lengths
[i
] += bytes
;
1800 /* For now, this only works on Win32/WinCE and some Unix-like */
1801 /* systems. If you have something else, don't define */
1803 /* We assume ANSI C to support this feature. */
1805 #if !defined(MSWIN32) && !defined(MSWINCE)
1808 #include <sys/mman.h>
1809 #include <sys/stat.h>
1810 #include <sys/types.h>
1814 /* Compute a page aligned starting address for the unmap */
1815 /* operation on a block of size bytes starting at start. */
1816 /* Return 0 if the block is too small to make this feasible. */
1817 ptr_t
GC_unmap_start(ptr_t start
, word bytes
)
1819 ptr_t result
= start
;
1820 /* Round start to next page boundary. */
1821 result
+= GC_page_size
- 1;
1822 result
= (ptr_t
)((word
)result
& ~(GC_page_size
- 1));
1823 if (result
+ GC_page_size
> start
+ bytes
) return 0;
1827 /* Compute end address for an unmap operation on the indicated */
1829 ptr_t
GC_unmap_end(ptr_t start
, word bytes
)
1831 ptr_t end_addr
= start
+ bytes
;
1832 end_addr
= (ptr_t
)((word
)end_addr
& ~(GC_page_size
- 1));
1836 /* Under Win32/WinCE we commit (map) and decommit (unmap) */
1837 /* memory using VirtualAlloc and VirtualFree. These functions */
1838 /* work on individual allocations of virtual memory, made */
1839 /* previously using VirtualAlloc with the MEM_RESERVE flag. */
1840 /* The ranges we need to (de)commit may span several of these */
1841 /* allocations; therefore we use VirtualQuery to check */
1842 /* allocation lengths, and split up the range as necessary. */
1844 /* We assume that GC_remap is called on exactly the same range */
1845 /* as a previous call to GC_unmap. It is safe to consistently */
1846 /* round the endpoints in both places. */
1847 void GC_unmap(ptr_t start
, word bytes
)
1849 ptr_t start_addr
= GC_unmap_start(start
, bytes
);
1850 ptr_t end_addr
= GC_unmap_end(start
, bytes
);
1851 word len
= end_addr
- start_addr
;
1852 if (0 == start_addr
) return;
1853 # if defined(MSWIN32) || defined(MSWINCE)
1855 MEMORY_BASIC_INFORMATION mem_info
;
1857 if (VirtualQuery(start_addr
, &mem_info
, sizeof(mem_info
))
1858 != sizeof(mem_info
))
1859 ABORT("Weird VirtualQuery result");
1860 free_len
= (len
< mem_info
.RegionSize
) ? len
: mem_info
.RegionSize
;
1861 if (!VirtualFree(start_addr
, free_len
, MEM_DECOMMIT
))
1862 ABORT("VirtualFree failed");
1863 GC_unmapped_bytes
+= free_len
;
1864 start_addr
+= free_len
;
1868 /* We immediately remap it to prevent an intervening mmap from */
1869 /* accidentally grabbing the same address space. */
1872 result
= mmap(start_addr
, len
, PROT_NONE
,
1873 MAP_PRIVATE
| MAP_FIXED
| OPT_MAP_ANON
,
1874 zero_fd
, 0/* offset */);
1875 if (result
!= (void *)start_addr
) ABORT("mmap(...PROT_NONE...) failed");
1877 GC_unmapped_bytes
+= len
;
1882 void GC_remap(ptr_t start
, word bytes
)
1884 ptr_t start_addr
= GC_unmap_start(start
, bytes
);
1885 ptr_t end_addr
= GC_unmap_end(start
, bytes
);
1886 word len
= end_addr
- start_addr
;
1888 # if defined(MSWIN32) || defined(MSWINCE)
1891 if (0 == start_addr
) return;
1893 MEMORY_BASIC_INFORMATION mem_info
;
1895 if (VirtualQuery(start_addr
, &mem_info
, sizeof(mem_info
))
1896 != sizeof(mem_info
))
1897 ABORT("Weird VirtualQuery result");
1898 alloc_len
= (len
< mem_info
.RegionSize
) ? len
: mem_info
.RegionSize
;
1899 result
= VirtualAlloc(start_addr
, alloc_len
,
1901 PAGE_EXECUTE_READWRITE
);
1902 if (result
!= start_addr
) {
1903 ABORT("VirtualAlloc remapping failed");
1905 GC_unmapped_bytes
-= alloc_len
;
1906 start_addr
+= alloc_len
;
1910 /* It was already remapped with PROT_NONE. */
1913 if (0 == start_addr
) return;
1914 result
= mprotect(start_addr
, len
,
1915 PROT_READ
| PROT_WRITE
| OPT_PROT_EXEC
);
1918 "Mprotect failed at 0x%lx (length %ld) with errno %ld\n",
1919 start_addr
, len
, errno
);
1920 ABORT("Mprotect remapping failed");
1922 GC_unmapped_bytes
-= len
;
1926 /* Two adjacent blocks have already been unmapped and are about to */
1927 /* be merged. Unmap the whole block. This typically requires */
1928 /* that we unmap a small section in the middle that was not previously */
1929 /* unmapped due to alignment constraints. */
1930 void GC_unmap_gap(ptr_t start1
, word bytes1
, ptr_t start2
, word bytes2
)
1932 ptr_t start1_addr
= GC_unmap_start(start1
, bytes1
);
1933 ptr_t end1_addr
= GC_unmap_end(start1
, bytes1
);
1934 ptr_t start2_addr
= GC_unmap_start(start2
, bytes2
);
1935 ptr_t end2_addr
= GC_unmap_end(start2
, bytes2
);
1936 ptr_t start_addr
= end1_addr
;
1937 ptr_t end_addr
= start2_addr
;
1939 GC_ASSERT(start1
+ bytes1
== start2
);
1940 if (0 == start1_addr
) start_addr
= GC_unmap_start(start1
, bytes1
+ bytes2
);
1941 if (0 == start2_addr
) end_addr
= GC_unmap_end(start1
, bytes1
+ bytes2
);
1942 if (0 == start_addr
) return;
1943 len
= end_addr
- start_addr
;
1944 # if defined(MSWIN32) || defined(MSWINCE)
1946 MEMORY_BASIC_INFORMATION mem_info
;
1948 if (VirtualQuery(start_addr
, &mem_info
, sizeof(mem_info
))
1949 != sizeof(mem_info
))
1950 ABORT("Weird VirtualQuery result");
1951 free_len
= (len
< mem_info
.RegionSize
) ? len
: mem_info
.RegionSize
;
1952 if (!VirtualFree(start_addr
, free_len
, MEM_DECOMMIT
))
1953 ABORT("VirtualFree failed");
1954 GC_unmapped_bytes
+= free_len
;
1955 start_addr
+= free_len
;
1959 if (len
!= 0 && munmap(start_addr
, len
) != 0) ABORT("munmap failed");
1960 GC_unmapped_bytes
+= len
;
1964 #endif /* USE_MUNMAP */
1966 /* Routine for pushing any additional roots. In THREADS */
1967 /* environment, this is also responsible for marking from */
1968 /* thread stacks. */
1970 void (*GC_push_other_roots
)() = 0;
1974 PCR_ERes
GC_push_thread_stack(PCR_Th_T
*t
, PCR_Any dummy
)
1976 struct PCR_ThCtl_TInfoRep info
;
1979 info
.ti_stkLow
= info
.ti_stkHi
= 0;
1980 result
= PCR_ThCtl_GetInfo(t
, &info
);
1981 GC_push_all_stack((ptr_t
)(info
.ti_stkLow
), (ptr_t
)(info
.ti_stkHi
));
1985 /* Push the contents of an old object. We treat this as stack */
1986 /* data only becasue that makes it robust against mark stack */
1988 PCR_ERes
GC_push_old_obj(void *p
, size_t size
, PCR_Any data
)
1990 GC_push_all_stack((ptr_t
)p
, (ptr_t
)p
+ size
);
1991 return(PCR_ERes_okay
);
1995 void GC_default_push_other_roots
GC_PROTO((void))
1997 /* Traverse data allocated by previous memory managers. */
1999 extern struct PCR_MM_ProcsRep
* GC_old_allocator
;
2001 if ((*(GC_old_allocator
->mmp_enumerate
))(PCR_Bool_false
,
2004 ABORT("Old object enumeration failed");
2007 /* Traverse all thread stacks. */
2009 PCR_ThCtl_ApplyToAllOtherThreads(GC_push_thread_stack
,0))
2010 || PCR_ERes_IsErr(GC_push_thread_stack(PCR_Th_CurrThread(), 0))) {
2011 ABORT("Thread stack marking failed\n");
2019 # ifdef ALL_INTERIOR_POINTERS
2023 void GC_push_thread_structures
GC_PROTO((void))
2025 /* Not our responsibibility. */
2028 extern void ThreadF__ProcessStacks();
2030 void GC_push_thread_stack(start
, stop
)
2033 GC_push_all_stack((ptr_t
)start
, (ptr_t
)stop
+ sizeof(word
));
2036 /* Push routine with M3 specific calling convention. */
2037 GC_m3_push_root(dummy1
, p
, dummy2
, dummy3
)
2039 ptr_t dummy1
, dummy2
;
2044 GC_PUSH_ONE_STACK(q
, p
);
2047 /* M3 set equivalent to RTHeap.TracedRefTypes */
2048 typedef struct { int elts
[1]; } RefTypeSet
;
2049 RefTypeSet GC_TracedRefTypes
= {{0x1}};
2051 void GC_default_push_other_roots
GC_PROTO((void))
2053 /* Use the M3 provided routine for finding static roots. */
2054 /* This is a bit dubious, since it presumes no C roots. */
2055 /* We handle the collector roots explicitly in GC_push_roots */
2056 RTMain__GlobalMapProc(GC_m3_push_root
, 0, GC_TracedRefTypes
);
2057 if (GC_words_allocd
> 0) {
2058 ThreadF__ProcessStacks(GC_push_thread_stack
);
2060 /* Otherwise this isn't absolutely necessary, and we have */
2061 /* startup ordering problems. */
2064 # endif /* SRC_M3 */
2066 # if defined(GC_SOLARIS_THREADS) || defined(GC_PTHREADS) || \
2067 defined(GC_WIN32_THREADS)
2069 extern void GC_push_all_stacks();
2071 void GC_default_push_other_roots
GC_PROTO((void))
2073 GC_push_all_stacks();
2076 # endif /* GC_SOLARIS_THREADS || GC_PTHREADS */
2078 void (*GC_push_other_roots
) GC_PROTO((void)) = GC_default_push_other_roots
;
2080 #endif /* THREADS */
2083 * Routines for accessing dirty bits on virtual pages.
2084 * We plan to eventually implement four strategies for doing so:
2085 * DEFAULT_VDB: A simple dummy implementation that treats every page
2086 * as possibly dirty. This makes incremental collection
2087 * useless, but the implementation is still correct.
2088 * PCR_VDB: Use PPCRs virtual dirty bit facility.
2089 * PROC_VDB: Use the /proc facility for reading dirty bits. Only
2090 * works under some SVR4 variants. Even then, it may be
2091 * too slow to be entirely satisfactory. Requires reading
2092 * dirty bits for entire address space. Implementations tend
2093 * to assume that the client is a (slow) debugger.
2094 * MPROTECT_VDB:Protect pages and then catch the faults to keep track of
2095 * dirtied pages. The implementation (and implementability)
2096 * is highly system dependent. This usually fails when system
2097 * calls write to a protected page. We prevent the read system
2098 * call from doing so. It is the clients responsibility to
2099 * make sure that other system calls are similarly protected
2100 * or write only to the stack.
2102 GC_bool GC_dirty_maintained
= FALSE
;
2106 /* All of the following assume the allocation lock is held, and */
2107 /* signals are disabled. */
2109 /* The client asserts that unallocated pages in the heap are never */
2112 /* Initialize virtual dirty bit implementation. */
2113 void GC_dirty_init()
2116 GC_printf0("Initializing DEFAULT_VDB...\n");
2118 GC_dirty_maintained
= TRUE
;
2121 /* Retrieve system dirty bits for heap to a local buffer. */
2122 /* Restore the systems notion of which pages are dirty. */
2123 void GC_read_dirty()
2126 /* Is the HBLKSIZE sized page at h marked dirty in the local buffer? */
2127 /* If the actual page size is different, this returns TRUE if any */
2128 /* of the pages overlapping h are dirty. This routine may err on the */
2129 /* side of labelling pages as dirty (and this implementation does). */
2131 GC_bool
GC_page_was_dirty(h
)
2138 * The following two routines are typically less crucial. They matter
2139 * most with large dynamic libraries, or if we can't accurately identify
2140 * stacks, e.g. under Solaris 2.X. Otherwise the following default
2141 * versions are adequate.
2144 /* Could any valid GC heap pointer ever have been written to this page? */
2146 GC_bool
GC_page_was_ever_dirty(h
)
2152 /* Reset the n pages starting at h to "was never dirty" status. */
2153 void GC_is_fresh(h
, n
)
2160 /* I) hints that [h, h+nblocks) is about to be written. */
2161 /* II) guarantees that protection is removed. */
2162 /* (I) may speed up some dirty bit implementations. */
2163 /* (II) may be essential if we need to ensure that */
2164 /* pointer-free system call buffers in the heap are */
2165 /* not protected. */
2167 void GC_remove_protection(h
, nblocks
, is_ptrfree
)
2174 # endif /* DEFAULT_VDB */
2177 # ifdef MPROTECT_VDB
2180 * See DEFAULT_VDB for interface descriptions.
2184 * This implementation maintains dirty bits itself by catching write
2185 * faults and keeping track of them. We assume nobody else catches
2186 * SIGBUS or SIGSEGV. We assume no write faults occur in system calls.
2187 * This means that clients must ensure that system calls don't write
2188 * to the write-protected heap. Probably the best way to do this is to
2189 * ensure that system calls write at most to POINTERFREE objects in the
2190 * heap, and do even that only if we are on a platform on which those
2191 * are not protected. Another alternative is to wrap system calls
2192 * (see example for read below), but the current implementation holds
2193 * a lock across blocking calls, making it problematic for multithreaded
2195 * We assume the page size is a multiple of HBLKSIZE.
2196 * We prefer them to be the same. We avoid protecting POINTERFREE
2197 * objects only if they are the same.
2200 # if !defined(MSWIN32) && !defined(MSWINCE) && !defined(DARWIN)
2202 # include <sys/mman.h>
2203 # include <signal.h>
2204 # include <sys/syscall.h>
2206 # define PROTECT(addr, len) \
2207 if (mprotect((caddr_t)(addr), (size_t)(len), \
2208 PROT_READ | OPT_PROT_EXEC) < 0) { \
2209 ABORT("mprotect failed"); \
2211 # define UNPROTECT(addr, len) \
2212 if (mprotect((caddr_t)(addr), (size_t)(len), \
2213 PROT_WRITE | PROT_READ | OPT_PROT_EXEC ) < 0) { \
2214 ABORT("un-mprotect failed"); \
2220 /* Using vm_protect (mach syscall) over mprotect (BSD syscall) seems to
2221 decrease the likelihood of some of the problems described below. */
2222 #include <mach/vm_map.h>
2223 static mach_port_t GC_task_self
;
2224 #define PROTECT(addr,len) \
2225 if(vm_protect(GC_task_self,(vm_address_t)(addr),(vm_size_t)(len), \
2226 FALSE,VM_PROT_READ) != KERN_SUCCESS) { \
2227 ABORT("vm_portect failed"); \
2229 #define UNPROTECT(addr,len) \
2230 if(vm_protect(GC_task_self,(vm_address_t)(addr),(vm_size_t)(len), \
2231 FALSE,VM_PROT_READ|VM_PROT_WRITE) != KERN_SUCCESS) { \
2232 ABORT("vm_portect failed"); \
2237 # include <signal.h>
2240 static DWORD protect_junk
;
2241 # define PROTECT(addr, len) \
2242 if (!VirtualProtect((addr), (len), PAGE_EXECUTE_READ, \
2244 DWORD last_error = GetLastError(); \
2245 GC_printf1("Last error code: %lx\n", last_error); \
2246 ABORT("VirtualProtect failed"); \
2248 # define UNPROTECT(addr, len) \
2249 if (!VirtualProtect((addr), (len), PAGE_EXECUTE_READWRITE, \
2251 ABORT("un-VirtualProtect failed"); \
2253 # endif /* !DARWIN */
2254 # endif /* MSWIN32 || MSWINCE || DARWIN */
2256 #if defined(SUNOS4) || (defined(FREEBSD) && !defined(SUNOS5SIGS))
2257 typedef void (* SIG_PF
)();
2258 #endif /* SUNOS4 || (FREEBSD && !SUNOS5SIGS) */
2260 #if defined(SUNOS5SIGS) || defined(OSF1) || defined(LINUX) \
2263 typedef void (* SIG_PF
)(int);
2265 typedef void (* SIG_PF
)();
2267 #endif /* SUNOS5SIGS || OSF1 || LINUX || HURD */
2269 #if defined(MSWIN32)
2270 typedef LPTOP_LEVEL_EXCEPTION_FILTER SIG_PF
;
2272 # define SIG_DFL (LPTOP_LEVEL_EXCEPTION_FILTER) (-1)
2274 #if defined(MSWINCE)
2275 typedef LONG (WINAPI
*SIG_PF
)(struct _EXCEPTION_POINTERS
*);
2277 # define SIG_DFL (SIG_PF) (-1)
2280 #if defined(IRIX5) || defined(OSF1) || defined(HURD)
2281 typedef void (* REAL_SIG_PF
)(int, int, struct sigcontext
*);
2282 #endif /* IRIX5 || OSF1 || HURD */
2284 #if defined(SUNOS5SIGS)
2285 # if defined(HPUX) || defined(FREEBSD)
2286 # define SIGINFO_T siginfo_t
2288 # define SIGINFO_T struct siginfo
2291 typedef void (* REAL_SIG_PF
)(int, SIGINFO_T
*, void *);
2293 typedef void (* REAL_SIG_PF
)();
2295 #endif /* SUNOS5SIGS */
2298 # if __GLIBC__ > 2 || __GLIBC__ == 2 && __GLIBC_MINOR__ >= 2
2299 typedef struct sigcontext s_c
;
2300 # else /* glibc < 2.2 */
2301 # include <linux/version.h>
2302 # if (LINUX_VERSION_CODE >= 0x20100) && !defined(M68K) || defined(ALPHA) || defined(ARM32)
2303 typedef struct sigcontext s_c
;
2305 typedef struct sigcontext_struct s_c
;
2307 # endif /* glibc < 2.2 */
2308 # if defined(ALPHA) || defined(M68K)
2309 typedef void (* REAL_SIG_PF
)(int, int, s_c
*);
2311 # if defined(IA64) || defined(HP_PA) || defined(X86_64)
2312 typedef void (* REAL_SIG_PF
)(int, siginfo_t
*, s_c
*);
2314 /* According to SUSV3, the last argument should have type */
2315 /* void * or ucontext_t * */
2317 typedef void (* REAL_SIG_PF
)(int, s_c
);
2321 /* Retrieve fault address from sigcontext structure by decoding */
2323 char * get_fault_addr(s_c
*sc
) {
2327 instr
= *((unsigned *)(sc
->sc_pc
));
2328 faultaddr
= sc
->sc_regs
[(instr
>> 16) & 0x1f];
2329 faultaddr
+= (word
) (((int)instr
<< 16) >> 16);
2330 return (char *)faultaddr
;
2332 # endif /* !ALPHA */
2336 SIG_PF GC_old_bus_handler
;
2337 SIG_PF GC_old_segv_handler
; /* Also old MSWIN32 ACCESS_VIOLATION filter */
2338 #endif /* !DARWIN */
2340 #if defined(THREADS)
2341 /* We need to lock around the bitmap update in the write fault handler */
2342 /* in order to avoid the risk of losing a bit. We do this with a */
2343 /* test-and-set spin lock if we know how to do that. Otherwise we */
2344 /* check whether we are already in the handler and use the dumb but */
2345 /* safe fallback algorithm of setting all bits in the word. */
2346 /* Contention should be very rare, so we do the minimum to handle it */
2348 #ifdef GC_TEST_AND_SET_DEFINED
2349 static VOLATILE
unsigned int fault_handler_lock
= 0;
2350 void async_set_pht_entry_from_index(VOLATILE page_hash_table db
, int index
) {
2351 while (GC_test_and_set(&fault_handler_lock
)) {}
2352 /* Could also revert to set_pht_entry_from_index_safe if initial */
2353 /* GC_test_and_set fails. */
2354 set_pht_entry_from_index(db
, index
);
2355 GC_clear(&fault_handler_lock
);
2357 #else /* !GC_TEST_AND_SET_DEFINED */
2358 /* THIS IS INCORRECT! The dirty bit vector may be temporarily wrong, */
2359 /* just before we notice the conflict and correct it. We may end up */
2360 /* looking at it while it's wrong. But this requires contention */
2361 /* exactly when a GC is triggered, which seems far less likely to */
2362 /* fail than the old code, which had no reported failures. Thus we */
2363 /* leave it this way while we think of something better, or support */
2364 /* GC_test_and_set on the remaining platforms. */
2365 static VOLATILE word currently_updating
= 0;
2366 void async_set_pht_entry_from_index(VOLATILE page_hash_table db
, int index
) {
2367 unsigned int update_dummy
;
2368 currently_updating
= (word
)(&update_dummy
);
2369 set_pht_entry_from_index(db
, index
);
2370 /* If we get contention in the 10 or so instruction window here, */
2371 /* and we get stopped by a GC between the two updates, we lose! */
2372 if (currently_updating
!= (word
)(&update_dummy
)) {
2373 set_pht_entry_from_index_safe(db
, index
);
2374 /* We claim that if two threads concurrently try to update the */
2375 /* dirty bit vector, the first one to execute UPDATE_START */
2376 /* will see it changed when UPDATE_END is executed. (Note that */
2377 /* &update_dummy must differ in two distinct threads.) It */
2378 /* will then execute set_pht_entry_from_index_safe, thus */
2379 /* returning us to a safe state, though not soon enough. */
2382 #endif /* !GC_TEST_AND_SET_DEFINED */
2383 #else /* !THREADS */
2384 # define async_set_pht_entry_from_index(db, index) \
2385 set_pht_entry_from_index(db, index)
2386 #endif /* !THREADS */
2389 #if !defined(DARWIN)
2390 # if defined (SUNOS4) || (defined(FREEBSD) && !defined(SUNOS5SIGS))
2391 void GC_write_fault_handler(sig
, code
, scp
, addr
)
2393 struct sigcontext
*scp
;
2396 # define SIG_OK (sig == SIGSEGV || sig == SIGBUS)
2397 # define CODE_OK (FC_CODE(code) == FC_PROT \
2398 || (FC_CODE(code) == FC_OBJERR \
2399 && FC_ERRNO(code) == FC_PROT))
2402 # define SIG_OK (sig == SIGBUS)
2403 # define CODE_OK TRUE
2405 # endif /* SUNOS4 || (FREEBSD && !SUNOS5SIGS) */
2407 # if defined(IRIX5) || defined(OSF1) || defined(HURD)
2409 void GC_write_fault_handler(int sig
, int code
, struct sigcontext
*scp
)
2411 # define SIG_OK (sig == SIGSEGV)
2412 # define CODE_OK (code == 2 /* experimentally determined */)
2415 # define SIG_OK (sig == SIGSEGV)
2416 # define CODE_OK (code == EACCES)
2419 # define SIG_OK (sig == SIGBUS || sig == SIGSEGV)
2420 # define CODE_OK TRUE
2422 # endif /* IRIX5 || OSF1 || HURD */
2425 # if defined(ALPHA) || defined(M68K)
2426 void GC_write_fault_handler(int sig
, int code
, s_c
* sc
)
2428 # if defined(IA64) || defined(HP_PA) || defined(X86_64)
2429 void GC_write_fault_handler(int sig
, siginfo_t
* si
, s_c
* scp
)
2432 void GC_write_fault_handler(int sig
, int a2
, int a3
, int a4
, s_c sc
)
2434 void GC_write_fault_handler(int sig
, s_c sc
)
2438 # define SIG_OK (sig == SIGSEGV)
2439 # define CODE_OK TRUE
2440 /* Empirically c.trapno == 14, on IA32, but is that useful? */
2441 /* Should probably consider alignment issues on other */
2442 /* architectures. */
2445 # if defined(SUNOS5SIGS)
2447 void GC_write_fault_handler(int sig
, SIGINFO_T
*scp
, void * context
)
2449 void GC_write_fault_handler(sig
, scp
, context
)
2455 # define SIG_OK (sig == SIGSEGV || sig == SIGBUS)
2456 # define CODE_OK (scp -> si_code == SEGV_ACCERR) \
2457 || (scp -> si_code == BUS_ADRERR) \
2458 || (scp -> si_code == BUS_UNKNOWN) \
2459 || (scp -> si_code == SEGV_UNKNOWN) \
2460 || (scp -> si_code == BUS_OBJERR)
2463 # define SIG_OK (sig == SIGBUS)
2464 # define CODE_OK (scp -> si_code == BUS_PAGE_FAULT)
2466 # define SIG_OK (sig == SIGSEGV)
2467 # define CODE_OK (scp -> si_code == SEGV_ACCERR)
2470 # endif /* SUNOS5SIGS */
2472 # if defined(MSWIN32) || defined(MSWINCE)
2473 LONG WINAPI
GC_write_fault_handler(struct _EXCEPTION_POINTERS
*exc_info
)
2474 # define SIG_OK (exc_info -> ExceptionRecord -> ExceptionCode == \
2475 STATUS_ACCESS_VIOLATION)
2476 # define CODE_OK (exc_info -> ExceptionRecord -> ExceptionInformation[0] == 1)
2478 # endif /* MSWIN32 || MSWINCE */
2480 register unsigned i
;
2482 char *addr
= (char *) code
;
2485 char * addr
= (char *) (size_t) (scp
-> sc_badvaddr
);
2487 # if defined(OSF1) && defined(ALPHA)
2488 char * addr
= (char *) (scp
-> sc_traparg_a0
);
2491 char * addr
= (char *) (scp
-> si_addr
);
2495 char * addr
= (char *) (sc
.cr2
);
2500 struct sigcontext
*scp
= (struct sigcontext
*)(sc
);
2502 int format
= (scp
->sc_formatvec
>> 12) & 0xf;
2503 unsigned long *framedata
= (unsigned long *)(scp
+ 1);
2506 if (format
== 0xa || format
== 0xb) {
2509 } else if (format
== 7) {
2512 if (framedata
[1] & 0x08000000) {
2513 /* correct addr on misaligned access */
2514 ea
= (ea
+4095)&(~4095);
2516 } else if (format
== 4) {
2519 if (framedata
[1] & 0x08000000) {
2520 /* correct addr on misaligned access */
2521 ea
= (ea
+4095)&(~4095);
2527 char * addr
= get_fault_addr(sc
);
2529 # if defined(IA64) || defined(HP_PA) || defined(X86_64)
2530 char * addr
= si
-> si_addr
;
2531 /* I believe this is claimed to work on all platforms for */
2532 /* Linux 2.3.47 and later. Hopefully we don't have to */
2533 /* worry about earlier kernels on IA64. */
2535 # if defined(POWERPC)
2536 char * addr
= (char *) (sc
.regs
->dar
);
2539 char * addr
= (char *)sc
.fault_address
;
2542 char * addr
= (char *)sc
.regs
.csraddr
;
2544 --> architecture
not supported
2553 # if defined(MSWIN32) || defined(MSWINCE)
2554 char * addr
= (char *) (exc_info
-> ExceptionRecord
2555 -> ExceptionInformation
[1]);
2556 # define sig SIGSEGV
2559 if (SIG_OK
&& CODE_OK
) {
2560 register struct hblk
* h
=
2561 (struct hblk
*)((word
)addr
& ~(GC_page_size
-1));
2562 GC_bool in_allocd_block
;
2565 /* Address is only within the correct physical page. */
2566 in_allocd_block
= FALSE
;
2567 for (i
= 0; i
< divHBLKSZ(GC_page_size
); i
++) {
2568 if (HDR(h
+i
) != 0) {
2569 in_allocd_block
= TRUE
;
2573 in_allocd_block
= (HDR(addr
) != 0);
2575 if (!in_allocd_block
) {
2576 /* FIXME - We should make sure that we invoke the */
2577 /* old handler with the appropriate calling */
2578 /* sequence, which often depends on SA_SIGINFO. */
2580 /* Heap blocks now begin and end on page boundaries */
2583 if (sig
== SIGSEGV
) {
2584 old_handler
= GC_old_segv_handler
;
2586 old_handler
= GC_old_bus_handler
;
2588 if (old_handler
== SIG_DFL
) {
2589 # if !defined(MSWIN32) && !defined(MSWINCE)
2590 GC_err_printf1("Segfault at 0x%lx\n", addr
);
2591 ABORT("Unexpected bus error or segmentation fault");
2593 return(EXCEPTION_CONTINUE_SEARCH
);
2596 # if defined (SUNOS4) \
2597 || (defined(FREEBSD) && !defined(SUNOS5SIGS))
2598 (*old_handler
) (sig
, code
, scp
, addr
);
2601 # if defined (SUNOS5SIGS)
2603 * FIXME: For FreeBSD, this code should check if the
2604 * old signal handler used the traditional BSD style and
2605 * if so call it using that style.
2607 (*(REAL_SIG_PF
)old_handler
) (sig
, scp
, context
);
2610 # if defined (LINUX)
2611 # if defined(ALPHA) || defined(M68K)
2612 (*(REAL_SIG_PF
)old_handler
) (sig
, code
, sc
);
2614 # if defined(IA64) || defined(HP_PA) || defined(X86_64)
2615 (*(REAL_SIG_PF
)old_handler
) (sig
, si
, scp
);
2617 (*(REAL_SIG_PF
)old_handler
) (sig
, sc
);
2622 # if defined (IRIX5) || defined(OSF1) || defined(HURD)
2623 (*(REAL_SIG_PF
)old_handler
) (sig
, code
, scp
);
2627 return((*old_handler
)(exc_info
));
2631 UNPROTECT(h
, GC_page_size
);
2632 /* We need to make sure that no collection occurs between */
2633 /* the UNPROTECT and the setting of the dirty bit. Otherwise */
2634 /* a write by a third thread might go unnoticed. Reversing */
2635 /* the order is just as bad, since we would end up unprotecting */
2636 /* a page in a GC cycle during which it's not marked. */
2637 /* Currently we do this by disabling the thread stopping */
2638 /* signals while this handler is running. An alternative might */
2639 /* be to record the fact that we're about to unprotect, or */
2640 /* have just unprotected a page in the GC's thread structure, */
2641 /* and then to have the thread stopping code set the dirty */
2642 /* flag, if necessary. */
2643 for (i
= 0; i
< divHBLKSZ(GC_page_size
); i
++) {
2644 register int index
= PHT_HASH(h
+i
);
2646 async_set_pht_entry_from_index(GC_dirty_pages
, index
);
2649 /* These reset the signal handler each time by default. */
2650 signal(SIGSEGV
, (SIG_PF
) GC_write_fault_handler
);
2652 /* The write may not take place before dirty bits are read. */
2653 /* But then we'll fault again ... */
2654 # if defined(MSWIN32) || defined(MSWINCE)
2655 return(EXCEPTION_CONTINUE_EXECUTION
);
2660 #if defined(MSWIN32) || defined(MSWINCE)
2661 return EXCEPTION_CONTINUE_SEARCH
;
2663 GC_err_printf1("Segfault at 0x%lx\n", addr
);
2664 ABORT("Unexpected bus error or segmentation fault");
2667 #endif /* !DARWIN */
2670 * We hold the allocation lock. We expect block h to be written
2671 * shortly. Ensure that all pages containing any part of the n hblks
2672 * starting at h are no longer protected. If is_ptrfree is false,
2673 * also ensure that they will subsequently appear to be dirty.
2675 void GC_remove_protection(h
, nblocks
, is_ptrfree
)
2680 struct hblk
* h_trunc
; /* Truncated to page boundary */
2681 struct hblk
* h_end
; /* Page boundary following block end */
2682 struct hblk
* current
;
2683 GC_bool found_clean
;
2685 if (!GC_dirty_maintained
) return;
2686 h_trunc
= (struct hblk
*)((word
)h
& ~(GC_page_size
-1));
2687 h_end
= (struct hblk
*)(((word
)(h
+ nblocks
) + GC_page_size
-1)
2688 & ~(GC_page_size
-1));
2689 found_clean
= FALSE
;
2690 for (current
= h_trunc
; current
< h_end
; ++current
) {
2691 int index
= PHT_HASH(current
);
2693 if (!is_ptrfree
|| current
< h
|| current
>= h
+ nblocks
) {
2694 async_set_pht_entry_from_index(GC_dirty_pages
, index
);
2697 UNPROTECT(h_trunc
, (ptr_t
)h_end
- (ptr_t
)h_trunc
);
2700 #if !defined(DARWIN)
2701 void GC_dirty_init()
2703 # if defined(SUNOS5SIGS) || defined(IRIX5) || defined(LINUX) || \
2704 defined(OSF1) || defined(HURD)
2705 struct sigaction act
, oldact
;
2706 /* We should probably specify SA_SIGINFO for Linux, and handle */
2707 /* the different architectures more uniformly. */
2708 # if defined(IRIX5) || defined(LINUX) && !defined(X86_64) \
2709 || defined(OSF1) || defined(HURD)
2710 act
.sa_flags
= SA_RESTART
;
2711 act
.sa_handler
= (SIG_PF
)GC_write_fault_handler
;
2713 act
.sa_flags
= SA_RESTART
| SA_SIGINFO
;
2714 act
.sa_sigaction
= GC_write_fault_handler
;
2716 (void)sigemptyset(&act
.sa_mask
);
2718 /* Arrange to postpone SIG_SUSPEND while we're in a write fault */
2719 /* handler. This effectively makes the handler atomic w.r.t. */
2720 /* stopping the world for GC. */
2721 (void)sigaddset(&act
.sa_mask
, SIG_SUSPEND
);
2722 # endif /* SIG_SUSPEND */
2725 GC_printf0("Inititalizing mprotect virtual dirty bit implementation\n");
2727 GC_dirty_maintained
= TRUE
;
2728 if (GC_page_size
% HBLKSIZE
!= 0) {
2729 GC_err_printf0("Page size not multiple of HBLKSIZE\n");
2730 ABORT("Page size not multiple of HBLKSIZE");
2732 # if defined(SUNOS4) || (defined(FREEBSD) && !defined(SUNOS5SIGS))
2733 GC_old_bus_handler
= signal(SIGBUS
, GC_write_fault_handler
);
2734 if (GC_old_bus_handler
== SIG_IGN
) {
2735 GC_err_printf0("Previously ignored bus error!?");
2736 GC_old_bus_handler
= SIG_DFL
;
2738 if (GC_old_bus_handler
!= SIG_DFL
) {
2740 GC_err_printf0("Replaced other SIGBUS handler\n");
2744 # if defined(SUNOS4)
2745 GC_old_segv_handler
= signal(SIGSEGV
, (SIG_PF
)GC_write_fault_handler
);
2746 if (GC_old_segv_handler
== SIG_IGN
) {
2747 GC_err_printf0("Previously ignored segmentation violation!?");
2748 GC_old_segv_handler
= SIG_DFL
;
2750 if (GC_old_segv_handler
!= SIG_DFL
) {
2752 GC_err_printf0("Replaced other SIGSEGV handler\n");
2756 # if (defined(SUNOS5SIGS) && !defined(FREEBSD)) || defined(IRIX5) \
2757 || defined(LINUX) || defined(OSF1) || defined(HURD)
2758 /* SUNOS5SIGS includes HPUX */
2759 # if defined(GC_IRIX_THREADS)
2760 sigaction(SIGSEGV
, 0, &oldact
);
2761 sigaction(SIGSEGV
, &act
, 0);
2764 int res
= sigaction(SIGSEGV
, &act
, &oldact
);
2765 if (res
!= 0) ABORT("Sigaction failed");
2768 # if defined(_sigargs) || defined(HURD) || !defined(SA_SIGINFO)
2769 /* This is Irix 5.x, not 6.x. Irix 5.x does not have */
2771 GC_old_segv_handler
= oldact
.sa_handler
;
2772 # else /* Irix 6.x or SUNOS5SIGS or LINUX */
2773 if (oldact
.sa_flags
& SA_SIGINFO
) {
2774 GC_old_segv_handler
= (SIG_PF
)(oldact
.sa_sigaction
);
2776 GC_old_segv_handler
= oldact
.sa_handler
;
2779 if (GC_old_segv_handler
== SIG_IGN
) {
2780 GC_err_printf0("Previously ignored segmentation violation!?");
2781 GC_old_segv_handler
= SIG_DFL
;
2783 if (GC_old_segv_handler
!= SIG_DFL
) {
2785 GC_err_printf0("Replaced other SIGSEGV handler\n");
2788 # endif /* (SUNOS5SIGS && !FREEBSD) || IRIX5 || LINUX || OSF1 || HURD */
2789 # if defined(HPUX) || defined(LINUX) || defined(HURD) \
2790 || (defined(FREEBSD) && defined(SUNOS5SIGS))
2791 sigaction(SIGBUS
, &act
, &oldact
);
2792 GC_old_bus_handler
= oldact
.sa_handler
;
2793 if (GC_old_bus_handler
== SIG_IGN
) {
2794 GC_err_printf0("Previously ignored bus error!?");
2795 GC_old_bus_handler
= SIG_DFL
;
2797 if (GC_old_bus_handler
!= SIG_DFL
) {
2799 GC_err_printf0("Replaced other SIGBUS handler\n");
2802 # endif /* HPUX || LINUX || HURD || (FREEBSD && SUNOS5SIGS) */
2803 # if defined(MSWIN32)
2804 GC_old_segv_handler
= SetUnhandledExceptionFilter(GC_write_fault_handler
);
2805 if (GC_old_segv_handler
!= NULL
) {
2807 GC_err_printf0("Replaced other UnhandledExceptionFilter\n");
2810 GC_old_segv_handler
= SIG_DFL
;
2814 #endif /* !DARWIN */
2816 int GC_incremental_protection_needs()
2818 if (GC_page_size
== HBLKSIZE
) {
2819 return GC_PROTECTS_POINTER_HEAP
;
2821 return GC_PROTECTS_POINTER_HEAP
| GC_PROTECTS_PTRFREE_HEAP
;
2825 #define HAVE_INCREMENTAL_PROTECTION_NEEDS
2827 #define IS_PTRFREE(hhdr) ((hhdr)->hb_descr == 0)
2829 #define PAGE_ALIGNED(x) !((word)(x) & (GC_page_size - 1))
2830 void GC_protect_heap()
2834 struct hblk
* current
;
2835 struct hblk
* current_start
; /* Start of block to be protected. */
2836 struct hblk
* limit
;
2838 GC_bool protect_all
=
2839 (0 != (GC_incremental_protection_needs() & GC_PROTECTS_PTRFREE_HEAP
));
2840 for (i
= 0; i
< GC_n_heap_sects
; i
++) {
2841 start
= GC_heap_sects
[i
].hs_start
;
2842 len
= GC_heap_sects
[i
].hs_bytes
;
2844 PROTECT(start
, len
);
2846 GC_ASSERT(PAGE_ALIGNED(len
))
2847 GC_ASSERT(PAGE_ALIGNED(start
))
2848 current_start
= current
= (struct hblk
*)start
;
2849 limit
= (struct hblk
*)(start
+ len
);
2850 while (current
< limit
) {
2855 GC_ASSERT(PAGE_ALIGNED(current
));
2856 GET_HDR(current
, hhdr
);
2857 if (IS_FORWARDING_ADDR_OR_NIL(hhdr
)) {
2858 /* This can happen only if we're at the beginning of a */
2859 /* heap segment, and a block spans heap segments. */
2860 /* We will handle that block as part of the preceding */
2862 GC_ASSERT(current_start
== current
);
2863 current_start
= ++current
;
2866 if (HBLK_IS_FREE(hhdr
)) {
2867 GC_ASSERT(PAGE_ALIGNED(hhdr
-> hb_sz
));
2868 nhblks
= divHBLKSZ(hhdr
-> hb_sz
);
2869 is_ptrfree
= TRUE
; /* dirty on alloc */
2871 nhblks
= OBJ_SZ_TO_BLOCKS(hhdr
-> hb_sz
);
2872 is_ptrfree
= IS_PTRFREE(hhdr
);
2875 if (current_start
< current
) {
2876 PROTECT(current_start
, (ptr_t
)current
- (ptr_t
)current_start
);
2878 current_start
= (current
+= nhblks
);
2883 if (current_start
< current
) {
2884 PROTECT(current_start
, (ptr_t
)current
- (ptr_t
)current_start
);
2890 /* We assume that either the world is stopped or its OK to lose dirty */
2891 /* bits while this is happenning (as in GC_enable_incremental). */
2892 void GC_read_dirty()
2894 BCOPY((word
*)GC_dirty_pages
, GC_grungy_pages
,
2895 (sizeof GC_dirty_pages
));
2896 BZERO((word
*)GC_dirty_pages
, (sizeof GC_dirty_pages
));
2900 GC_bool
GC_page_was_dirty(h
)
2903 register word index
= PHT_HASH(h
);
2905 return(HDR(h
) == 0 || get_pht_entry_from_index(GC_grungy_pages
, index
));
2909 * Acquiring the allocation lock here is dangerous, since this
2910 * can be called from within GC_call_with_alloc_lock, and the cord
2911 * package does so. On systems that allow nested lock acquisition, this
2913 * On other systems, SET_LOCK_HOLDER and friends must be suitably defined.
2916 static GC_bool syscall_acquired_lock
= FALSE
; /* Protected by GC lock. */
2918 void GC_begin_syscall()
2920 if (!I_HOLD_LOCK()) {
2922 syscall_acquired_lock
= TRUE
;
2926 void GC_end_syscall()
2928 if (syscall_acquired_lock
) {
2929 syscall_acquired_lock
= FALSE
;
2934 void GC_unprotect_range(addr
, len
)
2938 struct hblk
* start_block
;
2939 struct hblk
* end_block
;
2940 register struct hblk
*h
;
2943 if (!GC_dirty_maintained
) return;
2944 obj_start
= GC_base(addr
);
2945 if (obj_start
== 0) return;
2946 if (GC_base(addr
+ len
- 1) != obj_start
) {
2947 ABORT("GC_unprotect_range(range bigger than object)");
2949 start_block
= (struct hblk
*)((word
)addr
& ~(GC_page_size
- 1));
2950 end_block
= (struct hblk
*)((word
)(addr
+ len
- 1) & ~(GC_page_size
- 1));
2951 end_block
+= GC_page_size
/HBLKSIZE
- 1;
2952 for (h
= start_block
; h
<= end_block
; h
++) {
2953 register word index
= PHT_HASH(h
);
2955 async_set_pht_entry_from_index(GC_dirty_pages
, index
);
2957 UNPROTECT(start_block
,
2958 ((ptr_t
)end_block
- (ptr_t
)start_block
) + HBLKSIZE
);
2963 /* We no longer wrap read by default, since that was causing too many */
2964 /* problems. It is preferred that the client instead avoids writing */
2965 /* to the write-protected heap with a system call. */
2966 /* This still serves as sample code if you do want to wrap system calls.*/
2968 #if !defined(MSWIN32) && !defined(MSWINCE) && !defined(GC_USE_LD_WRAP)
2969 /* Replacement for UNIX system call. */
2970 /* Other calls that write to the heap should be handled similarly. */
2971 /* Note that this doesn't work well for blocking reads: It will hold */
2972 /* the allocation lock for the entire duration of the call. Multithreaded */
2973 /* clients should really ensure that it won't block, either by setting */
2974 /* the descriptor nonblocking, or by calling select or poll first, to */
2975 /* make sure that input is available. */
2976 /* Another, preferred alternative is to ensure that system calls never */
2977 /* write to the protected heap (see above). */
2978 # if defined(__STDC__) && !defined(SUNOS4)
2979 # include <unistd.h>
2980 # include <sys/uio.h>
2981 ssize_t
read(int fd
, void *buf
, size_t nbyte
)
2984 int read(fd
, buf
, nbyte
)
2986 int GC_read(fd
, buf
, nbyte
)
2996 GC_unprotect_range(buf
, (word
)nbyte
);
2997 # if defined(IRIX5) || defined(GC_LINUX_THREADS)
2998 /* Indirect system call may not always be easily available. */
2999 /* We could call _read, but that would interfere with the */
3000 /* libpthread interception of read. */
3001 /* On Linux, we have to be careful with the linuxthreads */
3002 /* read interception. */
3007 iov
.iov_len
= nbyte
;
3008 result
= readv(fd
, &iov
, 1);
3012 result
= __read(fd
, buf
, nbyte
);
3014 /* The two zero args at the end of this list are because one
3015 IA-64 syscall() implementation actually requires six args
3016 to be passed, even though they aren't always used. */
3017 result
= syscall(SYS_read
, fd
, buf
, nbyte
, 0, 0);
3023 #endif /* !MSWIN32 && !MSWINCE && !GC_LINUX_THREADS */
3025 #if defined(GC_USE_LD_WRAP) && !defined(THREADS)
3026 /* We use the GNU ld call wrapping facility. */
3027 /* This requires that the linker be invoked with "--wrap read". */
3028 /* This can be done by passing -Wl,"--wrap read" to gcc. */
3029 /* I'm not sure that this actually wraps whatever version of read */
3030 /* is called by stdio. That code also mentions __read. */
3031 # include <unistd.h>
3032 ssize_t
__wrap_read(int fd
, void *buf
, size_t nbyte
)
3037 GC_unprotect_range(buf
, (word
)nbyte
);
3038 result
= __real_read(fd
, buf
, nbyte
);
3043 /* We should probably also do this for __read, or whatever stdio */
3044 /* actually calls. */
3050 GC_bool
GC_page_was_ever_dirty(h
)
3056 /* Reset the n pages starting at h to "was never dirty" status. */
3058 void GC_is_fresh(h
, n
)
3064 # endif /* MPROTECT_VDB */
3069 * See DEFAULT_VDB for interface descriptions.
3073 * This implementaion assumes a Solaris 2.X like /proc pseudo-file-system
3074 * from which we can read page modified bits. This facility is far from
3075 * optimal (e.g. we would like to get the info for only some of the
3076 * address space), but it avoids intercepting system calls.
3080 #include <sys/types.h>
3081 #include <sys/signal.h>
3082 #include <sys/fault.h>
3083 #include <sys/syscall.h>
3084 #include <sys/procfs.h>
3085 #include <sys/stat.h>
3087 #define INITIAL_BUF_SZ 16384
3088 word GC_proc_buf_size
= INITIAL_BUF_SZ
;
3091 #ifdef GC_SOLARIS_THREADS
3092 /* We don't have exact sp values for threads. So we count on */
3093 /* occasionally declaring stack pages to be fresh. Thus we */
3094 /* need a real implementation of GC_is_fresh. We can't clear */
3095 /* entries in GC_written_pages, since that would declare all */
3096 /* pages with the given hash address to be fresh. */
3097 # define MAX_FRESH_PAGES 8*1024 /* Must be power of 2 */
3098 struct hblk
** GC_fresh_pages
; /* A direct mapped cache. */
3099 /* Collisions are dropped. */
3101 # define FRESH_PAGE_SLOT(h) (divHBLKSZ((word)(h)) & (MAX_FRESH_PAGES-1))
3102 # define ADD_FRESH_PAGE(h) \
3103 GC_fresh_pages[FRESH_PAGE_SLOT(h)] = (h)
3104 # define PAGE_IS_FRESH(h) \
3105 (GC_fresh_pages[FRESH_PAGE_SLOT(h)] == (h) && (h) != 0)
3108 /* Add all pages in pht2 to pht1 */
3109 void GC_or_pages(pht1
, pht2
)
3110 page_hash_table pht1
, pht2
;
3114 for (i
= 0; i
< PHT_SIZE
; i
++) pht1
[i
] |= pht2
[i
];
3119 void GC_dirty_init()
3124 GC_dirty_maintained
= TRUE
;
3125 if (GC_words_allocd
!= 0 || GC_words_allocd_before_gc
!= 0) {
3128 for (i
= 0; i
< PHT_SIZE
; i
++) GC_written_pages
[i
] = (word
)(-1);
3130 GC_printf1("Allocated words:%lu:all pages may have been written\n",
3132 (GC_words_allocd
+ GC_words_allocd_before_gc
));
3135 sprintf(buf
, "/proc/%d", getpid());
3136 fd
= open(buf
, O_RDONLY
);
3138 ABORT("/proc open failed");
3140 GC_proc_fd
= syscall(SYS_ioctl
, fd
, PIOCOPENPD
, 0);
3142 syscall(SYS_fcntl
, GC_proc_fd
, F_SETFD
, FD_CLOEXEC
);
3143 if (GC_proc_fd
< 0) {
3144 ABORT("/proc ioctl failed");
3146 GC_proc_buf
= GC_scratch_alloc(GC_proc_buf_size
);
3147 # ifdef GC_SOLARIS_THREADS
3148 GC_fresh_pages
= (struct hblk
**)
3149 GC_scratch_alloc(MAX_FRESH_PAGES
* sizeof (struct hblk
*));
3150 if (GC_fresh_pages
== 0) {
3151 GC_err_printf0("No space for fresh pages\n");
3154 BZERO(GC_fresh_pages
, MAX_FRESH_PAGES
* sizeof (struct hblk
*));
3158 /* Ignore write hints. They don't help us here. */
3160 void GC_remove_protection(h
, nblocks
, is_ptrfree
)
3167 #ifdef GC_SOLARIS_THREADS
3168 # define READ(fd,buf,nbytes) syscall(SYS_read, fd, buf, nbytes)
3170 # define READ(fd,buf,nbytes) read(fd, buf, nbytes)
3173 void GC_read_dirty()
3175 unsigned long ps
, np
;
3178 struct prasmap
* map
;
3180 ptr_t current_addr
, limit
;
3184 BZERO(GC_grungy_pages
, (sizeof GC_grungy_pages
));
3187 if (READ(GC_proc_fd
, bufp
, GC_proc_buf_size
) <= 0) {
3189 GC_printf1("/proc read failed: GC_proc_buf_size = %lu\n",
3193 /* Retry with larger buffer. */
3194 word new_size
= 2 * GC_proc_buf_size
;
3195 char * new_buf
= GC_scratch_alloc(new_size
);
3198 GC_proc_buf
= bufp
= new_buf
;
3199 GC_proc_buf_size
= new_size
;
3201 if (READ(GC_proc_fd
, bufp
, GC_proc_buf_size
) <= 0) {
3202 WARN("Insufficient space for /proc read\n", 0);
3204 memset(GC_grungy_pages
, 0xff, sizeof (page_hash_table
));
3205 memset(GC_written_pages
, 0xff, sizeof(page_hash_table
));
3206 # ifdef GC_SOLARIS_THREADS
3207 BZERO(GC_fresh_pages
,
3208 MAX_FRESH_PAGES
* sizeof (struct hblk
*));
3214 /* Copy dirty bits into GC_grungy_pages */
3215 nmaps
= ((struct prpageheader
*)bufp
) -> pr_nmap
;
3216 /* printf( "nmaps = %d, PG_REFERENCED = %d, PG_MODIFIED = %d\n",
3217 nmaps, PG_REFERENCED, PG_MODIFIED); */
3218 bufp
= bufp
+ sizeof(struct prpageheader
);
3219 for (i
= 0; i
< nmaps
; i
++) {
3220 map
= (struct prasmap
*)bufp
;
3221 vaddr
= (ptr_t
)(map
-> pr_vaddr
);
3222 ps
= map
-> pr_pagesize
;
3223 np
= map
-> pr_npage
;
3224 /* printf("vaddr = 0x%X, ps = 0x%X, np = 0x%X\n", vaddr, ps, np); */
3225 limit
= vaddr
+ ps
* np
;
3226 bufp
+= sizeof (struct prasmap
);
3227 for (current_addr
= vaddr
;
3228 current_addr
< limit
; current_addr
+= ps
){
3229 if ((*bufp
++) & PG_MODIFIED
) {
3230 register struct hblk
* h
= (struct hblk
*) current_addr
;
3232 while ((ptr_t
)h
< current_addr
+ ps
) {
3233 register word index
= PHT_HASH(h
);
3235 set_pht_entry_from_index(GC_grungy_pages
, index
);
3236 # ifdef GC_SOLARIS_THREADS
3238 register int slot
= FRESH_PAGE_SLOT(h
);
3240 if (GC_fresh_pages
[slot
] == h
) {
3241 GC_fresh_pages
[slot
] = 0;
3249 bufp
+= sizeof(long) - 1;
3250 bufp
= (char *)((unsigned long)bufp
& ~(sizeof(long)-1));
3252 /* Update GC_written_pages. */
3253 GC_or_pages(GC_written_pages
, GC_grungy_pages
);
3254 # ifdef GC_SOLARIS_THREADS
3255 /* Make sure that old stacks are considered completely clean */
3256 /* unless written again. */
3257 GC_old_stacks_are_fresh();
3263 GC_bool
GC_page_was_dirty(h
)
3266 register word index
= PHT_HASH(h
);
3267 register GC_bool result
;
3269 result
= get_pht_entry_from_index(GC_grungy_pages
, index
);
3270 # ifdef GC_SOLARIS_THREADS
3271 if (result
&& PAGE_IS_FRESH(h
)) result
= FALSE
;
3272 /* This happens only if page was declared fresh since */
3273 /* the read_dirty call, e.g. because it's in an unused */
3274 /* thread stack. It's OK to treat it as clean, in */
3275 /* that case. And it's consistent with */
3276 /* GC_page_was_ever_dirty. */
3281 GC_bool
GC_page_was_ever_dirty(h
)
3284 register word index
= PHT_HASH(h
);
3285 register GC_bool result
;
3287 result
= get_pht_entry_from_index(GC_written_pages
, index
);
3288 # ifdef GC_SOLARIS_THREADS
3289 if (result
&& PAGE_IS_FRESH(h
)) result
= FALSE
;
3294 /* Caller holds allocation lock. */
3295 void GC_is_fresh(h
, n
)
3300 register word index
;
3302 # ifdef GC_SOLARIS_THREADS
3305 if (GC_fresh_pages
!= 0) {
3306 for (i
= 0; i
< n
; i
++) {
3307 ADD_FRESH_PAGE(h
+ i
);
3313 # endif /* PROC_VDB */
3318 # include "vd/PCR_VD.h"
3320 # define NPAGES (32*1024) /* 128 MB */
3322 PCR_VD_DB GC_grungy_bits
[NPAGES
];
3324 ptr_t GC_vd_base
; /* Address corresponding to GC_grungy_bits[0] */
3325 /* HBLKSIZE aligned. */
3327 void GC_dirty_init()
3329 GC_dirty_maintained
= TRUE
;
3330 /* For the time being, we assume the heap generally grows up */
3331 GC_vd_base
= GC_heap_sects
[0].hs_start
;
3332 if (GC_vd_base
== 0) {
3333 ABORT("Bad initial heap segment");
3335 if (PCR_VD_Start(HBLKSIZE
, GC_vd_base
, NPAGES
*HBLKSIZE
)
3337 ABORT("dirty bit initialization failed");
3341 void GC_read_dirty()
3343 /* lazily enable dirty bits on newly added heap sects */
3345 static int onhs
= 0;
3346 int nhs
= GC_n_heap_sects
;
3347 for( ; onhs
< nhs
; onhs
++ ) {
3348 PCR_VD_WriteProtectEnable(
3349 GC_heap_sects
[onhs
].hs_start
,
3350 GC_heap_sects
[onhs
].hs_bytes
);
3355 if (PCR_VD_Clear(GC_vd_base
, NPAGES
*HBLKSIZE
, GC_grungy_bits
)
3357 ABORT("dirty bit read failed");
3361 GC_bool
GC_page_was_dirty(h
)
3364 if((ptr_t
)h
< GC_vd_base
|| (ptr_t
)h
>= GC_vd_base
+ NPAGES
*HBLKSIZE
) {
3367 return(GC_grungy_bits
[h
- (struct hblk
*)GC_vd_base
] & PCR_VD_DB_dirtyBit
);
3371 void GC_remove_protection(h
, nblocks
, is_ptrfree
)
3376 PCR_VD_WriteProtectDisable(h
, nblocks
*HBLKSIZE
);
3377 PCR_VD_WriteProtectEnable(h
, nblocks
*HBLKSIZE
);
3380 # endif /* PCR_VDB */
3382 #if defined(MPROTECT_VDB) && defined(DARWIN)
3383 /* The following sources were used as a *reference* for this exception handling
3385 1. Apple's mach/xnu documentation
3386 2. Timothy J. Wood's "Mach Exception Handlers 101" post to the
3387 omnigroup's macosx-dev list.
3388 www.omnigroup.com/mailman/archive/macosx-dev/2000-June/014178.html
3389 3. macosx-nat.c from Apple's GDB source code.
3392 /* The bug that caused all this trouble should now be fixed. This should
3393 eventually be removed if all goes well. */
3394 /* define BROKEN_EXCEPTION_HANDLING */
3396 #include <mach/mach.h>
3397 #include <mach/mach_error.h>
3398 #include <mach/thread_status.h>
3399 #include <mach/exception.h>
3400 #include <mach/task.h>
3401 #include <pthread.h>
3403 /* These are not defined in any header, although they are documented */
3404 extern boolean_t
exc_server(mach_msg_header_t
*,mach_msg_header_t
*);
3405 extern kern_return_t
exception_raise(
3406 mach_port_t
,mach_port_t
,mach_port_t
,
3407 exception_type_t
,exception_data_t
,mach_msg_type_number_t
);
3408 extern kern_return_t
exception_raise_state(
3409 mach_port_t
,mach_port_t
,mach_port_t
,
3410 exception_type_t
,exception_data_t
,mach_msg_type_number_t
,
3411 thread_state_flavor_t
*,thread_state_t
,mach_msg_type_number_t
,
3412 thread_state_t
,mach_msg_type_number_t
*);
3413 extern kern_return_t
exception_raise_state_identity(
3414 mach_port_t
,mach_port_t
,mach_port_t
,
3415 exception_type_t
,exception_data_t
,mach_msg_type_number_t
,
3416 thread_state_flavor_t
*,thread_state_t
,mach_msg_type_number_t
,
3417 thread_state_t
,mach_msg_type_number_t
*);
3420 #define MAX_EXCEPTION_PORTS 16
3423 mach_msg_type_number_t count
;
3424 exception_mask_t masks
[MAX_EXCEPTION_PORTS
];
3425 exception_handler_t ports
[MAX_EXCEPTION_PORTS
];
3426 exception_behavior_t behaviors
[MAX_EXCEPTION_PORTS
];
3427 thread_state_flavor_t flavors
[MAX_EXCEPTION_PORTS
];
3431 mach_port_t exception
;
3432 #if defined(THREADS)
3438 mach_msg_header_t head
;
3442 GC_MP_NORMAL
, GC_MP_DISCARDING
, GC_MP_STOPPED
3443 } GC_mprotect_state_t
;
3445 /* FIXME: 1 and 2 seem to be safe to use in the msgh_id field,
3446 but it isn't documented. Use the source and see if they
3451 /* These values are only used on the reply port */
3454 #if defined(THREADS)
3456 GC_mprotect_state_t GC_mprotect_state
;
3458 /* The following should ONLY be called when the world is stopped */
3459 static void GC_mprotect_thread_notify(mach_msg_id_t id
) {
3462 mach_msg_trailer_t trailer
;
3464 mach_msg_return_t r
;
3466 buf
.msg
.head
.msgh_bits
=
3467 MACH_MSGH_BITS(MACH_MSG_TYPE_MAKE_SEND
,0);
3468 buf
.msg
.head
.msgh_size
= sizeof(buf
.msg
);
3469 buf
.msg
.head
.msgh_remote_port
= GC_ports
.exception
;
3470 buf
.msg
.head
.msgh_local_port
= MACH_PORT_NULL
;
3471 buf
.msg
.head
.msgh_id
= id
;
3475 MACH_SEND_MSG
|MACH_RCV_MSG
|MACH_RCV_LARGE
,
3479 MACH_MSG_TIMEOUT_NONE
,
3481 if(r
!= MACH_MSG_SUCCESS
)
3482 ABORT("mach_msg failed in GC_mprotect_thread_notify");
3483 if(buf
.msg
.head
.msgh_id
!= ID_ACK
)
3484 ABORT("invalid ack in GC_mprotect_thread_notify");
3487 /* Should only be called by the mprotect thread */
3488 static void GC_mprotect_thread_reply() {
3490 mach_msg_return_t r
;
3492 msg
.head
.msgh_bits
=
3493 MACH_MSGH_BITS(MACH_MSG_TYPE_MAKE_SEND
,0);
3494 msg
.head
.msgh_size
= sizeof(msg
);
3495 msg
.head
.msgh_remote_port
= GC_ports
.reply
;
3496 msg
.head
.msgh_local_port
= MACH_PORT_NULL
;
3497 msg
.head
.msgh_id
= ID_ACK
;
3505 MACH_MSG_TIMEOUT_NONE
,
3507 if(r
!= MACH_MSG_SUCCESS
)
3508 ABORT("mach_msg failed in GC_mprotect_thread_reply");
3511 void GC_mprotect_stop() {
3512 GC_mprotect_thread_notify(ID_STOP
);
3514 void GC_mprotect_resume() {
3515 GC_mprotect_thread_notify(ID_RESUME
);
3518 #else /* !THREADS */
3519 /* The compiler should optimize away any GC_mprotect_state computations */
3520 #define GC_mprotect_state GC_MP_NORMAL
3523 static void *GC_mprotect_thread(void *arg
) {
3524 mach_msg_return_t r
;
3525 /* These two structures contain some private kernel data. We don't need to
3526 access any of it so we don't bother defining a proper struct. The
3527 correct definitions are in the xnu source code. */
3529 mach_msg_header_t head
;
3533 mach_msg_header_t head
;
3534 mach_msg_body_t msgh_body
;
3540 GC_darwin_register_mach_handler_thread(mach_thread_self());
3545 MACH_RCV_MSG
|MACH_RCV_LARGE
|
3546 (GC_mprotect_state
== GC_MP_DISCARDING
? MACH_RCV_TIMEOUT
: 0),
3550 GC_mprotect_state
== GC_MP_DISCARDING
? 0 : MACH_MSG_TIMEOUT_NONE
,
3553 id
= r
== MACH_MSG_SUCCESS
? msg
.head
.msgh_id
: -1;
3555 #if defined(THREADS)
3556 if(GC_mprotect_state
== GC_MP_DISCARDING
) {
3557 if(r
== MACH_RCV_TIMED_OUT
) {
3558 GC_mprotect_state
= GC_MP_STOPPED
;
3559 GC_mprotect_thread_reply();
3562 if(r
== MACH_MSG_SUCCESS
&& (id
== ID_STOP
|| id
== ID_RESUME
))
3563 ABORT("out of order mprotect thread request");
3567 if(r
!= MACH_MSG_SUCCESS
) {
3568 GC_err_printf2("mach_msg failed with %d %s\n",
3569 (int)r
,mach_error_string(r
));
3570 ABORT("mach_msg failed");
3574 #if defined(THREADS)
3576 if(GC_mprotect_state
!= GC_MP_NORMAL
)
3577 ABORT("Called mprotect_stop when state wasn't normal");
3578 GC_mprotect_state
= GC_MP_DISCARDING
;
3581 if(GC_mprotect_state
!= GC_MP_STOPPED
)
3582 ABORT("Called mprotect_resume when state wasn't stopped");
3583 GC_mprotect_state
= GC_MP_NORMAL
;
3584 GC_mprotect_thread_reply();
3586 #endif /* THREADS */
3588 /* Handle the message (calls catch_exception_raise) */
3589 if(!exc_server(&msg
.head
,&reply
.head
))
3590 ABORT("exc_server failed");
3591 /* Send the reply */
3595 reply
.head
.msgh_size
,
3598 MACH_MSG_TIMEOUT_NONE
,
3600 if(r
!= MACH_MSG_SUCCESS
) {
3601 /* This will fail if the thread dies, but the thread shouldn't
3603 #ifdef BROKEN_EXCEPTION_HANDLING
3605 "mach_msg failed with %d %s while sending exc reply\n",
3606 (int)r
,mach_error_string(r
));
3608 ABORT("mach_msg failed while sending exception reply");
3617 /* All this SIGBUS code shouldn't be necessary. All protection faults should
3618 be going throught the mach exception handler. However, it seems a SIGBUS is
3619 occasionally sent for some unknown reason. Even more odd, it seems to be
3620 meaningless and safe to ignore. */
3621 #ifdef BROKEN_EXCEPTION_HANDLING
3623 typedef void (* SIG_PF
)();
3624 static SIG_PF GC_old_bus_handler
;
3626 /* Updates to this aren't atomic, but the SIGBUSs seem pretty rare.
3627 Even if this doesn't get updated property, it isn't really a problem */
3628 static int GC_sigbus_count
;
3630 static void GC_darwin_sigbus(int num
,siginfo_t
*sip
,void *context
) {
3631 if(num
!= SIGBUS
) ABORT("Got a non-sigbus signal in the sigbus handler");
3633 /* Ugh... some seem safe to ignore, but too many in a row probably means
3634 trouble. GC_sigbus_count is reset for each mach exception that is
3636 if(GC_sigbus_count
>= 8) {
3637 ABORT("Got more than 8 SIGBUSs in a row!");
3640 GC_err_printf0("GC: WARNING: Ignoring SIGBUS.\n");
3643 #endif /* BROKEN_EXCEPTION_HANDLING */
3645 void GC_dirty_init() {
3649 pthread_attr_t attr
;
3650 exception_mask_t mask
;
3653 GC_printf0("Inititalizing mach/darwin mprotect virtual dirty bit "
3654 "implementation\n");
3656 # ifdef BROKEN_EXCEPTION_HANDLING
3657 GC_err_printf0("GC: WARNING: Enabling workarounds for various darwin "
3658 "exception handling bugs.\n");
3660 GC_dirty_maintained
= TRUE
;
3661 if (GC_page_size
% HBLKSIZE
!= 0) {
3662 GC_err_printf0("Page size not multiple of HBLKSIZE\n");
3663 ABORT("Page size not multiple of HBLKSIZE");
3666 GC_task_self
= me
= mach_task_self();
3668 r
= mach_port_allocate(me
,MACH_PORT_RIGHT_RECEIVE
,&GC_ports
.exception
);
3669 if(r
!= KERN_SUCCESS
) ABORT("mach_port_allocate failed (exception port)");
3671 r
= mach_port_insert_right(me
,GC_ports
.exception
,GC_ports
.exception
,
3672 MACH_MSG_TYPE_MAKE_SEND
);
3673 if(r
!= KERN_SUCCESS
)
3674 ABORT("mach_port_insert_right failed (exception port)");
3676 #if defined(THREADS)
3677 r
= mach_port_allocate(me
,MACH_PORT_RIGHT_RECEIVE
,&GC_ports
.reply
);
3678 if(r
!= KERN_SUCCESS
) ABORT("mach_port_allocate failed (reply port)");
3681 /* The exceptions we want to catch */
3682 mask
= EXC_MASK_BAD_ACCESS
;
3684 r
= task_get_exception_ports(
3687 GC_old_exc_ports
.masks
,
3688 &GC_old_exc_ports
.count
,
3689 GC_old_exc_ports
.ports
,
3690 GC_old_exc_ports
.behaviors
,
3691 GC_old_exc_ports
.flavors
3693 if(r
!= KERN_SUCCESS
) ABORT("task_get_exception_ports failed");
3695 r
= task_set_exception_ports(
3700 GC_MACH_THREAD_STATE
3702 if(r
!= KERN_SUCCESS
) ABORT("task_set_exception_ports failed");
3704 if(pthread_attr_init(&attr
) != 0) ABORT("pthread_attr_init failed");
3705 if(pthread_attr_setdetachstate(&attr
,PTHREAD_CREATE_DETACHED
) != 0)
3706 ABORT("pthread_attr_setdetachedstate failed");
3708 # undef pthread_create
3709 /* This will call the real pthread function, not our wrapper */
3710 if(pthread_create(&thread
,&attr
,GC_mprotect_thread
,NULL
) != 0)
3711 ABORT("pthread_create failed");
3712 pthread_attr_destroy(&attr
);
3714 /* Setup the sigbus handler for ignoring the meaningless SIGBUSs */
3715 #ifdef BROKEN_EXCEPTION_HANDLING
3717 struct sigaction sa
, oldsa
;
3718 sa
.sa_handler
= (SIG_PF
)GC_darwin_sigbus
;
3719 sigemptyset(&sa
.sa_mask
);
3720 sa
.sa_flags
= SA_RESTART
|SA_SIGINFO
;
3721 if(sigaction(SIGBUS
,&sa
,&oldsa
) < 0) ABORT("sigaction");
3722 GC_old_bus_handler
= (SIG_PF
)oldsa
.sa_handler
;
3723 if (GC_old_bus_handler
!= SIG_DFL
) {
3725 GC_err_printf0("Replaced other SIGBUS handler\n");
3729 #endif /* BROKEN_EXCEPTION_HANDLING */
3732 /* The source code for Apple's GDB was used as a reference for the exception
3733 forwarding code. This code is similar to be GDB code only because there is
3734 only one way to do it. */
3735 static kern_return_t
GC_forward_exception(
3738 exception_type_t exception
,
3739 exception_data_t data
,
3740 mach_msg_type_number_t data_count
3745 exception_behavior_t behavior
;
3746 thread_state_flavor_t flavor
;
3748 thread_state_t thread_state
;
3749 mach_msg_type_number_t thread_state_count
= THREAD_STATE_MAX
;
3751 for(i
=0;i
<GC_old_exc_ports
.count
;i
++)
3752 if(GC_old_exc_ports
.masks
[i
] & (1 << exception
))
3754 if(i
==GC_old_exc_ports
.count
) ABORT("No handler for exception!");
3756 port
= GC_old_exc_ports
.ports
[i
];
3757 behavior
= GC_old_exc_ports
.behaviors
[i
];
3758 flavor
= GC_old_exc_ports
.flavors
[i
];
3760 if(behavior
!= EXCEPTION_DEFAULT
) {
3761 r
= thread_get_state(thread
,flavor
,thread_state
,&thread_state_count
);
3762 if(r
!= KERN_SUCCESS
)
3763 ABORT("thread_get_state failed in forward_exception");
3767 case EXCEPTION_DEFAULT
:
3768 r
= exception_raise(port
,thread
,task
,exception
,data
,data_count
);
3770 case EXCEPTION_STATE
:
3771 r
= exception_raise_state(port
,thread
,task
,exception
,data
,
3772 data_count
,&flavor
,thread_state
,thread_state_count
,
3773 thread_state
,&thread_state_count
);
3775 case EXCEPTION_STATE_IDENTITY
:
3776 r
= exception_raise_state_identity(port
,thread
,task
,exception
,data
,
3777 data_count
,&flavor
,thread_state
,thread_state_count
,
3778 thread_state
,&thread_state_count
);
3781 r
= KERN_FAILURE
; /* make gcc happy */
3782 ABORT("forward_exception: unknown behavior");
3786 if(behavior
!= EXCEPTION_DEFAULT
) {
3787 r
= thread_set_state(thread
,flavor
,thread_state
,thread_state_count
);
3788 if(r
!= KERN_SUCCESS
)
3789 ABORT("thread_set_state failed in forward_exception");
3795 #define FWD() GC_forward_exception(thread,task,exception,code,code_count)
3797 /* This violates the namespace rules but there isn't anything that can be done
3798 about it. The exception handling stuff is hard coded to call this */
3800 catch_exception_raise(
3801 mach_port_t exception_port
,mach_port_t thread
,mach_port_t task
,
3802 exception_type_t exception
,exception_data_t code
,
3803 mach_msg_type_number_t code_count
3809 # if defined(POWERPC)
3810 # if CPP_WORDSZ == 32
3811 thread_state_flavor_t flavor
= PPC_EXCEPTION_STATE
;
3812 mach_msg_type_number_t exc_state_count
= PPC_EXCEPTION_STATE_COUNT
;
3813 ppc_exception_state_t exc_state
;
3815 thread_state_flavor_t flavor
= PPC_EXCEPTION_STATE64
;
3816 mach_msg_type_number_t exc_state_count
= PPC_EXCEPTION_STATE64_COUNT
;
3817 ppc_exception_state64_t exc_state
;
3819 # elif defined(I386) || defined(X86_64)
3820 # if CPP_WORDSZ == 32
3821 thread_state_flavor_t flavor
= x86_EXCEPTION_STATE32
;
3822 mach_msg_type_number_t exc_state_count
= x86_EXCEPTION_STATE32_COUNT
;
3823 x86_exception_state32_t exc_state
;
3825 thread_state_flavor_t flavor
= x86_EXCEPTION_STATE64
;
3826 mach_msg_type_number_t exc_state_count
= x86_EXCEPTION_STATE64_COUNT
;
3827 x86_exception_state64_t exc_state
;
3830 # error FIXME for non-ppc darwin
3834 if(exception
!= EXC_BAD_ACCESS
|| code
[0] != KERN_PROTECTION_FAILURE
) {
3835 #ifdef DEBUG_EXCEPTION_HANDLING
3836 /* We aren't interested, pass it on to the old handler */
3837 GC_printf3("Exception: 0x%x Code: 0x%x 0x%x in catch....\n",
3839 code_count
> 0 ? code
[0] : -1,
3840 code_count
> 1 ? code
[1] : -1);
3845 r
= thread_get_state(thread
,flavor
,
3846 (natural_t
*)&exc_state
,&exc_state_count
);
3847 if(r
!= KERN_SUCCESS
) {
3848 /* The thread is supposed to be suspended while the exception handler
3849 is called. This shouldn't fail. */
3850 #ifdef BROKEN_EXCEPTION_HANDLING
3851 GC_err_printf0("thread_get_state failed in "
3852 "catch_exception_raise\n");
3853 return KERN_SUCCESS
;
3855 ABORT("thread_get_state failed in catch_exception_raise");
3859 /* This is the address that caused the fault */
3860 #if defined(POWERPC)
3861 addr
= (char*) exc_state
. THREAD_FLD(dar
);
3862 #elif defined (I386) || defined (X86_64)
3863 addr
= (char*) exc_state
. THREAD_FLD(faultvaddr
);
3865 # error FIXME for non POWERPC/I386
3868 if((HDR(addr
)) == 0) {
3869 /* Ugh... just like the SIGBUS problem above, it seems we get a bogus
3870 KERN_PROTECTION_FAILURE every once and a while. We wait till we get
3871 a bunch in a row before doing anything about it. If a "real" fault
3872 ever occurres it'll just keep faulting over and over and we'll hit
3873 the limit pretty quickly. */
3874 #ifdef BROKEN_EXCEPTION_HANDLING
3875 static char *last_fault
;
3876 static int last_fault_count
;
3878 if(addr
!= last_fault
) {
3880 last_fault_count
= 0;
3882 if(++last_fault_count
< 32) {
3883 if(last_fault_count
== 1)
3885 "GC: WARNING: Ignoring KERN_PROTECTION_FAILURE at %p\n",
3887 return KERN_SUCCESS
;
3890 GC_err_printf1("Unexpected KERN_PROTECTION_FAILURE at %p\n",addr
);
3891 /* Can't pass it along to the signal handler because that is
3892 ignoring SIGBUS signals. We also shouldn't call ABORT here as
3893 signals don't always work too well from the exception handler. */
3894 GC_err_printf0("Aborting\n");
3896 #else /* BROKEN_EXCEPTION_HANDLING */
3897 /* Pass it along to the next exception handler
3898 (which should call SIGBUS/SIGSEGV) */
3900 #endif /* !BROKEN_EXCEPTION_HANDLING */
3903 #ifdef BROKEN_EXCEPTION_HANDLING
3904 /* Reset the number of consecutive SIGBUSs */
3905 GC_sigbus_count
= 0;
3908 if(GC_mprotect_state
== GC_MP_NORMAL
) { /* common case */
3909 h
= (struct hblk
*)((word
)addr
& ~(GC_page_size
-1));
3910 UNPROTECT(h
, GC_page_size
);
3911 for (i
= 0; i
< divHBLKSZ(GC_page_size
); i
++) {
3912 register int index
= PHT_HASH(h
+i
);
3913 async_set_pht_entry_from_index(GC_dirty_pages
, index
);
3915 } else if(GC_mprotect_state
== GC_MP_DISCARDING
) {
3916 /* Lie to the thread for now. No sense UNPROTECT()ing the memory
3917 when we're just going to PROTECT() it again later. The thread
3918 will just fault again once it resumes */
3920 /* Shouldn't happen, i don't think */
3921 GC_printf0("KERN_PROTECTION_FAILURE while world is stopped\n");
3924 return KERN_SUCCESS
;
3928 /* These should never be called, but just in case... */
3929 kern_return_t
catch_exception_raise_state(mach_port_name_t exception_port
,
3930 int exception
, exception_data_t code
, mach_msg_type_number_t codeCnt
,
3931 int flavor
, thread_state_t old_state
, int old_stateCnt
,
3932 thread_state_t new_state
, int new_stateCnt
)
3934 ABORT("catch_exception_raise_state");
3935 return(KERN_INVALID_ARGUMENT
);
3937 kern_return_t
catch_exception_raise_state_identity(
3938 mach_port_name_t exception_port
, mach_port_t thread
, mach_port_t task
,
3939 int exception
, exception_data_t code
, mach_msg_type_number_t codeCnt
,
3940 int flavor
, thread_state_t old_state
, int old_stateCnt
,
3941 thread_state_t new_state
, int new_stateCnt
)
3943 ABORT("catch_exception_raise_state_identity");
3944 return(KERN_INVALID_ARGUMENT
);
3948 #endif /* DARWIN && MPROTECT_VDB */
3950 # ifndef HAVE_INCREMENTAL_PROTECTION_NEEDS
3951 int GC_incremental_protection_needs()
3953 return GC_PROTECTS_NONE
;
3955 # endif /* !HAVE_INCREMENTAL_PROTECTION_NEEDS */
3958 * Call stack save code for debugging.
3959 * Should probably be in mach_dep.c, but that requires reorganization.
3962 /* I suspect the following works for most X86 *nix variants, so */
3963 /* long as the frame pointer is explicitly stored. In the case of gcc, */
3964 /* compiler flags (e.g. -fomit-frame-pointer) determine whether it is. */
3965 #if defined(I386) && defined(LINUX) && defined(SAVE_CALL_CHAIN)
3966 # include <features.h>
3969 struct frame
*fr_savfp
;
3971 long fr_arg
[NARGS
]; /* All the arguments go here. */
3977 # include <features.h>
3982 struct frame
*fr_savfp
;
3991 # if defined(SUNOS4)
3992 # include <machine/frame.h>
3994 # if defined (DRSNX)
3995 # include <sys/sparc/frame.h>
3997 # if defined(OPENBSD)
4000 # if defined(FREEBSD) || defined(NETBSD)
4001 # include <machine/frame.h>
4003 # include <sys/frame.h>
4010 --> We only know how to to get the first
6 arguments
4014 #ifdef NEED_CALLINFO
4015 /* Fill in the pc and argument information for up to NFRAMES of my */
4016 /* callers. Ignore my frame and my callers frame. */
4019 # include <unistd.h>
4022 #endif /* NEED_CALLINFO */
4024 #if defined(GC_HAVE_BUILTIN_BACKTRACE)
4025 # include <execinfo.h>
4028 #ifdef SAVE_CALL_CHAIN
4030 #if NARGS == 0 && NFRAMES % 2 == 0 /* No padding */ \
4031 && defined(GC_HAVE_BUILTIN_BACKTRACE)
4033 #ifdef REDIRECT_MALLOC
4034 /* Deal with possible malloc calls in backtrace by omitting */
4035 /* the infinitely recursing backtrace. */
4037 __thread
/* If your compiler doesn't understand this */
4038 /* you could use something like pthread_getspecific. */
4040 GC_in_save_callers
= FALSE
;
4043 void GC_save_callers (info
)
4044 struct callinfo info
[NFRAMES
];
4046 void * tmp_info
[NFRAMES
+ 1];
4048 # define IGNORE_FRAMES 1
4050 /* We retrieve NFRAMES+1 pc values, but discard the first, since it */
4051 /* points to our own frame. */
4052 # ifdef REDIRECT_MALLOC
4053 if (GC_in_save_callers
) {
4054 info
[0].ci_pc
= (word
)(&GC_save_callers
);
4055 for (i
= 1; i
< NFRAMES
; ++i
) info
[i
].ci_pc
= 0;
4058 GC_in_save_callers
= TRUE
;
4060 GC_ASSERT(sizeof(struct callinfo
) == sizeof(void *));
4061 npcs
= backtrace((void **)tmp_info
, NFRAMES
+ IGNORE_FRAMES
);
4062 BCOPY(tmp_info
+IGNORE_FRAMES
, info
, (npcs
- IGNORE_FRAMES
) * sizeof(void *));
4063 for (i
= npcs
- IGNORE_FRAMES
; i
< NFRAMES
; ++i
) info
[i
].ci_pc
= 0;
4064 # ifdef REDIRECT_MALLOC
4065 GC_in_save_callers
= FALSE
;
4069 #else /* No builtin backtrace; do it ourselves */
4071 #if (defined(OPENBSD) || defined(NETBSD) || defined(FREEBSD)) && defined(SPARC)
4072 # define FR_SAVFP fr_fp
4073 # define FR_SAVPC fr_pc
4075 # define FR_SAVFP fr_savfp
4076 # define FR_SAVPC fr_savpc
4079 #if defined(SPARC) && (defined(__arch64__) || defined(__sparcv9))
4085 void GC_save_callers (info
)
4086 struct callinfo info
[NFRAMES
];
4088 struct frame
*frame
;
4092 /* We assume this is turned on only with gcc as the compiler. */
4093 asm("movl %%ebp,%0" : "=r"(frame
));
4096 frame
= (struct frame
*) GC_save_regs_in_stack ();
4097 fp
= (struct frame
*)((long) frame
-> FR_SAVFP
+ BIAS
);
4100 for (; (!(fp HOTTER_THAN frame
) && !(GC_stackbottom
HOTTER_THAN (ptr_t
)fp
)
4101 && (nframes
< NFRAMES
));
4102 fp
= (struct frame
*)((long) fp
-> FR_SAVFP
+ BIAS
), nframes
++) {
4105 info
[nframes
].ci_pc
= fp
->FR_SAVPC
;
4107 for (i
= 0; i
< NARGS
; i
++) {
4108 info
[nframes
].ci_arg
[i
] = ~(fp
->fr_arg
[i
]);
4110 # endif /* NARGS > 0 */
4112 if (nframes
< NFRAMES
) info
[nframes
].ci_pc
= 0;
4115 #endif /* No builtin backtrace */
4117 #endif /* SAVE_CALL_CHAIN */
4119 #ifdef NEED_CALLINFO
4121 /* Print info to stderr. We do NOT hold the allocation lock */
4122 void GC_print_callers (info
)
4123 struct callinfo info
[NFRAMES
];
4126 static int reentry_count
= 0;
4127 GC_bool stop
= FALSE
;
4129 /* FIXME: This should probably use a different lock, so that we */
4130 /* become callable with or without the allocation lock. */
4136 GC_err_printf0("\tCaller at allocation:\n");
4138 GC_err_printf0("\tCall chain at allocation:\n");
4140 for (i
= 0; i
< NFRAMES
&& !stop
; i
++) {
4141 if (info
[i
].ci_pc
== 0) break;
4146 GC_err_printf0("\t\targs: ");
4147 for (j
= 0; j
< NARGS
; j
++) {
4148 if (j
!= 0) GC_err_printf0(", ");
4149 GC_err_printf2("%d (0x%X)", ~(info
[i
].ci_arg
[j
]),
4150 ~(info
[i
].ci_arg
[j
]));
4152 GC_err_printf0("\n");
4155 if (reentry_count
> 1) {
4156 /* We were called during an allocation during */
4157 /* a previous GC_print_callers call; punt. */
4158 GC_err_printf1("\t\t##PC##= 0x%lx\n", info
[i
].ci_pc
);
4165 # if defined(GC_HAVE_BUILTIN_BACKTRACE) \
4166 && !defined(GC_BACKTRACE_SYMBOLS_BROKEN)
4168 backtrace_symbols((void **)(&(info
[i
].ci_pc
)), 1);
4169 char *name
= sym_name
[0];
4173 sprintf(buf
, "##PC##= 0x%lx", info
[i
].ci_pc
);
4175 # if defined(LINUX) && !defined(SMALL_CONFIG)
4176 /* Try for a line number. */
4179 static char exe_name
[EXE_SZ
];
4181 char cmd_buf
[CMD_SZ
];
4182 # define RESULT_SZ 200
4183 static char result_buf
[RESULT_SZ
];
4186 # define PRELOAD_SZ 200
4187 char preload_buf
[PRELOAD_SZ
];
4188 static GC_bool found_exe_name
= FALSE
;
4189 static GC_bool will_fail
= FALSE
;
4191 /* Try to get it via a hairy and expensive scheme. */
4192 /* First we get the name of the executable: */
4193 if (will_fail
) goto out
;
4194 if (!found_exe_name
) {
4195 ret_code
= readlink("/proc/self/exe", exe_name
, EXE_SZ
);
4196 if (ret_code
< 0 || ret_code
>= EXE_SZ
4197 || exe_name
[0] != '/') {
4198 will_fail
= TRUE
; /* Dont try again. */
4201 exe_name
[ret_code
] = '\0';
4202 found_exe_name
= TRUE
;
4204 /* Then we use popen to start addr2line -e <exe> <addr> */
4205 /* There are faster ways to do this, but hopefully this */
4206 /* isn't time critical. */
4207 sprintf(cmd_buf
, "/usr/bin/addr2line -f -e %s 0x%lx", exe_name
,
4208 (unsigned long)info
[i
].ci_pc
);
4209 old_preload
= getenv ("LD_PRELOAD");
4210 if (0 != old_preload
) {
4211 if (strlen (old_preload
) >= PRELOAD_SZ
) {
4215 strcpy (preload_buf
, old_preload
);
4216 unsetenv ("LD_PRELOAD");
4218 pipe
= popen(cmd_buf
, "r");
4219 if (0 != old_preload
4220 && 0 != setenv ("LD_PRELOAD", preload_buf
, 0)) {
4221 WARN("Failed to reset LD_PRELOAD\n", 0);
4224 || (result_len
= fread(result_buf
, 1, RESULT_SZ
- 1, pipe
))
4226 if (pipe
!= NULL
) pclose(pipe
);
4230 if (result_buf
[result_len
- 1] == '\n') --result_len
;
4231 result_buf
[result_len
] = 0;
4232 if (result_buf
[0] == '?'
4233 || result_buf
[result_len
-2] == ':'
4234 && result_buf
[result_len
-1] == '0') {
4238 /* Get rid of embedded newline, if any. Test for "main" */
4240 char * nl
= strchr(result_buf
, '\n');
4241 if (nl
!= NULL
&& nl
< result_buf
+ result_len
) {
4244 if (strncmp(result_buf
, "main", nl
- result_buf
) == 0) {
4248 if (result_len
< RESULT_SZ
- 25) {
4249 /* Add in hex address */
4250 sprintf(result_buf
+ result_len
, " [0x%lx]",
4251 (unsigned long)info
[i
].ci_pc
);
4258 GC_err_printf1("\t\t%s\n", name
);
4259 # if defined(GC_HAVE_BUILTIN_BACKTRACE) \
4260 && !defined(GC_BACKTRACE_SYMBOLS_BROKEN)
4261 free(sym_name
); /* May call GC_free; that's OK */
4270 #endif /* NEED_CALLINFO */
4274 #if defined(LINUX) && defined(__ELF__) && !defined(SMALL_CONFIG)
4276 /* Dump /proc/self/maps to GC_stderr, to enable looking up names for
4277 addresses in FIND_LEAK output. */
4279 static word
dump_maps(char *maps
)
4281 GC_err_write(maps
, strlen(maps
));
4285 void GC_print_address_map()
4287 GC_err_printf0("---------- Begin address map ----------\n");
4288 GC_apply_to_maps(dump_maps
);
4289 GC_err_printf0("---------- End address map ----------\n");