ifmedia/ifconfig: Take flowcontrol as an alias for rxpause|txpause
[dragonfly.git] / contrib / tcsh-6 / tc.alloc.c
blobb87f9e9f716b887bd2bdb03338ccbea4f0a9feef
1 /* $Header: /p/tcsh/cvsroot/tcsh/tc.alloc.c,v 3.53 2015/02/22 16:31:54 christos Exp $ */
2 /*
3 * tc.alloc.c (Caltech) 2/21/82
4 * Chris Kingsley, kingsley@cit-20.
6 * This is a very fast storage allocator. It allocates blocks of a small
7 * number of different sizes, and keeps free lists of each size. Blocks that
8 * don't exactly fit are passed up to the next larger size. In this
9 * implementation, the available sizes are 2^n-4 (or 2^n-12) bytes long.
10 * This is designed for use in a program that uses vast quantities of memory,
11 * but bombs when it runs out.
13 /*-
14 * Copyright (c) 1980, 1991 The Regents of the University of California.
15 * All rights reserved.
17 * Redistribution and use in source and binary forms, with or without
18 * modification, are permitted provided that the following conditions
19 * are met:
20 * 1. Redistributions of source code must retain the above copyright
21 * notice, this list of conditions and the following disclaimer.
22 * 2. Redistributions in binary form must reproduce the above copyright
23 * notice, this list of conditions and the following disclaimer in the
24 * documentation and/or other materials provided with the distribution.
25 * 3. Neither the name of the University nor the names of its contributors
26 * may be used to endorse or promote products derived from this software
27 * without specific prior written permission.
29 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
30 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
31 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
32 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
33 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
34 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
35 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
36 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
37 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
38 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
39 * SUCH DAMAGE.
41 #include "sh.h"
42 #ifdef HAVE_MALLINFO
43 #include <malloc.h>
44 #endif
45 #if defined(HAVE_SBRK) && !defined(__APPLE__)
46 #define USE_SBRK
47 #endif
49 RCSID("$tcsh: tc.alloc.c,v 3.53 2015/02/22 16:31:54 christos Exp $")
51 #define RCHECK
52 #define DEBUG
54 static char *memtop = NULL; /* PWP: top of current memory */
55 static char *membot = NULL; /* PWP: bottom of allocatable memory */
57 int dont_free = 0;
59 #ifdef WINNT_NATIVE
60 # define malloc fmalloc
61 # define free ffree
62 # define calloc fcalloc
63 # define realloc frealloc
64 #endif /* WINNT_NATIVE */
66 #if !defined(DEBUG) || defined(SYSMALLOC)
67 static void
68 out_of_memory (void)
70 static const char msg[] = "Out of memory\n";
72 TCSH_IGNORE(write(didfds ? 2 : SHDIAG, msg, strlen(msg)));
73 _exit(1);
75 #endif
77 #ifndef SYSMALLOC
79 #ifdef SX
80 extern void* sbrk();
81 #endif
83 * Lots of os routines are busted and try to free invalid pointers.
84 * Although our free routine is smart enough and it will pick bad
85 * pointers most of the time, in cases where we know we are going to get
86 * a bad pointer, we'd rather leak.
89 #ifndef NULL
90 #define NULL 0
91 #endif
93 typedef unsigned char U_char; /* we don't really have signed chars */
94 typedef unsigned int U_int;
95 typedef unsigned short U_short;
96 typedef unsigned long U_long;
100 * The overhead on a block is at least 4 bytes. When free, this space
101 * contains a pointer to the next free block, and the bottom two bits must
102 * be zero. When in use, the first byte is set to MAGIC, and the second
103 * byte is the size index. The remaining bytes are for alignment.
104 * If range checking is enabled and the size of the block fits
105 * in two bytes, then the top two bytes hold the size of the requested block
106 * plus the range checking words, and the header word MINUS ONE.
110 #define MEMALIGN(a) (((a) + ROUNDUP) & ~ROUNDUP)
112 union overhead {
113 union overhead *ov_next; /* when free */
114 struct {
115 U_char ovu_magic; /* magic number */
116 U_char ovu_index; /* bucket # */
117 #ifdef RCHECK
118 U_short ovu_size; /* actual block size */
119 U_int ovu_rmagic; /* range magic number */
120 #endif
121 } ovu;
122 #define ov_magic ovu.ovu_magic
123 #define ov_index ovu.ovu_index
124 #define ov_size ovu.ovu_size
125 #define ov_rmagic ovu.ovu_rmagic
128 #define MAGIC 0xfd /* magic # on accounting info */
129 #define RMAGIC 0x55555555 /* magic # on range info */
130 #ifdef RCHECK
131 #define RSLOP sizeof (U_int)
132 #else
133 #define RSLOP 0
134 #endif
137 #define ROUNDUP 7
140 * nextf[i] is the pointer to the next free block of size 2^(i+3). The
141 * smallest allocatable block is 8 bytes. The overhead information
142 * precedes the data area returned to the user.
144 #define NBUCKETS ((sizeof(long) << 3) - 3)
145 static union overhead *nextf[NBUCKETS] IZERO_STRUCT;
148 * nmalloc[i] is the difference between the number of mallocs and frees
149 * for a given block size.
151 static U_int nmalloc[NBUCKETS] IZERO_STRUCT;
153 #ifndef lint
154 static int findbucket (union overhead *, int);
155 static void morecore (int);
156 #endif
159 #ifdef DEBUG
160 # define CHECK(a, str, p) \
161 if (a) { \
162 xprintf(str, p); \
163 xprintf(" (memtop = %p membot = %p)\n", memtop, membot); \
164 abort(); \
166 #else
167 # define CHECK(a, str, p) \
168 if (a) { \
169 xprintf(str, p); \
170 xprintf(" (memtop = %p membot = %p)\n", memtop, membot); \
171 return; \
173 #endif
175 memalign_t
176 malloc(size_t nbytes)
178 #ifndef lint
179 union overhead *p;
180 int bucket = 0;
181 unsigned shiftr;
184 * Convert amount of memory requested into closest block size stored in
185 * hash buckets which satisfies request. Account for space used per block
186 * for accounting.
188 #ifdef SUNOS4
190 * SunOS localtime() overwrites the 9th byte on an 8 byte malloc()....
191 * so we get one more...
192 * From Michael Schroeder: This is not true. It depends on the
193 * timezone string. In Europe it can overwrite the 13th byte on a
194 * 12 byte malloc.
195 * So we punt and we always allocate an extra byte.
197 nbytes++;
198 #endif
200 nbytes = MEMALIGN(MEMALIGN(sizeof(union overhead)) + nbytes + RSLOP);
201 shiftr = (nbytes - 1) >> 2;
203 /* apart from this loop, this is O(1) */
204 while ((shiftr >>= 1) != 0)
205 bucket++;
207 * If nothing in hash bucket right now, request more memory from the
208 * system.
210 if (nextf[bucket] == NULL)
211 morecore(bucket);
212 if ((p = nextf[bucket]) == NULL) {
213 child++;
214 #ifndef DEBUG
215 out_of_memory();
216 #else
217 showall(NULL, NULL);
218 xprintf(CGETS(19, 1, "nbytes=%zu: Out of memory\n"), nbytes);
219 abort();
220 #endif
221 /* fool lint */
222 return ((memalign_t) 0);
224 /* remove from linked list */
225 nextf[bucket] = nextf[bucket]->ov_next;
226 p->ov_magic = MAGIC;
227 p->ov_index = bucket;
228 nmalloc[bucket]++;
229 #ifdef RCHECK
231 * Record allocated size of block and bound space with magic numbers.
233 p->ov_size = (p->ov_index <= 13) ? nbytes - 1 : 0;
234 p->ov_rmagic = RMAGIC;
235 *((U_int *) (((caddr_t) p) + nbytes - RSLOP)) = RMAGIC;
236 #endif
237 return ((memalign_t) (((caddr_t) p) + MEMALIGN(sizeof(union overhead))));
238 #else
239 if (nbytes)
240 return ((memalign_t) 0);
241 else
242 return ((memalign_t) 0);
243 #endif /* !lint */
246 #ifndef lint
248 * Allocate more memory to the indicated bucket.
250 static void
251 morecore(int bucket)
253 union overhead *op;
254 int rnu; /* 2^rnu bytes will be requested */
255 int nblks; /* become nblks blocks of the desired size */
256 int siz;
258 if (nextf[bucket])
259 return;
261 * Insure memory is allocated on a page boundary. Should make getpageize
262 * call?
264 op = (union overhead *) sbrk(0);
265 memtop = (char *) op;
266 if (membot == NULL)
267 membot = memtop;
268 if ((long) op & 0x3ff) {
269 memtop = sbrk((int) (1024 - ((long) op & 0x3ff)));
270 memtop += (long) (1024 - ((long) op & 0x3ff));
273 /* take 2k unless the block is bigger than that */
274 rnu = (bucket <= 8) ? 11 : bucket + 3;
275 nblks = 1 << (rnu - (bucket + 3)); /* how many blocks to get */
276 memtop = sbrk(1 << rnu); /* PWP */
277 op = (union overhead *) memtop;
278 /* no more room! */
279 if ((long) op == -1)
280 return;
281 memtop += (long) (1 << rnu);
283 * Round up to minimum allocation size boundary and deduct from block count
284 * to reflect.
286 if (((U_long) op) & ROUNDUP) {
287 op = (union overhead *) (((U_long) op + (ROUNDUP + 1)) & ~ROUNDUP);
288 nblks--;
291 * Add new memory allocated to that on free list for this hash bucket.
293 nextf[bucket] = op;
294 siz = 1 << (bucket + 3);
295 while (--nblks > 0) {
296 op->ov_next = (union overhead *) (((caddr_t) op) + siz);
297 op = (union overhead *) (((caddr_t) op) + siz);
299 op->ov_next = NULL;
302 #endif
304 void
305 free(ptr_t cp)
307 #ifndef lint
308 int size;
309 union overhead *op;
312 * the don't free flag is there so that we avoid os bugs in routines
313 * that free invalid pointers!
315 if (cp == NULL || dont_free)
316 return;
317 CHECK(!memtop || !membot,
318 CGETS(19, 2, "free(%p) called before any allocations."), cp);
319 CHECK(cp > (ptr_t) memtop,
320 CGETS(19, 3, "free(%p) above top of memory."), cp);
321 CHECK(cp < (ptr_t) membot,
322 CGETS(19, 4, "free(%p) below bottom of memory."), cp);
323 op = (union overhead *) (((caddr_t) cp) - MEMALIGN(sizeof(union overhead)));
324 CHECK(op->ov_magic != MAGIC,
325 CGETS(19, 5, "free(%p) bad block."), cp);
327 #ifdef RCHECK
328 if (op->ov_index <= 13)
329 CHECK(*(U_int *) ((caddr_t) op + op->ov_size + 1 - RSLOP) != RMAGIC,
330 CGETS(19, 6, "free(%p) bad range check."), cp);
331 #endif
332 CHECK(op->ov_index >= NBUCKETS,
333 CGETS(19, 7, "free(%p) bad block index."), cp);
334 size = op->ov_index;
335 op->ov_next = nextf[size];
336 nextf[size] = op;
338 nmalloc[size]--;
340 #else
341 if (cp == NULL)
342 return;
343 #endif
346 memalign_t
347 calloc(size_t i, size_t j)
349 #ifndef lint
350 char *cp;
352 i *= j;
353 cp = xmalloc(i);
354 memset(cp, 0, i);
356 return ((memalign_t) cp);
357 #else
358 if (i && j)
359 return ((memalign_t) 0);
360 else
361 return ((memalign_t) 0);
362 #endif
366 * When a program attempts "storage compaction" as mentioned in the
367 * old malloc man page, it realloc's an already freed block. Usually
368 * this is the last block it freed; occasionally it might be farther
369 * back. We have to search all the free lists for the block in order
370 * to determine its bucket: 1st we make one pass thru the lists
371 * checking only the first block in each; if that fails we search
372 * ``realloc_srchlen'' blocks in each list for a match (the variable
373 * is extern so the caller can modify it). If that fails we just copy
374 * however many bytes was given to realloc() and hope it's not huge.
376 #ifndef lint
377 /* 4 should be plenty, -1 =>'s whole list */
378 static int realloc_srchlen = 4;
379 #endif /* lint */
381 memalign_t
382 realloc(ptr_t cp, size_t nbytes)
384 #ifndef lint
385 U_int onb;
386 union overhead *op;
387 ptr_t res;
388 int i;
389 int was_alloced = 0;
391 if (cp == NULL)
392 return (malloc(nbytes));
393 op = (union overhead *) (((caddr_t) cp) - MEMALIGN(sizeof(union overhead)));
394 if (op->ov_magic == MAGIC) {
395 was_alloced++;
396 i = op->ov_index;
398 else
400 * Already free, doing "compaction".
402 * Search for the old block of memory on the free list. First, check the
403 * most common case (last element free'd), then (this failing) the last
404 * ``realloc_srchlen'' items free'd. If all lookups fail, then assume
405 * the size of the memory block being realloc'd is the smallest
406 * possible.
408 if ((i = findbucket(op, 1)) < 0 &&
409 (i = findbucket(op, realloc_srchlen)) < 0)
410 i = 0;
412 onb = MEMALIGN(nbytes + MEMALIGN(sizeof(union overhead)) + RSLOP);
414 /* avoid the copy if same size block */
415 if (was_alloced && (onb <= (U_int) (1 << (i + 3))) &&
416 (onb > (U_int) (1 << (i + 2)))) {
417 #ifdef RCHECK
418 /* JMR: formerly this wasn't updated ! */
419 nbytes = MEMALIGN(MEMALIGN(sizeof(union overhead))+nbytes+RSLOP);
420 *((U_int *) (((caddr_t) op) + nbytes - RSLOP)) = RMAGIC;
421 op->ov_rmagic = RMAGIC;
422 op->ov_size = (op->ov_index <= 13) ? nbytes - 1 : 0;
423 #endif
424 return ((memalign_t) cp);
426 if ((res = malloc(nbytes)) == NULL)
427 return ((memalign_t) NULL);
428 if (cp != res) { /* common optimization */
430 * christos: this used to copy nbytes! It should copy the
431 * smaller of the old and new size
433 onb = (1 << (i + 3)) - MEMALIGN(sizeof(union overhead)) - RSLOP;
434 (void) memmove(res, cp, onb < nbytes ? onb : nbytes);
436 if (was_alloced)
437 free(cp);
438 return ((memalign_t) res);
439 #else
440 if (cp && nbytes)
441 return ((memalign_t) 0);
442 else
443 return ((memalign_t) 0);
444 #endif /* !lint */
448 * On linux, _nss_nis_setnetgrent() calls this function to determine
449 * the usable size of the pointer passed, but this is not a portable
450 * API, so we cannot use our malloc replacement without providing one.
451 * Thanks a lot glibc!
453 #ifdef __linux__
454 #define M_U_S_CONST
455 #else
456 #define M_U_S_CONST
457 #endif
458 size_t malloc_usable_size(M_U_S_CONST void *);
459 size_t
460 malloc_usable_size(M_U_S_CONST void *ptr)
462 const union overhead *op = (const union overhead *)
463 (((const char *) ptr) - MEMALIGN(sizeof(*op)));
464 if (op->ov_magic == MAGIC)
465 return 1 << (op->ov_index + 3);
466 else
467 return 0;
471 #ifndef lint
473 * Search ``srchlen'' elements of each free list for a block whose
474 * header starts at ``freep''. If srchlen is -1 search the whole list.
475 * Return bucket number, or -1 if not found.
477 static int
478 findbucket(union overhead *freep, int srchlen)
480 union overhead *p;
481 size_t i;
482 int j;
484 for (i = 0; i < NBUCKETS; i++) {
485 j = 0;
486 for (p = nextf[i]; p && j != srchlen; p = p->ov_next) {
487 if (p == freep)
488 return (i);
489 j++;
492 return (-1);
495 #endif
498 #else /* SYSMALLOC */
501 ** ``Protected versions'' of malloc, realloc, calloc, and free
503 ** On many systems:
505 ** 1. malloc(0) is bad
506 ** 2. free(0) is bad
507 ** 3. realloc(0, n) is bad
508 ** 4. realloc(n, 0) is bad
510 ** Also we call our error routine if we run out of memory.
512 memalign_t
513 smalloc(size_t n)
515 ptr_t ptr;
517 n = n ? n : 1;
519 #ifdef USE_SBRK
520 if (membot == NULL)
521 membot = sbrk(0);
522 #endif /* USE_SBRK */
524 if ((ptr = malloc(n)) == NULL)
525 out_of_memory();
526 #ifndef USE_SBRK
527 if (memtop < ((char *) ptr) + n)
528 memtop = ((char *) ptr) + n;
529 if (membot == NULL)
530 membot = ptr;
531 #endif /* !USE_SBRK */
532 return ((memalign_t) ptr);
535 memalign_t
536 srealloc(ptr_t p, size_t n)
538 ptr_t ptr;
540 n = n ? n : 1;
542 #ifdef USE_SBRK
543 if (membot == NULL)
544 membot = sbrk(0);
545 #endif /* USE_SBRK */
547 if ((ptr = (p ? realloc(p, n) : malloc(n))) == NULL)
548 out_of_memory();
549 #ifndef USE_SBRK
550 if (memtop < ((char *) ptr) + n)
551 memtop = ((char *) ptr) + n;
552 if (membot == NULL)
553 membot = ptr;
554 #endif /* !USE_SBRK */
555 return ((memalign_t) ptr);
558 memalign_t
559 scalloc(size_t s, size_t n)
561 ptr_t ptr;
563 n *= s;
564 n = n ? n : 1;
566 #ifdef USE_SBRK
567 if (membot == NULL)
568 membot = sbrk(0);
569 #endif /* USE_SBRK */
571 if ((ptr = malloc(n)) == NULL)
572 out_of_memory();
574 memset (ptr, 0, n);
576 #ifndef USE_SBRK
577 if (memtop < ((char *) ptr) + n)
578 memtop = ((char *) ptr) + n;
579 if (membot == NULL)
580 membot = ptr;
581 #endif /* !USE_SBRK */
583 return ((memalign_t) ptr);
586 void
587 sfree(ptr_t p)
589 if (p && !dont_free)
590 free(p);
593 #endif /* SYSMALLOC */
596 * mstats - print out statistics about malloc
598 * Prints two lines of numbers, one showing the length of the free list
599 * for each size category, the second showing the number of mallocs -
600 * frees for each size category.
602 /*ARGSUSED*/
603 void
604 showall(Char **v, struct command *c)
606 #ifndef SYSMALLOC
607 size_t i, j;
608 union overhead *p;
609 int totfree = 0, totused = 0;
611 xprintf(CGETS(19, 8, "%s current memory allocation:\nfree:\t"), progname);
612 for (i = 0; i < NBUCKETS; i++) {
613 for (j = 0, p = nextf[i]; p; p = p->ov_next, j++)
614 continue;
615 xprintf(" %4zd", j);
616 totfree += j * (1 << (i + 3));
618 xprintf("\n%s:\t", CGETS(19, 9, "used"));
619 for (i = 0; i < NBUCKETS; i++) {
620 xprintf(" %4d", nmalloc[i]);
621 totused += nmalloc[i] * (1 << (i + 3));
623 xprintf(CGETS(19, 10, "\n\tTotal in use: %d, total free: %d\n"),
624 totused, totfree);
625 xprintf(CGETS(19, 11,
626 "\tAllocated memory from 0x%lx to 0x%lx. Real top at 0x%lx\n"),
627 (unsigned long) membot, (unsigned long) memtop,
628 (unsigned long) sbrk(0));
629 #else /* SYSMALLOC */
630 #ifndef HAVE_MALLINFO
631 #ifdef USE_SBRK
632 memtop = sbrk(0);
633 #endif /* USE_SBRK */
634 xprintf(CGETS(19, 12, "Allocated memory from 0x%lx to 0x%lx (%ld).\n"),
635 (unsigned long) membot, (unsigned long) memtop,
636 (unsigned long) (memtop - membot));
637 #else /* HAVE_MALLINFO */
638 struct mallinfo mi;
640 mi = mallinfo();
641 xprintf(CGETS(19, 13, "%s current memory allocation:\n"), progname);
642 xprintf(CGETS(19, 14, "Total space allocated from system: %d\n"), mi.arena);
643 xprintf(CGETS(19, 15, "Number of non-inuse chunks: %d\n"), mi.ordblks);
644 xprintf(CGETS(19, 16, "Number of mmapped regions: %d\n"), mi.hblks);
645 xprintf(CGETS(19, 17, "Total space in mmapped regions: %d\n"), mi.hblkhd);
646 xprintf(CGETS(19, 18, "Total allocated space: %d\n"), mi.uordblks);
647 xprintf(CGETS(19, 19, "Total non-inuse space: %d\n"), mi.fordblks);
648 xprintf(CGETS(19, 20, "Top-most, releasable space: %d\n"), mi.keepcost);
649 #endif /* HAVE_MALLINFO */
650 #endif /* SYSMALLOC */
651 USE(c);
652 USE(v);