1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 22 /* 23 * Copyright 2008 Sun Microsystems, Inc. All rights reserved. 24 * Use is subject to license terms. 25 */ 26 27 #pragma ident "%Z%%M% %I% %E% SMI" 28 29 /* Copyright (c) 1988 AT&T */ 30 /* All Rights Reserved */ 31 32 /* 33 * Memory management: malloc(), realloc(), free(). 34 * 35 * The following #-parameters may be redefined: 36 * SEGMENTED: if defined, memory requests are assumed to be 37 * non-contiguous across calls of GETCORE's. 38 * GETCORE: a function to get more core memory. If not SEGMENTED, 39 * GETCORE(0) is assumed to return the next available 40 * address. Default is 'sbrk'. 41 * ERRCORE: the error code as returned by GETCORE. 42 * Default is (char *)(-1). 43 * CORESIZE: a desired unit (measured in bytes) to be used 44 * with GETCORE. Default is (1024*ALIGN). 45 * 46 * This algorithm is based on a best fit strategy with lists of 47 * free elts maintained in a self-adjusting binary tree. Each list 48 * contains all elts of the same size. The tree is ordered by size. 49 * For results on self-adjusting trees, see the paper: 50 * Self-Adjusting Binary Trees, 51 * DD Sleator & RE Tarjan, JACM 1985. 52 * 53 * The header of a block contains the size of the data part in bytes. 54 * Since the size of a block is 0%4, the low two bits of the header 55 * are free and used as follows: 56 * 57 * BIT0: 1 for busy (block is in use), 0 for free. 58 * BIT1: if the block is busy, this bit is 1 if the 59 * preceding block in contiguous memory is free. 60 * Otherwise, it is always 0. 61 */ 62 63 #include "synonyms.h" 64 #include "mallint.h" 65 #include "mtlib.h" 66 67 /* 68 * Some abusers of the system (notably java1.2) acquire __malloc_lock 69 * in order to prevent threads from holding it while they are being 70 * suspended via thr_suspend() or thr_suspend_allmutators(). 71 * This never worked when alternate malloc() libraries were used 72 * because they don't use __malloc_lock for their locking strategy. 73 * We leave __malloc_lock as an external variable to satisfy these 74 * old programs, but we define a new lock, private to libc, to do the 75 * real locking: libc_malloc_lock. This puts libc's malloc() package 76 * on the same footing as all other malloc packages. 77 */ 78 mutex_t __malloc_lock = DEFAULTMUTEX; 79 mutex_t libc_malloc_lock = DEFAULTMUTEX; 80 81 static TREE *Root, /* root of the free tree */ 82 *Bottom, /* the last free chunk in the arena */ 83 *_morecore(size_t); /* function to get more core */ 84 85 static char *Baddr; /* current high address of the arena */ 86 static char *Lfree; /* last freed block with data intact */ 87 88 static void t_delete(TREE *); 89 static void t_splay(TREE *); 90 static void realfree(void *); 91 static void cleanfree(void *); 92 static void *_malloc_unlocked(size_t); 93 94 #define FREESIZE (1<<5) /* size for preserving free blocks until next malloc */ 95 #define FREEMASK FREESIZE-1 96 97 static void *flist[FREESIZE]; /* list of blocks to be freed on next malloc */ 98 static int freeidx; /* index of free blocks in flist % FREESIZE */ 99 100 /* 101 * Interfaces used only by atfork_init() functions. 102 */ 103 void 104 malloc_locks(void) 105 { 106 (void) mutex_lock(&libc_malloc_lock); 107 } 108 109 void 110 malloc_unlocks(void) 111 { 112 (void) mutex_unlock(&libc_malloc_lock); 113 } 114 115 /* 116 * Allocation of small blocks 117 */ 118 static TREE *List[MINSIZE/WORDSIZE-1]; /* lists of small blocks */ 119 120 static void * 121 _smalloc(size_t size) 122 { 123 TREE *tp; 124 size_t i; 125 126 ASSERT(size % WORDSIZE == 0); 127 /* want to return a unique pointer on malloc(0) */ 128 if (size == 0) 129 size = WORDSIZE; 130 131 /* list to use */ 132 i = size / WORDSIZE - 1; 133 134 if (List[i] == NULL) { 135 TREE *np; 136 int n; 137 /* number of blocks to get at one time */ 138 #define NPS (WORDSIZE*8) 139 ASSERT((size + WORDSIZE) * NPS >= MINSIZE); 140 141 /* get NPS of these block types */ 142 if ((List[i] = _malloc_unlocked((size + WORDSIZE) * NPS)) == 0) 143 return (0); 144 145 /* make them into a link list */ 146 for (n = 0, np = List[i]; n < NPS; ++n) { 147 tp = np; 148 SIZE(tp) = size; 149 np = NEXT(tp); 150 AFTER(tp) = np; 151 } 152 AFTER(tp) = NULL; 153 } 154 155 /* allocate from the head of the queue */ 156 tp = List[i]; 157 List[i] = AFTER(tp); 158 SETBIT0(SIZE(tp)); 159 return (DATA(tp)); 160 } 161 162 void * 163 malloc(size_t size) 164 { 165 void *ret; 166 167 if (!primary_link_map) { 168 errno = ENOTSUP; 169 return (NULL); 170 } 171 assert_no_libc_locks_held(); 172 (void) mutex_lock(&libc_malloc_lock); 173 ret = _malloc_unlocked(size); 174 (void) mutex_unlock(&libc_malloc_lock); 175 return (ret); 176 } 177 178 static void * 179 _malloc_unlocked(size_t size) 180 { 181 size_t n; 182 TREE *tp, *sp; 183 size_t o_bit1; 184 185 COUNT(nmalloc); 186 ASSERT(WORDSIZE == ALIGN); 187 188 /* check for size that could overflow calculations */ 189 if (size > MAX_MALLOC) { 190 errno = ENOMEM; 191 return (NULL); 192 } 193 194 /* make sure that size is 0 mod ALIGN */ 195 ROUND(size); 196 197 /* see if the last free block can be used */ 198 if (Lfree) { 199 sp = BLOCK(Lfree); 200 n = SIZE(sp); 201 CLRBITS01(n); 202 if (n == size) { 203 /* 204 * exact match, use it as is 205 */ 206 freeidx = (freeidx + FREESIZE - 1) & 207 FREEMASK; /* 1 back */ 208 flist[freeidx] = Lfree = NULL; 209 return (DATA(sp)); 210 } else if (size >= MINSIZE && n > size) { 211 /* 212 * got a big enough piece 213 */ 214 freeidx = (freeidx + FREESIZE - 1) & 215 FREEMASK; /* 1 back */ 216 flist[freeidx] = Lfree = NULL; 217 o_bit1 = SIZE(sp) & BIT1; 218 SIZE(sp) = n; 219 goto leftover; 220 } 221 } 222 o_bit1 = 0; 223 224 /* perform free's of space since last malloc */ 225 cleanfree(NULL); 226 227 /* small blocks */ 228 if (size < MINSIZE) 229 return (_smalloc(size)); 230 231 /* search for an elt of the right size */ 232 sp = NULL; 233 n = 0; 234 if (Root) { 235 tp = Root; 236 for (;;) { 237 /* branch left */ 238 if (SIZE(tp) >= size) { 239 if (n == 0 || n >= SIZE(tp)) { 240 sp = tp; 241 n = SIZE(tp); 242 } 243 if (LEFT(tp)) 244 tp = LEFT(tp); 245 else 246 break; 247 } else { /* branch right */ 248 if (RIGHT(tp)) 249 tp = RIGHT(tp); 250 else 251 break; 252 } 253 } 254 255 if (sp) { 256 t_delete(sp); 257 } else if (tp != Root) { 258 /* make the searched-to element the root */ 259 t_splay(tp); 260 Root = tp; 261 } 262 } 263 264 /* if found none fitted in the tree */ 265 if (!sp) { 266 if (Bottom && size <= SIZE(Bottom)) { 267 sp = Bottom; 268 CLRBITS01(SIZE(sp)); 269 } else if ((sp = _morecore(size)) == NULL) /* no more memory */ 270 return (NULL); 271 } 272 273 /* tell the forward neighbor that we're busy */ 274 CLRBIT1(SIZE(NEXT(sp))); 275 276 ASSERT(ISBIT0(SIZE(NEXT(sp)))); 277 278 leftover: 279 /* if the leftover is enough for a new free piece */ 280 if ((n = (SIZE(sp) - size)) >= MINSIZE + WORDSIZE) { 281 n -= WORDSIZE; 282 SIZE(sp) = size; 283 tp = NEXT(sp); 284 SIZE(tp) = n|BIT0; 285 realfree(DATA(tp)); 286 } else if (BOTTOM(sp)) 287 Bottom = NULL; 288 289 /* return the allocated space */ 290 SIZE(sp) |= BIT0 | o_bit1; 291 return (DATA(sp)); 292 } 293 294 295 /* 296 * realloc(). 297 * 298 * If the block size is increasing, we try forward merging first. 299 * This is not best-fit but it avoids some data recopying. 300 */ 301 void * 302 realloc(void *old, size_t size) 303 { 304 TREE *tp, *np; 305 size_t ts; 306 char *new; 307 308 if (!primary_link_map) { 309 errno = ENOTSUP; 310 return (NULL); 311 } 312 313 assert_no_libc_locks_held(); 314 COUNT(nrealloc); 315 316 /* check for size that could overflow calculations */ 317 if (size > MAX_MALLOC) { 318 errno = ENOMEM; 319 return (NULL); 320 } 321 322 /* pointer to the block */ 323 (void) mutex_lock(&libc_malloc_lock); 324 if (old == NULL) { 325 new = _malloc_unlocked(size); 326 (void) mutex_unlock(&libc_malloc_lock); 327 return (new); 328 } 329 330 /* perform free's of space since last malloc */ 331 cleanfree(old); 332 333 /* make sure that size is 0 mod ALIGN */ 334 ROUND(size); 335 336 tp = BLOCK(old); 337 ts = SIZE(tp); 338 339 /* if the block was freed, data has been destroyed. */ 340 if (!ISBIT0(ts)) { 341 (void) mutex_unlock(&libc_malloc_lock); 342 return (NULL); 343 } 344 345 /* nothing to do */ 346 CLRBITS01(SIZE(tp)); 347 if (size == SIZE(tp)) { 348 SIZE(tp) = ts; 349 (void) mutex_unlock(&libc_malloc_lock); 350 return (old); 351 } 352 353 /* special cases involving small blocks */ 354 if (size < MINSIZE || SIZE(tp) < MINSIZE) { 355 /* free is size is zero */ 356 if (size == 0) { 357 SETOLD01(SIZE(tp), ts); 358 _free_unlocked(old); 359 (void) mutex_unlock(&libc_malloc_lock); 360 return (NULL); 361 } else { 362 goto call_malloc; 363 } 364 } 365 366 /* block is increasing in size, try merging the next block */ 367 if (size > SIZE(tp)) { 368 np = NEXT(tp); 369 if (!ISBIT0(SIZE(np))) { 370 ASSERT(SIZE(np) >= MINSIZE); 371 ASSERT(!ISBIT1(SIZE(np))); 372 SIZE(tp) += SIZE(np) + WORDSIZE; 373 if (np != Bottom) 374 t_delete(np); 375 else 376 Bottom = NULL; 377 CLRBIT1(SIZE(NEXT(np))); 378 } 379 380 #ifndef SEGMENTED 381 /* not enough & at TRUE end of memory, try extending core */ 382 if (size > SIZE(tp) && BOTTOM(tp) && GETCORE(0) == Baddr) { 383 Bottom = tp; 384 if ((tp = _morecore(size)) == NULL) { 385 tp = Bottom; 386 Bottom = NULL; 387 } 388 } 389 #endif 390 } 391 392 /* got enough space to use */ 393 if (size <= SIZE(tp)) { 394 size_t n; 395 396 chop_big: 397 if ((n = (SIZE(tp) - size)) >= MINSIZE + WORDSIZE) { 398 n -= WORDSIZE; 399 SIZE(tp) = size; 400 np = NEXT(tp); 401 SIZE(np) = n|BIT0; 402 realfree(DATA(np)); 403 } else if (BOTTOM(tp)) 404 Bottom = NULL; 405 406 /* the previous block may be free */ 407 SETOLD01(SIZE(tp), ts); 408 (void) mutex_unlock(&libc_malloc_lock); 409 return (old); 410 } 411 412 /* call malloc to get a new block */ 413 call_malloc: 414 SETOLD01(SIZE(tp), ts); 415 if ((new = _malloc_unlocked(size)) != NULL) { 416 CLRBITS01(ts); 417 if (ts > size) 418 ts = size; 419 MEMCOPY(new, old, ts); 420 _free_unlocked(old); 421 (void) mutex_unlock(&libc_malloc_lock); 422 return (new); 423 } 424 425 /* 426 * Attempt special case recovery allocations since malloc() failed: 427 * 428 * 1. size <= SIZE(tp) < MINSIZE 429 * Simply return the existing block 430 * 2. SIZE(tp) < size < MINSIZE 431 * malloc() may have failed to allocate the chunk of 432 * small blocks. Try asking for MINSIZE bytes. 433 * 3. size < MINSIZE <= SIZE(tp) 434 * malloc() may have failed as with 2. Change to 435 * MINSIZE allocation which is taken from the beginning 436 * of the current block. 437 * 4. MINSIZE <= SIZE(tp) < size 438 * If the previous block is free and the combination of 439 * these two blocks has at least size bytes, then merge 440 * the two blocks copying the existing contents backwards. 441 */ 442 CLRBITS01(SIZE(tp)); 443 if (SIZE(tp) < MINSIZE) { 444 if (size < SIZE(tp)) { /* case 1. */ 445 SETOLD01(SIZE(tp), ts); 446 (void) mutex_unlock(&libc_malloc_lock); 447 return (old); 448 } else if (size < MINSIZE) { /* case 2. */ 449 size = MINSIZE; 450 goto call_malloc; 451 } 452 } else if (size < MINSIZE) { /* case 3. */ 453 size = MINSIZE; 454 goto chop_big; 455 } else if (ISBIT1(ts) && 456 (SIZE(np = LAST(tp)) + SIZE(tp) + WORDSIZE) >= size) { 457 ASSERT(!ISBIT0(SIZE(np))); 458 t_delete(np); 459 SIZE(np) += SIZE(tp) + WORDSIZE; 460 /* 461 * Since the copy may overlap, use memmove() if available. 462 * Otherwise, copy by hand. 463 */ 464 (void) memmove(DATA(np), old, SIZE(tp)); 465 old = DATA(np); 466 tp = np; 467 CLRBIT1(ts); 468 goto chop_big; 469 } 470 SETOLD01(SIZE(tp), ts); 471 (void) mutex_unlock(&libc_malloc_lock); 472 return (NULL); 473 } 474 475 476 /* 477 * realfree(). 478 * 479 * Coalescing of adjacent free blocks is done first. 480 * Then, the new free block is leaf-inserted into the free tree 481 * without splaying. This strategy does not guarantee the amortized 482 * O(nlogn) behaviour for the insert/delete/find set of operations 483 * on the tree. In practice, however, free is much more infrequent 484 * than malloc/realloc and the tree searches performed by these 485 * functions adequately keep the tree in balance. 486 */ 487 static void 488 realfree(void *old) 489 { 490 TREE *tp, *sp, *np; 491 size_t ts, size; 492 493 COUNT(nfree); 494 495 /* pointer to the block */ 496 tp = BLOCK(old); 497 ts = SIZE(tp); 498 if (!ISBIT0(ts)) 499 return; 500 CLRBITS01(SIZE(tp)); 501 502 /* small block, put it in the right linked list */ 503 if (SIZE(tp) < MINSIZE) { 504 ASSERT(SIZE(tp) / WORDSIZE >= 1); 505 ts = SIZE(tp) / WORDSIZE - 1; 506 AFTER(tp) = List[ts]; 507 List[ts] = tp; 508 return; 509 } 510 511 /* see if coalescing with next block is warranted */ 512 np = NEXT(tp); 513 if (!ISBIT0(SIZE(np))) { 514 if (np != Bottom) 515 t_delete(np); 516 SIZE(tp) += SIZE(np) + WORDSIZE; 517 } 518 519 /* the same with the preceding block */ 520 if (ISBIT1(ts)) { 521 np = LAST(tp); 522 ASSERT(!ISBIT0(SIZE(np))); 523 ASSERT(np != Bottom); 524 t_delete(np); 525 SIZE(np) += SIZE(tp) + WORDSIZE; 526 tp = np; 527 } 528 529 /* initialize tree info */ 530 PARENT(tp) = LEFT(tp) = RIGHT(tp) = LINKFOR(tp) = NULL; 531 532 /* the last word of the block contains self's address */ 533 *(SELFP(tp)) = tp; 534 535 /* set bottom block, or insert in the free tree */ 536 if (BOTTOM(tp)) 537 Bottom = tp; 538 else { 539 /* search for the place to insert */ 540 if (Root) { 541 size = SIZE(tp); 542 np = Root; 543 for (;;) { 544 if (SIZE(np) > size) { 545 if (LEFT(np)) 546 np = LEFT(np); 547 else { 548 LEFT(np) = tp; 549 PARENT(tp) = np; 550 break; 551 } 552 } else if (SIZE(np) < size) { 553 if (RIGHT(np)) 554 np = RIGHT(np); 555 else { 556 RIGHT(np) = tp; 557 PARENT(tp) = np; 558 break; 559 } 560 } else { 561 if ((sp = PARENT(np)) != NULL) { 562 if (np == LEFT(sp)) 563 LEFT(sp) = tp; 564 else 565 RIGHT(sp) = tp; 566 PARENT(tp) = sp; 567 } else 568 Root = tp; 569 570 /* insert to head of list */ 571 if ((sp = LEFT(np)) != NULL) 572 PARENT(sp) = tp; 573 LEFT(tp) = sp; 574 575 if ((sp = RIGHT(np)) != NULL) 576 PARENT(sp) = tp; 577 RIGHT(tp) = sp; 578 579 /* doubly link list */ 580 LINKFOR(tp) = np; 581 LINKBAK(np) = tp; 582 SETNOTREE(np); 583 584 break; 585 } 586 } 587 } else 588 Root = tp; 589 } 590 591 /* tell next block that this one is free */ 592 SETBIT1(SIZE(NEXT(tp))); 593 594 ASSERT(ISBIT0(SIZE(NEXT(tp)))); 595 } 596 597 /* 598 * Get more core. Gaps in memory are noted as busy blocks. 599 */ 600 static TREE * 601 _morecore(size_t size) 602 { 603 TREE *tp; 604 ssize_t n, offset; 605 char *addr; 606 ssize_t nsize; 607 608 /* compute new amount of memory to get */ 609 tp = Bottom; 610 n = (ssize_t)size + 2 * WORDSIZE; 611 addr = GETCORE(0); 612 613 if (addr == ERRCORE) 614 return (NULL); 615 616 /* need to pad size out so that addr is aligned */ 617 if ((((uintptr_t)addr) % ALIGN) != 0) 618 offset = ALIGN - (uintptr_t)addr % ALIGN; 619 else 620 offset = 0; 621 622 #ifndef SEGMENTED 623 /* if not segmented memory, what we need may be smaller */ 624 if (addr == Baddr) { 625 n -= WORDSIZE; 626 if (tp != NULL) 627 n -= SIZE(tp); 628 } 629 #endif 630 631 /* get a multiple of CORESIZE */ 632 n = ((n - 1) / CORESIZE + 1) * CORESIZE; 633 nsize = n + offset; 634 635 /* check if nsize request could overflow in GETCORE */ 636 if ((size_t)nsize > MAX_MALLOC - (uintptr_t)addr) { 637 errno = ENOMEM; 638 return (NULL); 639 } 640 641 if ((size_t)nsize <= MAX_GETCORE) { 642 if (GETCORE(nsize) == ERRCORE) 643 return (NULL); 644 } else { 645 intptr_t delta; 646 /* 647 * the value required is too big for GETCORE() to deal with 648 * in one go, so use GETCORE() at most 2 times instead. 649 * Argument to GETCORE() must be multiple of ALIGN. 650 * If not, GETCORE(-MAX_GETCORE) will not return brk point 651 * to previous value, but will be ALIGN more. 652 * This would leave a small hole. 653 */ 654 delta = MAX_GETCORE; 655 while (delta > 0) { 656 if (GETCORE(delta) == ERRCORE) { 657 if (addr != GETCORE(0)) 658 (void) GETCORE(-MAX_GETCORE); 659 return (NULL); 660 } 661 nsize -= MAX_GETCORE; 662 delta = nsize; 663 } 664 } 665 666 /* contiguous memory */ 667 if (addr == Baddr) { 668 ASSERT(offset == 0); 669 if (tp) { 670 addr = (char *)tp; 671 n += SIZE(tp) + 2 * WORDSIZE; 672 } else { 673 addr = Baddr - WORDSIZE; 674 n += WORDSIZE; 675 } 676 } else 677 addr += offset; 678 679 /* new bottom address */ 680 Baddr = addr + n; 681 682 /* new bottom block */ 683 tp = (TREE *)(uintptr_t)addr; 684 SIZE(tp) = n - 2 * WORDSIZE; 685 ASSERT((SIZE(tp) % ALIGN) == 0); 686 687 /* reserved the last word to head any noncontiguous memory */ 688 SETBIT0(SIZE(NEXT(tp))); 689 690 /* non-contiguous memory, free old bottom block */ 691 if (Bottom && Bottom != tp) { 692 SETBIT0(SIZE(Bottom)); 693 realfree(DATA(Bottom)); 694 } 695 696 return (tp); 697 } 698 699 700 /* 701 * Tree rotation functions (BU: bottom-up, TD: top-down) 702 */ 703 704 #define LEFT1(x, y) \ 705 if ((RIGHT(x) = LEFT(y)) != NULL) PARENT(RIGHT(x)) = x;\ 706 if ((PARENT(y) = PARENT(x)) != NULL)\ 707 if (LEFT(PARENT(x)) == x) LEFT(PARENT(y)) = y;\ 708 else RIGHT(PARENT(y)) = y;\ 709 LEFT(y) = x; PARENT(x) = y 710 711 #define RIGHT1(x, y) \ 712 if ((LEFT(x) = RIGHT(y)) != NULL) PARENT(LEFT(x)) = x;\ 713 if ((PARENT(y) = PARENT(x)) != NULL)\ 714 if (LEFT(PARENT(x)) == x) LEFT(PARENT(y)) = y;\ 715 else RIGHT(PARENT(y)) = y;\ 716 RIGHT(y) = x; PARENT(x) = y 717 718 #define BULEFT2(x, y, z) \ 719 if ((RIGHT(x) = LEFT(y)) != NULL) PARENT(RIGHT(x)) = x;\ 720 if ((RIGHT(y) = LEFT(z)) != NULL) PARENT(RIGHT(y)) = y;\ 721 if ((PARENT(z) = PARENT(x)) != NULL)\ 722 if (LEFT(PARENT(x)) == x) LEFT(PARENT(z)) = z;\ 723 else RIGHT(PARENT(z)) = z;\ 724 LEFT(z) = y; PARENT(y) = z; LEFT(y) = x; PARENT(x) = y 725 726 #define BURIGHT2(x, y, z) \ 727 if ((LEFT(x) = RIGHT(y)) != NULL) PARENT(LEFT(x)) = x;\ 728 if ((LEFT(y) = RIGHT(z)) != NULL) PARENT(LEFT(y)) = y;\ 729 if ((PARENT(z) = PARENT(x)) != NULL)\ 730 if (LEFT(PARENT(x)) == x) LEFT(PARENT(z)) = z;\ 731 else RIGHT(PARENT(z)) = z;\ 732 RIGHT(z) = y; PARENT(y) = z; RIGHT(y) = x; PARENT(x) = y 733 734 #define TDLEFT2(x, y, z) \ 735 if ((RIGHT(y) = LEFT(z)) != NULL) PARENT(RIGHT(y)) = y;\ 736 if ((PARENT(z) = PARENT(x)) != NULL)\ 737 if (LEFT(PARENT(x)) == x) LEFT(PARENT(z)) = z;\ 738 else RIGHT(PARENT(z)) = z;\ 739 PARENT(x) = z; LEFT(z) = x; 740 741 #define TDRIGHT2(x, y, z) \ 742 if ((LEFT(y) = RIGHT(z)) != NULL) PARENT(LEFT(y)) = y;\ 743 if ((PARENT(z) = PARENT(x)) != NULL)\ 744 if (LEFT(PARENT(x)) == x) LEFT(PARENT(z)) = z;\ 745 else RIGHT(PARENT(z)) = z;\ 746 PARENT(x) = z; RIGHT(z) = x; 747 748 /* 749 * Delete a tree element 750 */ 751 static void 752 t_delete(TREE *op) 753 { 754 TREE *tp, *sp, *gp; 755 756 /* if this is a non-tree node */ 757 if (ISNOTREE(op)) { 758 tp = LINKBAK(op); 759 if ((sp = LINKFOR(op)) != NULL) 760 LINKBAK(sp) = tp; 761 LINKFOR(tp) = sp; 762 return; 763 } 764 765 /* make op the root of the tree */ 766 if (PARENT(op)) 767 t_splay(op); 768 769 /* if this is the start of a list */ 770 if ((tp = LINKFOR(op)) != NULL) { 771 PARENT(tp) = NULL; 772 if ((sp = LEFT(op)) != NULL) 773 PARENT(sp) = tp; 774 LEFT(tp) = sp; 775 776 if ((sp = RIGHT(op)) != NULL) 777 PARENT(sp) = tp; 778 RIGHT(tp) = sp; 779 780 Root = tp; 781 return; 782 } 783 784 /* if op has a non-null left subtree */ 785 if ((tp = LEFT(op)) != NULL) { 786 PARENT(tp) = NULL; 787 788 if (RIGHT(op)) { 789 /* make the right-end of the left subtree its root */ 790 while ((sp = RIGHT(tp)) != NULL) { 791 if ((gp = RIGHT(sp)) != NULL) { 792 TDLEFT2(tp, sp, gp); 793 tp = gp; 794 } else { 795 LEFT1(tp, sp); 796 tp = sp; 797 } 798 } 799 800 /* hook the right subtree of op to the above elt */ 801 RIGHT(tp) = RIGHT(op); 802 PARENT(RIGHT(tp)) = tp; 803 } 804 } else if ((tp = RIGHT(op)) != NULL) /* no left subtree */ 805 PARENT(tp) = NULL; 806 807 Root = tp; 808 } 809 810 /* 811 * Bottom up splaying (simple version). 812 * The basic idea is to roughly cut in half the 813 * path from Root to tp and make tp the new root. 814 */ 815 static void 816 t_splay(TREE *tp) 817 { 818 TREE *pp, *gp; 819 820 /* iterate until tp is the root */ 821 while ((pp = PARENT(tp)) != NULL) { 822 /* grandparent of tp */ 823 gp = PARENT(pp); 824 825 /* x is a left child */ 826 if (LEFT(pp) == tp) { 827 if (gp && LEFT(gp) == pp) { 828 BURIGHT2(gp, pp, tp); 829 } else { 830 RIGHT1(pp, tp); 831 } 832 } else { 833 ASSERT(RIGHT(pp) == tp); 834 if (gp && RIGHT(gp) == pp) { 835 BULEFT2(gp, pp, tp); 836 } else { 837 LEFT1(pp, tp); 838 } 839 } 840 } 841 } 842 843 844 /* 845 * free(). 846 * Performs a delayed free of the block pointed to 847 * by old. The pointer to old is saved on a list, flist, 848 * until the next malloc or realloc. At that time, all the 849 * blocks pointed to in flist are actually freed via 850 * realfree(). This allows the contents of free blocks to 851 * remain undisturbed until the next malloc or realloc. 852 */ 853 void 854 free(void *old) 855 { 856 if (!primary_link_map) { 857 errno = ENOTSUP; 858 return; 859 } 860 assert_no_libc_locks_held(); 861 (void) mutex_lock(&libc_malloc_lock); 862 _free_unlocked(old); 863 (void) mutex_unlock(&libc_malloc_lock); 864 } 865 866 867 void 868 _free_unlocked(void *old) 869 { 870 int i; 871 872 if (old == NULL) 873 return; 874 875 /* 876 * Make sure the same data block is not freed twice. 877 * 3 cases are checked. It returns immediately if either 878 * one of the conditions is true. 879 * 1. Last freed. 880 * 2. Not in use or freed already. 881 * 3. In the free list. 882 */ 883 if (old == Lfree) 884 return; 885 if (!ISBIT0(SIZE(BLOCK(old)))) 886 return; 887 for (i = 0; i < freeidx; i++) 888 if (old == flist[i]) 889 return; 890 891 if (flist[freeidx] != NULL) 892 realfree(flist[freeidx]); 893 flist[freeidx] = Lfree = old; 894 freeidx = (freeidx + 1) & FREEMASK; /* one forward */ 895 } 896 897 /* 898 * cleanfree() frees all the blocks pointed to be flist. 899 * 900 * realloc() should work if it is called with a pointer 901 * to a block that was freed since the last call to malloc() or 902 * realloc(). If cleanfree() is called from realloc(), ptr 903 * is set to the old block and that block should not be 904 * freed since it is actually being reallocated. 905 */ 906 static void 907 cleanfree(void *ptr) 908 { 909 char **flp; 910 911 flp = (char **)&(flist[freeidx]); 912 for (;;) { 913 if (flp == (char **)&(flist[0])) 914 flp = (char **)&(flist[FREESIZE]); 915 if (*--flp == NULL) 916 break; 917 if (*flp != ptr) 918 realfree(*flp); 919 *flp = NULL; 920 } 921 freeidx = 0; 922 Lfree = NULL; 923 } 924