1 /*- 2 * Copyright (c) 1987, 1991, 1993 3 * The Regents of the University of California. All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright 9 * notice, this list of conditions and the following disclaimer. 10 * 2. Redistributions in binary form must reproduce the above copyright 11 * notice, this list of conditions and the following disclaimer in the 12 * documentation and/or other materials provided with the distribution. 13 * 4. Neither the name of the University nor the names of its contributors 14 * may be used to endorse or promote products derived from this software 15 * without specific prior written permission. 16 * 17 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND 18 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 19 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 20 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE 21 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 22 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 23 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 24 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 25 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 26 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 27 * SUCH DAMAGE. 28 * 29 * @(#)kern_malloc.c 8.3 (Berkeley) 1/4/94 30 */ 31 32 #include <sys/cdefs.h> 33 __FBSDID("$FreeBSD$"); 34 35 #include "opt_vm.h" 36 37 #include <sys/param.h> 38 #include <sys/systm.h> 39 #include <sys/kdb.h> 40 #include <sys/kernel.h> 41 #include <sys/lock.h> 42 #include <sys/malloc.h> 43 #include <sys/mbuf.h> 44 #include <sys/mutex.h> 45 #include <sys/vmmeter.h> 46 #include <sys/proc.h> 47 #include <sys/sysctl.h> 48 #include <sys/time.h> 49 50 #include <vm/vm.h> 51 #include <vm/pmap.h> 52 #include <vm/vm_param.h> 53 #include <vm/vm_kern.h> 54 #include <vm/vm_extern.h> 55 #include <vm/vm_map.h> 56 #include <vm/vm_page.h> 57 #include <vm/uma.h> 58 #include <vm/uma_int.h> 59 #include <vm/uma_dbg.h> 60 61 #ifdef DEBUG_MEMGUARD 62 #include <vm/memguard.h> 63 #endif 64 65 #if defined(INVARIANTS) && defined(__i386__) 66 #include <machine/cpu.h> 67 #endif 68 69 /* 70 * When realloc() is called, if the new size is sufficiently smaller than 71 * the old size, realloc() will allocate a new, smaller block to avoid 72 * wasting memory. 'Sufficiently smaller' is defined as: newsize <= 73 * oldsize / 2^n, where REALLOC_FRACTION defines the value of 'n'. 74 */ 75 #ifndef REALLOC_FRACTION 76 #define REALLOC_FRACTION 1 /* new block if <= half the size */ 77 #endif 78 79 MALLOC_DEFINE(M_CACHE, "cache", "Various Dynamically allocated caches"); 80 MALLOC_DEFINE(M_DEVBUF, "devbuf", "device driver memory"); 81 MALLOC_DEFINE(M_TEMP, "temp", "misc temporary data buffers"); 82 83 MALLOC_DEFINE(M_IP6OPT, "ip6opt", "IPv6 options"); 84 MALLOC_DEFINE(M_IP6NDP, "ip6ndp", "IPv6 Neighbor Discovery"); 85 86 static void kmeminit(void *); 87 SYSINIT(kmem, SI_SUB_KMEM, SI_ORDER_FIRST, kmeminit, NULL) 88 89 static MALLOC_DEFINE(M_FREE, "free", "should be on free list"); 90 91 static struct malloc_type *kmemstatistics; 92 static char *kmembase; 93 static char *kmemlimit; 94 95 #define KMEM_ZSHIFT 4 96 #define KMEM_ZBASE 16 97 #define KMEM_ZMASK (KMEM_ZBASE - 1) 98 99 #define KMEM_ZMAX PAGE_SIZE 100 #define KMEM_ZSIZE (KMEM_ZMAX >> KMEM_ZSHIFT) 101 static u_int8_t kmemsize[KMEM_ZSIZE + 1]; 102 103 /* These won't be powers of two for long */ 104 struct { 105 int kz_size; 106 char *kz_name; 107 uma_zone_t kz_zone; 108 } kmemzones[] = { 109 {16, "16", NULL}, 110 {32, "32", NULL}, 111 {64, "64", NULL}, 112 {128, "128", NULL}, 113 {256, "256", NULL}, 114 {512, "512", NULL}, 115 {1024, "1024", NULL}, 116 {2048, "2048", NULL}, 117 {4096, "4096", NULL}, 118 #if PAGE_SIZE > 4096 119 {8192, "8192", NULL}, 120 #if PAGE_SIZE > 8192 121 {16384, "16384", NULL}, 122 #if PAGE_SIZE > 16384 123 {32768, "32768", NULL}, 124 #if PAGE_SIZE > 32768 125 {65536, "65536", NULL}, 126 #if PAGE_SIZE > 65536 127 #error "Unsupported PAGE_SIZE" 128 #endif /* 65536 */ 129 #endif /* 32768 */ 130 #endif /* 16384 */ 131 #endif /* 8192 */ 132 #endif /* 4096 */ 133 {0, NULL}, 134 }; 135 136 #ifdef DEBUG_MEMGUARD 137 u_int vm_memguard_divisor; 138 SYSCTL_UINT(_vm, OID_AUTO, memguard_divisor, CTLFLAG_RD, &vm_memguard_divisor, 139 0, "(kmem_size/memguard_divisor) == memguard submap size"); 140 #endif 141 142 u_int vm_kmem_size; 143 SYSCTL_UINT(_vm, OID_AUTO, kmem_size, CTLFLAG_RD, &vm_kmem_size, 0, 144 "Size of kernel memory"); 145 146 u_int vm_kmem_size_max; 147 SYSCTL_UINT(_vm, OID_AUTO, kmem_size_max, CTLFLAG_RD, &vm_kmem_size_max, 0, 148 "Maximum size of kernel memory"); 149 150 u_int vm_kmem_size_scale; 151 SYSCTL_UINT(_vm, OID_AUTO, kmem_size_scale, CTLFLAG_RD, &vm_kmem_size_scale, 0, 152 "Scale factor for kernel memory size"); 153 154 /* 155 * The malloc_mtx protects the kmemstatistics linked list. 156 */ 157 158 struct mtx malloc_mtx; 159 160 #ifdef MALLOC_PROFILE 161 uint64_t krequests[KMEM_ZSIZE + 1]; 162 163 static int sysctl_kern_mprof(SYSCTL_HANDLER_ARGS); 164 #endif 165 166 static int sysctl_kern_malloc(SYSCTL_HANDLER_ARGS); 167 168 /* time_uptime of last malloc(9) failure */ 169 static time_t t_malloc_fail; 170 171 #ifdef MALLOC_MAKE_FAILURES 172 /* 173 * Causes malloc failures every (n) mallocs with M_NOWAIT. If set to 0, 174 * doesn't cause failures. 175 */ 176 SYSCTL_NODE(_debug, OID_AUTO, malloc, CTLFLAG_RD, 0, 177 "Kernel malloc debugging options"); 178 179 static int malloc_failure_rate; 180 static int malloc_nowait_count; 181 static int malloc_failure_count; 182 SYSCTL_INT(_debug_malloc, OID_AUTO, failure_rate, CTLFLAG_RW, 183 &malloc_failure_rate, 0, "Every (n) mallocs with M_NOWAIT will fail"); 184 TUNABLE_INT("debug.malloc.failure_rate", &malloc_failure_rate); 185 SYSCTL_INT(_debug_malloc, OID_AUTO, failure_count, CTLFLAG_RD, 186 &malloc_failure_count, 0, "Number of imposed M_NOWAIT malloc failures"); 187 #endif 188 189 int 190 malloc_last_fail(void) 191 { 192 193 return (time_uptime - t_malloc_fail); 194 } 195 196 /* 197 * Add this to the informational malloc_type bucket. 198 */ 199 static void 200 malloc_type_zone_allocated(struct malloc_type *ksp, unsigned long size, 201 int zindx) 202 { 203 mtx_lock(&ksp->ks_mtx); 204 ksp->ks_calls++; 205 if (zindx != -1) 206 ksp->ks_size |= 1 << zindx; 207 if (size != 0) { 208 ksp->ks_memuse += size; 209 ksp->ks_inuse++; 210 if (ksp->ks_memuse > ksp->ks_maxused) 211 ksp->ks_maxused = ksp->ks_memuse; 212 } 213 mtx_unlock(&ksp->ks_mtx); 214 } 215 216 void 217 malloc_type_allocated(struct malloc_type *ksp, unsigned long size) 218 { 219 malloc_type_zone_allocated(ksp, size, -1); 220 } 221 222 /* 223 * Remove this allocation from the informational malloc_type bucket. 224 */ 225 void 226 malloc_type_freed(struct malloc_type *ksp, unsigned long size) 227 { 228 mtx_lock(&ksp->ks_mtx); 229 KASSERT(size <= ksp->ks_memuse, 230 ("malloc(9)/free(9) confusion.\n%s", 231 "Probably freeing with wrong type, but maybe not here.")); 232 ksp->ks_memuse -= size; 233 ksp->ks_inuse--; 234 mtx_unlock(&ksp->ks_mtx); 235 } 236 237 /* 238 * malloc: 239 * 240 * Allocate a block of memory. 241 * 242 * If M_NOWAIT is set, this routine will not block and return NULL if 243 * the allocation fails. 244 */ 245 void * 246 malloc(size, type, flags) 247 unsigned long size; 248 struct malloc_type *type; 249 int flags; 250 { 251 int indx; 252 caddr_t va; 253 uma_zone_t zone; 254 uma_keg_t keg; 255 #ifdef DIAGNOSTIC 256 unsigned long osize = size; 257 #endif 258 259 #ifdef INVARIANTS 260 /* 261 * To make sure that WAITOK or NOWAIT is set, but not more than 262 * one, and check against the API botches that are common. 263 */ 264 indx = flags & (M_WAITOK | M_NOWAIT | M_DONTWAIT | M_TRYWAIT); 265 if (indx != M_NOWAIT && indx != M_WAITOK) { 266 static struct timeval lasterr; 267 static int curerr, once; 268 if (once == 0 && ppsratecheck(&lasterr, &curerr, 1)) { 269 printf("Bad malloc flags: %x\n", indx); 270 kdb_backtrace(); 271 flags |= M_WAITOK; 272 once++; 273 } 274 } 275 #endif 276 #if 0 277 if (size == 0) 278 kdb_enter("zero size malloc"); 279 #endif 280 #ifdef MALLOC_MAKE_FAILURES 281 if ((flags & M_NOWAIT) && (malloc_failure_rate != 0)) { 282 atomic_add_int(&malloc_nowait_count, 1); 283 if ((malloc_nowait_count % malloc_failure_rate) == 0) { 284 atomic_add_int(&malloc_failure_count, 1); 285 t_malloc_fail = time_uptime; 286 return (NULL); 287 } 288 } 289 #endif 290 if (flags & M_WAITOK) 291 KASSERT(curthread->td_intr_nesting_level == 0, 292 ("malloc(M_WAITOK) in interrupt context")); 293 294 #ifdef DEBUG_MEMGUARD 295 /* XXX CHANGEME! */ 296 if (type == M_SUBPROC) 297 return memguard_alloc(size, flags); 298 #endif 299 300 if (size <= KMEM_ZMAX) { 301 if (size & KMEM_ZMASK) 302 size = (size & ~KMEM_ZMASK) + KMEM_ZBASE; 303 indx = kmemsize[size >> KMEM_ZSHIFT]; 304 zone = kmemzones[indx].kz_zone; 305 keg = zone->uz_keg; 306 #ifdef MALLOC_PROFILE 307 krequests[size >> KMEM_ZSHIFT]++; 308 #endif 309 va = uma_zalloc(zone, flags); 310 if (va != NULL) 311 size = keg->uk_size; 312 malloc_type_zone_allocated(type, va == NULL ? 0 : size, indx); 313 } else { 314 size = roundup(size, PAGE_SIZE); 315 zone = NULL; 316 keg = NULL; 317 va = uma_large_malloc(size, flags); 318 malloc_type_allocated(type, va == NULL ? 0 : size); 319 } 320 if (flags & M_WAITOK) 321 KASSERT(va != NULL, ("malloc(M_WAITOK) returned NULL")); 322 else if (va == NULL) 323 t_malloc_fail = time_uptime; 324 #ifdef DIAGNOSTIC 325 if (va != NULL && !(flags & M_ZERO)) { 326 memset(va, 0x70, osize); 327 } 328 #endif 329 return ((void *) va); 330 } 331 332 /* 333 * free: 334 * 335 * Free a block of memory allocated by malloc. 336 * 337 * This routine may not block. 338 */ 339 void 340 free(addr, type) 341 void *addr; 342 struct malloc_type *type; 343 { 344 uma_slab_t slab; 345 u_long size; 346 347 /* free(NULL, ...) does nothing */ 348 if (addr == NULL) 349 return; 350 351 #ifdef DEBUG_MEMGUARD 352 /* XXX CHANGEME! */ 353 if (type == M_SUBPROC) { 354 memguard_free(addr); 355 return; 356 } 357 #endif 358 359 KASSERT(type->ks_memuse > 0, 360 ("malloc(9)/free(9) confusion.\n%s", 361 "Probably freeing with wrong type, but maybe not here.")); 362 size = 0; 363 364 slab = vtoslab((vm_offset_t)addr & (~UMA_SLAB_MASK)); 365 366 if (slab == NULL) 367 panic("free: address %p(%p) has not been allocated.\n", 368 addr, (void *)((u_long)addr & (~UMA_SLAB_MASK))); 369 370 371 if (!(slab->us_flags & UMA_SLAB_MALLOC)) { 372 #ifdef INVARIANTS 373 struct malloc_type **mtp = addr; 374 #endif 375 size = slab->us_keg->uk_size; 376 #ifdef INVARIANTS 377 /* 378 * Cache a pointer to the malloc_type that most recently freed 379 * this memory here. This way we know who is most likely to 380 * have stepped on it later. 381 * 382 * This code assumes that size is a multiple of 8 bytes for 383 * 64 bit machines 384 */ 385 mtp = (struct malloc_type **) 386 ((unsigned long)mtp & ~UMA_ALIGN_PTR); 387 mtp += (size - sizeof(struct malloc_type *)) / 388 sizeof(struct malloc_type *); 389 *mtp = type; 390 #endif 391 uma_zfree_arg(LIST_FIRST(&slab->us_keg->uk_zones), addr, slab); 392 } else { 393 size = slab->us_size; 394 uma_large_free(slab); 395 } 396 malloc_type_freed(type, size); 397 } 398 399 /* 400 * realloc: change the size of a memory block 401 */ 402 void * 403 realloc(addr, size, type, flags) 404 void *addr; 405 unsigned long size; 406 struct malloc_type *type; 407 int flags; 408 { 409 uma_slab_t slab; 410 unsigned long alloc; 411 void *newaddr; 412 413 /* realloc(NULL, ...) is equivalent to malloc(...) */ 414 if (addr == NULL) 415 return (malloc(size, type, flags)); 416 417 #ifdef DEBUG_MEMGUARD 418 /* XXX: CHANGEME! */ 419 if (type == M_SUBPROC) { 420 slab = NULL; 421 alloc = size; 422 } else { 423 #endif 424 425 slab = vtoslab((vm_offset_t)addr & ~(UMA_SLAB_MASK)); 426 427 /* Sanity check */ 428 KASSERT(slab != NULL, 429 ("realloc: address %p out of range", (void *)addr)); 430 431 /* Get the size of the original block */ 432 if (slab->us_keg) 433 alloc = slab->us_keg->uk_size; 434 else 435 alloc = slab->us_size; 436 437 /* Reuse the original block if appropriate */ 438 if (size <= alloc 439 && (size > (alloc >> REALLOC_FRACTION) || alloc == MINALLOCSIZE)) 440 return (addr); 441 442 #ifdef DEBUG_MEMGUARD 443 } 444 #endif 445 446 /* Allocate a new, bigger (or smaller) block */ 447 if ((newaddr = malloc(size, type, flags)) == NULL) 448 return (NULL); 449 450 /* Copy over original contents */ 451 bcopy(addr, newaddr, min(size, alloc)); 452 free(addr, type); 453 return (newaddr); 454 } 455 456 /* 457 * reallocf: same as realloc() but free memory on failure. 458 */ 459 void * 460 reallocf(addr, size, type, flags) 461 void *addr; 462 unsigned long size; 463 struct malloc_type *type; 464 int flags; 465 { 466 void *mem; 467 468 if ((mem = realloc(addr, size, type, flags)) == NULL) 469 free(addr, type); 470 return (mem); 471 } 472 473 /* 474 * Initialize the kernel memory allocator 475 */ 476 /* ARGSUSED*/ 477 static void 478 kmeminit(dummy) 479 void *dummy; 480 { 481 u_int8_t indx; 482 u_long mem_size; 483 int i; 484 485 mtx_init(&malloc_mtx, "malloc", NULL, MTX_DEF); 486 487 /* 488 * Try to auto-tune the kernel memory size, so that it is 489 * more applicable for a wider range of machine sizes. 490 * On an X86, a VM_KMEM_SIZE_SCALE value of 4 is good, while 491 * a VM_KMEM_SIZE of 12MB is a fair compromise. The 492 * VM_KMEM_SIZE_MAX is dependent on the maximum KVA space 493 * available, and on an X86 with a total KVA space of 256MB, 494 * try to keep VM_KMEM_SIZE_MAX at 80MB or below. 495 * 496 * Note that the kmem_map is also used by the zone allocator, 497 * so make sure that there is enough space. 498 */ 499 vm_kmem_size = VM_KMEM_SIZE + nmbclusters * PAGE_SIZE; 500 mem_size = cnt.v_page_count; 501 502 #if defined(VM_KMEM_SIZE_SCALE) 503 vm_kmem_size_scale = VM_KMEM_SIZE_SCALE; 504 #endif 505 TUNABLE_INT_FETCH("vm.kmem_size_scale", &vm_kmem_size_scale); 506 if (vm_kmem_size_scale > 0 && 507 (mem_size / vm_kmem_size_scale) > (vm_kmem_size / PAGE_SIZE)) 508 vm_kmem_size = (mem_size / vm_kmem_size_scale) * PAGE_SIZE; 509 510 #if defined(VM_KMEM_SIZE_MAX) 511 vm_kmem_size_max = VM_KMEM_SIZE_MAX; 512 #endif 513 TUNABLE_INT_FETCH("vm.kmem_size_max", &vm_kmem_size_max); 514 if (vm_kmem_size_max > 0 && vm_kmem_size >= vm_kmem_size_max) 515 vm_kmem_size = vm_kmem_size_max; 516 517 /* Allow final override from the kernel environment */ 518 #ifndef BURN_BRIDGES 519 if (TUNABLE_INT_FETCH("kern.vm.kmem.size", &vm_kmem_size) != 0) 520 printf("kern.vm.kmem.size is now called vm.kmem_size!\n"); 521 #endif 522 TUNABLE_INT_FETCH("vm.kmem_size", &vm_kmem_size); 523 524 /* 525 * Limit kmem virtual size to twice the physical memory. 526 * This allows for kmem map sparseness, but limits the size 527 * to something sane. Be careful to not overflow the 32bit 528 * ints while doing the check. 529 */ 530 if (((vm_kmem_size / 2) / PAGE_SIZE) > cnt.v_page_count) 531 vm_kmem_size = 2 * cnt.v_page_count * PAGE_SIZE; 532 533 /* 534 * Tune settings based on the kernel map's size at this time. 535 */ 536 init_param3(vm_kmem_size / PAGE_SIZE); 537 538 kmem_map = kmem_suballoc(kernel_map, (vm_offset_t *)&kmembase, 539 (vm_offset_t *)&kmemlimit, vm_kmem_size); 540 kmem_map->system_map = 1; 541 542 #ifdef DEBUG_MEMGUARD 543 /* 544 * Initialize MemGuard if support compiled in. MemGuard is a 545 * replacement allocator used for detecting tamper-after-free 546 * scenarios as they occur. It is only used for debugging. 547 */ 548 vm_memguard_divisor = 10; 549 TUNABLE_INT_FETCH("vm.memguard_divisor", &vm_memguard_divisor); 550 551 /* Pick a conservative value if provided value sucks. */ 552 if ((vm_memguard_divisor <= 0) || 553 ((vm_kmem_size / vm_memguard_divisor) == 0)) 554 vm_memguard_divisor = 10; 555 memguard_init(kmem_map, vm_kmem_size / vm_memguard_divisor); 556 #endif 557 558 uma_startup2(); 559 560 for (i = 0, indx = 0; kmemzones[indx].kz_size != 0; indx++) { 561 int size = kmemzones[indx].kz_size; 562 char *name = kmemzones[indx].kz_name; 563 564 kmemzones[indx].kz_zone = uma_zcreate(name, size, 565 #ifdef INVARIANTS 566 mtrash_ctor, mtrash_dtor, mtrash_init, mtrash_fini, 567 #else 568 NULL, NULL, NULL, NULL, 569 #endif 570 UMA_ALIGN_PTR, UMA_ZONE_MALLOC); 571 572 for (;i <= size; i+= KMEM_ZBASE) 573 kmemsize[i >> KMEM_ZSHIFT] = indx; 574 575 } 576 } 577 578 void 579 malloc_init(data) 580 void *data; 581 { 582 struct malloc_type *type = (struct malloc_type *)data; 583 584 mtx_lock(&malloc_mtx); 585 if (type->ks_magic != M_MAGIC) 586 panic("malloc type lacks magic"); 587 588 if (cnt.v_page_count == 0) 589 panic("malloc_init not allowed before vm init"); 590 591 if (type->ks_next != NULL) 592 return; 593 594 type->ks_next = kmemstatistics; 595 kmemstatistics = type; 596 mtx_init(&type->ks_mtx, type->ks_shortdesc, "Malloc Stats", MTX_DEF); 597 mtx_unlock(&malloc_mtx); 598 } 599 600 void 601 malloc_uninit(data) 602 void *data; 603 { 604 struct malloc_type *type = (struct malloc_type *)data; 605 struct malloc_type *t; 606 607 mtx_lock(&malloc_mtx); 608 mtx_lock(&type->ks_mtx); 609 if (type->ks_magic != M_MAGIC) 610 panic("malloc type lacks magic"); 611 612 if (cnt.v_page_count == 0) 613 panic("malloc_uninit not allowed before vm init"); 614 615 if (type == kmemstatistics) 616 kmemstatistics = type->ks_next; 617 else { 618 for (t = kmemstatistics; t->ks_next != NULL; t = t->ks_next) { 619 if (t->ks_next == type) { 620 t->ks_next = type->ks_next; 621 break; 622 } 623 } 624 } 625 type->ks_next = NULL; 626 mtx_destroy(&type->ks_mtx); 627 mtx_unlock(&malloc_mtx); 628 } 629 630 static int 631 sysctl_kern_malloc(SYSCTL_HANDLER_ARGS) 632 { 633 struct malloc_type *type; 634 int linesize = 128; 635 int curline; 636 int bufsize; 637 int first; 638 int error; 639 char *buf; 640 char *p; 641 int cnt; 642 int len; 643 int i; 644 645 cnt = 0; 646 647 mtx_lock(&malloc_mtx); 648 for (type = kmemstatistics; type != NULL; type = type->ks_next) 649 cnt++; 650 651 mtx_unlock(&malloc_mtx); 652 bufsize = linesize * (cnt + 1); 653 p = buf = (char *)malloc(bufsize, M_TEMP, M_WAITOK|M_ZERO); 654 mtx_lock(&malloc_mtx); 655 656 len = snprintf(p, linesize, 657 "\n Type InUse MemUse HighUse Requests Size(s)\n"); 658 p += len; 659 660 for (type = kmemstatistics; cnt != 0 && type != NULL; 661 type = type->ks_next, cnt--) { 662 if (type->ks_calls == 0) 663 continue; 664 665 curline = linesize - 2; /* Leave room for the \n */ 666 len = snprintf(p, curline, "%13s%6lu%6luK%7luK%9llu", 667 type->ks_shortdesc, 668 type->ks_inuse, 669 (type->ks_memuse + 1023) / 1024, 670 (type->ks_maxused + 1023) / 1024, 671 (long long unsigned)type->ks_calls); 672 curline -= len; 673 p += len; 674 675 first = 1; 676 for (i = 0; i < sizeof(kmemzones) / sizeof(kmemzones[0]) - 1; 677 i++) { 678 if (type->ks_size & (1 << i)) { 679 if (first) 680 len = snprintf(p, curline, " "); 681 else 682 len = snprintf(p, curline, ","); 683 curline -= len; 684 p += len; 685 686 len = snprintf(p, curline, 687 "%s", kmemzones[i].kz_name); 688 curline -= len; 689 p += len; 690 691 first = 0; 692 } 693 } 694 695 len = snprintf(p, 2, "\n"); 696 p += len; 697 } 698 699 mtx_unlock(&malloc_mtx); 700 error = SYSCTL_OUT(req, buf, p - buf); 701 702 free(buf, M_TEMP); 703 return (error); 704 } 705 706 SYSCTL_OID(_kern, OID_AUTO, malloc, CTLTYPE_STRING|CTLFLAG_RD, 707 NULL, 0, sysctl_kern_malloc, "A", "Malloc Stats"); 708 709 #ifdef MALLOC_PROFILE 710 711 static int 712 sysctl_kern_mprof(SYSCTL_HANDLER_ARGS) 713 { 714 int linesize = 64; 715 uint64_t count; 716 uint64_t waste; 717 uint64_t mem; 718 int bufsize; 719 int error; 720 char *buf; 721 int rsize; 722 int size; 723 char *p; 724 int len; 725 int i; 726 727 bufsize = linesize * (KMEM_ZSIZE + 1); 728 bufsize += 128; /* For the stats line */ 729 bufsize += 128; /* For the banner line */ 730 waste = 0; 731 mem = 0; 732 733 p = buf = (char *)malloc(bufsize, M_TEMP, M_WAITOK|M_ZERO); 734 len = snprintf(p, bufsize, 735 "\n Size Requests Real Size\n"); 736 bufsize -= len; 737 p += len; 738 739 for (i = 0; i < KMEM_ZSIZE; i++) { 740 size = i << KMEM_ZSHIFT; 741 rsize = kmemzones[kmemsize[i]].kz_size; 742 count = (long long unsigned)krequests[i]; 743 744 len = snprintf(p, bufsize, "%6d%28llu%11d\n", 745 size, (unsigned long long)count, rsize); 746 bufsize -= len; 747 p += len; 748 749 if ((rsize * count) > (size * count)) 750 waste += (rsize * count) - (size * count); 751 mem += (rsize * count); 752 } 753 754 len = snprintf(p, bufsize, 755 "\nTotal memory used:\t%30llu\nTotal Memory wasted:\t%30llu\n", 756 (unsigned long long)mem, (unsigned long long)waste); 757 p += len; 758 759 error = SYSCTL_OUT(req, buf, p - buf); 760 761 free(buf, M_TEMP); 762 return (error); 763 } 764 765 SYSCTL_OID(_kern, OID_AUTO, mprof, CTLTYPE_STRING|CTLFLAG_RD, 766 NULL, 0, sysctl_kern_mprof, "A", "Malloc Profiling"); 767 #endif /* MALLOC_PROFILE */ 768