xref: /freebsd/sys/kern/kern_malloc.c (revision 6b806d21d144c25f4fad714e1c0cf780f5e27d7e)
1 /*-
2  * Copyright (c) 1987, 1991, 1993
3  *	The Regents of the University of California.  All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions
7  * are met:
8  * 1. Redistributions of source code must retain the above copyright
9  *    notice, this list of conditions and the following disclaimer.
10  * 2. Redistributions in binary form must reproduce the above copyright
11  *    notice, this list of conditions and the following disclaimer in the
12  *    documentation and/or other materials provided with the distribution.
13  * 4. Neither the name of the University nor the names of its contributors
14  *    may be used to endorse or promote products derived from this software
15  *    without specific prior written permission.
16  *
17  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
18  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
19  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
20  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
21  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
22  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
23  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
24  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
25  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
26  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
27  * SUCH DAMAGE.
28  *
29  *	@(#)kern_malloc.c	8.3 (Berkeley) 1/4/94
30  */
31 
32 #include <sys/cdefs.h>
33 __FBSDID("$FreeBSD$");
34 
35 #include "opt_vm.h"
36 
37 #include <sys/param.h>
38 #include <sys/systm.h>
39 #include <sys/kdb.h>
40 #include <sys/kernel.h>
41 #include <sys/lock.h>
42 #include <sys/malloc.h>
43 #include <sys/mbuf.h>
44 #include <sys/mutex.h>
45 #include <sys/vmmeter.h>
46 #include <sys/proc.h>
47 #include <sys/sysctl.h>
48 #include <sys/time.h>
49 
50 #include <vm/vm.h>
51 #include <vm/pmap.h>
52 #include <vm/vm_param.h>
53 #include <vm/vm_kern.h>
54 #include <vm/vm_extern.h>
55 #include <vm/vm_map.h>
56 #include <vm/vm_page.h>
57 #include <vm/uma.h>
58 #include <vm/uma_int.h>
59 #include <vm/uma_dbg.h>
60 
61 #ifdef DEBUG_MEMGUARD
62 #include <vm/memguard.h>
63 #endif
64 
65 #if defined(INVARIANTS) && defined(__i386__)
66 #include <machine/cpu.h>
67 #endif
68 
69 /*
70  * When realloc() is called, if the new size is sufficiently smaller than
71  * the old size, realloc() will allocate a new, smaller block to avoid
72  * wasting memory. 'Sufficiently smaller' is defined as: newsize <=
73  * oldsize / 2^n, where REALLOC_FRACTION defines the value of 'n'.
74  */
75 #ifndef REALLOC_FRACTION
76 #define	REALLOC_FRACTION	1	/* new block if <= half the size */
77 #endif
78 
79 MALLOC_DEFINE(M_CACHE, "cache", "Various Dynamically allocated caches");
80 MALLOC_DEFINE(M_DEVBUF, "devbuf", "device driver memory");
81 MALLOC_DEFINE(M_TEMP, "temp", "misc temporary data buffers");
82 
83 MALLOC_DEFINE(M_IP6OPT, "ip6opt", "IPv6 options");
84 MALLOC_DEFINE(M_IP6NDP, "ip6ndp", "IPv6 Neighbor Discovery");
85 
86 static void kmeminit(void *);
87 SYSINIT(kmem, SI_SUB_KMEM, SI_ORDER_FIRST, kmeminit, NULL)
88 
89 static MALLOC_DEFINE(M_FREE, "free", "should be on free list");
90 
91 static struct malloc_type *kmemstatistics;
92 static char *kmembase;
93 static char *kmemlimit;
94 
95 #define KMEM_ZSHIFT	4
96 #define KMEM_ZBASE	16
97 #define KMEM_ZMASK	(KMEM_ZBASE - 1)
98 
99 #define KMEM_ZMAX	PAGE_SIZE
100 #define KMEM_ZSIZE	(KMEM_ZMAX >> KMEM_ZSHIFT)
101 static u_int8_t kmemsize[KMEM_ZSIZE + 1];
102 
103 /* These won't be powers of two for long */
104 struct {
105 	int kz_size;
106 	char *kz_name;
107 	uma_zone_t kz_zone;
108 } kmemzones[] = {
109 	{16, "16", NULL},
110 	{32, "32", NULL},
111 	{64, "64", NULL},
112 	{128, "128", NULL},
113 	{256, "256", NULL},
114 	{512, "512", NULL},
115 	{1024, "1024", NULL},
116 	{2048, "2048", NULL},
117 	{4096, "4096", NULL},
118 #if PAGE_SIZE > 4096
119 	{8192, "8192", NULL},
120 #if PAGE_SIZE > 8192
121 	{16384, "16384", NULL},
122 #if PAGE_SIZE > 16384
123 	{32768, "32768", NULL},
124 #if PAGE_SIZE > 32768
125 	{65536, "65536", NULL},
126 #if PAGE_SIZE > 65536
127 #error	"Unsupported PAGE_SIZE"
128 #endif	/* 65536 */
129 #endif	/* 32768 */
130 #endif	/* 16384 */
131 #endif	/* 8192 */
132 #endif	/* 4096 */
133 	{0, NULL},
134 };
135 
136 #ifdef DEBUG_MEMGUARD
137 u_int vm_memguard_divisor;
138 SYSCTL_UINT(_vm, OID_AUTO, memguard_divisor, CTLFLAG_RD, &vm_memguard_divisor,
139     0, "(kmem_size/memguard_divisor) == memguard submap size");
140 #endif
141 
142 u_int vm_kmem_size;
143 SYSCTL_UINT(_vm, OID_AUTO, kmem_size, CTLFLAG_RD, &vm_kmem_size, 0,
144     "Size of kernel memory");
145 
146 u_int vm_kmem_size_max;
147 SYSCTL_UINT(_vm, OID_AUTO, kmem_size_max, CTLFLAG_RD, &vm_kmem_size_max, 0,
148     "Maximum size of kernel memory");
149 
150 u_int vm_kmem_size_scale;
151 SYSCTL_UINT(_vm, OID_AUTO, kmem_size_scale, CTLFLAG_RD, &vm_kmem_size_scale, 0,
152     "Scale factor for kernel memory size");
153 
154 /*
155  * The malloc_mtx protects the kmemstatistics linked list.
156  */
157 
158 struct mtx malloc_mtx;
159 
160 #ifdef MALLOC_PROFILE
161 uint64_t krequests[KMEM_ZSIZE + 1];
162 
163 static int sysctl_kern_mprof(SYSCTL_HANDLER_ARGS);
164 #endif
165 
166 static int sysctl_kern_malloc(SYSCTL_HANDLER_ARGS);
167 
168 /* time_uptime of last malloc(9) failure */
169 static time_t t_malloc_fail;
170 
171 #ifdef MALLOC_MAKE_FAILURES
172 /*
173  * Causes malloc failures every (n) mallocs with M_NOWAIT.  If set to 0,
174  * doesn't cause failures.
175  */
176 SYSCTL_NODE(_debug, OID_AUTO, malloc, CTLFLAG_RD, 0,
177     "Kernel malloc debugging options");
178 
179 static int malloc_failure_rate;
180 static int malloc_nowait_count;
181 static int malloc_failure_count;
182 SYSCTL_INT(_debug_malloc, OID_AUTO, failure_rate, CTLFLAG_RW,
183     &malloc_failure_rate, 0, "Every (n) mallocs with M_NOWAIT will fail");
184 TUNABLE_INT("debug.malloc.failure_rate", &malloc_failure_rate);
185 SYSCTL_INT(_debug_malloc, OID_AUTO, failure_count, CTLFLAG_RD,
186     &malloc_failure_count, 0, "Number of imposed M_NOWAIT malloc failures");
187 #endif
188 
189 int
190 malloc_last_fail(void)
191 {
192 
193 	return (time_uptime - t_malloc_fail);
194 }
195 
196 /*
197  * Add this to the informational malloc_type bucket.
198  */
199 static void
200 malloc_type_zone_allocated(struct malloc_type *ksp, unsigned long size,
201     int zindx)
202 {
203 	mtx_lock(&ksp->ks_mtx);
204 	ksp->ks_calls++;
205 	if (zindx != -1)
206 		ksp->ks_size |= 1 << zindx;
207 	if (size != 0) {
208 		ksp->ks_memuse += size;
209 		ksp->ks_inuse++;
210 		if (ksp->ks_memuse > ksp->ks_maxused)
211 			ksp->ks_maxused = ksp->ks_memuse;
212 	}
213 	mtx_unlock(&ksp->ks_mtx);
214 }
215 
216 void
217 malloc_type_allocated(struct malloc_type *ksp, unsigned long size)
218 {
219 	malloc_type_zone_allocated(ksp, size, -1);
220 }
221 
222 /*
223  * Remove this allocation from the informational malloc_type bucket.
224  */
225 void
226 malloc_type_freed(struct malloc_type *ksp, unsigned long size)
227 {
228 	mtx_lock(&ksp->ks_mtx);
229 	KASSERT(size <= ksp->ks_memuse,
230 		("malloc(9)/free(9) confusion.\n%s",
231 		 "Probably freeing with wrong type, but maybe not here."));
232 	ksp->ks_memuse -= size;
233 	ksp->ks_inuse--;
234 	mtx_unlock(&ksp->ks_mtx);
235 }
236 
237 /*
238  *	malloc:
239  *
240  *	Allocate a block of memory.
241  *
242  *	If M_NOWAIT is set, this routine will not block and return NULL if
243  *	the allocation fails.
244  */
245 void *
246 malloc(size, type, flags)
247 	unsigned long size;
248 	struct malloc_type *type;
249 	int flags;
250 {
251 	int indx;
252 	caddr_t va;
253 	uma_zone_t zone;
254 	uma_keg_t keg;
255 #ifdef DIAGNOSTIC
256 	unsigned long osize = size;
257 #endif
258 
259 #ifdef INVARIANTS
260 	/*
261 	 * To make sure that WAITOK or NOWAIT is set, but not more than
262 	 * one, and check against the API botches that are common.
263 	 */
264 	indx = flags & (M_WAITOK | M_NOWAIT | M_DONTWAIT | M_TRYWAIT);
265 	if (indx != M_NOWAIT && indx != M_WAITOK) {
266 		static	struct timeval lasterr;
267 		static	int curerr, once;
268 		if (once == 0 && ppsratecheck(&lasterr, &curerr, 1)) {
269 			printf("Bad malloc flags: %x\n", indx);
270 			kdb_backtrace();
271 			flags |= M_WAITOK;
272 			once++;
273 		}
274 	}
275 #endif
276 #if 0
277 	if (size == 0)
278 		kdb_enter("zero size malloc");
279 #endif
280 #ifdef MALLOC_MAKE_FAILURES
281 	if ((flags & M_NOWAIT) && (malloc_failure_rate != 0)) {
282 		atomic_add_int(&malloc_nowait_count, 1);
283 		if ((malloc_nowait_count % malloc_failure_rate) == 0) {
284 			atomic_add_int(&malloc_failure_count, 1);
285 			t_malloc_fail = time_uptime;
286 			return (NULL);
287 		}
288 	}
289 #endif
290 	if (flags & M_WAITOK)
291 		KASSERT(curthread->td_intr_nesting_level == 0,
292 		   ("malloc(M_WAITOK) in interrupt context"));
293 
294 #ifdef DEBUG_MEMGUARD
295 	/* XXX CHANGEME! */
296 	if (type == M_SUBPROC)
297 		return memguard_alloc(size, flags);
298 #endif
299 
300 	if (size <= KMEM_ZMAX) {
301 		if (size & KMEM_ZMASK)
302 			size = (size & ~KMEM_ZMASK) + KMEM_ZBASE;
303 		indx = kmemsize[size >> KMEM_ZSHIFT];
304 		zone = kmemzones[indx].kz_zone;
305 		keg = zone->uz_keg;
306 #ifdef MALLOC_PROFILE
307 		krequests[size >> KMEM_ZSHIFT]++;
308 #endif
309 		va = uma_zalloc(zone, flags);
310 		if (va != NULL)
311 			size = keg->uk_size;
312 		malloc_type_zone_allocated(type, va == NULL ? 0 : size, indx);
313 	} else {
314 		size = roundup(size, PAGE_SIZE);
315 		zone = NULL;
316 		keg = NULL;
317 		va = uma_large_malloc(size, flags);
318 		malloc_type_allocated(type, va == NULL ? 0 : size);
319 	}
320 	if (flags & M_WAITOK)
321 		KASSERT(va != NULL, ("malloc(M_WAITOK) returned NULL"));
322 	else if (va == NULL)
323 		t_malloc_fail = time_uptime;
324 #ifdef DIAGNOSTIC
325 	if (va != NULL && !(flags & M_ZERO)) {
326 		memset(va, 0x70, osize);
327 	}
328 #endif
329 	return ((void *) va);
330 }
331 
332 /*
333  *	free:
334  *
335  *	Free a block of memory allocated by malloc.
336  *
337  *	This routine may not block.
338  */
339 void
340 free(addr, type)
341 	void *addr;
342 	struct malloc_type *type;
343 {
344 	uma_slab_t slab;
345 	u_long size;
346 
347 	/* free(NULL, ...) does nothing */
348 	if (addr == NULL)
349 		return;
350 
351 #ifdef DEBUG_MEMGUARD
352 	/* XXX CHANGEME! */
353 	if (type == M_SUBPROC) {
354 		memguard_free(addr);
355 		return;
356 	}
357 #endif
358 
359 	KASSERT(type->ks_memuse > 0,
360 		("malloc(9)/free(9) confusion.\n%s",
361 		 "Probably freeing with wrong type, but maybe not here."));
362 	size = 0;
363 
364 	slab = vtoslab((vm_offset_t)addr & (~UMA_SLAB_MASK));
365 
366 	if (slab == NULL)
367 		panic("free: address %p(%p) has not been allocated.\n",
368 		    addr, (void *)((u_long)addr & (~UMA_SLAB_MASK)));
369 
370 
371 	if (!(slab->us_flags & UMA_SLAB_MALLOC)) {
372 #ifdef INVARIANTS
373 		struct malloc_type **mtp = addr;
374 #endif
375 		size = slab->us_keg->uk_size;
376 #ifdef INVARIANTS
377 		/*
378 		 * Cache a pointer to the malloc_type that most recently freed
379 		 * this memory here.  This way we know who is most likely to
380 		 * have stepped on it later.
381 		 *
382 		 * This code assumes that size is a multiple of 8 bytes for
383 		 * 64 bit machines
384 		 */
385 		mtp = (struct malloc_type **)
386 		    ((unsigned long)mtp & ~UMA_ALIGN_PTR);
387 		mtp += (size - sizeof(struct malloc_type *)) /
388 		    sizeof(struct malloc_type *);
389 		*mtp = type;
390 #endif
391 		uma_zfree_arg(LIST_FIRST(&slab->us_keg->uk_zones), addr, slab);
392 	} else {
393 		size = slab->us_size;
394 		uma_large_free(slab);
395 	}
396 	malloc_type_freed(type, size);
397 }
398 
399 /*
400  *	realloc: change the size of a memory block
401  */
402 void *
403 realloc(addr, size, type, flags)
404 	void *addr;
405 	unsigned long size;
406 	struct malloc_type *type;
407 	int flags;
408 {
409 	uma_slab_t slab;
410 	unsigned long alloc;
411 	void *newaddr;
412 
413 	/* realloc(NULL, ...) is equivalent to malloc(...) */
414 	if (addr == NULL)
415 		return (malloc(size, type, flags));
416 
417 #ifdef DEBUG_MEMGUARD
418 /* XXX: CHANGEME! */
419 if (type == M_SUBPROC) {
420 	slab = NULL;
421 	alloc = size;
422 } else {
423 #endif
424 
425 	slab = vtoslab((vm_offset_t)addr & ~(UMA_SLAB_MASK));
426 
427 	/* Sanity check */
428 	KASSERT(slab != NULL,
429 	    ("realloc: address %p out of range", (void *)addr));
430 
431 	/* Get the size of the original block */
432 	if (slab->us_keg)
433 		alloc = slab->us_keg->uk_size;
434 	else
435 		alloc = slab->us_size;
436 
437 	/* Reuse the original block if appropriate */
438 	if (size <= alloc
439 	    && (size > (alloc >> REALLOC_FRACTION) || alloc == MINALLOCSIZE))
440 		return (addr);
441 
442 #ifdef DEBUG_MEMGUARD
443 }
444 #endif
445 
446 	/* Allocate a new, bigger (or smaller) block */
447 	if ((newaddr = malloc(size, type, flags)) == NULL)
448 		return (NULL);
449 
450 	/* Copy over original contents */
451 	bcopy(addr, newaddr, min(size, alloc));
452 	free(addr, type);
453 	return (newaddr);
454 }
455 
456 /*
457  *	reallocf: same as realloc() but free memory on failure.
458  */
459 void *
460 reallocf(addr, size, type, flags)
461 	void *addr;
462 	unsigned long size;
463 	struct malloc_type *type;
464 	int flags;
465 {
466 	void *mem;
467 
468 	if ((mem = realloc(addr, size, type, flags)) == NULL)
469 		free(addr, type);
470 	return (mem);
471 }
472 
473 /*
474  * Initialize the kernel memory allocator
475  */
476 /* ARGSUSED*/
477 static void
478 kmeminit(dummy)
479 	void *dummy;
480 {
481 	u_int8_t indx;
482 	u_long mem_size;
483 	int i;
484 
485 	mtx_init(&malloc_mtx, "malloc", NULL, MTX_DEF);
486 
487 	/*
488 	 * Try to auto-tune the kernel memory size, so that it is
489 	 * more applicable for a wider range of machine sizes.
490 	 * On an X86, a VM_KMEM_SIZE_SCALE value of 4 is good, while
491 	 * a VM_KMEM_SIZE of 12MB is a fair compromise.  The
492 	 * VM_KMEM_SIZE_MAX is dependent on the maximum KVA space
493 	 * available, and on an X86 with a total KVA space of 256MB,
494 	 * try to keep VM_KMEM_SIZE_MAX at 80MB or below.
495 	 *
496 	 * Note that the kmem_map is also used by the zone allocator,
497 	 * so make sure that there is enough space.
498 	 */
499 	vm_kmem_size = VM_KMEM_SIZE + nmbclusters * PAGE_SIZE;
500 	mem_size = cnt.v_page_count;
501 
502 #if defined(VM_KMEM_SIZE_SCALE)
503 	vm_kmem_size_scale = VM_KMEM_SIZE_SCALE;
504 #endif
505 	TUNABLE_INT_FETCH("vm.kmem_size_scale", &vm_kmem_size_scale);
506 	if (vm_kmem_size_scale > 0 &&
507 	    (mem_size / vm_kmem_size_scale) > (vm_kmem_size / PAGE_SIZE))
508 		vm_kmem_size = (mem_size / vm_kmem_size_scale) * PAGE_SIZE;
509 
510 #if defined(VM_KMEM_SIZE_MAX)
511 	vm_kmem_size_max = VM_KMEM_SIZE_MAX;
512 #endif
513 	TUNABLE_INT_FETCH("vm.kmem_size_max", &vm_kmem_size_max);
514 	if (vm_kmem_size_max > 0 && vm_kmem_size >= vm_kmem_size_max)
515 		vm_kmem_size = vm_kmem_size_max;
516 
517 	/* Allow final override from the kernel environment */
518 #ifndef BURN_BRIDGES
519 	if (TUNABLE_INT_FETCH("kern.vm.kmem.size", &vm_kmem_size) != 0)
520 		printf("kern.vm.kmem.size is now called vm.kmem_size!\n");
521 #endif
522 	TUNABLE_INT_FETCH("vm.kmem_size", &vm_kmem_size);
523 
524 	/*
525 	 * Limit kmem virtual size to twice the physical memory.
526 	 * This allows for kmem map sparseness, but limits the size
527 	 * to something sane. Be careful to not overflow the 32bit
528 	 * ints while doing the check.
529 	 */
530 	if (((vm_kmem_size / 2) / PAGE_SIZE) > cnt.v_page_count)
531 		vm_kmem_size = 2 * cnt.v_page_count * PAGE_SIZE;
532 
533 	/*
534 	 * Tune settings based on the kernel map's size at this time.
535 	 */
536 	init_param3(vm_kmem_size / PAGE_SIZE);
537 
538 	kmem_map = kmem_suballoc(kernel_map, (vm_offset_t *)&kmembase,
539 		(vm_offset_t *)&kmemlimit, vm_kmem_size);
540 	kmem_map->system_map = 1;
541 
542 #ifdef DEBUG_MEMGUARD
543 	/*
544 	 * Initialize MemGuard if support compiled in.  MemGuard is a
545 	 * replacement allocator used for detecting tamper-after-free
546 	 * scenarios as they occur.  It is only used for debugging.
547 	 */
548 	vm_memguard_divisor = 10;
549 	TUNABLE_INT_FETCH("vm.memguard_divisor", &vm_memguard_divisor);
550 
551 	/* Pick a conservative value if provided value sucks. */
552 	if ((vm_memguard_divisor <= 0) ||
553 	    ((vm_kmem_size / vm_memguard_divisor) == 0))
554 		vm_memguard_divisor = 10;
555 	memguard_init(kmem_map, vm_kmem_size / vm_memguard_divisor);
556 #endif
557 
558 	uma_startup2();
559 
560 	for (i = 0, indx = 0; kmemzones[indx].kz_size != 0; indx++) {
561 		int size = kmemzones[indx].kz_size;
562 		char *name = kmemzones[indx].kz_name;
563 
564 		kmemzones[indx].kz_zone = uma_zcreate(name, size,
565 #ifdef INVARIANTS
566 		    mtrash_ctor, mtrash_dtor, mtrash_init, mtrash_fini,
567 #else
568 		    NULL, NULL, NULL, NULL,
569 #endif
570 		    UMA_ALIGN_PTR, UMA_ZONE_MALLOC);
571 
572 		for (;i <= size; i+= KMEM_ZBASE)
573 			kmemsize[i >> KMEM_ZSHIFT] = indx;
574 
575 	}
576 }
577 
578 void
579 malloc_init(data)
580 	void *data;
581 {
582 	struct malloc_type *type = (struct malloc_type *)data;
583 
584 	mtx_lock(&malloc_mtx);
585 	if (type->ks_magic != M_MAGIC)
586 		panic("malloc type lacks magic");
587 
588 	if (cnt.v_page_count == 0)
589 		panic("malloc_init not allowed before vm init");
590 
591 	if (type->ks_next != NULL)
592 		return;
593 
594 	type->ks_next = kmemstatistics;
595 	kmemstatistics = type;
596 	mtx_init(&type->ks_mtx, type->ks_shortdesc, "Malloc Stats", MTX_DEF);
597 	mtx_unlock(&malloc_mtx);
598 }
599 
600 void
601 malloc_uninit(data)
602 	void *data;
603 {
604 	struct malloc_type *type = (struct malloc_type *)data;
605 	struct malloc_type *t;
606 
607 	mtx_lock(&malloc_mtx);
608 	mtx_lock(&type->ks_mtx);
609 	if (type->ks_magic != M_MAGIC)
610 		panic("malloc type lacks magic");
611 
612 	if (cnt.v_page_count == 0)
613 		panic("malloc_uninit not allowed before vm init");
614 
615 	if (type == kmemstatistics)
616 		kmemstatistics = type->ks_next;
617 	else {
618 		for (t = kmemstatistics; t->ks_next != NULL; t = t->ks_next) {
619 			if (t->ks_next == type) {
620 				t->ks_next = type->ks_next;
621 				break;
622 			}
623 		}
624 	}
625 	type->ks_next = NULL;
626 	mtx_destroy(&type->ks_mtx);
627 	mtx_unlock(&malloc_mtx);
628 }
629 
630 static int
631 sysctl_kern_malloc(SYSCTL_HANDLER_ARGS)
632 {
633 	struct malloc_type *type;
634 	int linesize = 128;
635 	int curline;
636 	int bufsize;
637 	int first;
638 	int error;
639 	char *buf;
640 	char *p;
641 	int cnt;
642 	int len;
643 	int i;
644 
645 	cnt = 0;
646 
647 	mtx_lock(&malloc_mtx);
648 	for (type = kmemstatistics; type != NULL; type = type->ks_next)
649 		cnt++;
650 
651 	mtx_unlock(&malloc_mtx);
652 	bufsize = linesize * (cnt + 1);
653 	p = buf = (char *)malloc(bufsize, M_TEMP, M_WAITOK|M_ZERO);
654 	mtx_lock(&malloc_mtx);
655 
656 	len = snprintf(p, linesize,
657 	    "\n        Type  InUse MemUse HighUse Requests  Size(s)\n");
658 	p += len;
659 
660 	for (type = kmemstatistics; cnt != 0 && type != NULL;
661 	    type = type->ks_next, cnt--) {
662 		if (type->ks_calls == 0)
663 			continue;
664 
665 		curline = linesize - 2;	/* Leave room for the \n */
666 		len = snprintf(p, curline, "%13s%6lu%6luK%7luK%9llu",
667 			type->ks_shortdesc,
668 			type->ks_inuse,
669 			(type->ks_memuse + 1023) / 1024,
670 			(type->ks_maxused + 1023) / 1024,
671 			(long long unsigned)type->ks_calls);
672 		curline -= len;
673 		p += len;
674 
675 		first = 1;
676 		for (i = 0; i < sizeof(kmemzones) / sizeof(kmemzones[0]) - 1;
677 		    i++) {
678 			if (type->ks_size & (1 << i)) {
679 				if (first)
680 					len = snprintf(p, curline, "  ");
681 				else
682 					len = snprintf(p, curline, ",");
683 				curline -= len;
684 				p += len;
685 
686 				len = snprintf(p, curline,
687 				    "%s", kmemzones[i].kz_name);
688 				curline -= len;
689 				p += len;
690 
691 				first = 0;
692 			}
693 		}
694 
695 		len = snprintf(p, 2, "\n");
696 		p += len;
697 	}
698 
699 	mtx_unlock(&malloc_mtx);
700 	error = SYSCTL_OUT(req, buf, p - buf);
701 
702 	free(buf, M_TEMP);
703 	return (error);
704 }
705 
706 SYSCTL_OID(_kern, OID_AUTO, malloc, CTLTYPE_STRING|CTLFLAG_RD,
707     NULL, 0, sysctl_kern_malloc, "A", "Malloc Stats");
708 
709 #ifdef MALLOC_PROFILE
710 
711 static int
712 sysctl_kern_mprof(SYSCTL_HANDLER_ARGS)
713 {
714 	int linesize = 64;
715 	uint64_t count;
716 	uint64_t waste;
717 	uint64_t mem;
718 	int bufsize;
719 	int error;
720 	char *buf;
721 	int rsize;
722 	int size;
723 	char *p;
724 	int len;
725 	int i;
726 
727 	bufsize = linesize * (KMEM_ZSIZE + 1);
728 	bufsize += 128; 	/* For the stats line */
729 	bufsize += 128; 	/* For the banner line */
730 	waste = 0;
731 	mem = 0;
732 
733 	p = buf = (char *)malloc(bufsize, M_TEMP, M_WAITOK|M_ZERO);
734 	len = snprintf(p, bufsize,
735 	    "\n  Size                    Requests  Real Size\n");
736 	bufsize -= len;
737 	p += len;
738 
739 	for (i = 0; i < KMEM_ZSIZE; i++) {
740 		size = i << KMEM_ZSHIFT;
741 		rsize = kmemzones[kmemsize[i]].kz_size;
742 		count = (long long unsigned)krequests[i];
743 
744 		len = snprintf(p, bufsize, "%6d%28llu%11d\n",
745 		    size, (unsigned long long)count, rsize);
746 		bufsize -= len;
747 		p += len;
748 
749 		if ((rsize * count) > (size * count))
750 			waste += (rsize * count) - (size * count);
751 		mem += (rsize * count);
752 	}
753 
754 	len = snprintf(p, bufsize,
755 	    "\nTotal memory used:\t%30llu\nTotal Memory wasted:\t%30llu\n",
756 	    (unsigned long long)mem, (unsigned long long)waste);
757 	p += len;
758 
759 	error = SYSCTL_OUT(req, buf, p - buf);
760 
761 	free(buf, M_TEMP);
762 	return (error);
763 }
764 
765 SYSCTL_OID(_kern, OID_AUTO, mprof, CTLTYPE_STRING|CTLFLAG_RD,
766     NULL, 0, sysctl_kern_mprof, "A", "Malloc Profiling");
767 #endif /* MALLOC_PROFILE */
768