xref: /freebsd/sys/contrib/openzfs/module/os/freebsd/spl/spl_kmem.c (revision d0abb9a6399accc9053e2808052be00a6754ecef)
1 // SPDX-License-Identifier: BSD-2-Clause
2 /*
3  * Copyright (c) 2006-2007 Pawel Jakub Dawidek <pjd@FreeBSD.org>
4  * All rights reserved.
5  *
6  * Redistribution and use in source and binary forms, with or without
7  * modification, are permitted provided that the following conditions
8  * are met:
9  * 1. Redistributions of source code must retain the above copyright
10  *    notice, this list of conditions and the following disclaimer.
11  * 2. Redistributions in binary form must reproduce the above copyright
12  *    notice, this list of conditions and the following disclaimer in the
13  *    documentation and/or other materials provided with the distribution.
14  *
15  * THIS SOFTWARE IS PROVIDED BY THE AUTHORS AND CONTRIBUTORS ``AS IS'' AND
16  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
17  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
18  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHORS OR CONTRIBUTORS BE LIABLE
19  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
20  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
21  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
22  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
23  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
24  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
25  * SUCH DAMAGE.
26  */
27 
28 #include <sys/types.h>
29 #include <sys/param.h>
30 #include <sys/byteorder.h>
31 #include <sys/kernel.h>
32 #include <sys/systm.h>
33 #include <sys/malloc.h>
34 #include <sys/kmem.h>
35 #include <sys/kmem_cache.h>
36 #include <sys/debug.h>
37 #include <sys/mutex.h>
38 #include <sys/vmmeter.h>
39 
40 
41 #include <vm/vm_page.h>
42 #include <vm/vm_object.h>
43 #include <vm/vm_kern.h>
44 #include <vm/vm_map.h>
45 
46 #ifdef KMEM_DEBUG
47 #include <sys/queue.h>
48 #include <sys/stack.h>
49 #endif
50 
51 #ifdef _KERNEL
52 MALLOC_DEFINE(M_SOLARIS, "solaris", "Solaris");
53 #else
54 #define	malloc(size, type, flags)	malloc(size)
55 #define	free(addr, type)		free(addr)
56 #endif
57 
58 #ifdef KMEM_DEBUG
59 struct kmem_item {
60 	struct stack	stack;
61 	LIST_ENTRY(kmem_item) next;
62 };
63 static LIST_HEAD(, kmem_item) kmem_items;
64 static struct mtx kmem_items_mtx;
65 MTX_SYSINIT(kmem_items_mtx, &kmem_items_mtx, "kmem_items", MTX_DEF);
66 #endif	/* KMEM_DEBUG */
67 
68 #include <sys/vmem.h>
69 
70 void *
zfs_kmem_alloc(size_t size,int kmflags)71 zfs_kmem_alloc(size_t size, int kmflags)
72 {
73 	void *p;
74 #ifdef KMEM_DEBUG
75 	struct kmem_item *i;
76 
77 	size += sizeof (struct kmem_item);
78 #endif
79 	p = malloc(MAX(size, 16), M_SOLARIS, kmflags);
80 #ifndef _KERNEL
81 	if (kmflags & KM_SLEEP)
82 		assert(p != NULL);
83 #endif
84 #ifdef KMEM_DEBUG
85 	if (p != NULL) {
86 		i = p;
87 		p = (uint8_t *)p + sizeof (struct kmem_item);
88 		stack_save(&i->stack);
89 		mtx_lock(&kmem_items_mtx);
90 		LIST_INSERT_HEAD(&kmem_items, i, next);
91 		mtx_unlock(&kmem_items_mtx);
92 	}
93 #endif
94 	return (p);
95 }
96 
97 void
zfs_kmem_free(void * buf,size_t size __unused)98 zfs_kmem_free(void *buf, size_t size __unused)
99 {
100 #ifdef KMEM_DEBUG
101 	if (buf == NULL) {
102 		printf("%s: attempt to free NULL\n", __func__);
103 		return;
104 	}
105 	struct kmem_item *i;
106 
107 	buf = (uint8_t *)buf - sizeof (struct kmem_item);
108 	mtx_lock(&kmem_items_mtx);
109 	LIST_FOREACH(i, &kmem_items, next) {
110 		if (i == buf)
111 			break;
112 	}
113 	ASSERT3P(i, !=, NULL);
114 	LIST_REMOVE(i, next);
115 	mtx_unlock(&kmem_items_mtx);
116 	memset(buf, 0xDC, MAX(size, 16));
117 #endif
118 	free(buf, M_SOLARIS);
119 }
120 
121 static uint64_t kmem_size_val;
122 
123 static void
kmem_size_init(void * unused __unused)124 kmem_size_init(void *unused __unused)
125 {
126 
127 	kmem_size_val = (uint64_t)vm_cnt.v_page_count * PAGE_SIZE;
128 	if (kmem_size_val > vm_kmem_size)
129 		kmem_size_val = vm_kmem_size;
130 }
131 SYSINIT(kmem_size_init, SI_SUB_KMEM, SI_ORDER_ANY, kmem_size_init, NULL);
132 
133 uint64_t
kmem_size(void)134 kmem_size(void)
135 {
136 
137 	return (kmem_size_val);
138 }
139 
140 static int
kmem_std_constructor(void * mem,int size __unused,void * private,int flags)141 kmem_std_constructor(void *mem, int size __unused, void *private, int flags)
142 {
143 	struct kmem_cache *cache = private;
144 
145 	return (cache->kc_constructor(mem, cache->kc_private, flags));
146 }
147 
148 static void
kmem_std_destructor(void * mem,int size __unused,void * private)149 kmem_std_destructor(void *mem, int size __unused, void *private)
150 {
151 	struct kmem_cache *cache = private;
152 
153 	cache->kc_destructor(mem, cache->kc_private);
154 }
155 
156 kmem_cache_t *
kmem_cache_create(const char * name,size_t bufsize,size_t align,int (* constructor)(void *,void *,int),void (* destructor)(void *,void *),void (* reclaim)(void *)__unused,void * private,vmem_t * vmp,int cflags)157 kmem_cache_create(const char *name, size_t bufsize, size_t align,
158     int (*constructor)(void *, void *, int), void (*destructor)(void *, void *),
159     void (*reclaim)(void *) __unused, void *private, vmem_t *vmp, int cflags)
160 {
161 	kmem_cache_t *cache;
162 
163 	ASSERT0P(vmp);
164 
165 	cache = kmem_alloc(sizeof (*cache), KM_SLEEP);
166 	strlcpy(cache->kc_name, name, sizeof (cache->kc_name));
167 	cache->kc_constructor = constructor;
168 	cache->kc_destructor = destructor;
169 	cache->kc_private = private;
170 #if defined(_KERNEL) && !defined(KMEM_DEBUG)
171 	cache->kc_zone = uma_zcreate(cache->kc_name, bufsize,
172 	    constructor != NULL ? kmem_std_constructor : NULL,
173 	    destructor != NULL ? kmem_std_destructor : NULL,
174 	    NULL, NULL, align > 0 ? align - 1 : 0, cflags);
175 #else
176 	cache->kc_size = bufsize;
177 #endif
178 
179 	return (cache);
180 }
181 
182 void
kmem_cache_destroy(kmem_cache_t * cache)183 kmem_cache_destroy(kmem_cache_t *cache)
184 {
185 #if defined(_KERNEL) && !defined(KMEM_DEBUG)
186 	uma_zdestroy(cache->kc_zone);
187 #endif
188 	kmem_free(cache, sizeof (*cache));
189 }
190 
191 void *
kmem_cache_alloc(kmem_cache_t * cache,int flags)192 kmem_cache_alloc(kmem_cache_t *cache, int flags)
193 {
194 #if defined(_KERNEL) && !defined(KMEM_DEBUG)
195 	return (uma_zalloc_arg(cache->kc_zone, cache, flags));
196 #else
197 	void *p;
198 
199 	p = kmem_alloc(cache->kc_size, flags);
200 	if (p != NULL && cache->kc_constructor != NULL)
201 		kmem_std_constructor(p, cache->kc_size, cache, flags);
202 	return (p);
203 #endif
204 }
205 
206 void
kmem_cache_free(kmem_cache_t * cache,void * buf)207 kmem_cache_free(kmem_cache_t *cache, void *buf)
208 {
209 #if defined(_KERNEL) && !defined(KMEM_DEBUG)
210 	uma_zfree_arg(cache->kc_zone, buf, cache);
211 #else
212 	if (cache->kc_destructor != NULL)
213 		kmem_std_destructor(buf, cache->kc_size, cache);
214 	kmem_free(buf, cache->kc_size);
215 #endif
216 }
217 
218 /*
219  * Allow our caller to determine if there are running reaps.
220  *
221  * This call is very conservative and may return B_TRUE even when
222  * reaping activity isn't active. If it returns B_FALSE, then reaping
223  * activity is definitely inactive.
224  */
225 boolean_t
kmem_cache_reap_active(void)226 kmem_cache_reap_active(void)
227 {
228 
229 	return (B_FALSE);
230 }
231 
232 /*
233  * Reap (almost) everything soon.
234  *
235  * Note: this does not wait for the reap-tasks to complete. Caller
236  * should use kmem_cache_reap_active() (above) and/or moderation to
237  * avoid scheduling too many reap-tasks.
238  */
239 #ifdef _KERNEL
240 void
kmem_cache_reap_soon(kmem_cache_t * cache)241 kmem_cache_reap_soon(kmem_cache_t *cache)
242 {
243 #ifndef KMEM_DEBUG
244 	uma_zone_reclaim(cache->kc_zone, UMA_RECLAIM_DRAIN);
245 #endif
246 }
247 
248 void
kmem_reap(void)249 kmem_reap(void)
250 {
251 	uma_reclaim(UMA_RECLAIM_TRIM);
252 }
253 #else
254 void
kmem_cache_reap_soon(kmem_cache_t * cache __unused)255 kmem_cache_reap_soon(kmem_cache_t *cache __unused)
256 {
257 }
258 
259 void
kmem_reap(void)260 kmem_reap(void)
261 {
262 }
263 #endif
264 
265 int
kmem_debugging(void)266 kmem_debugging(void)
267 {
268 	return (0);
269 }
270 
271 void *
calloc(size_t n,size_t s)272 calloc(size_t n, size_t s)
273 {
274 	return (kmem_zalloc(n * s, KM_NOSLEEP));
275 }
276 
277 char *
kmem_vasprintf(const char * fmt,va_list adx)278 kmem_vasprintf(const char *fmt, va_list adx)
279 {
280 	char *msg;
281 	va_list adx2;
282 
283 	va_copy(adx2, adx);
284 	msg = kmem_alloc(vsnprintf(NULL, 0, fmt, adx) + 1, KM_SLEEP);
285 	(void) vsprintf(msg, fmt, adx2);
286 	va_end(adx2);
287 
288 	return (msg);
289 }
290 
291 #include <vm/uma.h>
292 #include <vm/uma_int.h>
293 #ifdef KMEM_DEBUG
294 #error "KMEM_DEBUG not currently supported"
295 #endif
296 
297 uint64_t
spl_kmem_cache_inuse(kmem_cache_t * cache)298 spl_kmem_cache_inuse(kmem_cache_t *cache)
299 {
300 	return (uma_zone_get_cur(cache->kc_zone));
301 }
302 
303 uint64_t
spl_kmem_cache_entry_size(kmem_cache_t * cache)304 spl_kmem_cache_entry_size(kmem_cache_t *cache)
305 {
306 	return (cache->kc_zone->uz_size);
307 }
308 
309 /*
310  * Register a move callback for cache defragmentation.
311  * XXX: Unimplemented but harmless to stub out for now.
312  */
313 void
spl_kmem_cache_set_move(kmem_cache_t * skc,kmem_cbrc_t (move)(void *,void *,size_t,void *))314 spl_kmem_cache_set_move(kmem_cache_t *skc,
315     kmem_cbrc_t (move)(void *, void *, size_t, void *))
316 {
317 	ASSERT3P(move, !=, NULL);
318 }
319 
320 #ifdef KMEM_DEBUG
321 void kmem_show(void *);
322 void
kmem_show(void * dummy __unused)323 kmem_show(void *dummy __unused)
324 {
325 	struct kmem_item *i;
326 
327 	mtx_lock(&kmem_items_mtx);
328 	if (LIST_EMPTY(&kmem_items))
329 		printf("KMEM_DEBUG: No leaked elements.\n");
330 	else {
331 		printf("KMEM_DEBUG: Leaked elements:\n\n");
332 		LIST_FOREACH(i, &kmem_items, next) {
333 			printf("address=%p\n", i);
334 			stack_print_ddb(&i->stack);
335 			printf("\n");
336 		}
337 	}
338 	mtx_unlock(&kmem_items_mtx);
339 }
340 
341 SYSUNINIT(sol_kmem, SI_SUB_CPU, SI_ORDER_FIRST, kmem_show, NULL);
342 #endif	/* KMEM_DEBUG */
343