xref: /freebsd/sys/kern/subr_kobj.c (revision a2aef24aa3c8458e4036735dd6928b4ef77294e5)
1 /*-
2  * Copyright (c) 2000,2003 Doug Rabson
3  * All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions
7  * are met:
8  * 1. Redistributions of source code must retain the above copyright
9  *    notice, this list of conditions and the following disclaimer.
10  * 2. Redistributions in binary form must reproduce the above copyright
11  *    notice, this list of conditions and the following disclaimer in the
12  *    documentation and/or other materials provided with the distribution.
13  *
14  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
15  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
18  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
19  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
20  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
21  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
22  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
24  * SUCH DAMAGE.
25  */
26 
27 #include <sys/cdefs.h>
28 __FBSDID("$FreeBSD$");
29 
30 #include <sys/param.h>
31 #include <sys/kernel.h>
32 #include <sys/kobj.h>
33 #include <sys/lock.h>
34 #include <sys/malloc.h>
35 #include <sys/mutex.h>
36 #include <sys/sysctl.h>
37 #ifndef TEST
38 #include <sys/systm.h>
39 #endif
40 
41 #ifdef TEST
42 #include "usertest.h"
43 #endif
44 
45 static MALLOC_DEFINE(M_KOBJ, "kobj", "Kernel object structures");
46 
47 #ifdef KOBJ_STATS
48 
49 u_int kobj_lookup_hits;
50 u_int kobj_lookup_misses;
51 
52 SYSCTL_UINT(_kern, OID_AUTO, kobj_hits, CTLFLAG_RD,
53 	   &kobj_lookup_hits, 0, "");
54 SYSCTL_UINT(_kern, OID_AUTO, kobj_misses, CTLFLAG_RD,
55 	   &kobj_lookup_misses, 0, "");
56 
57 #endif
58 
59 static struct mtx kobj_mtx;
60 static int kobj_mutex_inited;
61 static int kobj_next_id = 1;
62 
63 #define	KOBJ_LOCK()		mtx_lock(&kobj_mtx)
64 #define	KOBJ_UNLOCK()		mtx_unlock(&kobj_mtx)
65 #define	KOBJ_ASSERT(what)	mtx_assert(&kobj_mtx, what);
66 
67 SYSCTL_INT(_kern, OID_AUTO, kobj_methodcount, CTLFLAG_RD,
68 	   &kobj_next_id, 0, "");
69 
70 static void
71 kobj_init_mutex(void *arg)
72 {
73 	if (!kobj_mutex_inited) {
74 		mtx_init(&kobj_mtx, "kobj", NULL, MTX_DEF);
75 		kobj_mutex_inited = 1;
76 	}
77 }
78 
79 SYSINIT(kobj, SI_SUB_LOCK, SI_ORDER_ANY, kobj_init_mutex, NULL);
80 
81 /*
82  * This method structure is used to initialise new caches. Since the
83  * desc pointer is NULL, it is guaranteed never to match any read
84  * descriptors.
85  */
86 static const struct kobj_method null_method = {
87 	0, 0,
88 };
89 
90 int
91 kobj_error_method(void)
92 {
93 
94 	return ENXIO;
95 }
96 
97 static void
98 kobj_class_compile_common(kobj_class_t cls, kobj_ops_t ops)
99 {
100 	kobj_method_t *m;
101 	int i;
102 
103 	/*
104 	 * Don't do anything if we are already compiled.
105 	 */
106 	if (cls->ops)
107 		return;
108 
109 	/*
110 	 * First register any methods which need it.
111 	 */
112 	for (i = 0, m = cls->methods; m->desc; i++, m++) {
113 		if (m->desc->id == 0)
114 			m->desc->id = kobj_next_id++;
115 	}
116 
117 	/*
118 	 * Then initialise the ops table.
119 	 */
120 	for (i = 0; i < KOBJ_CACHE_SIZE; i++)
121 		ops->cache[i] = &null_method;
122 	ops->cls = cls;
123 	cls->ops = ops;
124 }
125 
126 void
127 kobj_class_compile(kobj_class_t cls)
128 {
129 	kobj_ops_t ops;
130 
131 	KOBJ_ASSERT(MA_NOTOWNED);
132 
133 	/*
134 	 * Allocate space for the compiled ops table.
135 	 */
136 	ops = malloc(sizeof(struct kobj_ops), M_KOBJ, M_NOWAIT);
137 	if (!ops)
138 		panic("%s: out of memory", __func__);
139 
140 	KOBJ_LOCK();
141 
142 	/*
143 	 * We may have lost a race for kobj_class_compile here - check
144 	 * to make sure someone else hasn't already compiled this
145 	 * class.
146 	 */
147 	if (cls->ops) {
148 		KOBJ_UNLOCK();
149 		free(ops, M_KOBJ);
150 		return;
151 	}
152 
153 	kobj_class_compile_common(cls, ops);
154 	KOBJ_UNLOCK();
155 }
156 
157 void
158 kobj_class_compile_static(kobj_class_t cls, kobj_ops_t ops)
159 {
160 
161 	KASSERT(kobj_mutex_inited == 0,
162 	    ("%s: only supported during early cycles", __func__));
163 
164 	/*
165 	 * Increment refs to make sure that the ops table is not freed.
166 	 */
167 	cls->refs++;
168 	kobj_class_compile_common(cls, ops);
169 }
170 
171 static kobj_method_t*
172 kobj_lookup_method_class(kobj_class_t cls, kobjop_desc_t desc)
173 {
174 	kobj_method_t *methods = cls->methods;
175 	kobj_method_t *ce;
176 
177 	for (ce = methods; ce && ce->desc; ce++) {
178 		if (ce->desc == desc) {
179 			return ce;
180 		}
181 	}
182 
183 	return NULL;
184 }
185 
186 static kobj_method_t*
187 kobj_lookup_method_mi(kobj_class_t cls,
188 		      kobjop_desc_t desc)
189 {
190 	kobj_method_t *ce;
191 	kobj_class_t *basep;
192 
193 	ce = kobj_lookup_method_class(cls, desc);
194 	if (ce)
195 		return ce;
196 
197 	basep = cls->baseclasses;
198 	if (basep) {
199 		for (; *basep; basep++) {
200 			ce = kobj_lookup_method_mi(*basep, desc);
201 			if (ce)
202 				return ce;
203 		}
204 	}
205 
206 	return NULL;
207 }
208 
209 kobj_method_t*
210 kobj_lookup_method(kobj_class_t cls,
211 		   kobj_method_t **cep,
212 		   kobjop_desc_t desc)
213 {
214 	kobj_method_t *ce;
215 
216 	ce = kobj_lookup_method_mi(cls, desc);
217 	if (!ce)
218 		ce = &desc->deflt;
219 	if (cep)
220 		*cep = ce;
221 	return ce;
222 }
223 
224 void
225 kobj_class_free(kobj_class_t cls)
226 {
227 	void* ops = NULL;
228 
229 	KOBJ_ASSERT(MA_NOTOWNED);
230 	KOBJ_LOCK();
231 
232 	/*
233 	 * Protect against a race between kobj_create and
234 	 * kobj_delete.
235 	 */
236 	if (cls->refs == 0) {
237 		/*
238 		 * For now we don't do anything to unregister any methods
239 		 * which are no longer used.
240 		 */
241 
242 		/*
243 		 * Free memory and clean up.
244 		 */
245 		ops = cls->ops;
246 		cls->ops = NULL;
247 	}
248 
249 	KOBJ_UNLOCK();
250 
251 	if (ops)
252 		free(ops, M_KOBJ);
253 }
254 
255 kobj_t
256 kobj_create(kobj_class_t cls,
257 	    struct malloc_type *mtype,
258 	    int mflags)
259 {
260 	kobj_t obj;
261 
262 	/*
263 	 * Allocate and initialise the new object.
264 	 */
265 	obj = malloc(cls->size, mtype, mflags | M_ZERO);
266 	if (!obj)
267 		return NULL;
268 	kobj_init(obj, cls);
269 
270 	return obj;
271 }
272 
273 static void
274 kobj_init_common(kobj_t obj, kobj_class_t cls)
275 {
276 
277 	obj->ops = cls->ops;
278 	cls->refs++;
279 }
280 
281 void
282 kobj_init(kobj_t obj, kobj_class_t cls)
283 {
284 	KOBJ_ASSERT(MA_NOTOWNED);
285   retry:
286 	KOBJ_LOCK();
287 
288 	/*
289 	 * Consider compiling the class' method table.
290 	 */
291 	if (!cls->ops) {
292 		/*
293 		 * kobj_class_compile doesn't want the lock held
294 		 * because of the call to malloc - we drop the lock
295 		 * and re-try.
296 		 */
297 		KOBJ_UNLOCK();
298 		kobj_class_compile(cls);
299 		goto retry;
300 	}
301 
302 	kobj_init_common(obj, cls);
303 
304 	KOBJ_UNLOCK();
305 }
306 
307 void
308 kobj_init_static(kobj_t obj, kobj_class_t cls)
309 {
310 
311 	KASSERT(kobj_mutex_inited == 0,
312 	    ("%s: only supported during early cycles", __func__));
313 
314 	kobj_init_common(obj, cls);
315 }
316 
317 void
318 kobj_delete(kobj_t obj, struct malloc_type *mtype)
319 {
320 	kobj_class_t cls = obj->ops->cls;
321 	int refs;
322 
323 	/*
324 	 * Consider freeing the compiled method table for the class
325 	 * after its last instance is deleted. As an optimisation, we
326 	 * should defer this for a short while to avoid thrashing.
327 	 */
328 	KOBJ_ASSERT(MA_NOTOWNED);
329 	KOBJ_LOCK();
330 	cls->refs--;
331 	refs = cls->refs;
332 	KOBJ_UNLOCK();
333 
334 	if (!refs)
335 		kobj_class_free(cls);
336 
337 	obj->ops = NULL;
338 	if (mtype)
339 		free(obj, mtype);
340 }
341