xref: /freebsd/sys/fs/devfs/devfs_rule.c (revision 19fae0f66023a97a9b464b3beeeabb2081f575b3)
1 /*-
2  * SPDX-License-Identifier: BSD-2-Clause
3  *
4  * Copyright (c) 2002 Dima Dorfman.
5  * All rights reserved.
6  *
7  * Redistribution and use in source and binary forms, with or without
8  * modification, are permitted provided that the following conditions
9  * are met:
10  * 1. Redistributions of source code must retain the above copyright
11  *    notice, this list of conditions and the following disclaimer.
12  * 2. Redistributions in binary form must reproduce the above copyright
13  *    notice, this list of conditions and the following disclaimer in the
14  *    documentation and/or other materials provided with the distribution.
15  *
16  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
17  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
18  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
19  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
20  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
21  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
22  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
23  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
24  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
25  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
26  * SUCH DAMAGE.
27  *
28  * $FreeBSD$
29  */
30 
31 /*
32  * DEVFS ruleset implementation.
33  *
34  * A note on terminology: To "run" a rule on a dirent is to take the
35  * prescribed action; to "apply" a rule is to check whether it matches
36  * a dirent and run if if it does.
37  *
38  * A note on locking: Only foreign entry points (non-static functions)
39  * should deal with locking.  Everything else assumes we already hold
40  * the required kind of lock.
41  *
42  * A note on namespace: devfs_rules_* are the non-static functions for
43  * the entire "ruleset" subsystem, devfs_rule_* are the static
44  * functions that operate on rules, and devfs_ruleset_* are the static
45  * functions that operate on rulesets.  The line between the last two
46  * isn't always clear, but the guideline is still useful.
47  *
48  * A note on "special" identifiers: Ruleset 0 is the NULL, or empty,
49  * ruleset; it cannot be deleted or changed in any way.  This may be
50  * assumed inside the code; e.g., a ruleset of 0 may be interpeted to
51  * mean "no ruleset".  The interpretation of rule 0 is
52  * command-dependent, but in no case is there a real rule with number
53  * 0.
54  *
55  * A note on errno codes: To make it easier for the userland to tell
56  * what went wrong, we sometimes use errno codes that are not entirely
57  * appropriate for the error but that would be less ambiguous than the
58  * appropriate "generic" code.  For example, when we can't find a
59  * ruleset, we return ESRCH instead of ENOENT (except in
60  * DEVFSIO_{R,S}GETNEXT, where a nonexistent ruleset means "end of
61  * list", and the userland expects ENOENT to be this indicator); this
62  * way, when an operation fails, it's clear that what couldn't be
63  * found is a ruleset and not a rule (well, it's clear to those who
64  * know the convention).
65  */
66 
67 #include <sys/param.h>
68 #include <sys/systm.h>
69 #include <sys/conf.h>
70 #include <sys/kernel.h>
71 #include <sys/malloc.h>
72 #include <sys/priv.h>
73 #include <sys/dirent.h>
74 #include <sys/ioccom.h>
75 #include <sys/lock.h>
76 #include <sys/sx.h>
77 
78 #include <fs/devfs/devfs.h>
79 #include <fs/devfs/devfs_int.h>
80 
81 /*
82  * Kernel version of devfs_rule.
83  */
84 struct devfs_krule {
85 	TAILQ_ENTRY(devfs_krule)	dk_list;
86 	struct devfs_ruleset		*dk_ruleset;
87 	struct devfs_rule		dk_rule;
88 };
89 
90 TAILQ_HEAD(rulehead, devfs_krule);
91 static MALLOC_DEFINE(M_DEVFSRULE, "DEVFS_RULE", "DEVFS rule storage");
92 
93 /*
94  * Structure to describe a ruleset.
95  */
96 struct devfs_ruleset {
97 	TAILQ_ENTRY(devfs_ruleset)	ds_list;
98 	struct rulehead			ds_rules;
99 	devfs_rsnum			ds_number;
100 	int				ds_refcount;
101 };
102 
103 static devfs_rid devfs_rid_input(devfs_rid rid, struct devfs_mount *dm);
104 
105 static void devfs_rule_applyde_recursive(struct devfs_krule *dk,
106 		struct devfs_mount *dm, struct devfs_dirent *de);
107 static void devfs_rule_applydm(struct devfs_krule *dk, struct devfs_mount *dm);
108 static int  devfs_rule_autonumber(struct devfs_ruleset *ds, devfs_rnum *rnp);
109 static struct devfs_krule *devfs_rule_byid(devfs_rid rid);
110 static int  devfs_rule_delete(struct devfs_krule *dkp);
111 static struct cdev *devfs_rule_getdev(struct devfs_dirent *de);
112 static int  devfs_rule_input(struct devfs_rule *dr, struct devfs_mount *dm);
113 static int  devfs_rule_insert(struct devfs_rule *dr);
114 static int  devfs_rule_match(struct devfs_krule *dk, struct devfs_mount *dm,
115 		struct devfs_dirent *de);
116 static int  devfs_rule_matchpath(struct devfs_krule *dk, struct devfs_mount *dm,
117 		struct devfs_dirent *de);
118 static void devfs_rule_run(struct devfs_krule *dk, struct devfs_mount *dm,
119 		struct devfs_dirent *de, unsigned depth);
120 
121 static void devfs_ruleset_applyde(struct devfs_ruleset *ds,
122 		struct devfs_mount *dm, struct devfs_dirent *de,
123 		unsigned depth);
124 static void devfs_ruleset_applydm(struct devfs_ruleset *ds,
125 		struct devfs_mount *dm);
126 static struct devfs_ruleset *devfs_ruleset_bynum(devfs_rsnum rsnum);
127 static struct devfs_ruleset *devfs_ruleset_create(devfs_rsnum rsnum);
128 static void devfs_ruleset_reap(struct devfs_ruleset *dsp);
129 static int  devfs_ruleset_use(devfs_rsnum rsnum, struct devfs_mount *dm);
130 
131 static struct sx sx_rules;
132 SX_SYSINIT(sx_rules, &sx_rules, "DEVFS ruleset lock");
133 
134 static TAILQ_HEAD(, devfs_ruleset) devfs_rulesets =
135     TAILQ_HEAD_INITIALIZER(devfs_rulesets);
136 
137 /*
138  * Called to apply the proper rules for 'de' before it can be
139  * exposed to the userland.  This should be called with an exclusive
140  * lock on dm in case we need to run anything.
141  */
142 void
143 devfs_rules_apply(struct devfs_mount *dm, struct devfs_dirent *de)
144 {
145 	struct devfs_ruleset *ds;
146 
147 	sx_assert(&dm->dm_lock, SX_XLOCKED);
148 
149 	if (dm->dm_ruleset == 0)
150 		return;
151 	sx_slock(&sx_rules);
152 	ds = devfs_ruleset_bynum(dm->dm_ruleset);
153 	KASSERT(ds != NULL, ("mount-point has NULL ruleset"));
154 	devfs_ruleset_applyde(ds, dm, de, devfs_rule_depth);
155 	sx_sunlock(&sx_rules);
156 }
157 
158 /*
159  * Rule subsystem ioctl hook.
160  */
161 int
162 devfs_rules_ioctl(struct devfs_mount *dm, u_long cmd, caddr_t data, struct thread *td)
163 {
164 	struct devfs_ruleset *ds;
165 	struct devfs_krule *dk;
166 	struct devfs_rule *dr;
167 	devfs_rsnum rsnum;
168 	devfs_rnum rnum;
169 	devfs_rid rid;
170 	int error;
171 
172 	sx_assert(&dm->dm_lock, SX_XLOCKED);
173 
174 	/*
175 	 * XXX: This returns an error regardless of whether we actually
176 	 * support the cmd or not.
177 	 *
178 	 * We could make this privileges finer grained if desired.
179 	 */
180 	error = priv_check(td, PRIV_DEVFS_RULE);
181 	if (error)
182 		return (error);
183 
184 	sx_xlock(&sx_rules);
185 
186 	switch (cmd) {
187 	case DEVFSIO_RADD:
188 		dr = (struct devfs_rule *)data;
189 		error = devfs_rule_input(dr, dm);
190 		if (error != 0)
191 			break;
192 		dk = devfs_rule_byid(dr->dr_id);
193 		if (dk != NULL) {
194 			error = EEXIST;
195 			break;
196 		}
197 		if (rid2rsn(dr->dr_id) == 0) {
198 			error = EIO;
199 			break;
200 		}
201 		error = devfs_rule_insert(dr);
202 		break;
203 	case DEVFSIO_RAPPLY:
204 		dr = (struct devfs_rule *)data;
205 		error = devfs_rule_input(dr, dm);
206 		if (error != 0)
207 			break;
208 
209 		/*
210 		 * This is one of many possible hackish
211 		 * implementations.  The primary contender is an
212 		 * implementation where the rule we read in is
213 		 * temporarily inserted into some ruleset, perhaps
214 		 * with a hypothetical DRO_NOAUTO flag so that it
215 		 * doesn't get used where it isn't intended, and
216 		 * applied in the normal way.  This can be done in the
217 		 * userland (DEVFSIO_ADD, DEVFSIO_APPLYID,
218 		 * DEVFSIO_DEL) or in the kernel; either way it breaks
219 		 * some corner case assumptions in other parts of the
220 		 * code (not that this implementation doesn't do
221 		 * that).
222 		 */
223 		if (dr->dr_iacts & DRA_INCSET &&
224 		    devfs_ruleset_bynum(dr->dr_incset) == NULL) {
225 			error = ESRCH;
226 			break;
227 		}
228 		dk = malloc(sizeof(*dk), M_TEMP, M_WAITOK | M_ZERO);
229 		memcpy(&dk->dk_rule, dr, sizeof(*dr));
230 		devfs_rule_applydm(dk, dm);
231 		free(dk, M_TEMP);
232 		break;
233 	case DEVFSIO_RAPPLYID:
234 		rid = *(devfs_rid *)data;
235 		rid = devfs_rid_input(rid, dm);
236 		dk = devfs_rule_byid(rid);
237 		if (dk == NULL) {
238 			error = ENOENT;
239 			break;
240 		}
241 		devfs_rule_applydm(dk, dm);
242 		break;
243 	case DEVFSIO_RDEL:
244 		rid = *(devfs_rid *)data;
245 		rid = devfs_rid_input(rid, dm);
246 		dk = devfs_rule_byid(rid);
247 		if (dk == NULL) {
248 			error = ENOENT;
249 			break;
250 		}
251 		ds = dk->dk_ruleset;
252 		error = devfs_rule_delete(dk);
253 		break;
254 	case DEVFSIO_RGETNEXT:
255 		dr = (struct devfs_rule *)data;
256 		error = devfs_rule_input(dr, dm);
257 		if (error != 0)
258 			break;
259 		/*
260 		 * We can't use devfs_rule_byid() here since that
261 		 * requires the rule specified to exist, but we want
262 		 * getnext(N) to work whether there is a rule N or not
263 		 * (specifically, getnext(0) must work, but we should
264 		 * never have a rule 0 since the add command
265 		 * interprets 0 to mean "auto-number").
266 		 */
267 		ds = devfs_ruleset_bynum(rid2rsn(dr->dr_id));
268 		if (ds == NULL) {
269 			error = ENOENT;
270 			break;
271 		}
272 		rnum = rid2rn(dr->dr_id);
273 		TAILQ_FOREACH(dk, &ds->ds_rules, dk_list) {
274 			if (rid2rn(dk->dk_rule.dr_id) > rnum)
275 				break;
276 		}
277 		if (dk == NULL) {
278 			error = ENOENT;
279 			break;
280 		}
281 		memcpy(dr, &dk->dk_rule, sizeof(*dr));
282 		break;
283 	case DEVFSIO_SUSE:
284 		rsnum = *(devfs_rsnum *)data;
285 		error = devfs_ruleset_use(rsnum, dm);
286 		break;
287 	case DEVFSIO_SAPPLY:
288 		rsnum = *(devfs_rsnum *)data;
289 		rsnum = rid2rsn(devfs_rid_input(mkrid(rsnum, 0), dm));
290 		ds = devfs_ruleset_bynum(rsnum);
291 		if (ds == NULL) {
292 			error = ESRCH;
293 			break;
294 		}
295 		devfs_ruleset_applydm(ds, dm);
296 		break;
297 	case DEVFSIO_SGETNEXT:
298 		rsnum = *(devfs_rsnum *)data;
299 		TAILQ_FOREACH(ds, &devfs_rulesets, ds_list) {
300 			if (ds->ds_number > rsnum)
301 				break;
302 		}
303 		if (ds == NULL) {
304 			error = ENOENT;
305 			break;
306 		}
307 		*(devfs_rsnum *)data = ds->ds_number;
308 		break;
309 	default:
310 		error = ENOIOCTL;
311 		break;
312 	}
313 
314 	sx_xunlock(&sx_rules);
315 	return (error);
316 }
317 
318 /*
319  * Adjust the rule identifier to use the ruleset of dm if one isn't
320  * explicitly specified.
321  *
322  * Note that after this operation, rid2rsn(rid) might still be 0, and
323  * that's okay; ruleset 0 is a valid ruleset, but when it's read in
324  * from the userland, it means "current ruleset for this mount-point".
325  */
326 static devfs_rid
327 devfs_rid_input(devfs_rid rid, struct devfs_mount *dm)
328 {
329 
330 	if (rid2rsn(rid) == 0)
331 		return (mkrid(dm->dm_ruleset, rid2rn(rid)));
332 	else
333 		return (rid);
334 }
335 
336 /*
337  * Apply dk to de and everything under de.
338  *
339  * XXX: This method needs a function call for every nested
340  * subdirectory in a devfs mount.  If we plan to have many of these,
341  * we might eventually run out of kernel stack space.
342  * XXX: a linear search could be done through the cdev list instead.
343  */
344 static void
345 devfs_rule_applyde_recursive(struct devfs_krule *dk, struct devfs_mount *dm,
346     struct devfs_dirent *de)
347 {
348 	struct devfs_dirent *de2;
349 
350 	TAILQ_FOREACH(de2, &de->de_dlist, de_list)
351 		devfs_rule_applyde_recursive(dk, dm, de2);
352 	devfs_rule_run(dk, dm, de, devfs_rule_depth);
353 }
354 
355 /*
356  * Apply dk to all entires in dm.
357  */
358 static void
359 devfs_rule_applydm(struct devfs_krule *dk, struct devfs_mount *dm)
360 {
361 
362 	devfs_rule_applyde_recursive(dk, dm, dm->dm_rootdir);
363 }
364 
365 /*
366  * Automatically select a number for a new rule in ds, and write the
367  * result into rnump.
368  */
369 static int
370 devfs_rule_autonumber(struct devfs_ruleset *ds, devfs_rnum *rnump)
371 {
372 	struct devfs_krule *dk;
373 
374 	/* Find the last rule. */
375 	dk = TAILQ_LAST(&ds->ds_rules, rulehead);
376 	if (dk == NULL)
377 		*rnump = 100;
378 	else {
379 		*rnump = rid2rn(dk->dk_rule.dr_id) + 100;
380 		/* Detect overflow. */
381 		if (*rnump < rid2rn(dk->dk_rule.dr_id))
382 			return (ERANGE);
383 	}
384 	KASSERT(devfs_rule_byid(mkrid(ds->ds_number, *rnump)) == NULL,
385 	    ("autonumbering resulted in an already existing rule"));
386 	return (0);
387 }
388 
389 /*
390  * Find a krule by id.
391  */
392 static struct devfs_krule *
393 devfs_rule_byid(devfs_rid rid)
394 {
395 	struct devfs_ruleset *ds;
396 	struct devfs_krule *dk;
397 	devfs_rnum rn;
398 
399 	rn = rid2rn(rid);
400 	ds = devfs_ruleset_bynum(rid2rsn(rid));
401 	if (ds == NULL)
402 		return (NULL);
403 	TAILQ_FOREACH(dk, &ds->ds_rules, dk_list) {
404 		if (rid2rn(dk->dk_rule.dr_id) == rn)
405 			return (dk);
406 		else if (rid2rn(dk->dk_rule.dr_id) > rn)
407 			break;
408 	}
409 	return (NULL);
410 }
411 
412 /*
413  * Remove dkp from any lists it may be on and remove memory associated
414  * with it.
415  */
416 static int
417 devfs_rule_delete(struct devfs_krule *dk)
418 {
419 	struct devfs_ruleset *ds;
420 
421 	if (dk->dk_rule.dr_iacts & DRA_INCSET) {
422 		ds = devfs_ruleset_bynum(dk->dk_rule.dr_incset);
423 		KASSERT(ds != NULL, ("DRA_INCSET but bad dr_incset"));
424 		--ds->ds_refcount;
425 		devfs_ruleset_reap(ds);
426 	}
427 	ds = dk->dk_ruleset;
428 	TAILQ_REMOVE(&ds->ds_rules, dk, dk_list);
429 	devfs_ruleset_reap(ds);
430 	free(dk, M_DEVFSRULE);
431 	return (0);
432 }
433 
434 /*
435  * Get a struct cdev *corresponding to de so we can try to match rules based
436  * on it.  If this routine returns NULL, there is no struct cdev *associated
437  * with the dirent (symlinks and directories don't have dev_ts), and
438  * the caller should assume that any critera dependent on a dev_t
439  * don't match.
440  */
441 static struct cdev *
442 devfs_rule_getdev(struct devfs_dirent *de)
443 {
444 
445 	if (de->de_cdp == NULL)
446 		return (NULL);
447 	if (de->de_cdp->cdp_flags & CDP_ACTIVE)
448 		return (&de->de_cdp->cdp_c);
449 	else
450 		return (NULL);
451 }
452 
453 /*
454  * Do what we need to do to a rule that we just loaded from the
455  * userland.  In particular, we need to check the magic, and adjust
456  * the ruleset appropriate if desired.
457  */
458 static int
459 devfs_rule_input(struct devfs_rule *dr, struct devfs_mount *dm)
460 {
461 
462 	if (dr->dr_magic != DEVFS_MAGIC)
463 		return (ERPCMISMATCH);
464 	dr->dr_id = devfs_rid_input(dr->dr_id, dm);
465 	return (0);
466 }
467 
468 /*
469  * Import dr into the appropriate place in the kernel (i.e., make a
470  * krule).  The value of dr is copied, so the pointer may be destroyed
471  * after this call completes.
472  */
473 static int
474 devfs_rule_insert(struct devfs_rule *dr)
475 {
476 	struct devfs_ruleset *ds, *dsi;
477 	struct devfs_krule *k1;
478 	struct devfs_krule *dk;
479 	devfs_rsnum rsnum;
480 	devfs_rnum dkrn;
481 	int error;
482 
483 	/*
484 	 * This stuff seems out of place here, but we want to do it as
485 	 * soon as possible so that if it fails, we don't have to roll
486 	 * back any changes we already made (e.g., ruleset creation).
487 	 */
488 	if (dr->dr_iacts & DRA_INCSET) {
489 		dsi = devfs_ruleset_bynum(dr->dr_incset);
490 		if (dsi == NULL)
491 			return (ESRCH);
492 	} else
493 		dsi = NULL;
494 
495 	rsnum = rid2rsn(dr->dr_id);
496 	KASSERT(rsnum != 0, ("Inserting into ruleset zero"));
497 
498 	ds = devfs_ruleset_bynum(rsnum);
499 	if (ds == NULL)
500 		ds = devfs_ruleset_create(rsnum);
501 	dkrn = rid2rn(dr->dr_id);
502 	if (dkrn == 0) {
503 		error = devfs_rule_autonumber(ds, &dkrn);
504 		if (error != 0) {
505 			devfs_ruleset_reap(ds);
506 			return (error);
507 		}
508 	}
509 
510 	dk = malloc(sizeof(*dk), M_DEVFSRULE, M_WAITOK | M_ZERO);
511 	dk->dk_ruleset = ds;
512 	if (dsi != NULL)
513 		++dsi->ds_refcount;
514 	/* XXX: Inspect dr? */
515 	memcpy(&dk->dk_rule, dr, sizeof(*dr));
516 	dk->dk_rule.dr_id = mkrid(rid2rsn(dk->dk_rule.dr_id), dkrn);
517 
518 	TAILQ_FOREACH(k1, &ds->ds_rules, dk_list) {
519 		if (rid2rn(k1->dk_rule.dr_id) > dkrn) {
520 			TAILQ_INSERT_BEFORE(k1, dk, dk_list);
521 			break;
522 		}
523 	}
524 	if (k1 == NULL)
525 		TAILQ_INSERT_TAIL(&ds->ds_rules, dk, dk_list);
526 	return (0);
527 }
528 
529 /*
530  * Determine whether dk matches de.  Returns 1 if dk should be run on
531  * de; 0, otherwise.
532  */
533 static int
534 devfs_rule_match(struct devfs_krule *dk, struct devfs_mount *dm,
535     struct devfs_dirent *de)
536 {
537 	struct devfs_rule *dr = &dk->dk_rule;
538 	struct cdev *dev;
539 	struct cdevsw *dsw;
540 	int ref;
541 
542 	dev = devfs_rule_getdev(de);
543 	/*
544 	 * At this point, if dev is NULL, we should assume that any
545 	 * criteria that depend on it don't match.  We should *not*
546 	 * just ignore them (i.e., act like they weren't specified),
547 	 * since that makes a rule that only has criteria dependent on
548 	 * the struct cdev *match all symlinks and directories.
549 	 *
550 	 * Note also that the following tests are somewhat reversed:
551 	 * They're actually testing to see whether the condition does
552 	 * *not* match, since the default is to assume the rule should
553 	 * be run (such as if there are no conditions).
554 	 */
555 	if (dr->dr_icond & DRC_DSWFLAGS) {
556 		if (dev == NULL)
557 			return (0);
558 		dsw = dev_refthread(dev, &ref);
559 		if (dsw == NULL)
560 			return (0);
561 		if ((dsw->d_flags & dr->dr_dswflags) == 0) {
562 			dev_relthread(dev, ref);
563 			return (0);
564 		}
565 		dev_relthread(dev, ref);
566 	}
567 	if (dr->dr_icond & DRC_PATHPTRN)
568 		if (!devfs_rule_matchpath(dk, dm, de))
569 			return (0);
570 
571 	return (1);
572 }
573 
574 /*
575  * Determine whether dk matches de on account of dr_pathptrn.
576  */
577 static int
578 devfs_rule_matchpath(struct devfs_krule *dk, struct devfs_mount *dm,
579     struct devfs_dirent *de)
580 {
581 	struct devfs_rule *dr = &dk->dk_rule;
582 	struct cdev *dev;
583 	int match;
584 	char *pname, *specname;
585 
586 	specname = NULL;
587 	dev = devfs_rule_getdev(de);
588 	if (dev != NULL)
589 		pname = dev->si_name;
590 	else if (de->de_dirent->d_type == DT_LNK ||
591 	    (de->de_dirent->d_type == DT_DIR && de != dm->dm_rootdir &&
592 	    (de->de_flags & (DE_DOT | DE_DOTDOT)) == 0)) {
593 		specname = malloc(SPECNAMELEN + 1, M_TEMP, M_WAITOK);
594 		pname = devfs_fqpn(specname, dm, de, NULL);
595 	} else
596 		return (0);
597 
598 	KASSERT(pname != NULL, ("devfs_rule_matchpath: NULL pname"));
599 	match = fnmatch(dr->dr_pathptrn, pname, FNM_PATHNAME) == 0;
600 	free(specname, M_TEMP);
601 	return (match);
602 }
603 
604 /*
605  * Run dk on de.
606  */
607 static void
608 devfs_rule_run(struct devfs_krule *dk,  struct devfs_mount *dm,
609     struct devfs_dirent *de, unsigned depth)
610 {
611 	struct devfs_rule *dr = &dk->dk_rule;
612 	struct devfs_ruleset *ds;
613 
614 	if (!devfs_rule_match(dk, dm, de))
615 		return;
616 	if (dr->dr_iacts & DRA_BACTS) {
617 		if (dr->dr_bacts & DRB_HIDE)
618 			de->de_flags |= DE_WHITEOUT;
619 		if (dr->dr_bacts & DRB_UNHIDE)
620 			de->de_flags &= ~DE_WHITEOUT;
621 	}
622 	if (dr->dr_iacts & DRA_UID)
623 		de->de_uid = dr->dr_uid;
624 	if (dr->dr_iacts & DRA_GID)
625 		de->de_gid = dr->dr_gid;
626 	if (dr->dr_iacts & DRA_MODE)
627 		de->de_mode = dr->dr_mode;
628 	if (dr->dr_iacts & DRA_INCSET) {
629 		/*
630 		 * XXX: we should tell the user if the depth is exceeded here
631 		 * XXX: but it is not obvious how to.  A return value will
632 		 * XXX: not work as this is called when devices are created
633 		 * XXX: long time after the rules were instantiated.
634 		 * XXX: a printf() would probably give too much noise, or
635 		 * XXX: DoS the machine.  I guess a rate-limited message
636 		 * XXX: might work.
637 		 */
638 		if (depth > 0) {
639 			ds = devfs_ruleset_bynum(dk->dk_rule.dr_incset);
640 			KASSERT(ds != NULL, ("DRA_INCSET but bad dr_incset"));
641 			devfs_ruleset_applyde(ds, dm, de, depth - 1);
642 		}
643 	}
644 }
645 
646 /*
647  * Apply all the rules in ds to de.
648  */
649 static void
650 devfs_ruleset_applyde(struct devfs_ruleset *ds, struct devfs_mount *dm,
651     struct devfs_dirent *de, unsigned depth)
652 {
653 	struct devfs_krule *dk;
654 
655 	TAILQ_FOREACH(dk, &ds->ds_rules, dk_list)
656 		devfs_rule_run(dk, dm, de, depth);
657 }
658 
659 /*
660  * Apply all the rules in ds to all the entires in dm.
661  */
662 static void
663 devfs_ruleset_applydm(struct devfs_ruleset *ds, struct devfs_mount *dm)
664 {
665 	struct devfs_krule *dk;
666 
667 	/*
668 	 * XXX: Does it matter whether we do
669 	 *
670 	 *	foreach(dk in ds)
671 	 *		foreach(de in dm)
672 	 *			apply(dk to de)
673 	 *
674 	 * as opposed to
675 	 *
676 	 *	foreach(de in dm)
677 	 *		foreach(dk in ds)
678 	 *			apply(dk to de)
679 	 *
680 	 * The end result is obviously the same, but does the order
681 	 * matter?
682 	 */
683 	TAILQ_FOREACH(dk, &ds->ds_rules, dk_list)
684 		devfs_rule_applydm(dk, dm);
685 }
686 
687 /*
688  * Find a ruleset by number.
689  */
690 static struct devfs_ruleset *
691 devfs_ruleset_bynum(devfs_rsnum rsnum)
692 {
693 	struct devfs_ruleset *ds;
694 
695 	TAILQ_FOREACH(ds, &devfs_rulesets, ds_list) {
696 		if (ds->ds_number == rsnum)
697 			return (ds);
698 	}
699 	return (NULL);
700 }
701 
702 /*
703  * Create a new ruleset.
704  */
705 static struct devfs_ruleset *
706 devfs_ruleset_create(devfs_rsnum rsnum)
707 {
708 	struct devfs_ruleset *s1;
709 	struct devfs_ruleset *ds;
710 
711 	KASSERT(rsnum != 0, ("creating ruleset zero"));
712 
713 	KASSERT(devfs_ruleset_bynum(rsnum) == NULL,
714 	    ("creating already existent ruleset %d", rsnum));
715 
716 	ds = malloc(sizeof(*ds), M_DEVFSRULE, M_WAITOK | M_ZERO);
717 	ds->ds_number = rsnum;
718 	TAILQ_INIT(&ds->ds_rules);
719 
720 	TAILQ_FOREACH(s1, &devfs_rulesets, ds_list) {
721 		if (s1->ds_number > rsnum) {
722 			TAILQ_INSERT_BEFORE(s1, ds, ds_list);
723 			break;
724 		}
725 	}
726 	if (s1 == NULL)
727 		TAILQ_INSERT_TAIL(&devfs_rulesets, ds, ds_list);
728 	return (ds);
729 }
730 
731 /*
732  * Remove a ruleset from the system if it's empty and not used
733  * anywhere.  This should be called after every time a rule is deleted
734  * from this ruleset or the reference count is decremented.
735  */
736 static void
737 devfs_ruleset_reap(struct devfs_ruleset *ds)
738 {
739 
740 	KASSERT(ds->ds_number != 0, ("reaping ruleset zero "));
741 
742 	if (!TAILQ_EMPTY(&ds->ds_rules) || ds->ds_refcount != 0)
743 		return;
744 
745 	TAILQ_REMOVE(&devfs_rulesets, ds, ds_list);
746 	free(ds, M_DEVFSRULE);
747 }
748 
749 /*
750  * Make rsnum the active ruleset for dm.
751  */
752 static int
753 devfs_ruleset_use(devfs_rsnum rsnum, struct devfs_mount *dm)
754 {
755 	struct devfs_ruleset *cds, *ds;
756 
757 	if (dm->dm_ruleset != 0) {
758 		cds = devfs_ruleset_bynum(dm->dm_ruleset);
759 		--cds->ds_refcount;
760 		devfs_ruleset_reap(cds);
761 	}
762 
763 	if (rsnum == 0) {
764 		dm->dm_ruleset = 0;
765 		return (0);
766 	}
767 
768 	ds = devfs_ruleset_bynum(rsnum);
769 	if (ds == NULL)
770 		ds = devfs_ruleset_create(rsnum);
771 	/* These should probably be made atomic somehow. */
772 	++ds->ds_refcount;
773 	dm->dm_ruleset = rsnum;
774 
775 	return (0);
776 }
777 
778 void
779 devfs_rules_cleanup(struct devfs_mount *dm)
780 {
781 	struct devfs_ruleset *ds;
782 
783 	sx_assert(&dm->dm_lock, SX_XLOCKED);
784 	if (dm->dm_ruleset != 0) {
785 		ds = devfs_ruleset_bynum(dm->dm_ruleset);
786 		--ds->ds_refcount;
787 		devfs_ruleset_reap(ds);
788 	}
789 }
790 
791 /*
792  * Make rsnum the active ruleset for dm (locked)
793  */
794 void
795 devfs_ruleset_set(devfs_rsnum rsnum, struct devfs_mount *dm)
796 {
797 
798 	sx_assert(&dm->dm_lock, SX_XLOCKED);
799 
800 	sx_xlock(&sx_rules);
801 	devfs_ruleset_use(rsnum, dm);
802 	sx_xunlock(&sx_rules);
803 }
804 
805 /*
806  * Apply the current active ruleset on a mount
807  */
808 void
809 devfs_ruleset_apply(struct devfs_mount *dm)
810 {
811 	struct devfs_ruleset *ds;
812 
813 	sx_assert(&dm->dm_lock, SX_XLOCKED);
814 
815 	sx_xlock(&sx_rules);
816 	if (dm->dm_ruleset == 0) {
817 		sx_xunlock(&sx_rules);
818 		return;
819 	}
820 	ds = devfs_ruleset_bynum(dm->dm_ruleset);
821 	if (ds != NULL)
822 		devfs_ruleset_applydm(ds, dm);
823 	sx_xunlock(&sx_rules);
824 }
825