xref: /freebsd/sys/kern/kern_procctl.c (revision 058ac3e8063366dafa634d9107642e12b038bf09)
1 /*-
2  * Copyright (c) 2014 John Baldwin
3  * Copyright (c) 2014, 2016 The FreeBSD Foundation
4  *
5  * Portions of this software were developed by Konstantin Belousov
6  * under sponsorship from the FreeBSD Foundation.
7  *
8  * Redistribution and use in source and binary forms, with or without
9  * modification, are permitted provided that the following conditions
10  * are met:
11  * 1. Redistributions of source code must retain the above copyright
12  *    notice, this list of conditions and the following disclaimer.
13  * 2. Redistributions in binary form must reproduce the above copyright
14  *    notice, this list of conditions and the following disclaimer in the
15  *    documentation and/or other materials provided with the distribution.
16  *
17  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
18  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
19  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
20  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
21  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
22  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
23  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
24  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
25  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
26  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
27  * SUCH DAMAGE.
28  */
29 
30 #include <sys/cdefs.h>
31 __FBSDID("$FreeBSD$");
32 
33 #include <sys/param.h>
34 #include <sys/_unrhdr.h>
35 #include <sys/systm.h>
36 #include <sys/capsicum.h>
37 #include <sys/lock.h>
38 #include <sys/mman.h>
39 #include <sys/mutex.h>
40 #include <sys/priv.h>
41 #include <sys/proc.h>
42 #include <sys/procctl.h>
43 #include <sys/sx.h>
44 #include <sys/syscallsubr.h>
45 #include <sys/sysproto.h>
46 #include <sys/taskqueue.h>
47 #include <sys/wait.h>
48 
49 #include <vm/vm.h>
50 #include <vm/pmap.h>
51 #include <vm/vm_map.h>
52 #include <vm/vm_extern.h>
53 
54 static int
55 protect_setchild(struct thread *td, struct proc *p, int flags)
56 {
57 
58 	PROC_LOCK_ASSERT(p, MA_OWNED);
59 	if (p->p_flag & P_SYSTEM || p_cansched(td, p) != 0)
60 		return (0);
61 	if (flags & PPROT_SET) {
62 		p->p_flag |= P_PROTECTED;
63 		if (flags & PPROT_INHERIT)
64 			p->p_flag2 |= P2_INHERIT_PROTECTED;
65 	} else {
66 		p->p_flag &= ~P_PROTECTED;
67 		p->p_flag2 &= ~P2_INHERIT_PROTECTED;
68 	}
69 	return (1);
70 }
71 
72 static int
73 protect_setchildren(struct thread *td, struct proc *top, int flags)
74 {
75 	struct proc *p;
76 	int ret;
77 
78 	p = top;
79 	ret = 0;
80 	sx_assert(&proctree_lock, SX_LOCKED);
81 	for (;;) {
82 		ret |= protect_setchild(td, p, flags);
83 		PROC_UNLOCK(p);
84 		/*
85 		 * If this process has children, descend to them next,
86 		 * otherwise do any siblings, and if done with this level,
87 		 * follow back up the tree (but not past top).
88 		 */
89 		if (!LIST_EMPTY(&p->p_children))
90 			p = LIST_FIRST(&p->p_children);
91 		else for (;;) {
92 			if (p == top) {
93 				PROC_LOCK(p);
94 				return (ret);
95 			}
96 			if (LIST_NEXT(p, p_sibling)) {
97 				p = LIST_NEXT(p, p_sibling);
98 				break;
99 			}
100 			p = p->p_pptr;
101 		}
102 		PROC_LOCK(p);
103 	}
104 }
105 
106 static int
107 protect_set(struct thread *td, struct proc *p, void *data)
108 {
109 	int error, flags, ret;
110 
111 	flags = *(int *)data;
112 	switch (PPROT_OP(flags)) {
113 	case PPROT_SET:
114 	case PPROT_CLEAR:
115 		break;
116 	default:
117 		return (EINVAL);
118 	}
119 
120 	if ((PPROT_FLAGS(flags) & ~(PPROT_DESCEND | PPROT_INHERIT)) != 0)
121 		return (EINVAL);
122 
123 	error = priv_check(td, PRIV_VM_MADV_PROTECT);
124 	if (error)
125 		return (error);
126 
127 	if (flags & PPROT_DESCEND)
128 		ret = protect_setchildren(td, p, flags);
129 	else
130 		ret = protect_setchild(td, p, flags);
131 	if (ret == 0)
132 		return (EPERM);
133 	return (0);
134 }
135 
136 static int
137 reap_acquire(struct thread *td, struct proc *p, void *data __unused)
138 {
139 
140 	sx_assert(&proctree_lock, SX_XLOCKED);
141 	if (p != td->td_proc)
142 		return (EPERM);
143 	if ((p->p_treeflag & P_TREE_REAPER) != 0)
144 		return (EBUSY);
145 	p->p_treeflag |= P_TREE_REAPER;
146 	/*
147 	 * We do not reattach existing children and the whole tree
148 	 * under them to us, since p->p_reaper already seen them.
149 	 */
150 	return (0);
151 }
152 
153 static int
154 reap_release(struct thread *td, struct proc *p, void *data __unused)
155 {
156 
157 	sx_assert(&proctree_lock, SX_XLOCKED);
158 	if (p != td->td_proc)
159 		return (EPERM);
160 	if (p == initproc)
161 		return (EINVAL);
162 	if ((p->p_treeflag & P_TREE_REAPER) == 0)
163 		return (EINVAL);
164 	reaper_abandon_children(p, false);
165 	return (0);
166 }
167 
168 static int
169 reap_status(struct thread *td, struct proc *p, void *data)
170 {
171 	struct proc *reap, *p2, *first_p;
172 	struct procctl_reaper_status *rs;
173 
174 	rs = data;
175 	sx_assert(&proctree_lock, SX_LOCKED);
176 	if ((p->p_treeflag & P_TREE_REAPER) == 0) {
177 		reap = p->p_reaper;
178 	} else {
179 		reap = p;
180 		rs->rs_flags |= REAPER_STATUS_OWNED;
181 	}
182 	if (reap == initproc)
183 		rs->rs_flags |= REAPER_STATUS_REALINIT;
184 	rs->rs_reaper = reap->p_pid;
185 	rs->rs_descendants = 0;
186 	rs->rs_children = 0;
187 	if (!LIST_EMPTY(&reap->p_reaplist)) {
188 		first_p = LIST_FIRST(&reap->p_children);
189 		if (first_p == NULL)
190 			first_p = LIST_FIRST(&reap->p_reaplist);
191 		rs->rs_pid = first_p->p_pid;
192 		LIST_FOREACH(p2, &reap->p_reaplist, p_reapsibling) {
193 			if (proc_realparent(p2) == reap)
194 				rs->rs_children++;
195 			rs->rs_descendants++;
196 		}
197 	} else {
198 		rs->rs_pid = -1;
199 	}
200 	return (0);
201 }
202 
203 static int
204 reap_getpids(struct thread *td, struct proc *p, void *data)
205 {
206 	struct proc *reap, *p2;
207 	struct procctl_reaper_pidinfo *pi, *pip;
208 	struct procctl_reaper_pids *rp;
209 	u_int i, n;
210 	int error;
211 
212 	rp = data;
213 	sx_assert(&proctree_lock, SX_LOCKED);
214 	PROC_UNLOCK(p);
215 	reap = (p->p_treeflag & P_TREE_REAPER) == 0 ? p->p_reaper : p;
216 	n = i = 0;
217 	error = 0;
218 	LIST_FOREACH(p2, &reap->p_reaplist, p_reapsibling)
219 		n++;
220 	sx_unlock(&proctree_lock);
221 	if (rp->rp_count < n)
222 		n = rp->rp_count;
223 	pi = malloc(n * sizeof(*pi), M_TEMP, M_WAITOK);
224 	sx_slock(&proctree_lock);
225 	LIST_FOREACH(p2, &reap->p_reaplist, p_reapsibling) {
226 		if (i == n)
227 			break;
228 		pip = &pi[i];
229 		bzero(pip, sizeof(*pip));
230 		pip->pi_pid = p2->p_pid;
231 		pip->pi_subtree = p2->p_reapsubtree;
232 		pip->pi_flags = REAPER_PIDINFO_VALID;
233 		if (proc_realparent(p2) == reap)
234 			pip->pi_flags |= REAPER_PIDINFO_CHILD;
235 		if ((p2->p_treeflag & P_TREE_REAPER) != 0)
236 			pip->pi_flags |= REAPER_PIDINFO_REAPER;
237 		i++;
238 	}
239 	sx_sunlock(&proctree_lock);
240 	error = copyout(pi, rp->rp_pids, i * sizeof(*pi));
241 	free(pi, M_TEMP);
242 	sx_slock(&proctree_lock);
243 	PROC_LOCK(p);
244 	return (error);
245 }
246 
247 struct reap_kill_proc_work {
248 	struct ucred *cr;
249 	struct proc *target;
250 	ksiginfo_t *ksi;
251 	struct procctl_reaper_kill *rk;
252 	int *error;
253 	struct task t;
254 };
255 
256 static void
257 reap_kill_proc_locked(struct reap_kill_proc_work *w)
258 {
259 	int error1;
260 	bool need_stop;
261 
262 	PROC_LOCK_ASSERT(w->target, MA_OWNED);
263 	PROC_ASSERT_HELD(w->target);
264 
265 	error1 = cr_cansignal(w->cr, w->target, w->rk->rk_sig);
266 	if (error1 != 0) {
267 		if (*w->error == ESRCH) {
268 			w->rk->rk_fpid = w->target->p_pid;
269 			*w->error = error1;
270 		}
271 		return;
272 	}
273 
274 	/*
275 	 * The need_stop indicates if the target process needs to be
276 	 * suspended before being signalled.  This is needed when we
277 	 * guarantee that all processes in subtree are signalled,
278 	 * avoiding the race with some process not yet fully linked
279 	 * into all structures during fork, ignored by iterator, and
280 	 * then escaping signalling.
281 	 *
282 	 * The thread cannot usefully stop itself anyway, and if other
283 	 * thread of the current process forks while the current
284 	 * thread signals the whole subtree, it is an application
285 	 * race.
286 	 */
287 	if ((w->target->p_flag & (P_KPROC | P_SYSTEM | P_STOPPED)) == 0)
288 		need_stop = thread_single(w->target, SINGLE_ALLPROC) == 0;
289 	else
290 		need_stop = false;
291 
292 	(void)pksignal(w->target, w->rk->rk_sig, w->ksi);
293 	w->rk->rk_killed++;
294 	*w->error = error1;
295 
296 	if (need_stop)
297 		thread_single_end(w->target, SINGLE_ALLPROC);
298 }
299 
300 static void
301 reap_kill_proc_work(void *arg, int pending __unused)
302 {
303 	struct reap_kill_proc_work *w;
304 
305 	w = arg;
306 	PROC_LOCK(w->target);
307 	if ((w->target->p_flag2 & P2_WEXIT) == 0)
308 		reap_kill_proc_locked(w);
309 	PROC_UNLOCK(w->target);
310 
311 	sx_xlock(&proctree_lock);
312 	w->target = NULL;
313 	wakeup(&w->target);
314 	sx_xunlock(&proctree_lock);
315 }
316 
317 struct reap_kill_tracker {
318 	struct proc *parent;
319 	TAILQ_ENTRY(reap_kill_tracker) link;
320 };
321 
322 TAILQ_HEAD(reap_kill_tracker_head, reap_kill_tracker);
323 
324 static void
325 reap_kill_sched(struct reap_kill_tracker_head *tracker, struct proc *p2)
326 {
327 	struct reap_kill_tracker *t;
328 
329 	PROC_LOCK(p2);
330 	if ((p2->p_flag2 & P2_WEXIT) != 0) {
331 		PROC_UNLOCK(p2);
332 		return;
333 	}
334 	_PHOLD_LITE(p2);
335 	PROC_UNLOCK(p2);
336 	t = malloc(sizeof(struct reap_kill_tracker), M_TEMP, M_WAITOK);
337 	t->parent = p2;
338 	TAILQ_INSERT_TAIL(tracker, t, link);
339 }
340 
341 static void
342 reap_kill_sched_free(struct reap_kill_tracker *t)
343 {
344 	PRELE(t->parent);
345 	free(t, M_TEMP);
346 }
347 
348 static void
349 reap_kill_children(struct thread *td, struct proc *reaper,
350     struct procctl_reaper_kill *rk, ksiginfo_t *ksi, int *error)
351 {
352 	struct proc *p2;
353 	int error1;
354 
355 	LIST_FOREACH(p2, &reaper->p_children, p_sibling) {
356 		PROC_LOCK(p2);
357 		if ((p2->p_flag2 & P2_WEXIT) == 0) {
358 			error1 = p_cansignal(td, p2, rk->rk_sig);
359 			if (error1 != 0) {
360 				if (*error == ESRCH) {
361 					rk->rk_fpid = p2->p_pid;
362 					*error = error1;
363 				}
364 
365 				/*
366 				 * Do not end the loop on error,
367 				 * signal everything we can.
368 				 */
369 			} else {
370 				(void)pksignal(p2, rk->rk_sig, ksi);
371 				rk->rk_killed++;
372 			}
373 		}
374 		PROC_UNLOCK(p2);
375 	}
376 }
377 
378 static bool
379 reap_kill_subtree_once(struct thread *td, struct proc *p, struct proc *reaper,
380     struct unrhdr *pids, struct reap_kill_proc_work *w)
381 {
382 	struct reap_kill_tracker_head tracker;
383 	struct reap_kill_tracker *t;
384 	struct proc *p2;
385 	int r, xlocked;
386 	bool res, st;
387 
388 	res = false;
389 	TAILQ_INIT(&tracker);
390 	reap_kill_sched(&tracker, reaper);
391 	while ((t = TAILQ_FIRST(&tracker)) != NULL) {
392 		TAILQ_REMOVE(&tracker, t, link);
393 
394 		/*
395 		 * Since reap_kill_proc() drops proctree_lock sx, it
396 		 * is possible that the tracked reaper is no longer.
397 		 * In this case the subtree is reparented to the new
398 		 * reaper, which should handle it.
399 		 */
400 		if ((t->parent->p_treeflag & P_TREE_REAPER) == 0) {
401 			reap_kill_sched_free(t);
402 			res = true;
403 			continue;
404 		}
405 
406 		LIST_FOREACH(p2, &t->parent->p_reaplist, p_reapsibling) {
407 			if (t->parent == reaper &&
408 			    (w->rk->rk_flags & REAPER_KILL_SUBTREE) != 0 &&
409 			    p2->p_reapsubtree != w->rk->rk_subtree)
410 				continue;
411 			if ((p2->p_treeflag & P_TREE_REAPER) != 0)
412 				reap_kill_sched(&tracker, p2);
413 			if (alloc_unr_specific(pids, p2->p_pid) != p2->p_pid)
414 				continue;
415 			if (p2 == td->td_proc) {
416 				if ((p2->p_flag & P_HADTHREADS) != 0 &&
417 				    (p2->p_flag2 & P2_WEXIT) == 0) {
418 					xlocked = sx_xlocked(&proctree_lock);
419 					sx_unlock(&proctree_lock);
420 					st = true;
421 				} else {
422 					st = false;
423 				}
424 				PROC_LOCK(p2);
425 				if (st)
426 					r = thread_single(p2, SINGLE_NO_EXIT);
427 				(void)pksignal(p2, w->rk->rk_sig, w->ksi);
428 				w->rk->rk_killed++;
429 				if (st && r == 0)
430 					thread_single_end(p2, SINGLE_NO_EXIT);
431 				PROC_UNLOCK(p2);
432 				if (st) {
433 					if (xlocked)
434 						sx_xlock(&proctree_lock);
435 					else
436 						sx_slock(&proctree_lock);
437 				}
438 			} else {
439 				PROC_LOCK(p2);
440 				if ((p2->p_flag2 & P2_WEXIT) == 0) {
441 					_PHOLD_LITE(p2);
442 					PROC_UNLOCK(p2);
443 					w->target = p2;
444 					taskqueue_enqueue(taskqueue_thread,
445 					    &w->t);
446 					while (w->target != NULL) {
447 						sx_sleep(&w->target,
448 						    &proctree_lock, PWAIT,
449 						    "reapst", 0);
450 					}
451 					PROC_LOCK(p2);
452 					_PRELE(p2);
453 				}
454 				PROC_UNLOCK(p2);
455 			}
456 			res = true;
457 		}
458 		reap_kill_sched_free(t);
459 	}
460 	return (res);
461 }
462 
463 static void
464 reap_kill_subtree(struct thread *td, struct proc *p, struct proc *reaper,
465     struct reap_kill_proc_work *w)
466 {
467 	struct unrhdr pids;
468 
469 	/*
470 	 * pids records processes which were already signalled, to
471 	 * avoid doubling signals to them if iteration needs to be
472 	 * repeated.
473 	 */
474 	init_unrhdr(&pids, 1, PID_MAX, UNR_NO_MTX);
475 	PROC_LOCK(td->td_proc);
476 	if ((td->td_proc->p_flag2 & P2_WEXIT) != 0) {
477 		PROC_UNLOCK(td->td_proc);
478 		goto out;
479 	}
480 	PROC_UNLOCK(td->td_proc);
481 	while (reap_kill_subtree_once(td, p, reaper, &pids, w))
482 	       ;
483 out:
484 	clean_unrhdr(&pids);
485 	clear_unrhdr(&pids);
486 }
487 
488 static bool
489 reap_kill_sapblk(struct thread *td __unused, void *data)
490 {
491 	struct procctl_reaper_kill *rk;
492 
493 	rk = data;
494 	return ((rk->rk_flags & REAPER_KILL_CHILDREN) == 0);
495 }
496 
497 static int
498 reap_kill(struct thread *td, struct proc *p, void *data)
499 {
500 	struct reap_kill_proc_work w;
501 	struct proc *reaper;
502 	ksiginfo_t ksi;
503 	struct procctl_reaper_kill *rk;
504 	int error;
505 
506 	rk = data;
507 	sx_assert(&proctree_lock, SX_LOCKED);
508 	if (IN_CAPABILITY_MODE(td))
509 		return (ECAPMODE);
510 	if (rk->rk_sig <= 0 || rk->rk_sig > _SIG_MAXSIG ||
511 	    (rk->rk_flags & ~(REAPER_KILL_CHILDREN |
512 	    REAPER_KILL_SUBTREE)) != 0 || (rk->rk_flags &
513 	    (REAPER_KILL_CHILDREN | REAPER_KILL_SUBTREE)) ==
514 	    (REAPER_KILL_CHILDREN | REAPER_KILL_SUBTREE))
515 		return (EINVAL);
516 	PROC_UNLOCK(p);
517 	reaper = (p->p_treeflag & P_TREE_REAPER) == 0 ? p->p_reaper : p;
518 	ksiginfo_init(&ksi);
519 	ksi.ksi_signo = rk->rk_sig;
520 	ksi.ksi_code = SI_USER;
521 	ksi.ksi_pid = td->td_proc->p_pid;
522 	ksi.ksi_uid = td->td_ucred->cr_ruid;
523 	error = ESRCH;
524 	rk->rk_killed = 0;
525 	rk->rk_fpid = -1;
526 	if ((rk->rk_flags & REAPER_KILL_CHILDREN) != 0) {
527 		reap_kill_children(td, reaper, rk, &ksi, &error);
528 	} else {
529 		w.cr = crhold(td->td_ucred);
530 		w.ksi = &ksi;
531 		w.rk = rk;
532 		w.error = &error;
533 		TASK_INIT(&w.t, 0, reap_kill_proc_work, &w);
534 
535 		/*
536 		 * Prevent swapout, since w, ksi, and possibly rk, are
537 		 * allocated on the stack.  We sleep in
538 		 * reap_kill_subtree_once() waiting for task to
539 		 * complete single-threading.
540 		 */
541 		PHOLD(td->td_proc);
542 
543 		reap_kill_subtree(td, p, reaper, &w);
544 		PRELE(td->td_proc);
545 		crfree(w.cr);
546 	}
547 	PROC_LOCK(p);
548 	return (error);
549 }
550 
551 static int
552 trace_ctl(struct thread *td, struct proc *p, void *data)
553 {
554 	int state;
555 
556 	PROC_LOCK_ASSERT(p, MA_OWNED);
557 	state = *(int *)data;
558 
559 	/*
560 	 * Ktrace changes p_traceflag from or to zero under the
561 	 * process lock, so the test does not need to acquire ktrace
562 	 * mutex.
563 	 */
564 	if ((p->p_flag & P_TRACED) != 0 || p->p_traceflag != 0)
565 		return (EBUSY);
566 
567 	switch (state) {
568 	case PROC_TRACE_CTL_ENABLE:
569 		if (td->td_proc != p)
570 			return (EPERM);
571 		p->p_flag2 &= ~(P2_NOTRACE | P2_NOTRACE_EXEC);
572 		break;
573 	case PROC_TRACE_CTL_DISABLE_EXEC:
574 		p->p_flag2 |= P2_NOTRACE_EXEC | P2_NOTRACE;
575 		break;
576 	case PROC_TRACE_CTL_DISABLE:
577 		if ((p->p_flag2 & P2_NOTRACE_EXEC) != 0) {
578 			KASSERT((p->p_flag2 & P2_NOTRACE) != 0,
579 			    ("dandling P2_NOTRACE_EXEC"));
580 			if (td->td_proc != p)
581 				return (EPERM);
582 			p->p_flag2 &= ~P2_NOTRACE_EXEC;
583 		} else {
584 			p->p_flag2 |= P2_NOTRACE;
585 		}
586 		break;
587 	default:
588 		return (EINVAL);
589 	}
590 	return (0);
591 }
592 
593 static int
594 trace_status(struct thread *td, struct proc *p, void *data)
595 {
596 	int *status;
597 
598 	status = data;
599 	if ((p->p_flag2 & P2_NOTRACE) != 0) {
600 		KASSERT((p->p_flag & P_TRACED) == 0,
601 		    ("%d traced but tracing disabled", p->p_pid));
602 		*status = -1;
603 	} else if ((p->p_flag & P_TRACED) != 0) {
604 		*status = p->p_pptr->p_pid;
605 	} else {
606 		*status = 0;
607 	}
608 	return (0);
609 }
610 
611 static int
612 trapcap_ctl(struct thread *td, struct proc *p, void *data)
613 {
614 	int state;
615 
616 	PROC_LOCK_ASSERT(p, MA_OWNED);
617 	state = *(int *)data;
618 
619 	switch (state) {
620 	case PROC_TRAPCAP_CTL_ENABLE:
621 		p->p_flag2 |= P2_TRAPCAP;
622 		break;
623 	case PROC_TRAPCAP_CTL_DISABLE:
624 		p->p_flag2 &= ~P2_TRAPCAP;
625 		break;
626 	default:
627 		return (EINVAL);
628 	}
629 	return (0);
630 }
631 
632 static int
633 trapcap_status(struct thread *td, struct proc *p, void *data)
634 {
635 	int *status;
636 
637 	status = data;
638 	*status = (p->p_flag2 & P2_TRAPCAP) != 0 ? PROC_TRAPCAP_CTL_ENABLE :
639 	    PROC_TRAPCAP_CTL_DISABLE;
640 	return (0);
641 }
642 
643 static int
644 no_new_privs_ctl(struct thread *td, struct proc *p, void *data)
645 {
646 	int state;
647 
648 	PROC_LOCK_ASSERT(p, MA_OWNED);
649 	state = *(int *)data;
650 
651 	if (state != PROC_NO_NEW_PRIVS_ENABLE)
652 		return (EINVAL);
653 	p->p_flag2 |= P2_NO_NEW_PRIVS;
654 	return (0);
655 }
656 
657 static int
658 no_new_privs_status(struct thread *td, struct proc *p, void *data)
659 {
660 
661 	*(int *)data = (p->p_flag2 & P2_NO_NEW_PRIVS) != 0 ?
662 	    PROC_NO_NEW_PRIVS_ENABLE : PROC_NO_NEW_PRIVS_DISABLE;
663 	return (0);
664 }
665 
666 static int
667 protmax_ctl(struct thread *td, struct proc *p, void *data)
668 {
669 	int state;
670 
671 	PROC_LOCK_ASSERT(p, MA_OWNED);
672 	state = *(int *)data;
673 
674 	switch (state) {
675 	case PROC_PROTMAX_FORCE_ENABLE:
676 		p->p_flag2 &= ~P2_PROTMAX_DISABLE;
677 		p->p_flag2 |= P2_PROTMAX_ENABLE;
678 		break;
679 	case PROC_PROTMAX_FORCE_DISABLE:
680 		p->p_flag2 |= P2_PROTMAX_DISABLE;
681 		p->p_flag2 &= ~P2_PROTMAX_ENABLE;
682 		break;
683 	case PROC_PROTMAX_NOFORCE:
684 		p->p_flag2 &= ~(P2_PROTMAX_ENABLE | P2_PROTMAX_DISABLE);
685 		break;
686 	default:
687 		return (EINVAL);
688 	}
689 	return (0);
690 }
691 
692 static int
693 protmax_status(struct thread *td, struct proc *p, void *data)
694 {
695 	int d;
696 
697 	switch (p->p_flag2 & (P2_PROTMAX_ENABLE | P2_PROTMAX_DISABLE)) {
698 	case 0:
699 		d = PROC_PROTMAX_NOFORCE;
700 		break;
701 	case P2_PROTMAX_ENABLE:
702 		d = PROC_PROTMAX_FORCE_ENABLE;
703 		break;
704 	case P2_PROTMAX_DISABLE:
705 		d = PROC_PROTMAX_FORCE_DISABLE;
706 		break;
707 	}
708 	if (kern_mmap_maxprot(p, PROT_READ) == PROT_READ)
709 		d |= PROC_PROTMAX_ACTIVE;
710 	*(int *)data = d;
711 	return (0);
712 }
713 
714 static int
715 aslr_ctl(struct thread *td, struct proc *p, void *data)
716 {
717 	int state;
718 
719 	PROC_LOCK_ASSERT(p, MA_OWNED);
720 	state = *(int *)data;
721 
722 	switch (state) {
723 	case PROC_ASLR_FORCE_ENABLE:
724 		p->p_flag2 &= ~P2_ASLR_DISABLE;
725 		p->p_flag2 |= P2_ASLR_ENABLE;
726 		break;
727 	case PROC_ASLR_FORCE_DISABLE:
728 		p->p_flag2 |= P2_ASLR_DISABLE;
729 		p->p_flag2 &= ~P2_ASLR_ENABLE;
730 		break;
731 	case PROC_ASLR_NOFORCE:
732 		p->p_flag2 &= ~(P2_ASLR_ENABLE | P2_ASLR_DISABLE);
733 		break;
734 	default:
735 		return (EINVAL);
736 	}
737 	return (0);
738 }
739 
740 static int
741 aslr_status(struct thread *td, struct proc *p, void *data)
742 {
743 	struct vmspace *vm;
744 	int d;
745 
746 	switch (p->p_flag2 & (P2_ASLR_ENABLE | P2_ASLR_DISABLE)) {
747 	case 0:
748 		d = PROC_ASLR_NOFORCE;
749 		break;
750 	case P2_ASLR_ENABLE:
751 		d = PROC_ASLR_FORCE_ENABLE;
752 		break;
753 	case P2_ASLR_DISABLE:
754 		d = PROC_ASLR_FORCE_DISABLE;
755 		break;
756 	}
757 	if ((p->p_flag & P_WEXIT) == 0) {
758 		_PHOLD(p);
759 		PROC_UNLOCK(p);
760 		vm = vmspace_acquire_ref(p);
761 		if (vm != NULL) {
762 			if ((vm->vm_map.flags & MAP_ASLR) != 0)
763 				d |= PROC_ASLR_ACTIVE;
764 			vmspace_free(vm);
765 		}
766 		PROC_LOCK(p);
767 		_PRELE(p);
768 	}
769 	*(int *)data = d;
770 	return (0);
771 }
772 
773 static int
774 stackgap_ctl(struct thread *td, struct proc *p, void *data)
775 {
776 	int state;
777 
778 	PROC_LOCK_ASSERT(p, MA_OWNED);
779 	state = *(int *)data;
780 
781 	if ((state & ~(PROC_STACKGAP_ENABLE | PROC_STACKGAP_DISABLE |
782 	    PROC_STACKGAP_ENABLE_EXEC | PROC_STACKGAP_DISABLE_EXEC)) != 0)
783 		return (EINVAL);
784 	switch (state & (PROC_STACKGAP_ENABLE | PROC_STACKGAP_DISABLE)) {
785 	case PROC_STACKGAP_ENABLE:
786 		if ((p->p_flag2 & P2_STKGAP_DISABLE) != 0)
787 			return (EINVAL);
788 		break;
789 	case PROC_STACKGAP_DISABLE:
790 		p->p_flag2 |= P2_STKGAP_DISABLE;
791 		break;
792 	case 0:
793 		break;
794 	default:
795 		return (EINVAL);
796 	}
797 	switch (state & (PROC_STACKGAP_ENABLE_EXEC |
798 	    PROC_STACKGAP_DISABLE_EXEC)) {
799 	case PROC_STACKGAP_ENABLE_EXEC:
800 		p->p_flag2 &= ~P2_STKGAP_DISABLE_EXEC;
801 		break;
802 	case PROC_STACKGAP_DISABLE_EXEC:
803 		p->p_flag2 |= P2_STKGAP_DISABLE_EXEC;
804 		break;
805 	case 0:
806 		break;
807 	default:
808 		return (EINVAL);
809 	}
810 	return (0);
811 }
812 
813 static int
814 stackgap_status(struct thread *td, struct proc *p, void *data)
815 {
816 	int d;
817 
818 	PROC_LOCK_ASSERT(p, MA_OWNED);
819 
820 	d = (p->p_flag2 & P2_STKGAP_DISABLE) != 0 ? PROC_STACKGAP_DISABLE :
821 	    PROC_STACKGAP_ENABLE;
822 	d |= (p->p_flag2 & P2_STKGAP_DISABLE_EXEC) != 0 ?
823 	    PROC_STACKGAP_DISABLE_EXEC : PROC_STACKGAP_ENABLE_EXEC;
824 	*(int *)data = d;
825 	return (0);
826 }
827 
828 static int
829 wxmap_ctl(struct thread *td, struct proc *p, void *data)
830 {
831 	struct vmspace *vm;
832 	vm_map_t map;
833 	int state;
834 
835 	PROC_LOCK_ASSERT(p, MA_OWNED);
836 	if ((p->p_flag & P_WEXIT) != 0)
837 		return (ESRCH);
838 	state = *(int *)data;
839 
840 	switch (state) {
841 	case PROC_WX_MAPPINGS_PERMIT:
842 		p->p_flag2 |= P2_WXORX_DISABLE;
843 		_PHOLD(p);
844 		PROC_UNLOCK(p);
845 		vm = vmspace_acquire_ref(p);
846 		if (vm != NULL) {
847 			map = &vm->vm_map;
848 			vm_map_lock(map);
849 			map->flags &= ~MAP_WXORX;
850 			vm_map_unlock(map);
851 			vmspace_free(vm);
852 		}
853 		PROC_LOCK(p);
854 		_PRELE(p);
855 		break;
856 	case PROC_WX_MAPPINGS_DISALLOW_EXEC:
857 		p->p_flag2 |= P2_WXORX_ENABLE_EXEC;
858 		break;
859 	default:
860 		return (EINVAL);
861 	}
862 
863 	return (0);
864 }
865 
866 static int
867 wxmap_status(struct thread *td, struct proc *p, void *data)
868 {
869 	struct vmspace *vm;
870 	int d;
871 
872 	PROC_LOCK_ASSERT(p, MA_OWNED);
873 	if ((p->p_flag & P_WEXIT) != 0)
874 		return (ESRCH);
875 
876 	d = 0;
877 	if ((p->p_flag2 & P2_WXORX_DISABLE) != 0)
878 		d |= PROC_WX_MAPPINGS_PERMIT;
879 	if ((p->p_flag2 & P2_WXORX_ENABLE_EXEC) != 0)
880 		d |= PROC_WX_MAPPINGS_DISALLOW_EXEC;
881 	_PHOLD(p);
882 	PROC_UNLOCK(p);
883 	vm = vmspace_acquire_ref(p);
884 	if (vm != NULL) {
885 		if ((vm->vm_map.flags & MAP_WXORX) != 0)
886 			d |= PROC_WXORX_ENFORCE;
887 		vmspace_free(vm);
888 	}
889 	PROC_LOCK(p);
890 	_PRELE(p);
891 	*(int *)data = d;
892 	return (0);
893 }
894 
895 static int
896 pdeathsig_ctl(struct thread *td, struct proc *p, void *data)
897 {
898 	int signum;
899 
900 	signum = *(int *)data;
901 	if (p != td->td_proc || (signum != 0 && !_SIG_VALID(signum)))
902 		return (EINVAL);
903 	p->p_pdeathsig = signum;
904 	return (0);
905 }
906 
907 static int
908 pdeathsig_status(struct thread *td, struct proc *p, void *data)
909 {
910 	if (p != td->td_proc)
911 		return (EINVAL);
912 	*(int *)data = p->p_pdeathsig;
913 	return (0);
914 }
915 
916 enum {
917 	PCTL_SLOCKED,
918 	PCTL_XLOCKED,
919 	PCTL_UNLOCKED,
920 };
921 
922 struct procctl_cmd_info {
923 	int lock_tree;
924 	bool one_proc : 1;
925 	bool esrch_is_einval : 1;
926 	bool copyout_on_error : 1;
927 	bool no_nonnull_data : 1;
928 	bool need_candebug : 1;
929 	int copyin_sz;
930 	int copyout_sz;
931 	int (*exec)(struct thread *, struct proc *, void *);
932 	bool (*sapblk)(struct thread *, void *);
933 };
934 static const struct procctl_cmd_info procctl_cmds_info[] = {
935 	[PROC_SPROTECT] =
936 	    { .lock_tree = PCTL_SLOCKED, .one_proc = false,
937 	      .esrch_is_einval = false, .no_nonnull_data = false,
938 	      .need_candebug = false,
939 	      .copyin_sz = sizeof(int), .copyout_sz = 0,
940 	      .exec = protect_set, .copyout_on_error = false, },
941 	[PROC_REAP_ACQUIRE] =
942 	    { .lock_tree = PCTL_XLOCKED, .one_proc = true,
943 	      .esrch_is_einval = false, .no_nonnull_data = true,
944 	      .need_candebug = false,
945 	      .copyin_sz = 0, .copyout_sz = 0,
946 	      .exec = reap_acquire, .copyout_on_error = false, },
947 	[PROC_REAP_RELEASE] =
948 	    { .lock_tree = PCTL_XLOCKED, .one_proc = true,
949 	      .esrch_is_einval = false, .no_nonnull_data = true,
950 	      .need_candebug = false,
951 	      .copyin_sz = 0, .copyout_sz = 0,
952 	      .exec = reap_release, .copyout_on_error = false, },
953 	[PROC_REAP_STATUS] =
954 	    { .lock_tree = PCTL_SLOCKED, .one_proc = true,
955 	      .esrch_is_einval = false, .no_nonnull_data = false,
956 	      .need_candebug = false,
957 	      .copyin_sz = 0,
958 	      .copyout_sz = sizeof(struct procctl_reaper_status),
959 	      .exec = reap_status, .copyout_on_error = false, },
960 	[PROC_REAP_GETPIDS] =
961 	    { .lock_tree = PCTL_SLOCKED, .one_proc = true,
962 	      .esrch_is_einval = false, .no_nonnull_data = false,
963 	      .need_candebug = false,
964 	      .copyin_sz = sizeof(struct procctl_reaper_pids),
965 	      .copyout_sz = 0,
966 	      .exec = reap_getpids, .copyout_on_error = false, },
967 	[PROC_REAP_KILL] =
968 	    { .lock_tree = PCTL_SLOCKED, .one_proc = true,
969 	      .esrch_is_einval = false, .no_nonnull_data = false,
970 	      .need_candebug = false,
971 	      .copyin_sz = sizeof(struct procctl_reaper_kill),
972 	      .copyout_sz = sizeof(struct procctl_reaper_kill),
973 	      .exec = reap_kill, .copyout_on_error = true,
974 	      .sapblk = reap_kill_sapblk, },
975 	[PROC_TRACE_CTL] =
976 	    { .lock_tree = PCTL_SLOCKED, .one_proc = false,
977 	      .esrch_is_einval = false, .no_nonnull_data = false,
978 	      .need_candebug = true,
979 	      .copyin_sz = sizeof(int), .copyout_sz = 0,
980 	      .exec = trace_ctl, .copyout_on_error = false, },
981 	[PROC_TRACE_STATUS] =
982 	    { .lock_tree = PCTL_UNLOCKED, .one_proc = true,
983 	      .esrch_is_einval = false, .no_nonnull_data = false,
984 	      .need_candebug = false,
985 	      .copyin_sz = 0, .copyout_sz = sizeof(int),
986 	      .exec = trace_status, .copyout_on_error = false, },
987 	[PROC_TRAPCAP_CTL] =
988 	    { .lock_tree = PCTL_SLOCKED, .one_proc = false,
989 	      .esrch_is_einval = false, .no_nonnull_data = false,
990 	      .need_candebug = true,
991 	      .copyin_sz = sizeof(int), .copyout_sz = 0,
992 	      .exec = trapcap_ctl, .copyout_on_error = false, },
993 	[PROC_TRAPCAP_STATUS] =
994 	    { .lock_tree = PCTL_UNLOCKED, .one_proc = true,
995 	      .esrch_is_einval = false, .no_nonnull_data = false,
996 	      .need_candebug = false,
997 	      .copyin_sz = 0, .copyout_sz = sizeof(int),
998 	      .exec = trapcap_status, .copyout_on_error = false, },
999 	[PROC_PDEATHSIG_CTL] =
1000 	    { .lock_tree = PCTL_UNLOCKED, .one_proc = true,
1001 	      .esrch_is_einval = true, .no_nonnull_data = false,
1002 	      .need_candebug = false,
1003 	      .copyin_sz = sizeof(int), .copyout_sz = 0,
1004 	      .exec = pdeathsig_ctl, .copyout_on_error = false, },
1005 	[PROC_PDEATHSIG_STATUS] =
1006 	    { .lock_tree = PCTL_UNLOCKED, .one_proc = true,
1007 	      .esrch_is_einval = true, .no_nonnull_data = false,
1008 	      .need_candebug = false,
1009 	      .copyin_sz = 0, .copyout_sz = sizeof(int),
1010 	      .exec = pdeathsig_status, .copyout_on_error = false, },
1011 	[PROC_ASLR_CTL] =
1012 	    { .lock_tree = PCTL_UNLOCKED, .one_proc = true,
1013 	      .esrch_is_einval = false, .no_nonnull_data = false,
1014 	      .need_candebug = true,
1015 	      .copyin_sz = sizeof(int), .copyout_sz = 0,
1016 	      .exec = aslr_ctl, .copyout_on_error = false, },
1017 	[PROC_ASLR_STATUS] =
1018 	    { .lock_tree = PCTL_UNLOCKED, .one_proc = true,
1019 	      .esrch_is_einval = false, .no_nonnull_data = false,
1020 	      .need_candebug = false,
1021 	      .copyin_sz = 0, .copyout_sz = sizeof(int),
1022 	      .exec = aslr_status, .copyout_on_error = false, },
1023 	[PROC_PROTMAX_CTL] =
1024 	    { .lock_tree = PCTL_UNLOCKED, .one_proc = true,
1025 	      .esrch_is_einval = false, .no_nonnull_data = false,
1026 	      .need_candebug = true,
1027 	      .copyin_sz = sizeof(int), .copyout_sz = 0,
1028 	      .exec = protmax_ctl, .copyout_on_error = false, },
1029 	[PROC_PROTMAX_STATUS] =
1030 	    { .lock_tree = PCTL_UNLOCKED, .one_proc = true,
1031 	      .esrch_is_einval = false, .no_nonnull_data = false,
1032 	      .need_candebug = false,
1033 	      .copyin_sz = 0, .copyout_sz = sizeof(int),
1034 	      .exec = protmax_status, .copyout_on_error = false, },
1035 	[PROC_STACKGAP_CTL] =
1036 	    { .lock_tree = PCTL_UNLOCKED, .one_proc = true,
1037 	      .esrch_is_einval = false, .no_nonnull_data = false,
1038 	      .need_candebug = true,
1039 	      .copyin_sz = sizeof(int), .copyout_sz = 0,
1040 	      .exec = stackgap_ctl, .copyout_on_error = false, },
1041 	[PROC_STACKGAP_STATUS] =
1042 	    { .lock_tree = PCTL_UNLOCKED, .one_proc = true,
1043 	      .esrch_is_einval = false, .no_nonnull_data = false,
1044 	      .need_candebug = false,
1045 	      .copyin_sz = 0, .copyout_sz = sizeof(int),
1046 	      .exec = stackgap_status, .copyout_on_error = false, },
1047 	[PROC_NO_NEW_PRIVS_CTL] =
1048 	    { .lock_tree = PCTL_SLOCKED, .one_proc = true,
1049 	      .esrch_is_einval = false, .no_nonnull_data = false,
1050 	      .need_candebug = true,
1051 	      .copyin_sz = sizeof(int), .copyout_sz = 0,
1052 	      .exec = no_new_privs_ctl, .copyout_on_error = false, },
1053 	[PROC_NO_NEW_PRIVS_STATUS] =
1054 	    { .lock_tree = PCTL_UNLOCKED, .one_proc = true,
1055 	      .esrch_is_einval = false, .no_nonnull_data = false,
1056 	      .need_candebug = false,
1057 	      .copyin_sz = 0, .copyout_sz = sizeof(int),
1058 	      .exec = no_new_privs_status, .copyout_on_error = false, },
1059 	[PROC_WXMAP_CTL] =
1060 	    { .lock_tree = PCTL_UNLOCKED, .one_proc = true,
1061 	      .esrch_is_einval = false, .no_nonnull_data = false,
1062 	      .need_candebug = true,
1063 	      .copyin_sz = sizeof(int), .copyout_sz = 0,
1064 	      .exec = wxmap_ctl, .copyout_on_error = false, },
1065 	[PROC_WXMAP_STATUS] =
1066 	    { .lock_tree = PCTL_UNLOCKED, .one_proc = true,
1067 	      .esrch_is_einval = false, .no_nonnull_data = false,
1068 	      .need_candebug = false,
1069 	      .copyin_sz = 0, .copyout_sz = sizeof(int),
1070 	      .exec = wxmap_status, .copyout_on_error = false, },
1071 };
1072 
1073 int
1074 sys_procctl(struct thread *td, struct procctl_args *uap)
1075 {
1076 	union {
1077 		struct procctl_reaper_status rs;
1078 		struct procctl_reaper_pids rp;
1079 		struct procctl_reaper_kill rk;
1080 		int flags;
1081 	} x;
1082 	const struct procctl_cmd_info *cmd_info;
1083 	int error, error1;
1084 
1085 	if (uap->com >= PROC_PROCCTL_MD_MIN)
1086 		return (cpu_procctl(td, uap->idtype, uap->id,
1087 		    uap->com, uap->data));
1088 	if (uap->com == 0 || uap->com >= nitems(procctl_cmds_info))
1089 		return (EINVAL);
1090 	cmd_info = &procctl_cmds_info[uap->com];
1091 	bzero(&x, sizeof(x));
1092 
1093 	if (cmd_info->copyin_sz > 0) {
1094 		error = copyin(uap->data, &x, cmd_info->copyin_sz);
1095 		if (error != 0)
1096 			return (error);
1097 	} else if (cmd_info->no_nonnull_data && uap->data != NULL) {
1098 		return (EINVAL);
1099 	}
1100 
1101 	error = kern_procctl(td, uap->idtype, uap->id, uap->com, &x);
1102 
1103 	if (cmd_info->copyout_sz > 0 && (error == 0 ||
1104 	    cmd_info->copyout_on_error)) {
1105 		error1 = copyout(&x, uap->data, cmd_info->copyout_sz);
1106 		if (error == 0)
1107 			error = error1;
1108 	}
1109 	return (error);
1110 }
1111 
1112 static int
1113 kern_procctl_single(struct thread *td, struct proc *p, int com, void *data)
1114 {
1115 
1116 	PROC_LOCK_ASSERT(p, MA_OWNED);
1117 	return (procctl_cmds_info[com].exec(td, p, data));
1118 }
1119 
1120 int
1121 kern_procctl(struct thread *td, idtype_t idtype, id_t id, int com, void *data)
1122 {
1123 	struct pgrp *pg;
1124 	struct proc *p;
1125 	const struct procctl_cmd_info *cmd_info;
1126 	int error, first_error, ok;
1127 	bool sapblk;
1128 
1129 	MPASS(com > 0 && com < nitems(procctl_cmds_info));
1130 	cmd_info = &procctl_cmds_info[com];
1131 	if (idtype != P_PID && cmd_info->one_proc)
1132 		return (EINVAL);
1133 
1134 	sapblk = false;
1135 	if (cmd_info->sapblk != NULL) {
1136 		sapblk = cmd_info->sapblk(td, data);
1137 		if (sapblk && !stop_all_proc_block())
1138 			return (ERESTART);
1139 	}
1140 
1141 	switch (cmd_info->lock_tree) {
1142 	case PCTL_XLOCKED:
1143 		sx_xlock(&proctree_lock);
1144 		break;
1145 	case PCTL_SLOCKED:
1146 		sx_slock(&proctree_lock);
1147 		break;
1148 	default:
1149 		break;
1150 	}
1151 
1152 	switch (idtype) {
1153 	case P_PID:
1154 		if (id == 0) {
1155 			p = td->td_proc;
1156 			error = 0;
1157 			PROC_LOCK(p);
1158 		} else {
1159 			p = pfind(id);
1160 			if (p == NULL) {
1161 				error = cmd_info->esrch_is_einval ?
1162 				    EINVAL : ESRCH;
1163 				break;
1164 			}
1165 			error = cmd_info->need_candebug ? p_candebug(td, p) :
1166 			    p_cansee(td, p);
1167 		}
1168 		if (error == 0)
1169 			error = kern_procctl_single(td, p, com, data);
1170 		PROC_UNLOCK(p);
1171 		break;
1172 	case P_PGID:
1173 		/*
1174 		 * Attempt to apply the operation to all members of the
1175 		 * group.  Ignore processes in the group that can't be
1176 		 * seen.  Ignore errors so long as at least one process is
1177 		 * able to complete the request successfully.
1178 		 */
1179 		pg = pgfind(id);
1180 		if (pg == NULL) {
1181 			error = ESRCH;
1182 			break;
1183 		}
1184 		PGRP_UNLOCK(pg);
1185 		ok = 0;
1186 		first_error = 0;
1187 		LIST_FOREACH(p, &pg->pg_members, p_pglist) {
1188 			PROC_LOCK(p);
1189 			if (p->p_state == PRS_NEW ||
1190 			    p->p_state == PRS_ZOMBIE ||
1191 			    (cmd_info->need_candebug ? p_candebug(td, p) :
1192 			    p_cansee(td, p)) != 0) {
1193 				PROC_UNLOCK(p);
1194 				continue;
1195 			}
1196 			error = kern_procctl_single(td, p, com, data);
1197 			PROC_UNLOCK(p);
1198 			if (error == 0)
1199 				ok = 1;
1200 			else if (first_error == 0)
1201 				first_error = error;
1202 		}
1203 		if (ok)
1204 			error = 0;
1205 		else if (first_error != 0)
1206 			error = first_error;
1207 		else
1208 			/*
1209 			 * Was not able to see any processes in the
1210 			 * process group.
1211 			 */
1212 			error = ESRCH;
1213 		break;
1214 	default:
1215 		error = EINVAL;
1216 		break;
1217 	}
1218 
1219 	switch (cmd_info->lock_tree) {
1220 	case PCTL_XLOCKED:
1221 		sx_xunlock(&proctree_lock);
1222 		break;
1223 	case PCTL_SLOCKED:
1224 		sx_sunlock(&proctree_lock);
1225 		break;
1226 	default:
1227 		break;
1228 	}
1229 	if (sapblk)
1230 		stop_all_proc_unblock();
1231 	return (error);
1232 }
1233