xref: /freebsd/sys/kern/kern_condvar.c (revision 71fad9fdeefd5d874768802125f98ea6450cfa5c)
1238510fcSJason Evans /*-
2238510fcSJason Evans  * Copyright (c) 2000 Jake Burkholder <jake@freebsd.org>.
3238510fcSJason Evans  * All rights reserved.
4238510fcSJason Evans  *
5238510fcSJason Evans  * Redistribution and use in source and binary forms, with or without
6238510fcSJason Evans  * modification, are permitted provided that the following conditions
7238510fcSJason Evans  * are met:
8238510fcSJason Evans  * 1. Redistributions of source code must retain the above copyright
9238510fcSJason Evans  *    notice, this list of conditions and the following disclaimer.
10238510fcSJason Evans  * 2. Redistributions in binary form must reproduce the above copyright
11238510fcSJason Evans  *    notice, this list of conditions and the following disclaimer in the
12238510fcSJason Evans  *    documentation and/or other materials provided with the distribution.
13238510fcSJason Evans  *
14238510fcSJason Evans  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
15238510fcSJason Evans  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16238510fcSJason Evans  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17238510fcSJason Evans  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
18238510fcSJason Evans  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
19238510fcSJason Evans  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
20238510fcSJason Evans  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
21238510fcSJason Evans  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
22238510fcSJason Evans  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23238510fcSJason Evans  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
24238510fcSJason Evans  * SUCH DAMAGE.
25238510fcSJason Evans  *
26238510fcSJason Evans  * $FreeBSD$
27238510fcSJason Evans  */
28238510fcSJason Evans 
29238510fcSJason Evans #include "opt_ktrace.h"
30238510fcSJason Evans 
31238510fcSJason Evans #include <sys/param.h>
32238510fcSJason Evans #include <sys/systm.h>
33fb919e4dSMark Murray #include <sys/lock.h>
34fb919e4dSMark Murray #include <sys/mutex.h>
35238510fcSJason Evans #include <sys/proc.h>
36238510fcSJason Evans #include <sys/kernel.h>
37238510fcSJason Evans #include <sys/ktr.h>
38238510fcSJason Evans #include <sys/condvar.h>
39238510fcSJason Evans #include <sys/signalvar.h>
40238510fcSJason Evans #include <sys/resourcevar.h>
41238510fcSJason Evans #ifdef KTRACE
42238510fcSJason Evans #include <sys/uio.h>
43238510fcSJason Evans #include <sys/ktrace.h>
44238510fcSJason Evans #endif
45238510fcSJason Evans 
46238510fcSJason Evans /*
47238510fcSJason Evans  * Common sanity checks for cv_wait* functions.
48238510fcSJason Evans  */
49b40ce416SJulian Elischer #define	CV_ASSERT(cvp, mp, td) do {					\
50a48740b6SDavid E. O'Brien 	KASSERT((td) != NULL, ("%s: curthread NULL", __func__));	\
5171fad9fdSJulian Elischer 	KASSERT(TD_IS_RUNNING(td), ("%s: not TDS_RUNNING", __func__));	\
52a48740b6SDavid E. O'Brien 	KASSERT((cvp) != NULL, ("%s: cvp NULL", __func__));		\
53a48740b6SDavid E. O'Brien 	KASSERT((mp) != NULL, ("%s: mp NULL", __func__));		\
54238510fcSJason Evans 	mtx_assert((mp), MA_OWNED | MA_NOTRECURSED);			\
55238510fcSJason Evans } while (0)
56238510fcSJason Evans 
57e7876c09SDan Moschuk #ifdef INVARIANTS
58238510fcSJason Evans #define	CV_WAIT_VALIDATE(cvp, mp) do {					\
59238510fcSJason Evans 	if (TAILQ_EMPTY(&(cvp)->cv_waitq)) {				\
60238510fcSJason Evans 		/* Only waiter. */					\
61238510fcSJason Evans 		(cvp)->cv_mtx = (mp);					\
62238510fcSJason Evans 	} else {							\
63238510fcSJason Evans 		/*							\
64238510fcSJason Evans 		 * Other waiter; assert that we're using the		\
65238510fcSJason Evans 		 * same mutex.						\
66238510fcSJason Evans 		 */							\
67238510fcSJason Evans 		KASSERT((cvp)->cv_mtx == (mp),				\
68a48740b6SDavid E. O'Brien 		    ("%s: Multiple mutexes", __func__));		\
69238510fcSJason Evans 	}								\
70238510fcSJason Evans } while (0)
7171fad9fdSJulian Elischer 
72238510fcSJason Evans #define	CV_SIGNAL_VALIDATE(cvp) do {					\
73238510fcSJason Evans 	if (!TAILQ_EMPTY(&(cvp)->cv_waitq)) {				\
74238510fcSJason Evans 		KASSERT(mtx_owned((cvp)->cv_mtx),			\
75a48740b6SDavid E. O'Brien 		    ("%s: Mutex not owned", __func__));			\
76238510fcSJason Evans 	}								\
77238510fcSJason Evans } while (0)
7871fad9fdSJulian Elischer 
79238510fcSJason Evans #else
80238510fcSJason Evans #define	CV_WAIT_VALIDATE(cvp, mp)
81238510fcSJason Evans #define	CV_SIGNAL_VALIDATE(cvp)
82238510fcSJason Evans #endif
83238510fcSJason Evans 
84238510fcSJason Evans static void cv_timedwait_end(void *arg);
85e602ba25SJulian Elischer static void cv_check_upcall(struct thread *td);
86238510fcSJason Evans 
87238510fcSJason Evans /*
88238510fcSJason Evans  * Initialize a condition variable.  Must be called before use.
89238510fcSJason Evans  */
90238510fcSJason Evans void
91238510fcSJason Evans cv_init(struct cv *cvp, const char *desc)
92238510fcSJason Evans {
93238510fcSJason Evans 
94238510fcSJason Evans 	TAILQ_INIT(&cvp->cv_waitq);
95238510fcSJason Evans 	cvp->cv_mtx = NULL;
96238510fcSJason Evans 	cvp->cv_description = desc;
97238510fcSJason Evans }
98238510fcSJason Evans 
99238510fcSJason Evans /*
100238510fcSJason Evans  * Destroy a condition variable.  The condition variable must be re-initialized
101238510fcSJason Evans  * in order to be re-used.
102238510fcSJason Evans  */
103238510fcSJason Evans void
104238510fcSJason Evans cv_destroy(struct cv *cvp)
105238510fcSJason Evans {
106238510fcSJason Evans 
107a48740b6SDavid E. O'Brien 	KASSERT(cv_waitq_empty(cvp), ("%s: cv_waitq non-empty", __func__));
108238510fcSJason Evans }
109238510fcSJason Evans 
110238510fcSJason Evans /*
111238510fcSJason Evans  * Common code for cv_wait* functions.  All require sched_lock.
112238510fcSJason Evans  */
113238510fcSJason Evans 
114238510fcSJason Evans /*
115e602ba25SJulian Elischer  * Decide if we need to queue an upcall.
116e602ba25SJulian Elischer  * This is copied from msleep(), perhaps this should be a common function.
117e602ba25SJulian Elischer  */
118e602ba25SJulian Elischer static void
119e602ba25SJulian Elischer cv_check_upcall(struct thread *td)
120e602ba25SJulian Elischer {
121e602ba25SJulian Elischer 
122e602ba25SJulian Elischer 	/*
123e602ba25SJulian Elischer 	 * If we are capable of async syscalls and there isn't already
124e602ba25SJulian Elischer 	 * another one ready to return, start a new thread
125e602ba25SJulian Elischer 	 * and queue it as ready to run. Note that there is danger here
126e602ba25SJulian Elischer 	 * because we need to make sure that we don't sleep allocating
127e602ba25SJulian Elischer 	 * the thread (recursion here might be bad).
128e602ba25SJulian Elischer 	 * Hence the TDF_INMSLEEP flag.
129e602ba25SJulian Elischer 	 */
130e602ba25SJulian Elischer 	if ((td->td_proc->p_flag & P_KSES) && td->td_mailbox &&
131e602ba25SJulian Elischer 	    (td->td_flags & TDF_INMSLEEP) == 0) {
132e602ba25SJulian Elischer 		/*
133e602ba25SJulian Elischer 		 * If we have no queued work to do,
134e602ba25SJulian Elischer 		 * upcall to the UTS to see if it has more work.
135e602ba25SJulian Elischer 		 * We don't need to upcall now, just queue it.
136e602ba25SJulian Elischer 		 */
137e602ba25SJulian Elischer 		if (TAILQ_FIRST(&td->td_ksegrp->kg_runq) == NULL) {
138e602ba25SJulian Elischer 			/* Don't recurse here! */
139e602ba25SJulian Elischer 			td->td_flags |= TDF_INMSLEEP;
140e602ba25SJulian Elischer 			thread_schedule_upcall(td, td->td_kse);
141e602ba25SJulian Elischer 			td->td_flags &= ~TDF_INMSLEEP;
142e602ba25SJulian Elischer 		}
143e602ba25SJulian Elischer 	}
144e602ba25SJulian Elischer }
145e602ba25SJulian Elischer 
146e602ba25SJulian Elischer /*
147238510fcSJason Evans  * Switch context.
148238510fcSJason Evans  */
149238510fcSJason Evans static __inline void
150b40ce416SJulian Elischer cv_switch(struct thread *td)
151238510fcSJason Evans {
152238510fcSJason Evans 
153e602ba25SJulian Elischer 	cv_check_upcall(td);
15471fad9fdSJulian Elischer 	TD_SET_SLEEPING(td);
15571fad9fdSJulian Elischer 	td->td_proc->p_stats->p_ru.ru_nvcsw++;
156238510fcSJason Evans 	mi_switch();
1579ef3a985SJohn Baldwin 	CTR3(KTR_PROC, "cv_switch: resume thread %p (pid %d, %s)", td,
1589ef3a985SJohn Baldwin 	    td->td_proc->p_pid, td->td_proc->p_comm);
159238510fcSJason Evans }
160238510fcSJason Evans 
161238510fcSJason Evans /*
162238510fcSJason Evans  * Switch context, catching signals.
163238510fcSJason Evans  */
164238510fcSJason Evans static __inline int
165b40ce416SJulian Elischer cv_switch_catch(struct thread *td)
166238510fcSJason Evans {
1679ef3a985SJohn Baldwin 	struct proc *p;
168238510fcSJason Evans 	int sig;
169238510fcSJason Evans 
170238510fcSJason Evans 	/*
171238510fcSJason Evans 	 * We put ourselves on the sleep queue and start our timeout before
172628855e7SJulian Elischer 	 * calling cursig, as we could stop there, and a wakeup or a SIGCONT (or
173238510fcSJason Evans 	 * both) could occur while we were stopped.  A SIGCONT would cause us to
174e602ba25SJulian Elischer 	 * be marked as TDS_SLP without resuming us, thus we must be ready for
175628855e7SJulian Elischer 	 * sleep when cursig is called.  If the wakeup happens while we're
17671fad9fdSJulian Elischer 	 * stopped, td->td_wchan will be 0 upon return from cursig,
17771fad9fdSJulian Elischer 	 * and TD_ON_SLEEPQ() will return false.
178238510fcSJason Evans 	 */
179b40ce416SJulian Elischer 	td->td_flags |= TDF_SINTR;
1809ed346baSBosko Milekic 	mtx_unlock_spin(&sched_lock);
1819ef3a985SJohn Baldwin 	p = td->td_proc;
1829ef3a985SJohn Baldwin 	PROC_LOCK(p);
18371fad9fdSJulian Elischer 	sig = cursig(td);
184e602ba25SJulian Elischer 	if (thread_suspend_check(1))
185e602ba25SJulian Elischer 		sig = SIGSTOP;
1869ed346baSBosko Milekic 	mtx_lock_spin(&sched_lock);
187c86b6ff5SJohn Baldwin 	PROC_UNLOCK(p);
188238510fcSJason Evans 	if (sig != 0) {
18971fad9fdSJulian Elischer 		if (TD_ON_SLEEPQ(td))
190b40ce416SJulian Elischer 			cv_waitq_remove(td);
19171fad9fdSJulian Elischer 		TD_SET_RUNNING(td);
19271fad9fdSJulian Elischer 	} else if (TD_ON_SLEEPQ(td)) {
193b40ce416SJulian Elischer 		cv_switch(td);
194238510fcSJason Evans 	}
195b40ce416SJulian Elischer 	td->td_flags &= ~TDF_SINTR;
196238510fcSJason Evans 
197238510fcSJason Evans 	return sig;
198238510fcSJason Evans }
199238510fcSJason Evans 
200238510fcSJason Evans /*
201b40ce416SJulian Elischer  * Add a thread to the wait queue of a condition variable.
202238510fcSJason Evans  */
203238510fcSJason Evans static __inline void
204b40ce416SJulian Elischer cv_waitq_add(struct cv *cvp, struct thread *td)
205238510fcSJason Evans {
206238510fcSJason Evans 
207b40ce416SJulian Elischer 	td->td_flags |= TDF_CVWAITQ;
20871fad9fdSJulian Elischer 	TD_SET_ON_SLEEPQ(td);
209b40ce416SJulian Elischer 	td->td_wchan = cvp;
210b40ce416SJulian Elischer 	td->td_wmesg = cvp->cv_description;
211b40ce416SJulian Elischer 	td->td_ksegrp->kg_slptime = 0; /* XXXKSE */
2122c100766SJulian Elischer 	td->td_base_pri = td->td_priority;
213b40ce416SJulian Elischer 	CTR3(KTR_PROC, "cv_waitq_add: thread %p (pid %d, %s)", td,
2149ef3a985SJohn Baldwin 	    td->td_proc->p_pid, td->td_proc->p_comm);
215b40ce416SJulian Elischer 	TAILQ_INSERT_TAIL(&cvp->cv_waitq, td, td_slpq);
216238510fcSJason Evans }
217238510fcSJason Evans 
218238510fcSJason Evans /*
219b40ce416SJulian Elischer  * Wait on a condition variable.  The current thread is placed on the condition
220238510fcSJason Evans  * variable's wait queue and suspended.  A cv_signal or cv_broadcast on the same
221b40ce416SJulian Elischer  * condition variable will resume the thread.  The mutex is released before
222238510fcSJason Evans  * sleeping and will be held on return.  It is recommended that the mutex be
223238510fcSJason Evans  * held when cv_signal or cv_broadcast are called.
224238510fcSJason Evans  */
225238510fcSJason Evans void
226238510fcSJason Evans cv_wait(struct cv *cvp, struct mtx *mp)
227238510fcSJason Evans {
228b40ce416SJulian Elischer 	struct thread *td;
229238510fcSJason Evans 	WITNESS_SAVE_DECL(mp);
230238510fcSJason Evans 
231b40ce416SJulian Elischer 	td = curthread;
232238510fcSJason Evans #ifdef KTRACE
2339ba7fe1bSJohn Baldwin 	if (KTRPOINT(td, KTR_CSW))
2349ba7fe1bSJohn Baldwin 		ktrcsw(1, 0);
235238510fcSJason Evans #endif
236b40ce416SJulian Elischer 	CV_ASSERT(cvp, mp, td);
237c739adbfSJohn Baldwin 	WITNESS_SLEEP(0, &mp->mtx_object);
23819284646SJohn Baldwin 	WITNESS_SAVE(&mp->mtx_object, mp);
239238510fcSJason Evans 
240fe799533SAndrew Gallatin 	if (cold ) {
241238510fcSJason Evans 		/*
242fe799533SAndrew Gallatin 		 * During autoconfiguration, just give interrupts
243fe799533SAndrew Gallatin 		 * a chance, then just return.  Don't run any other
244fe799533SAndrew Gallatin 		 * thread or panic below, in case this is the idle
245fe799533SAndrew Gallatin 		 * process and already asleep.
246238510fcSJason Evans 		 */
247238510fcSJason Evans 		return;
248238510fcSJason Evans 	}
2494bc37205SJeffrey Hsu 
2504bc37205SJeffrey Hsu 	mtx_lock_spin(&sched_lock);
2514bc37205SJeffrey Hsu 
252238510fcSJason Evans 	CV_WAIT_VALIDATE(cvp, mp);
253238510fcSJason Evans 
254c86b6ff5SJohn Baldwin 	DROP_GIANT();
255c86b6ff5SJohn Baldwin 	mtx_unlock(mp);
256238510fcSJason Evans 
257b40ce416SJulian Elischer 	cv_waitq_add(cvp, td);
258b40ce416SJulian Elischer 	cv_switch(td);
259238510fcSJason Evans 
2609ed346baSBosko Milekic 	mtx_unlock_spin(&sched_lock);
261238510fcSJason Evans #ifdef KTRACE
2629ba7fe1bSJohn Baldwin 	if (KTRPOINT(td, KTR_CSW))
2639ba7fe1bSJohn Baldwin 		ktrcsw(0, 0);
264238510fcSJason Evans #endif
265238510fcSJason Evans 	PICKUP_GIANT();
2669ed346baSBosko Milekic 	mtx_lock(mp);
26719284646SJohn Baldwin 	WITNESS_RESTORE(&mp->mtx_object, mp);
268238510fcSJason Evans }
269238510fcSJason Evans 
270238510fcSJason Evans /*
271238510fcSJason Evans  * Wait on a condition variable, allowing interruption by signals.  Return 0 if
272b40ce416SJulian Elischer  * the thread was resumed with cv_signal or cv_broadcast, EINTR or ERESTART if
273238510fcSJason Evans  * a signal was caught.  If ERESTART is returned the system call should be
274238510fcSJason Evans  * restarted if possible.
275238510fcSJason Evans  */
276238510fcSJason Evans int
277238510fcSJason Evans cv_wait_sig(struct cv *cvp, struct mtx *mp)
278238510fcSJason Evans {
279b40ce416SJulian Elischer 	struct thread *td;
28066f769feSPeter Wemm 	struct proc *p;
281238510fcSJason Evans 	int rval;
282238510fcSJason Evans 	int sig;
283238510fcSJason Evans 	WITNESS_SAVE_DECL(mp);
284238510fcSJason Evans 
285b40ce416SJulian Elischer 	td = curthread;
2869ef3a985SJohn Baldwin 	p = td->td_proc;
287238510fcSJason Evans 	rval = 0;
288238510fcSJason Evans #ifdef KTRACE
2899ba7fe1bSJohn Baldwin 	if (KTRPOINT(td, KTR_CSW))
2909ba7fe1bSJohn Baldwin 		ktrcsw(1, 0);
291238510fcSJason Evans #endif
292b40ce416SJulian Elischer 	CV_ASSERT(cvp, mp, td);
293c739adbfSJohn Baldwin 	WITNESS_SLEEP(0, &mp->mtx_object);
29419284646SJohn Baldwin 	WITNESS_SAVE(&mp->mtx_object, mp);
295238510fcSJason Evans 
296238510fcSJason Evans 	if (cold || panicstr) {
297238510fcSJason Evans 		/*
298238510fcSJason Evans 		 * After a panic, or during autoconfiguration, just give
299238510fcSJason Evans 		 * interrupts a chance, then just return; don't run any other
300238510fcSJason Evans 		 * procs or panic below, in case this is the idle process and
301238510fcSJason Evans 		 * already asleep.
302238510fcSJason Evans 		 */
303238510fcSJason Evans 		return 0;
304238510fcSJason Evans 	}
3054bc37205SJeffrey Hsu 
3064bc37205SJeffrey Hsu 	mtx_lock_spin(&sched_lock);
3074bc37205SJeffrey Hsu 
308238510fcSJason Evans 	CV_WAIT_VALIDATE(cvp, mp);
309238510fcSJason Evans 
310c86b6ff5SJohn Baldwin 	DROP_GIANT();
311c86b6ff5SJohn Baldwin 	mtx_unlock(mp);
312238510fcSJason Evans 
313b40ce416SJulian Elischer 	cv_waitq_add(cvp, td);
314b40ce416SJulian Elischer 	sig = cv_switch_catch(td);
315238510fcSJason Evans 
3169ed346baSBosko Milekic 	mtx_unlock_spin(&sched_lock);
317238510fcSJason Evans 
3189ef3a985SJohn Baldwin 	PROC_LOCK(p);
319238510fcSJason Evans 	if (sig == 0)
320e602ba25SJulian Elischer 		sig = cursig(td);	/* XXXKSE */
321238510fcSJason Evans 	if (sig != 0) {
3229ef3a985SJohn Baldwin 		if (SIGISMEMBER(p->p_sigacts->ps_sigintr, sig))
323238510fcSJason Evans 			rval = EINTR;
324238510fcSJason Evans 		else
325238510fcSJason Evans 			rval = ERESTART;
326238510fcSJason Evans 	}
3279ef3a985SJohn Baldwin 	PROC_UNLOCK(p);
328e602ba25SJulian Elischer 	if (p->p_flag & P_WEXIT)
329e602ba25SJulian Elischer 		rval = EINTR;
330238510fcSJason Evans 
331238510fcSJason Evans #ifdef KTRACE
3329ba7fe1bSJohn Baldwin 	if (KTRPOINT(td, KTR_CSW))
3339ba7fe1bSJohn Baldwin 		ktrcsw(0, 0);
334238510fcSJason Evans #endif
3359ba7fe1bSJohn Baldwin 	PICKUP_GIANT();
3369ed346baSBosko Milekic 	mtx_lock(mp);
33719284646SJohn Baldwin 	WITNESS_RESTORE(&mp->mtx_object, mp);
338238510fcSJason Evans 
339238510fcSJason Evans 	return (rval);
340238510fcSJason Evans }
341238510fcSJason Evans 
342238510fcSJason Evans /*
343238510fcSJason Evans  * Wait on a condition variable for at most timo/hz seconds.  Returns 0 if the
344238510fcSJason Evans  * process was resumed by cv_signal or cv_broadcast, EWOULDBLOCK if the timeout
345238510fcSJason Evans  * expires.
346238510fcSJason Evans  */
347238510fcSJason Evans int
348238510fcSJason Evans cv_timedwait(struct cv *cvp, struct mtx *mp, int timo)
349238510fcSJason Evans {
350b40ce416SJulian Elischer 	struct thread *td;
351238510fcSJason Evans 	int rval;
352238510fcSJason Evans 	WITNESS_SAVE_DECL(mp);
353238510fcSJason Evans 
354b40ce416SJulian Elischer 	td = curthread;
355238510fcSJason Evans 	rval = 0;
356238510fcSJason Evans #ifdef KTRACE
3579ba7fe1bSJohn Baldwin 	if (KTRPOINT(td, KTR_CSW))
3589ba7fe1bSJohn Baldwin 		ktrcsw(1, 0);
359238510fcSJason Evans #endif
360b40ce416SJulian Elischer 	CV_ASSERT(cvp, mp, td);
361c739adbfSJohn Baldwin 	WITNESS_SLEEP(0, &mp->mtx_object);
36219284646SJohn Baldwin 	WITNESS_SAVE(&mp->mtx_object, mp);
363238510fcSJason Evans 
364238510fcSJason Evans 	if (cold || panicstr) {
365238510fcSJason Evans 		/*
366238510fcSJason Evans 		 * After a panic, or during autoconfiguration, just give
367238510fcSJason Evans 		 * interrupts a chance, then just return; don't run any other
368b40ce416SJulian Elischer 		 * thread or panic below, in case this is the idle process and
369238510fcSJason Evans 		 * already asleep.
370238510fcSJason Evans 		 */
371238510fcSJason Evans 		return 0;
372238510fcSJason Evans 	}
3734bc37205SJeffrey Hsu 
3744bc37205SJeffrey Hsu 	mtx_lock_spin(&sched_lock);
3754bc37205SJeffrey Hsu 
376238510fcSJason Evans 	CV_WAIT_VALIDATE(cvp, mp);
377238510fcSJason Evans 
378c86b6ff5SJohn Baldwin 	DROP_GIANT();
379c86b6ff5SJohn Baldwin 	mtx_unlock(mp);
380238510fcSJason Evans 
381b40ce416SJulian Elischer 	cv_waitq_add(cvp, td);
382b40ce416SJulian Elischer 	callout_reset(&td->td_slpcallout, timo, cv_timedwait_end, td);
383b40ce416SJulian Elischer 	cv_switch(td);
384238510fcSJason Evans 
385b40ce416SJulian Elischer 	if (td->td_flags & TDF_TIMEOUT) {
386b40ce416SJulian Elischer 		td->td_flags &= ~TDF_TIMEOUT;
387238510fcSJason Evans 		rval = EWOULDBLOCK;
388b40ce416SJulian Elischer 	} else if (td->td_flags & TDF_TIMOFAIL)
389b40ce416SJulian Elischer 		td->td_flags &= ~TDF_TIMOFAIL;
390b40ce416SJulian Elischer 	else if (callout_stop(&td->td_slpcallout) == 0) {
39191a4536fSJohn Baldwin 		/*
39291a4536fSJohn Baldwin 		 * Work around race with cv_timedwait_end similar to that
39391a4536fSJohn Baldwin 		 * between msleep and endtsleep.
394d5cb7e14SJulian Elischer 		 * Go back to sleep.
39591a4536fSJohn Baldwin 		 */
39671fad9fdSJulian Elischer 		TD_SET_SLEEPING(td);
397b40ce416SJulian Elischer 		td->td_proc->p_stats->p_ru.ru_nivcsw++;
39891a4536fSJohn Baldwin 		mi_switch();
39971fad9fdSJulian Elischer 		td->td_flags &= ~TDF_TIMOFAIL;
40091a4536fSJohn Baldwin 	}
401238510fcSJason Evans 
402e602ba25SJulian Elischer 	if (td->td_proc->p_flag & P_WEXIT)
403e602ba25SJulian Elischer 		rval = EWOULDBLOCK;
4049ed346baSBosko Milekic 	mtx_unlock_spin(&sched_lock);
405238510fcSJason Evans #ifdef KTRACE
4069ba7fe1bSJohn Baldwin 	if (KTRPOINT(td, KTR_CSW))
4079ba7fe1bSJohn Baldwin 		ktrcsw(0, 0);
408238510fcSJason Evans #endif
409238510fcSJason Evans 	PICKUP_GIANT();
4109ed346baSBosko Milekic 	mtx_lock(mp);
41119284646SJohn Baldwin 	WITNESS_RESTORE(&mp->mtx_object, mp);
412238510fcSJason Evans 
413238510fcSJason Evans 	return (rval);
414238510fcSJason Evans }
415238510fcSJason Evans 
416238510fcSJason Evans /*
417238510fcSJason Evans  * Wait on a condition variable for at most timo/hz seconds, allowing
418b40ce416SJulian Elischer  * interruption by signals.  Returns 0 if the thread was resumed by cv_signal
419238510fcSJason Evans  * or cv_broadcast, EWOULDBLOCK if the timeout expires, and EINTR or ERESTART if
420238510fcSJason Evans  * a signal was caught.
421238510fcSJason Evans  */
422238510fcSJason Evans int
423238510fcSJason Evans cv_timedwait_sig(struct cv *cvp, struct mtx *mp, int timo)
424238510fcSJason Evans {
425b40ce416SJulian Elischer 	struct thread *td;
4269ef3a985SJohn Baldwin 	struct proc *p;
427238510fcSJason Evans 	int rval;
428238510fcSJason Evans 	int sig;
429238510fcSJason Evans 	WITNESS_SAVE_DECL(mp);
430238510fcSJason Evans 
431b40ce416SJulian Elischer 	td = curthread;
4329ef3a985SJohn Baldwin 	p = td->td_proc;
433238510fcSJason Evans 	rval = 0;
434238510fcSJason Evans #ifdef KTRACE
4359ba7fe1bSJohn Baldwin 	if (KTRPOINT(td, KTR_CSW))
4369ba7fe1bSJohn Baldwin 		ktrcsw(1, 0);
437238510fcSJason Evans #endif
438b40ce416SJulian Elischer 	CV_ASSERT(cvp, mp, td);
439c739adbfSJohn Baldwin 	WITNESS_SLEEP(0, &mp->mtx_object);
44019284646SJohn Baldwin 	WITNESS_SAVE(&mp->mtx_object, mp);
441238510fcSJason Evans 
442238510fcSJason Evans 	if (cold || panicstr) {
443238510fcSJason Evans 		/*
444238510fcSJason Evans 		 * After a panic, or during autoconfiguration, just give
445238510fcSJason Evans 		 * interrupts a chance, then just return; don't run any other
446b40ce416SJulian Elischer 		 * thread or panic below, in case this is the idle process and
447238510fcSJason Evans 		 * already asleep.
448238510fcSJason Evans 		 */
449238510fcSJason Evans 		return 0;
450238510fcSJason Evans 	}
4514bc37205SJeffrey Hsu 
4524bc37205SJeffrey Hsu 	mtx_lock_spin(&sched_lock);
4534bc37205SJeffrey Hsu 
454238510fcSJason Evans 	CV_WAIT_VALIDATE(cvp, mp);
455238510fcSJason Evans 
456c86b6ff5SJohn Baldwin 	DROP_GIANT();
457c86b6ff5SJohn Baldwin 	mtx_unlock(mp);
458238510fcSJason Evans 
459b40ce416SJulian Elischer 	cv_waitq_add(cvp, td);
460b40ce416SJulian Elischer 	callout_reset(&td->td_slpcallout, timo, cv_timedwait_end, td);
461b40ce416SJulian Elischer 	sig = cv_switch_catch(td);
462238510fcSJason Evans 
463b40ce416SJulian Elischer 	if (td->td_flags & TDF_TIMEOUT) {
464b40ce416SJulian Elischer 		td->td_flags &= ~TDF_TIMEOUT;
465238510fcSJason Evans 		rval = EWOULDBLOCK;
466b40ce416SJulian Elischer 	} else if (td->td_flags & TDF_TIMOFAIL)
467b40ce416SJulian Elischer 		td->td_flags &= ~TDF_TIMOFAIL;
468b40ce416SJulian Elischer 	else if (callout_stop(&td->td_slpcallout) == 0) {
46991a4536fSJohn Baldwin 		/*
47091a4536fSJohn Baldwin 		 * Work around race with cv_timedwait_end similar to that
47191a4536fSJohn Baldwin 		 * between msleep and endtsleep.
472d5cb7e14SJulian Elischer 		 * Go back to sleep.
47391a4536fSJohn Baldwin 		 */
47471fad9fdSJulian Elischer 		TD_SET_SLEEPING(td);
475b40ce416SJulian Elischer 		td->td_proc->p_stats->p_ru.ru_nivcsw++;
47691a4536fSJohn Baldwin 		mi_switch();
47771fad9fdSJulian Elischer 		td->td_flags &= ~TDF_TIMOFAIL;
47891a4536fSJohn Baldwin 	}
4799ed346baSBosko Milekic 	mtx_unlock_spin(&sched_lock);
480238510fcSJason Evans 
4819ef3a985SJohn Baldwin 	PROC_LOCK(p);
482238510fcSJason Evans 	if (sig == 0)
483e602ba25SJulian Elischer 		sig = cursig(td);
484238510fcSJason Evans 	if (sig != 0) {
4859ef3a985SJohn Baldwin 		if (SIGISMEMBER(p->p_sigacts->ps_sigintr, sig))
486238510fcSJason Evans 			rval = EINTR;
487238510fcSJason Evans 		else
488238510fcSJason Evans 			rval = ERESTART;
489238510fcSJason Evans 	}
4909ef3a985SJohn Baldwin 	PROC_UNLOCK(p);
491238510fcSJason Evans 
492e602ba25SJulian Elischer 	if (p->p_flag & P_WEXIT)
493e602ba25SJulian Elischer 		rval = EINTR;
494e602ba25SJulian Elischer 
495238510fcSJason Evans #ifdef KTRACE
4969ba7fe1bSJohn Baldwin 	if (KTRPOINT(td, KTR_CSW))
4979ba7fe1bSJohn Baldwin 		ktrcsw(0, 0);
498238510fcSJason Evans #endif
4999ba7fe1bSJohn Baldwin 	PICKUP_GIANT();
5009ed346baSBosko Milekic 	mtx_lock(mp);
50119284646SJohn Baldwin 	WITNESS_RESTORE(&mp->mtx_object, mp);
502238510fcSJason Evans 
503238510fcSJason Evans 	return (rval);
504238510fcSJason Evans }
505238510fcSJason Evans 
506238510fcSJason Evans /*
507238510fcSJason Evans  * Common code for signal and broadcast.  Assumes waitq is not empty.  Must be
508238510fcSJason Evans  * called with sched_lock held.
509238510fcSJason Evans  */
510238510fcSJason Evans static __inline void
511238510fcSJason Evans cv_wakeup(struct cv *cvp)
512238510fcSJason Evans {
513b40ce416SJulian Elischer 	struct thread *td;
514238510fcSJason Evans 
515981808d1SJohn Baldwin 	mtx_assert(&sched_lock, MA_OWNED);
516b40ce416SJulian Elischer 	td = TAILQ_FIRST(&cvp->cv_waitq);
517a48740b6SDavid E. O'Brien 	KASSERT(td->td_wchan == cvp, ("%s: bogus wchan", __func__));
518a48740b6SDavid E. O'Brien 	KASSERT(td->td_flags & TDF_CVWAITQ, ("%s: not on waitq", __func__));
51971fad9fdSJulian Elischer 	cv_waitq_remove(td);
52071fad9fdSJulian Elischer 	TD_CLR_SLEEPING(td);
52171fad9fdSJulian Elischer 	setrunnable(td);
522238510fcSJason Evans }
523238510fcSJason Evans 
524238510fcSJason Evans /*
525b40ce416SJulian Elischer  * Signal a condition variable, wakes up one waiting thread.  Will also wakeup
526238510fcSJason Evans  * the swapper if the process is not in memory, so that it can bring the
527b40ce416SJulian Elischer  * sleeping process in.  Note that this may also result in additional threads
528238510fcSJason Evans  * being made runnable.  Should be called with the same mutex as was passed to
529238510fcSJason Evans  * cv_wait held.
530238510fcSJason Evans  */
531238510fcSJason Evans void
532238510fcSJason Evans cv_signal(struct cv *cvp)
533238510fcSJason Evans {
534238510fcSJason Evans 
535a48740b6SDavid E. O'Brien 	KASSERT(cvp != NULL, ("%s: cvp NULL", __func__));
5369ed346baSBosko Milekic 	mtx_lock_spin(&sched_lock);
537238510fcSJason Evans 	if (!TAILQ_EMPTY(&cvp->cv_waitq)) {
538238510fcSJason Evans 		CV_SIGNAL_VALIDATE(cvp);
539238510fcSJason Evans 		cv_wakeup(cvp);
540238510fcSJason Evans 	}
5419ed346baSBosko Milekic 	mtx_unlock_spin(&sched_lock);
542238510fcSJason Evans }
543238510fcSJason Evans 
544238510fcSJason Evans /*
545b40ce416SJulian Elischer  * Broadcast a signal to a condition variable.  Wakes up all waiting threads.
546238510fcSJason Evans  * Should be called with the same mutex as was passed to cv_wait held.
547238510fcSJason Evans  */
548238510fcSJason Evans void
549238510fcSJason Evans cv_broadcast(struct cv *cvp)
550238510fcSJason Evans {
551238510fcSJason Evans 
552a48740b6SDavid E. O'Brien 	KASSERT(cvp != NULL, ("%s: cvp NULL", __func__));
5539ed346baSBosko Milekic 	mtx_lock_spin(&sched_lock);
554238510fcSJason Evans 	CV_SIGNAL_VALIDATE(cvp);
555238510fcSJason Evans 	while (!TAILQ_EMPTY(&cvp->cv_waitq))
556238510fcSJason Evans 		cv_wakeup(cvp);
5579ed346baSBosko Milekic 	mtx_unlock_spin(&sched_lock);
558238510fcSJason Evans }
559238510fcSJason Evans 
560238510fcSJason Evans /*
561b40ce416SJulian Elischer  * Remove a thread from the wait queue of its condition variable.  This may be
562238510fcSJason Evans  * called externally.
563238510fcSJason Evans  */
564238510fcSJason Evans void
565b40ce416SJulian Elischer cv_waitq_remove(struct thread *td)
566238510fcSJason Evans {
567238510fcSJason Evans 	struct cv *cvp;
568238510fcSJason Evans 
56971fad9fdSJulian Elischer 	mtx_assert(&sched_lock, MA_OWNED);
570b40ce416SJulian Elischer 	if ((cvp = td->td_wchan) != NULL && td->td_flags & TDF_CVWAITQ) {
571b40ce416SJulian Elischer 		TAILQ_REMOVE(&cvp->cv_waitq, td, td_slpq);
572b40ce416SJulian Elischer 		td->td_flags &= ~TDF_CVWAITQ;
57371fad9fdSJulian Elischer 		TD_CLR_ON_SLEEPQ(td);
574238510fcSJason Evans 	}
575238510fcSJason Evans }
576238510fcSJason Evans 
577238510fcSJason Evans /*
578b40ce416SJulian Elischer  * Timeout function for cv_timedwait.  Put the thread on the runqueue and set
579238510fcSJason Evans  * its timeout flag.
580238510fcSJason Evans  */
581238510fcSJason Evans static void
582238510fcSJason Evans cv_timedwait_end(void *arg)
583238510fcSJason Evans {
584b40ce416SJulian Elischer 	struct thread *td;
585238510fcSJason Evans 
586b40ce416SJulian Elischer 	td = arg;
58771fad9fdSJulian Elischer 	CTR3(KTR_PROC, "cv_timedwait_end: thread %p (pid %d, %s)",
58871fad9fdSJulian Elischer 	    td, td->td_proc->p_pid, td->td_proc->p_comm);
5899ed346baSBosko Milekic 	mtx_lock_spin(&sched_lock);
59071fad9fdSJulian Elischer 	if (TD_ON_SLEEPQ(td)) {
591b40ce416SJulian Elischer 		cv_waitq_remove(td);
592b40ce416SJulian Elischer 		td->td_flags |= TDF_TIMEOUT;
59371fad9fdSJulian Elischer 	} else {
594b40ce416SJulian Elischer 		td->td_flags |= TDF_TIMOFAIL;
59571fad9fdSJulian Elischer 	}
59671fad9fdSJulian Elischer 	TD_CLR_SLEEPING(td);
59771fad9fdSJulian Elischer 	setrunnable(td);
5989ed346baSBosko Milekic 	mtx_unlock_spin(&sched_lock);
599238510fcSJason Evans }
600e602ba25SJulian Elischer 
601e602ba25SJulian Elischer /*
602e602ba25SJulian Elischer  * For now only abort interruptable waits.
603e602ba25SJulian Elischer  * The others will have to either complete on their own or have a timeout.
604e602ba25SJulian Elischer  */
605e602ba25SJulian Elischer void
606e602ba25SJulian Elischer cv_abort(struct thread *td)
607e602ba25SJulian Elischer {
608e602ba25SJulian Elischer 
609e602ba25SJulian Elischer 	CTR3(KTR_PROC, "cv_abort: thread %p (pid %d, %s)", td,
61071fad9fdSJulian Elischer 	    td->td_proc->p_pid, td->td_proc->p_comm);
611e602ba25SJulian Elischer 	mtx_lock_spin(&sched_lock);
612e602ba25SJulian Elischer 	if ((td->td_flags & (TDF_SINTR|TDF_TIMEOUT)) == TDF_SINTR) {
61371fad9fdSJulian Elischer 		if (TD_ON_SLEEPQ(td)) {
614e602ba25SJulian Elischer 			cv_waitq_remove(td);
615e602ba25SJulian Elischer 		}
61671fad9fdSJulian Elischer 		TD_CLR_SLEEPING(td);
61771fad9fdSJulian Elischer 		setrunnable(td);
618e602ba25SJulian Elischer 	}
619e602ba25SJulian Elischer 	mtx_unlock_spin(&sched_lock);
620e602ba25SJulian Elischer }
621e602ba25SJulian Elischer 
622