1238510fcSJason Evans /*- 2238510fcSJason Evans * Copyright (c) 2000 Jake Burkholder <jake@freebsd.org>. 3238510fcSJason Evans * All rights reserved. 4238510fcSJason Evans * 5238510fcSJason Evans * Redistribution and use in source and binary forms, with or without 6238510fcSJason Evans * modification, are permitted provided that the following conditions 7238510fcSJason Evans * are met: 8238510fcSJason Evans * 1. Redistributions of source code must retain the above copyright 9238510fcSJason Evans * notice, this list of conditions and the following disclaimer. 10238510fcSJason Evans * 2. Redistributions in binary form must reproduce the above copyright 11238510fcSJason Evans * notice, this list of conditions and the following disclaimer in the 12238510fcSJason Evans * documentation and/or other materials provided with the distribution. 13238510fcSJason Evans * 14238510fcSJason Evans * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 15238510fcSJason Evans * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 16238510fcSJason Evans * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 17238510fcSJason Evans * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 18238510fcSJason Evans * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 19238510fcSJason Evans * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 20238510fcSJason Evans * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 21238510fcSJason Evans * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 22238510fcSJason Evans * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 23238510fcSJason Evans * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 24238510fcSJason Evans * SUCH DAMAGE. 25238510fcSJason Evans * 26238510fcSJason Evans * $FreeBSD$ 27238510fcSJason Evans */ 28238510fcSJason Evans 29238510fcSJason Evans #include "opt_ktrace.h" 30238510fcSJason Evans 31238510fcSJason Evans #include <sys/param.h> 32238510fcSJason Evans #include <sys/systm.h> 33fb919e4dSMark Murray #include <sys/lock.h> 34fb919e4dSMark Murray #include <sys/mutex.h> 35238510fcSJason Evans #include <sys/proc.h> 36238510fcSJason Evans #include <sys/kernel.h> 37238510fcSJason Evans #include <sys/ktr.h> 38238510fcSJason Evans #include <sys/condvar.h> 39238510fcSJason Evans #include <sys/signalvar.h> 40238510fcSJason Evans #include <sys/resourcevar.h> 41238510fcSJason Evans #ifdef KTRACE 42238510fcSJason Evans #include <sys/uio.h> 43238510fcSJason Evans #include <sys/ktrace.h> 44238510fcSJason Evans #endif 45238510fcSJason Evans 46238510fcSJason Evans /* 47238510fcSJason Evans * Common sanity checks for cv_wait* functions. 48238510fcSJason Evans */ 49b40ce416SJulian Elischer #define CV_ASSERT(cvp, mp, td) do { \ 50a48740b6SDavid E. O'Brien KASSERT((td) != NULL, ("%s: curthread NULL", __func__)); \ 5171fad9fdSJulian Elischer KASSERT(TD_IS_RUNNING(td), ("%s: not TDS_RUNNING", __func__)); \ 52a48740b6SDavid E. O'Brien KASSERT((cvp) != NULL, ("%s: cvp NULL", __func__)); \ 53a48740b6SDavid E. O'Brien KASSERT((mp) != NULL, ("%s: mp NULL", __func__)); \ 54238510fcSJason Evans mtx_assert((mp), MA_OWNED | MA_NOTRECURSED); \ 55238510fcSJason Evans } while (0) 56238510fcSJason Evans 57e7876c09SDan Moschuk #ifdef INVARIANTS 58238510fcSJason Evans #define CV_WAIT_VALIDATE(cvp, mp) do { \ 59238510fcSJason Evans if (TAILQ_EMPTY(&(cvp)->cv_waitq)) { \ 60238510fcSJason Evans /* Only waiter. */ \ 61238510fcSJason Evans (cvp)->cv_mtx = (mp); \ 62238510fcSJason Evans } else { \ 63238510fcSJason Evans /* \ 64238510fcSJason Evans * Other waiter; assert that we're using the \ 65238510fcSJason Evans * same mutex. \ 66238510fcSJason Evans */ \ 67238510fcSJason Evans KASSERT((cvp)->cv_mtx == (mp), \ 68a48740b6SDavid E. O'Brien ("%s: Multiple mutexes", __func__)); \ 69238510fcSJason Evans } \ 70238510fcSJason Evans } while (0) 7171fad9fdSJulian Elischer 72238510fcSJason Evans #define CV_SIGNAL_VALIDATE(cvp) do { \ 73238510fcSJason Evans if (!TAILQ_EMPTY(&(cvp)->cv_waitq)) { \ 74238510fcSJason Evans KASSERT(mtx_owned((cvp)->cv_mtx), \ 75a48740b6SDavid E. O'Brien ("%s: Mutex not owned", __func__)); \ 76238510fcSJason Evans } \ 77238510fcSJason Evans } while (0) 7871fad9fdSJulian Elischer 79238510fcSJason Evans #else 80238510fcSJason Evans #define CV_WAIT_VALIDATE(cvp, mp) 81238510fcSJason Evans #define CV_SIGNAL_VALIDATE(cvp) 82238510fcSJason Evans #endif 83238510fcSJason Evans 84238510fcSJason Evans static void cv_timedwait_end(void *arg); 85e602ba25SJulian Elischer static void cv_check_upcall(struct thread *td); 86238510fcSJason Evans 87238510fcSJason Evans /* 88238510fcSJason Evans * Initialize a condition variable. Must be called before use. 89238510fcSJason Evans */ 90238510fcSJason Evans void 91238510fcSJason Evans cv_init(struct cv *cvp, const char *desc) 92238510fcSJason Evans { 93238510fcSJason Evans 94238510fcSJason Evans TAILQ_INIT(&cvp->cv_waitq); 95238510fcSJason Evans cvp->cv_mtx = NULL; 96238510fcSJason Evans cvp->cv_description = desc; 97238510fcSJason Evans } 98238510fcSJason Evans 99238510fcSJason Evans /* 100238510fcSJason Evans * Destroy a condition variable. The condition variable must be re-initialized 101238510fcSJason Evans * in order to be re-used. 102238510fcSJason Evans */ 103238510fcSJason Evans void 104238510fcSJason Evans cv_destroy(struct cv *cvp) 105238510fcSJason Evans { 106238510fcSJason Evans 107a48740b6SDavid E. O'Brien KASSERT(cv_waitq_empty(cvp), ("%s: cv_waitq non-empty", __func__)); 108238510fcSJason Evans } 109238510fcSJason Evans 110238510fcSJason Evans /* 111238510fcSJason Evans * Common code for cv_wait* functions. All require sched_lock. 112238510fcSJason Evans */ 113238510fcSJason Evans 114238510fcSJason Evans /* 115e602ba25SJulian Elischer * Decide if we need to queue an upcall. 116e602ba25SJulian Elischer * This is copied from msleep(), perhaps this should be a common function. 117e602ba25SJulian Elischer */ 118e602ba25SJulian Elischer static void 119e602ba25SJulian Elischer cv_check_upcall(struct thread *td) 120e602ba25SJulian Elischer { 121e602ba25SJulian Elischer 122e602ba25SJulian Elischer /* 123e602ba25SJulian Elischer * If we are capable of async syscalls and there isn't already 124e602ba25SJulian Elischer * another one ready to return, start a new thread 125e602ba25SJulian Elischer * and queue it as ready to run. Note that there is danger here 126e602ba25SJulian Elischer * because we need to make sure that we don't sleep allocating 127e602ba25SJulian Elischer * the thread (recursion here might be bad). 128e602ba25SJulian Elischer * Hence the TDF_INMSLEEP flag. 129e602ba25SJulian Elischer */ 130e602ba25SJulian Elischer if ((td->td_proc->p_flag & P_KSES) && td->td_mailbox && 131e602ba25SJulian Elischer (td->td_flags & TDF_INMSLEEP) == 0) { 132e602ba25SJulian Elischer /* 133e602ba25SJulian Elischer * If we have no queued work to do, 134e602ba25SJulian Elischer * upcall to the UTS to see if it has more work. 135e602ba25SJulian Elischer * We don't need to upcall now, just queue it. 136e602ba25SJulian Elischer */ 137e602ba25SJulian Elischer if (TAILQ_FIRST(&td->td_ksegrp->kg_runq) == NULL) { 138e602ba25SJulian Elischer /* Don't recurse here! */ 139e602ba25SJulian Elischer td->td_flags |= TDF_INMSLEEP; 140e602ba25SJulian Elischer thread_schedule_upcall(td, td->td_kse); 141e602ba25SJulian Elischer td->td_flags &= ~TDF_INMSLEEP; 142e602ba25SJulian Elischer } 143e602ba25SJulian Elischer } 144e602ba25SJulian Elischer } 145e602ba25SJulian Elischer 146e602ba25SJulian Elischer /* 147238510fcSJason Evans * Switch context. 148238510fcSJason Evans */ 149238510fcSJason Evans static __inline void 150b40ce416SJulian Elischer cv_switch(struct thread *td) 151238510fcSJason Evans { 152238510fcSJason Evans 153e602ba25SJulian Elischer cv_check_upcall(td); 15471fad9fdSJulian Elischer TD_SET_SLEEPING(td); 15571fad9fdSJulian Elischer td->td_proc->p_stats->p_ru.ru_nvcsw++; 156238510fcSJason Evans mi_switch(); 1579ef3a985SJohn Baldwin CTR3(KTR_PROC, "cv_switch: resume thread %p (pid %d, %s)", td, 1589ef3a985SJohn Baldwin td->td_proc->p_pid, td->td_proc->p_comm); 159238510fcSJason Evans } 160238510fcSJason Evans 161238510fcSJason Evans /* 162238510fcSJason Evans * Switch context, catching signals. 163238510fcSJason Evans */ 164238510fcSJason Evans static __inline int 165b40ce416SJulian Elischer cv_switch_catch(struct thread *td) 166238510fcSJason Evans { 1679ef3a985SJohn Baldwin struct proc *p; 168238510fcSJason Evans int sig; 169238510fcSJason Evans 170238510fcSJason Evans /* 171238510fcSJason Evans * We put ourselves on the sleep queue and start our timeout before 172628855e7SJulian Elischer * calling cursig, as we could stop there, and a wakeup or a SIGCONT (or 173238510fcSJason Evans * both) could occur while we were stopped. A SIGCONT would cause us to 174e602ba25SJulian Elischer * be marked as TDS_SLP without resuming us, thus we must be ready for 175628855e7SJulian Elischer * sleep when cursig is called. If the wakeup happens while we're 17671fad9fdSJulian Elischer * stopped, td->td_wchan will be 0 upon return from cursig, 17771fad9fdSJulian Elischer * and TD_ON_SLEEPQ() will return false. 178238510fcSJason Evans */ 179b40ce416SJulian Elischer td->td_flags |= TDF_SINTR; 1809ed346baSBosko Milekic mtx_unlock_spin(&sched_lock); 1819ef3a985SJohn Baldwin p = td->td_proc; 1829ef3a985SJohn Baldwin PROC_LOCK(p); 18371fad9fdSJulian Elischer sig = cursig(td); 184e602ba25SJulian Elischer if (thread_suspend_check(1)) 185e602ba25SJulian Elischer sig = SIGSTOP; 1869ed346baSBosko Milekic mtx_lock_spin(&sched_lock); 187c86b6ff5SJohn Baldwin PROC_UNLOCK(p); 188238510fcSJason Evans if (sig != 0) { 18971fad9fdSJulian Elischer if (TD_ON_SLEEPQ(td)) 190b40ce416SJulian Elischer cv_waitq_remove(td); 19171fad9fdSJulian Elischer TD_SET_RUNNING(td); 19271fad9fdSJulian Elischer } else if (TD_ON_SLEEPQ(td)) { 193b40ce416SJulian Elischer cv_switch(td); 194238510fcSJason Evans } 195b40ce416SJulian Elischer td->td_flags &= ~TDF_SINTR; 196238510fcSJason Evans 197238510fcSJason Evans return sig; 198238510fcSJason Evans } 199238510fcSJason Evans 200238510fcSJason Evans /* 201b40ce416SJulian Elischer * Add a thread to the wait queue of a condition variable. 202238510fcSJason Evans */ 203238510fcSJason Evans static __inline void 204b40ce416SJulian Elischer cv_waitq_add(struct cv *cvp, struct thread *td) 205238510fcSJason Evans { 206238510fcSJason Evans 207b40ce416SJulian Elischer td->td_flags |= TDF_CVWAITQ; 20871fad9fdSJulian Elischer TD_SET_ON_SLEEPQ(td); 209b40ce416SJulian Elischer td->td_wchan = cvp; 210b40ce416SJulian Elischer td->td_wmesg = cvp->cv_description; 211b40ce416SJulian Elischer td->td_ksegrp->kg_slptime = 0; /* XXXKSE */ 2122c100766SJulian Elischer td->td_base_pri = td->td_priority; 213b40ce416SJulian Elischer CTR3(KTR_PROC, "cv_waitq_add: thread %p (pid %d, %s)", td, 2149ef3a985SJohn Baldwin td->td_proc->p_pid, td->td_proc->p_comm); 215b40ce416SJulian Elischer TAILQ_INSERT_TAIL(&cvp->cv_waitq, td, td_slpq); 216238510fcSJason Evans } 217238510fcSJason Evans 218238510fcSJason Evans /* 219b40ce416SJulian Elischer * Wait on a condition variable. The current thread is placed on the condition 220238510fcSJason Evans * variable's wait queue and suspended. A cv_signal or cv_broadcast on the same 221b40ce416SJulian Elischer * condition variable will resume the thread. The mutex is released before 222238510fcSJason Evans * sleeping and will be held on return. It is recommended that the mutex be 223238510fcSJason Evans * held when cv_signal or cv_broadcast are called. 224238510fcSJason Evans */ 225238510fcSJason Evans void 226238510fcSJason Evans cv_wait(struct cv *cvp, struct mtx *mp) 227238510fcSJason Evans { 228b40ce416SJulian Elischer struct thread *td; 229238510fcSJason Evans WITNESS_SAVE_DECL(mp); 230238510fcSJason Evans 231b40ce416SJulian Elischer td = curthread; 232238510fcSJason Evans #ifdef KTRACE 2339ba7fe1bSJohn Baldwin if (KTRPOINT(td, KTR_CSW)) 2349ba7fe1bSJohn Baldwin ktrcsw(1, 0); 235238510fcSJason Evans #endif 236b40ce416SJulian Elischer CV_ASSERT(cvp, mp, td); 237c739adbfSJohn Baldwin WITNESS_SLEEP(0, &mp->mtx_object); 23819284646SJohn Baldwin WITNESS_SAVE(&mp->mtx_object, mp); 239238510fcSJason Evans 240fe799533SAndrew Gallatin if (cold ) { 241238510fcSJason Evans /* 242fe799533SAndrew Gallatin * During autoconfiguration, just give interrupts 243fe799533SAndrew Gallatin * a chance, then just return. Don't run any other 244fe799533SAndrew Gallatin * thread or panic below, in case this is the idle 245fe799533SAndrew Gallatin * process and already asleep. 246238510fcSJason Evans */ 247238510fcSJason Evans return; 248238510fcSJason Evans } 2494bc37205SJeffrey Hsu 2504bc37205SJeffrey Hsu mtx_lock_spin(&sched_lock); 2514bc37205SJeffrey Hsu 252238510fcSJason Evans CV_WAIT_VALIDATE(cvp, mp); 253238510fcSJason Evans 254c86b6ff5SJohn Baldwin DROP_GIANT(); 255c86b6ff5SJohn Baldwin mtx_unlock(mp); 256238510fcSJason Evans 257b40ce416SJulian Elischer cv_waitq_add(cvp, td); 258b40ce416SJulian Elischer cv_switch(td); 259238510fcSJason Evans 2609ed346baSBosko Milekic mtx_unlock_spin(&sched_lock); 261238510fcSJason Evans #ifdef KTRACE 2629ba7fe1bSJohn Baldwin if (KTRPOINT(td, KTR_CSW)) 2639ba7fe1bSJohn Baldwin ktrcsw(0, 0); 264238510fcSJason Evans #endif 265238510fcSJason Evans PICKUP_GIANT(); 2669ed346baSBosko Milekic mtx_lock(mp); 26719284646SJohn Baldwin WITNESS_RESTORE(&mp->mtx_object, mp); 268238510fcSJason Evans } 269238510fcSJason Evans 270238510fcSJason Evans /* 271238510fcSJason Evans * Wait on a condition variable, allowing interruption by signals. Return 0 if 272b40ce416SJulian Elischer * the thread was resumed with cv_signal or cv_broadcast, EINTR or ERESTART if 273238510fcSJason Evans * a signal was caught. If ERESTART is returned the system call should be 274238510fcSJason Evans * restarted if possible. 275238510fcSJason Evans */ 276238510fcSJason Evans int 277238510fcSJason Evans cv_wait_sig(struct cv *cvp, struct mtx *mp) 278238510fcSJason Evans { 279b40ce416SJulian Elischer struct thread *td; 28066f769feSPeter Wemm struct proc *p; 281238510fcSJason Evans int rval; 282238510fcSJason Evans int sig; 283238510fcSJason Evans WITNESS_SAVE_DECL(mp); 284238510fcSJason Evans 285b40ce416SJulian Elischer td = curthread; 2869ef3a985SJohn Baldwin p = td->td_proc; 287238510fcSJason Evans rval = 0; 288238510fcSJason Evans #ifdef KTRACE 2899ba7fe1bSJohn Baldwin if (KTRPOINT(td, KTR_CSW)) 2909ba7fe1bSJohn Baldwin ktrcsw(1, 0); 291238510fcSJason Evans #endif 292b40ce416SJulian Elischer CV_ASSERT(cvp, mp, td); 293c739adbfSJohn Baldwin WITNESS_SLEEP(0, &mp->mtx_object); 29419284646SJohn Baldwin WITNESS_SAVE(&mp->mtx_object, mp); 295238510fcSJason Evans 296238510fcSJason Evans if (cold || panicstr) { 297238510fcSJason Evans /* 298238510fcSJason Evans * After a panic, or during autoconfiguration, just give 299238510fcSJason Evans * interrupts a chance, then just return; don't run any other 300238510fcSJason Evans * procs or panic below, in case this is the idle process and 301238510fcSJason Evans * already asleep. 302238510fcSJason Evans */ 303238510fcSJason Evans return 0; 304238510fcSJason Evans } 3054bc37205SJeffrey Hsu 3064bc37205SJeffrey Hsu mtx_lock_spin(&sched_lock); 3074bc37205SJeffrey Hsu 308238510fcSJason Evans CV_WAIT_VALIDATE(cvp, mp); 309238510fcSJason Evans 310c86b6ff5SJohn Baldwin DROP_GIANT(); 311c86b6ff5SJohn Baldwin mtx_unlock(mp); 312238510fcSJason Evans 313b40ce416SJulian Elischer cv_waitq_add(cvp, td); 314b40ce416SJulian Elischer sig = cv_switch_catch(td); 315238510fcSJason Evans 3169ed346baSBosko Milekic mtx_unlock_spin(&sched_lock); 317238510fcSJason Evans 3189ef3a985SJohn Baldwin PROC_LOCK(p); 319238510fcSJason Evans if (sig == 0) 320e602ba25SJulian Elischer sig = cursig(td); /* XXXKSE */ 321238510fcSJason Evans if (sig != 0) { 3229ef3a985SJohn Baldwin if (SIGISMEMBER(p->p_sigacts->ps_sigintr, sig)) 323238510fcSJason Evans rval = EINTR; 324238510fcSJason Evans else 325238510fcSJason Evans rval = ERESTART; 326238510fcSJason Evans } 3279ef3a985SJohn Baldwin PROC_UNLOCK(p); 328e602ba25SJulian Elischer if (p->p_flag & P_WEXIT) 329e602ba25SJulian Elischer rval = EINTR; 330238510fcSJason Evans 331238510fcSJason Evans #ifdef KTRACE 3329ba7fe1bSJohn Baldwin if (KTRPOINT(td, KTR_CSW)) 3339ba7fe1bSJohn Baldwin ktrcsw(0, 0); 334238510fcSJason Evans #endif 3359ba7fe1bSJohn Baldwin PICKUP_GIANT(); 3369ed346baSBosko Milekic mtx_lock(mp); 33719284646SJohn Baldwin WITNESS_RESTORE(&mp->mtx_object, mp); 338238510fcSJason Evans 339238510fcSJason Evans return (rval); 340238510fcSJason Evans } 341238510fcSJason Evans 342238510fcSJason Evans /* 343238510fcSJason Evans * Wait on a condition variable for at most timo/hz seconds. Returns 0 if the 344238510fcSJason Evans * process was resumed by cv_signal or cv_broadcast, EWOULDBLOCK if the timeout 345238510fcSJason Evans * expires. 346238510fcSJason Evans */ 347238510fcSJason Evans int 348238510fcSJason Evans cv_timedwait(struct cv *cvp, struct mtx *mp, int timo) 349238510fcSJason Evans { 350b40ce416SJulian Elischer struct thread *td; 351238510fcSJason Evans int rval; 352238510fcSJason Evans WITNESS_SAVE_DECL(mp); 353238510fcSJason Evans 354b40ce416SJulian Elischer td = curthread; 355238510fcSJason Evans rval = 0; 356238510fcSJason Evans #ifdef KTRACE 3579ba7fe1bSJohn Baldwin if (KTRPOINT(td, KTR_CSW)) 3589ba7fe1bSJohn Baldwin ktrcsw(1, 0); 359238510fcSJason Evans #endif 360b40ce416SJulian Elischer CV_ASSERT(cvp, mp, td); 361c739adbfSJohn Baldwin WITNESS_SLEEP(0, &mp->mtx_object); 36219284646SJohn Baldwin WITNESS_SAVE(&mp->mtx_object, mp); 363238510fcSJason Evans 364238510fcSJason Evans if (cold || panicstr) { 365238510fcSJason Evans /* 366238510fcSJason Evans * After a panic, or during autoconfiguration, just give 367238510fcSJason Evans * interrupts a chance, then just return; don't run any other 368b40ce416SJulian Elischer * thread or panic below, in case this is the idle process and 369238510fcSJason Evans * already asleep. 370238510fcSJason Evans */ 371238510fcSJason Evans return 0; 372238510fcSJason Evans } 3734bc37205SJeffrey Hsu 3744bc37205SJeffrey Hsu mtx_lock_spin(&sched_lock); 3754bc37205SJeffrey Hsu 376238510fcSJason Evans CV_WAIT_VALIDATE(cvp, mp); 377238510fcSJason Evans 378c86b6ff5SJohn Baldwin DROP_GIANT(); 379c86b6ff5SJohn Baldwin mtx_unlock(mp); 380238510fcSJason Evans 381b40ce416SJulian Elischer cv_waitq_add(cvp, td); 382b40ce416SJulian Elischer callout_reset(&td->td_slpcallout, timo, cv_timedwait_end, td); 383b40ce416SJulian Elischer cv_switch(td); 384238510fcSJason Evans 385b40ce416SJulian Elischer if (td->td_flags & TDF_TIMEOUT) { 386b40ce416SJulian Elischer td->td_flags &= ~TDF_TIMEOUT; 387238510fcSJason Evans rval = EWOULDBLOCK; 388b40ce416SJulian Elischer } else if (td->td_flags & TDF_TIMOFAIL) 389b40ce416SJulian Elischer td->td_flags &= ~TDF_TIMOFAIL; 390b40ce416SJulian Elischer else if (callout_stop(&td->td_slpcallout) == 0) { 39191a4536fSJohn Baldwin /* 39291a4536fSJohn Baldwin * Work around race with cv_timedwait_end similar to that 39391a4536fSJohn Baldwin * between msleep and endtsleep. 394d5cb7e14SJulian Elischer * Go back to sleep. 39591a4536fSJohn Baldwin */ 39671fad9fdSJulian Elischer TD_SET_SLEEPING(td); 397b40ce416SJulian Elischer td->td_proc->p_stats->p_ru.ru_nivcsw++; 39891a4536fSJohn Baldwin mi_switch(); 39971fad9fdSJulian Elischer td->td_flags &= ~TDF_TIMOFAIL; 40091a4536fSJohn Baldwin } 401238510fcSJason Evans 402e602ba25SJulian Elischer if (td->td_proc->p_flag & P_WEXIT) 403e602ba25SJulian Elischer rval = EWOULDBLOCK; 4049ed346baSBosko Milekic mtx_unlock_spin(&sched_lock); 405238510fcSJason Evans #ifdef KTRACE 4069ba7fe1bSJohn Baldwin if (KTRPOINT(td, KTR_CSW)) 4079ba7fe1bSJohn Baldwin ktrcsw(0, 0); 408238510fcSJason Evans #endif 409238510fcSJason Evans PICKUP_GIANT(); 4109ed346baSBosko Milekic mtx_lock(mp); 41119284646SJohn Baldwin WITNESS_RESTORE(&mp->mtx_object, mp); 412238510fcSJason Evans 413238510fcSJason Evans return (rval); 414238510fcSJason Evans } 415238510fcSJason Evans 416238510fcSJason Evans /* 417238510fcSJason Evans * Wait on a condition variable for at most timo/hz seconds, allowing 418b40ce416SJulian Elischer * interruption by signals. Returns 0 if the thread was resumed by cv_signal 419238510fcSJason Evans * or cv_broadcast, EWOULDBLOCK if the timeout expires, and EINTR or ERESTART if 420238510fcSJason Evans * a signal was caught. 421238510fcSJason Evans */ 422238510fcSJason Evans int 423238510fcSJason Evans cv_timedwait_sig(struct cv *cvp, struct mtx *mp, int timo) 424238510fcSJason Evans { 425b40ce416SJulian Elischer struct thread *td; 4269ef3a985SJohn Baldwin struct proc *p; 427238510fcSJason Evans int rval; 428238510fcSJason Evans int sig; 429238510fcSJason Evans WITNESS_SAVE_DECL(mp); 430238510fcSJason Evans 431b40ce416SJulian Elischer td = curthread; 4329ef3a985SJohn Baldwin p = td->td_proc; 433238510fcSJason Evans rval = 0; 434238510fcSJason Evans #ifdef KTRACE 4359ba7fe1bSJohn Baldwin if (KTRPOINT(td, KTR_CSW)) 4369ba7fe1bSJohn Baldwin ktrcsw(1, 0); 437238510fcSJason Evans #endif 438b40ce416SJulian Elischer CV_ASSERT(cvp, mp, td); 439c739adbfSJohn Baldwin WITNESS_SLEEP(0, &mp->mtx_object); 44019284646SJohn Baldwin WITNESS_SAVE(&mp->mtx_object, mp); 441238510fcSJason Evans 442238510fcSJason Evans if (cold || panicstr) { 443238510fcSJason Evans /* 444238510fcSJason Evans * After a panic, or during autoconfiguration, just give 445238510fcSJason Evans * interrupts a chance, then just return; don't run any other 446b40ce416SJulian Elischer * thread or panic below, in case this is the idle process and 447238510fcSJason Evans * already asleep. 448238510fcSJason Evans */ 449238510fcSJason Evans return 0; 450238510fcSJason Evans } 4514bc37205SJeffrey Hsu 4524bc37205SJeffrey Hsu mtx_lock_spin(&sched_lock); 4534bc37205SJeffrey Hsu 454238510fcSJason Evans CV_WAIT_VALIDATE(cvp, mp); 455238510fcSJason Evans 456c86b6ff5SJohn Baldwin DROP_GIANT(); 457c86b6ff5SJohn Baldwin mtx_unlock(mp); 458238510fcSJason Evans 459b40ce416SJulian Elischer cv_waitq_add(cvp, td); 460b40ce416SJulian Elischer callout_reset(&td->td_slpcallout, timo, cv_timedwait_end, td); 461b40ce416SJulian Elischer sig = cv_switch_catch(td); 462238510fcSJason Evans 463b40ce416SJulian Elischer if (td->td_flags & TDF_TIMEOUT) { 464b40ce416SJulian Elischer td->td_flags &= ~TDF_TIMEOUT; 465238510fcSJason Evans rval = EWOULDBLOCK; 466b40ce416SJulian Elischer } else if (td->td_flags & TDF_TIMOFAIL) 467b40ce416SJulian Elischer td->td_flags &= ~TDF_TIMOFAIL; 468b40ce416SJulian Elischer else if (callout_stop(&td->td_slpcallout) == 0) { 46991a4536fSJohn Baldwin /* 47091a4536fSJohn Baldwin * Work around race with cv_timedwait_end similar to that 47191a4536fSJohn Baldwin * between msleep and endtsleep. 472d5cb7e14SJulian Elischer * Go back to sleep. 47391a4536fSJohn Baldwin */ 47471fad9fdSJulian Elischer TD_SET_SLEEPING(td); 475b40ce416SJulian Elischer td->td_proc->p_stats->p_ru.ru_nivcsw++; 47691a4536fSJohn Baldwin mi_switch(); 47771fad9fdSJulian Elischer td->td_flags &= ~TDF_TIMOFAIL; 47891a4536fSJohn Baldwin } 4799ed346baSBosko Milekic mtx_unlock_spin(&sched_lock); 480238510fcSJason Evans 4819ef3a985SJohn Baldwin PROC_LOCK(p); 482238510fcSJason Evans if (sig == 0) 483e602ba25SJulian Elischer sig = cursig(td); 484238510fcSJason Evans if (sig != 0) { 4859ef3a985SJohn Baldwin if (SIGISMEMBER(p->p_sigacts->ps_sigintr, sig)) 486238510fcSJason Evans rval = EINTR; 487238510fcSJason Evans else 488238510fcSJason Evans rval = ERESTART; 489238510fcSJason Evans } 4909ef3a985SJohn Baldwin PROC_UNLOCK(p); 491238510fcSJason Evans 492e602ba25SJulian Elischer if (p->p_flag & P_WEXIT) 493e602ba25SJulian Elischer rval = EINTR; 494e602ba25SJulian Elischer 495238510fcSJason Evans #ifdef KTRACE 4969ba7fe1bSJohn Baldwin if (KTRPOINT(td, KTR_CSW)) 4979ba7fe1bSJohn Baldwin ktrcsw(0, 0); 498238510fcSJason Evans #endif 4999ba7fe1bSJohn Baldwin PICKUP_GIANT(); 5009ed346baSBosko Milekic mtx_lock(mp); 50119284646SJohn Baldwin WITNESS_RESTORE(&mp->mtx_object, mp); 502238510fcSJason Evans 503238510fcSJason Evans return (rval); 504238510fcSJason Evans } 505238510fcSJason Evans 506238510fcSJason Evans /* 507238510fcSJason Evans * Common code for signal and broadcast. Assumes waitq is not empty. Must be 508238510fcSJason Evans * called with sched_lock held. 509238510fcSJason Evans */ 510238510fcSJason Evans static __inline void 511238510fcSJason Evans cv_wakeup(struct cv *cvp) 512238510fcSJason Evans { 513b40ce416SJulian Elischer struct thread *td; 514238510fcSJason Evans 515981808d1SJohn Baldwin mtx_assert(&sched_lock, MA_OWNED); 516b40ce416SJulian Elischer td = TAILQ_FIRST(&cvp->cv_waitq); 517a48740b6SDavid E. O'Brien KASSERT(td->td_wchan == cvp, ("%s: bogus wchan", __func__)); 518a48740b6SDavid E. O'Brien KASSERT(td->td_flags & TDF_CVWAITQ, ("%s: not on waitq", __func__)); 51971fad9fdSJulian Elischer cv_waitq_remove(td); 52071fad9fdSJulian Elischer TD_CLR_SLEEPING(td); 52171fad9fdSJulian Elischer setrunnable(td); 522238510fcSJason Evans } 523238510fcSJason Evans 524238510fcSJason Evans /* 525b40ce416SJulian Elischer * Signal a condition variable, wakes up one waiting thread. Will also wakeup 526238510fcSJason Evans * the swapper if the process is not in memory, so that it can bring the 527b40ce416SJulian Elischer * sleeping process in. Note that this may also result in additional threads 528238510fcSJason Evans * being made runnable. Should be called with the same mutex as was passed to 529238510fcSJason Evans * cv_wait held. 530238510fcSJason Evans */ 531238510fcSJason Evans void 532238510fcSJason Evans cv_signal(struct cv *cvp) 533238510fcSJason Evans { 534238510fcSJason Evans 535a48740b6SDavid E. O'Brien KASSERT(cvp != NULL, ("%s: cvp NULL", __func__)); 5369ed346baSBosko Milekic mtx_lock_spin(&sched_lock); 537238510fcSJason Evans if (!TAILQ_EMPTY(&cvp->cv_waitq)) { 538238510fcSJason Evans CV_SIGNAL_VALIDATE(cvp); 539238510fcSJason Evans cv_wakeup(cvp); 540238510fcSJason Evans } 5419ed346baSBosko Milekic mtx_unlock_spin(&sched_lock); 542238510fcSJason Evans } 543238510fcSJason Evans 544238510fcSJason Evans /* 545b40ce416SJulian Elischer * Broadcast a signal to a condition variable. Wakes up all waiting threads. 546238510fcSJason Evans * Should be called with the same mutex as was passed to cv_wait held. 547238510fcSJason Evans */ 548238510fcSJason Evans void 549238510fcSJason Evans cv_broadcast(struct cv *cvp) 550238510fcSJason Evans { 551238510fcSJason Evans 552a48740b6SDavid E. O'Brien KASSERT(cvp != NULL, ("%s: cvp NULL", __func__)); 5539ed346baSBosko Milekic mtx_lock_spin(&sched_lock); 554238510fcSJason Evans CV_SIGNAL_VALIDATE(cvp); 555238510fcSJason Evans while (!TAILQ_EMPTY(&cvp->cv_waitq)) 556238510fcSJason Evans cv_wakeup(cvp); 5579ed346baSBosko Milekic mtx_unlock_spin(&sched_lock); 558238510fcSJason Evans } 559238510fcSJason Evans 560238510fcSJason Evans /* 561b40ce416SJulian Elischer * Remove a thread from the wait queue of its condition variable. This may be 562238510fcSJason Evans * called externally. 563238510fcSJason Evans */ 564238510fcSJason Evans void 565b40ce416SJulian Elischer cv_waitq_remove(struct thread *td) 566238510fcSJason Evans { 567238510fcSJason Evans struct cv *cvp; 568238510fcSJason Evans 56971fad9fdSJulian Elischer mtx_assert(&sched_lock, MA_OWNED); 570b40ce416SJulian Elischer if ((cvp = td->td_wchan) != NULL && td->td_flags & TDF_CVWAITQ) { 571b40ce416SJulian Elischer TAILQ_REMOVE(&cvp->cv_waitq, td, td_slpq); 572b40ce416SJulian Elischer td->td_flags &= ~TDF_CVWAITQ; 57371fad9fdSJulian Elischer TD_CLR_ON_SLEEPQ(td); 574238510fcSJason Evans } 575238510fcSJason Evans } 576238510fcSJason Evans 577238510fcSJason Evans /* 578b40ce416SJulian Elischer * Timeout function for cv_timedwait. Put the thread on the runqueue and set 579238510fcSJason Evans * its timeout flag. 580238510fcSJason Evans */ 581238510fcSJason Evans static void 582238510fcSJason Evans cv_timedwait_end(void *arg) 583238510fcSJason Evans { 584b40ce416SJulian Elischer struct thread *td; 585238510fcSJason Evans 586b40ce416SJulian Elischer td = arg; 58771fad9fdSJulian Elischer CTR3(KTR_PROC, "cv_timedwait_end: thread %p (pid %d, %s)", 58871fad9fdSJulian Elischer td, td->td_proc->p_pid, td->td_proc->p_comm); 5899ed346baSBosko Milekic mtx_lock_spin(&sched_lock); 59071fad9fdSJulian Elischer if (TD_ON_SLEEPQ(td)) { 591b40ce416SJulian Elischer cv_waitq_remove(td); 592b40ce416SJulian Elischer td->td_flags |= TDF_TIMEOUT; 59371fad9fdSJulian Elischer } else { 594b40ce416SJulian Elischer td->td_flags |= TDF_TIMOFAIL; 59571fad9fdSJulian Elischer } 59671fad9fdSJulian Elischer TD_CLR_SLEEPING(td); 59771fad9fdSJulian Elischer setrunnable(td); 5989ed346baSBosko Milekic mtx_unlock_spin(&sched_lock); 599238510fcSJason Evans } 600e602ba25SJulian Elischer 601e602ba25SJulian Elischer /* 602e602ba25SJulian Elischer * For now only abort interruptable waits. 603e602ba25SJulian Elischer * The others will have to either complete on their own or have a timeout. 604e602ba25SJulian Elischer */ 605e602ba25SJulian Elischer void 606e602ba25SJulian Elischer cv_abort(struct thread *td) 607e602ba25SJulian Elischer { 608e602ba25SJulian Elischer 609e602ba25SJulian Elischer CTR3(KTR_PROC, "cv_abort: thread %p (pid %d, %s)", td, 61071fad9fdSJulian Elischer td->td_proc->p_pid, td->td_proc->p_comm); 611e602ba25SJulian Elischer mtx_lock_spin(&sched_lock); 612e602ba25SJulian Elischer if ((td->td_flags & (TDF_SINTR|TDF_TIMEOUT)) == TDF_SINTR) { 61371fad9fdSJulian Elischer if (TD_ON_SLEEPQ(td)) { 614e602ba25SJulian Elischer cv_waitq_remove(td); 615e602ba25SJulian Elischer } 61671fad9fdSJulian Elischer TD_CLR_SLEEPING(td); 61771fad9fdSJulian Elischer setrunnable(td); 618e602ba25SJulian Elischer } 619e602ba25SJulian Elischer mtx_unlock_spin(&sched_lock); 620e602ba25SJulian Elischer } 621e602ba25SJulian Elischer 622