Searched refs:td_lock (Results 1 – 14 of 14) sorted by relevance
/freebsd/sys/kern/ |
H A D | kern_mutex.c | 862 m = td->td_lock; in _thread_lock() 868 if (__predict_true(m == td->td_lock)) { in _thread_lock() 926 spin_time -= lockstat_nsecs(&td->td_lock->lock_object); in thread_lock_flags_() 932 m = td->td_lock; in thread_lock_flags_() 952 if (m != td->td_lock) { in thread_lock_flags_() 960 if (m == td->td_lock) in thread_lock_flags_() 988 lock = td->td_lock; in thread_lock_block() 990 td->td_lock = &blocked_lock; in thread_lock_block() 1000 KASSERT(td->td_lock == &blocked_lock, in thread_lock_unblock() 1002 td, td->td_lock, &blocked_lock)); in thread_lock_unblock() [all …]
|
H A D | sched_4bsd.c | 674 thread0.td_lock = &sched_lock; in schedinit() 818 childtd->td_lock = &sched_lock; in sched_fork_thread() 1043 * a new thread. Block the td_lock in order to avoid in sched_switch() 1046 if (td->td_lock != &sched_lock) { in sched_switch() 1056 MPASS(newtd->td_lock == &sched_lock); in sched_switch() 1114 td->td_lock = &sched_lock; in sched_switch() 1273 if (pcpu->pc_curthread->td_lock == &sched_lock) { in kick_other_cpu() 1338 if (td->td_lock != &sched_lock) { in sched_add() 1341 td->td_lock = &sched_lock; in sched_add() 1434 if (td->td_lock ! [all...] |
H A D | sched_ule.c | 872 if (td->td_lock == TDQ_LOCKPTR(tdq) && in sched_balance_group() 1013 td->td_lock = TDQ_LOCKPTR(to); in tdq_move() 1291 if (td->td_lock == TDQ_LOCKPTR(tdq)) { in sched_setcpu() 1544 thread0.td_lock = TDQ_LOCKPTR(tdq); in sched_setup() 1800 PCPU_GET(idlethread)->td_lock = TDQ_LOCKPTR(TDQ_SELF()); in schedinit_ap() 2163 * thread_lock_unblock() that does not assume td_lock is blocked. 2168 atomic_store_rel_ptr((volatile uintptr_t *)&td->td_lock, in thread_unblock_switch() 2426 child->td_lock = TDQ_LOCKPTR(tdq); in sched_fork_thread() 2796 if (td->td_lock != TDQ_LOCKPTR(tdq)) { in sched_add() 2799 td->td_lock in sched_add() [all...] |
H A D | subr_sleepqueue.c | 533 MPASS(td->td_lock != &sc->sc_lock); in sleepq_catch_signals() 835 td->td_lock == &sc->sc_lock) { in sleepq_remove_thread() 953 while (besttd->td_lock != &sc->sc_lock) { in sleepq_signal()
|
H A D | subr_turnstile.c | 189 if (td->td_lock != &top->ts_lock) in propagate_unlock_td() 231 if (td->td_lock != &ts->ts_lock) { in propagate_priority()
|
/freebsd/sys/arm64/arm64/ |
H A D | genassym.c | 75 ASSYM(TD_LOCK, offsetof(struct thread, td_lock));
|
/freebsd/sys/sys/ |
H A D | proc.h | 235 struct mtx *volatile td_lock; /* replaces sched lock */ member 405 mtx_assert((td)->td_lock, (type)) 409 struct mtx *__m = (td)->td_lock; \ 418 __m = (td)->td_lock; \ 426 __m = (td)->td_lock; \
|
H A D | mutex.h | 161 mtx_unlock_spin((tdp)->td_lock)
|
/freebsd/sys/arm/arm/ |
H A D | genassym.c | 91 ASSYM(TD_LOCK, offsetof(struct thread, td_lock));
|
/freebsd/sys/riscv/riscv/ |
H A D | genassym.c | 86 ASSYM(TD_LOCK, offsetof(struct thread, td_lock));
|
/freebsd/sys/i386/i386/ |
H A D | genassym.c | 85 ASSYM(TD_LOCK, offsetof(struct thread, td_lock));
|
/freebsd/sys/powerpc/powerpc/ |
H A D | genassym.c | 222 ASSYM(TD_LOCK, offsetof(struct thread, td_lock));
|
/freebsd/sys/ddb/ |
H A D | db_ps.c | 354 lock = (struct lock_object *)td->td_lock; in DB_SHOW_COMMAND()
|
/freebsd/sys/amd64/amd64/ |
H A D | genassym.c | 79 ASSYM(TD_LOCK, offsetof(struct thread, td_lock));
|