Home
last modified time | relevance | path

Searched refs:td_lock (Results 1 – 14 of 14) sorted by relevance

/freebsd/sys/kern/
H A Dkern_mutex.c862 m = td->td_lock; in _thread_lock()
868 if (__predict_true(m == td->td_lock)) { in _thread_lock()
926 spin_time -= lockstat_nsecs(&td->td_lock->lock_object); in thread_lock_flags_()
932 m = td->td_lock; in thread_lock_flags_()
952 if (m != td->td_lock) { in thread_lock_flags_()
960 if (m == td->td_lock) in thread_lock_flags_()
988 lock = td->td_lock; in thread_lock_block()
990 td->td_lock = &blocked_lock; in thread_lock_block()
1000 KASSERT(td->td_lock == &blocked_lock, in thread_lock_unblock()
1002 td, td->td_lock, &blocked_lock)); in thread_lock_unblock()
[all …]
H A Dsched_4bsd.c674 thread0.td_lock = &sched_lock; in schedinit()
818 childtd->td_lock = &sched_lock; in sched_fork_thread()
1043 * a new thread. Block the td_lock in order to avoid in sched_switch()
1046 if (td->td_lock != &sched_lock) { in sched_switch()
1056 MPASS(newtd->td_lock == &sched_lock); in sched_switch()
1114 td->td_lock = &sched_lock; in sched_switch()
1273 if (pcpu->pc_curthread->td_lock == &sched_lock) { in kick_other_cpu()
1338 if (td->td_lock != &sched_lock) { in sched_add()
1341 td->td_lock = &sched_lock; in sched_add()
1434 if (td->td_lock !
[all...]
H A Dsched_ule.c872 if (td->td_lock == TDQ_LOCKPTR(tdq) && in sched_balance_group()
1013 td->td_lock = TDQ_LOCKPTR(to); in tdq_move()
1291 if (td->td_lock == TDQ_LOCKPTR(tdq)) { in sched_setcpu()
1544 thread0.td_lock = TDQ_LOCKPTR(tdq); in sched_setup()
1800 PCPU_GET(idlethread)->td_lock = TDQ_LOCKPTR(TDQ_SELF()); in schedinit_ap()
2163 * thread_lock_unblock() that does not assume td_lock is blocked.
2168 atomic_store_rel_ptr((volatile uintptr_t *)&td->td_lock, in thread_unblock_switch()
2426 child->td_lock = TDQ_LOCKPTR(tdq); in sched_fork_thread()
2796 if (td->td_lock != TDQ_LOCKPTR(tdq)) { in sched_add()
2799 td->td_lock in sched_add()
[all...]
H A Dsubr_sleepqueue.c533 MPASS(td->td_lock != &sc->sc_lock); in sleepq_catch_signals()
835 td->td_lock == &sc->sc_lock) { in sleepq_remove_thread()
953 while (besttd->td_lock != &sc->sc_lock) { in sleepq_signal()
H A Dsubr_turnstile.c189 if (td->td_lock != &top->ts_lock) in propagate_unlock_td()
231 if (td->td_lock != &ts->ts_lock) { in propagate_priority()
/freebsd/sys/arm64/arm64/
H A Dgenassym.c75 ASSYM(TD_LOCK, offsetof(struct thread, td_lock));
/freebsd/sys/sys/
H A Dproc.h235 struct mtx *volatile td_lock; /* replaces sched lock */ member
405 mtx_assert((td)->td_lock, (type))
409 struct mtx *__m = (td)->td_lock; \
418 __m = (td)->td_lock; \
426 __m = (td)->td_lock; \
H A Dmutex.h161 mtx_unlock_spin((tdp)->td_lock)
/freebsd/sys/arm/arm/
H A Dgenassym.c91 ASSYM(TD_LOCK, offsetof(struct thread, td_lock));
/freebsd/sys/riscv/riscv/
H A Dgenassym.c86 ASSYM(TD_LOCK, offsetof(struct thread, td_lock));
/freebsd/sys/i386/i386/
H A Dgenassym.c85 ASSYM(TD_LOCK, offsetof(struct thread, td_lock));
/freebsd/sys/powerpc/powerpc/
H A Dgenassym.c222 ASSYM(TD_LOCK, offsetof(struct thread, td_lock));
/freebsd/sys/ddb/
H A Ddb_ps.c354 lock = (struct lock_object *)td->td_lock; in DB_SHOW_COMMAND()
/freebsd/sys/amd64/amd64/
H A Dgenassym.c79 ASSYM(TD_LOCK, offsetof(struct thread, td_lock));