Lines Matching refs:sleepy

19 	u8		sleepy; /* 1 if the previous vCPU was preempted or  member
76 static __always_inline int get_steal_spins(bool paravirt, bool sleepy) in get_steal_spins() argument
78 if (paravirt && sleepy) in get_steal_spins()
84 static __always_inline int get_remote_steal_spins(bool paravirt, bool sleepy) in get_remote_steal_spins() argument
86 if (paravirt && sleepy) in get_remote_steal_spins()
92 static __always_inline int get_head_spins(bool paravirt, bool sleepy) in get_head_spins() argument
94 if (paravirt && sleepy) in get_head_spins()
367 if (next->sleepy) in propagate_sleepy()
372 next->sleepy = 1; in propagate_sleepy()
391 if (node->sleepy || vcpu_is_preempted(prev_cpu)) { in yield_to_prev()
395 if (node->next && !node->next->sleepy) { in yield_to_prev()
404 node->next->sleepy = 1; in yield_to_prev()
411 node->sleepy = false; in yield_to_prev()
442 static __always_inline bool steal_break(u32 val, int iters, bool paravirt, bool sleepy) in steal_break() argument
444 if (iters >= get_steal_spins(paravirt, sleepy)) in steal_break()
448 (iters >= get_remote_steal_spins(paravirt, sleepy))) { in steal_break()
459 bool sleepy = false; in try_to_steal_lock() local
488 if (!sleepy) { in try_to_steal_lock()
491 sleepy = true; in try_to_steal_lock()
493 sleepy = true; in try_to_steal_lock()
505 sleepy = true; in try_to_steal_lock()
519 } while (!steal_break(val, iters, paravirt, sleepy)); in try_to_steal_lock()
532 bool sleepy = false; in queued_spin_lock_mcs_queue() local
558 node->sleepy = 0; in queued_spin_lock_mcs_queue()
618 if (!sleepy) { in queued_spin_lock_mcs_queue()
621 sleepy = true; in queued_spin_lock_mcs_queue()
623 sleepy = true; in queued_spin_lock_mcs_queue()
642 sleepy = true; in queued_spin_lock_mcs_queue()
650 if (!mustq && iters >= get_head_spins(paravirt, sleepy)) { in queued_spin_lock_mcs_queue()