Home
last modified time | relevance | path

Searched refs:qspinlock (Results 1 – 18 of 18) sorted by relevance

/linux/arch/x86/include/asm/
H A Dqspinlock.h14 static __always_inline u32 queued_fetch_set_pending_acquire(struct qspinlock *lock) in queued_fetch_set_pending_acquire()
31 extern void native_queued_spin_lock_slowpath(struct qspinlock *lock, u32 val);
33 extern void __pv_queued_spin_lock_slowpath(struct qspinlock *lock, u32 val);
34 extern void __raw_callee_save___pv_queued_spin_unlock(struct qspinlock *lock);
44 static inline void native_queued_spin_unlock(struct qspinlock *lock) in native_queued_spin_unlock()
49 static inline void queued_spin_lock_slowpath(struct qspinlock *lock, u32 val) in queued_spin_lock_slowpath()
54 static inline void queued_spin_unlock(struct qspinlock *lock) in queued_spin_unlock()
88 static inline bool virt_spin_lock(struct qspinlock *lock) in virt_spin_lock()
H A Dqspinlock_paravirt.h7 void __lockfunc __pv_queued_spin_unlock_slowpath(struct qspinlock *lock, u8 locked);
63 extern void __lockfunc __pv_queued_spin_unlock(struct qspinlock *lock);
H A Dparavirt_types.h213 struct qspinlock;
216 void (*queued_spin_lock_slowpath)(struct qspinlock *lock, u32 val);
/linux/arch/powerpc/include/asm/
H A Dqspinlock.h71 static __always_inline int queued_spin_is_locked(struct qspinlock *lock) in queued_spin_is_locked()
76 static __always_inline int queued_spin_value_unlocked(struct qspinlock lock) in queued_spin_value_unlocked()
81 static __always_inline int queued_spin_is_contended(struct qspinlock *lock) in queued_spin_is_contended()
92 static __always_inline int __queued_spin_trylock_nosteal(struct qspinlock *lock) in __queued_spin_trylock_nosteal()
114 static __always_inline int __queued_spin_trylock_steal(struct qspinlock *lock) in __queued_spin_trylock_steal()
138 static __always_inline int queued_spin_trylock(struct qspinlock *lock) in queued_spin_trylock()
146 void queued_spin_lock_slowpath(struct qspinlock *lock);
148 static __always_inline void queued_spin_lock(struct qspinlock *lock) in queued_spin_lock()
154 static inline void queued_spin_unlock(struct qspinlock *lock) in queued_spin_unlock()
H A Dqspinlock_types.h8 typedef struct qspinlock { struct
/linux/include/asm-generic/
H A Dqspinlock.h51 static __always_inline int queued_spin_is_locked(struct qspinlock *lock) in queued_spin_is_locked()
71 static __always_inline int queued_spin_value_unlocked(struct qspinlock lock) in queued_spin_value_unlocked()
81 static __always_inline int queued_spin_is_contended(struct qspinlock *lock) in queued_spin_is_contended()
90 static __always_inline int queued_spin_trylock(struct qspinlock *lock) in queued_spin_trylock()
100 extern void queued_spin_lock_slowpath(struct qspinlock *lock, u32 val);
107 static __always_inline void queued_spin_lock(struct qspinlock *lock) in queued_spin_lock()
123 static __always_inline void queued_spin_unlock(struct qspinlock *lock) in queued_spin_unlock()
133 static __always_inline bool virt_spin_lock(struct qspinlock *lock) in virt_spin_lock()
H A Dqspinlock_types.h14 typedef struct qspinlock { struct
/linux/kernel/locking/
H A Dqspinlock_paravirt.h81 static inline bool pv_hybrid_queued_unfair_trylock(struct qspinlock *lock) in pv_hybrid_queued_unfair_trylock()
110 static __always_inline void set_pending(struct qspinlock *lock) in set_pending()
120 static __always_inline bool trylock_clear_pending(struct qspinlock *lock) in trylock_clear_pending()
128 static __always_inline void set_pending(struct qspinlock *lock) in set_pending()
133 static __always_inline bool trylock_clear_pending(struct qspinlock *lock) in trylock_clear_pending()
168 struct qspinlock *lock;
208 static struct qspinlock **pv_hash(struct qspinlock *lock, struct pv_node *node) in pv_hash()
215 struct qspinlock *old = NULL; in pv_hash()
236 static struct pv_node *pv_unhash(struct qspinlock *lock) in pv_unhash()
357 static void pv_kick_node(struct qspinlock *lock, struct mcs_spinlock *node) in pv_kick_node()
[all …]
H A Dqspinlock.c24 #include <asm/qspinlock.h>
30 #include "qspinlock.h"
90 static __always_inline void __pv_kick_node(struct qspinlock *lock,
92 static __always_inline u32 __pv_wait_head_or_lock(struct qspinlock *lock,
130 void __lockfunc queued_spin_lock_slowpath(struct qspinlock *lock, u32 val) in decode_tail()
401 #include "qspinlock.c" in queued_spin_lock_slowpath()
/linux/arch/powerpc/lib/
H A Dqspinlock.c18 struct qspinlock *lock;
123 static __always_inline u32 trylock_clean_tail(struct qspinlock *lock, u32 tail) in trylock_clean_tail()
162 static __always_inline u32 publish_tail_cpu(struct qspinlock *lock, u32 tail) in publish_tail_cpu()
182 static __always_inline u32 set_mustq(struct qspinlock *lock) in set_mustq()
198 static __always_inline u32 clear_mustq(struct qspinlock *lock) in clear_mustq()
214 static __always_inline bool try_set_sleepy(struct qspinlock *lock, u32 old) in try_set_sleepy()
236 static __always_inline void seen_sleepy_owner(struct qspinlock *lock, u32 val) in seen_sleepy_owner()
261 static struct qnode *get_tail_qnode(struct qspinlock *lock, int prev_cpu) in get_tail_qnode()
285 static __always_inline bool __yield_to_locked_owner(struct qspinlock *lock, u32 val, bool paravirt,… in __yield_to_locked_owner()
338 static __always_inline bool yield_to_locked_owner(struct qspinlock *lock, u32 val, bool paravirt) in yield_to_locked_owner()
[all …]
H A DMakefile56 obj-$(CONFIG_SMP) += qspinlock.o
/linux/arch/mips/include/asm/
H A Dspinlock.h21 static inline void queued_spin_unlock(struct qspinlock *lock) in queued_spin_unlock()
/linux/arch/loongarch/include/asm/
H A Dqspinlock.h13 static inline bool virt_spin_lock(struct qspinlock *lock) in virt_spin_lock()
/linux/arch/x86/kernel/
H A Dparavirt-spinlocks.c12 __visible void __native_queued_spin_unlock(struct qspinlock *lock) in __native_queued_spin_unlock()
/linux/arch/xtensa/include/asm/
H A DKbuild8 generic-y += qspinlock.h
/linux/arch/riscv/include/asm/
H A DKbuild16 generic-y += qspinlock.h
/linux/arch/arm64/include/asm/
H A DKbuild15 generic-y += qspinlock.h
/linux/Documentation/admin-guide/
H A Dkernel-parameters.txt4550 Disables the qspinlock slow path using PV optimizations