xref: /linux/kernel/bpf/rqspinlock.h (revision 1260ed77798502de9c98020040d2995008de10cc)
1*164c2465SKumar Kartikeya Dwivedi /* SPDX-License-Identifier: GPL-2.0-or-later */
2*164c2465SKumar Kartikeya Dwivedi /*
3*164c2465SKumar Kartikeya Dwivedi  * Resilient Queued Spin Lock defines
4*164c2465SKumar Kartikeya Dwivedi  *
5*164c2465SKumar Kartikeya Dwivedi  * (C) Copyright 2024-2025 Meta Platforms, Inc. and affiliates.
6*164c2465SKumar Kartikeya Dwivedi  *
7*164c2465SKumar Kartikeya Dwivedi  * Authors: Kumar Kartikeya Dwivedi <memxor@gmail.com>
8*164c2465SKumar Kartikeya Dwivedi  */
9*164c2465SKumar Kartikeya Dwivedi #ifndef __LINUX_RQSPINLOCK_H
10*164c2465SKumar Kartikeya Dwivedi #define __LINUX_RQSPINLOCK_H
11*164c2465SKumar Kartikeya Dwivedi 
12*164c2465SKumar Kartikeya Dwivedi #include "../locking/qspinlock.h"
13*164c2465SKumar Kartikeya Dwivedi 
14*164c2465SKumar Kartikeya Dwivedi /*
15*164c2465SKumar Kartikeya Dwivedi  * try_cmpxchg_tail - Return result of cmpxchg of tail word with a new value
16*164c2465SKumar Kartikeya Dwivedi  * @lock: Pointer to queued spinlock structure
17*164c2465SKumar Kartikeya Dwivedi  * @tail: The tail to compare against
18*164c2465SKumar Kartikeya Dwivedi  * @new_tail: The new queue tail code word
19*164c2465SKumar Kartikeya Dwivedi  * Return: Bool to indicate whether the cmpxchg operation succeeded
20*164c2465SKumar Kartikeya Dwivedi  *
21*164c2465SKumar Kartikeya Dwivedi  * This is used by the head of the wait queue to clean up the queue.
22*164c2465SKumar Kartikeya Dwivedi  * Provides relaxed ordering, since observers only rely on initialized
23*164c2465SKumar Kartikeya Dwivedi  * state of the node which was made visible through the xchg_tail operation,
24*164c2465SKumar Kartikeya Dwivedi  * i.e. through the smp_wmb preceding xchg_tail.
25*164c2465SKumar Kartikeya Dwivedi  *
26*164c2465SKumar Kartikeya Dwivedi  * We avoid using 16-bit cmpxchg, which is not available on all architectures.
27*164c2465SKumar Kartikeya Dwivedi  */
28*164c2465SKumar Kartikeya Dwivedi static __always_inline bool try_cmpxchg_tail(struct qspinlock *lock, u32 tail, u32 new_tail)
29*164c2465SKumar Kartikeya Dwivedi {
30*164c2465SKumar Kartikeya Dwivedi 	u32 old, new;
31*164c2465SKumar Kartikeya Dwivedi 
32*164c2465SKumar Kartikeya Dwivedi 	old = atomic_read(&lock->val);
33*164c2465SKumar Kartikeya Dwivedi 	do {
34*164c2465SKumar Kartikeya Dwivedi 		/*
35*164c2465SKumar Kartikeya Dwivedi 		 * Is the tail part we compare to already stale? Fail.
36*164c2465SKumar Kartikeya Dwivedi 		 */
37*164c2465SKumar Kartikeya Dwivedi 		if ((old & _Q_TAIL_MASK) != tail)
38*164c2465SKumar Kartikeya Dwivedi 			return false;
39*164c2465SKumar Kartikeya Dwivedi 		/*
40*164c2465SKumar Kartikeya Dwivedi 		 * Encode latest locked/pending state for new tail.
41*164c2465SKumar Kartikeya Dwivedi 		 */
42*164c2465SKumar Kartikeya Dwivedi 		new = (old & _Q_LOCKED_PENDING_MASK) | new_tail;
43*164c2465SKumar Kartikeya Dwivedi 	} while (!atomic_try_cmpxchg_relaxed(&lock->val, &old, new));
44*164c2465SKumar Kartikeya Dwivedi 
45*164c2465SKumar Kartikeya Dwivedi 	return true;
46*164c2465SKumar Kartikeya Dwivedi }
47*164c2465SKumar Kartikeya Dwivedi 
48*164c2465SKumar Kartikeya Dwivedi #endif /* __LINUX_RQSPINLOCK_H */
49