1*164c2465SKumar Kartikeya Dwivedi /* SPDX-License-Identifier: GPL-2.0-or-later */ 2*164c2465SKumar Kartikeya Dwivedi /* 3*164c2465SKumar Kartikeya Dwivedi * Resilient Queued Spin Lock defines 4*164c2465SKumar Kartikeya Dwivedi * 5*164c2465SKumar Kartikeya Dwivedi * (C) Copyright 2024-2025 Meta Platforms, Inc. and affiliates. 6*164c2465SKumar Kartikeya Dwivedi * 7*164c2465SKumar Kartikeya Dwivedi * Authors: Kumar Kartikeya Dwivedi <memxor@gmail.com> 8*164c2465SKumar Kartikeya Dwivedi */ 9*164c2465SKumar Kartikeya Dwivedi #ifndef __LINUX_RQSPINLOCK_H 10*164c2465SKumar Kartikeya Dwivedi #define __LINUX_RQSPINLOCK_H 11*164c2465SKumar Kartikeya Dwivedi 12*164c2465SKumar Kartikeya Dwivedi #include "../locking/qspinlock.h" 13*164c2465SKumar Kartikeya Dwivedi 14*164c2465SKumar Kartikeya Dwivedi /* 15*164c2465SKumar Kartikeya Dwivedi * try_cmpxchg_tail - Return result of cmpxchg of tail word with a new value 16*164c2465SKumar Kartikeya Dwivedi * @lock: Pointer to queued spinlock structure 17*164c2465SKumar Kartikeya Dwivedi * @tail: The tail to compare against 18*164c2465SKumar Kartikeya Dwivedi * @new_tail: The new queue tail code word 19*164c2465SKumar Kartikeya Dwivedi * Return: Bool to indicate whether the cmpxchg operation succeeded 20*164c2465SKumar Kartikeya Dwivedi * 21*164c2465SKumar Kartikeya Dwivedi * This is used by the head of the wait queue to clean up the queue. 22*164c2465SKumar Kartikeya Dwivedi * Provides relaxed ordering, since observers only rely on initialized 23*164c2465SKumar Kartikeya Dwivedi * state of the node which was made visible through the xchg_tail operation, 24*164c2465SKumar Kartikeya Dwivedi * i.e. through the smp_wmb preceding xchg_tail. 25*164c2465SKumar Kartikeya Dwivedi * 26*164c2465SKumar Kartikeya Dwivedi * We avoid using 16-bit cmpxchg, which is not available on all architectures. 27*164c2465SKumar Kartikeya Dwivedi */ 28*164c2465SKumar Kartikeya Dwivedi static __always_inline bool try_cmpxchg_tail(struct qspinlock *lock, u32 tail, u32 new_tail) 29*164c2465SKumar Kartikeya Dwivedi { 30*164c2465SKumar Kartikeya Dwivedi u32 old, new; 31*164c2465SKumar Kartikeya Dwivedi 32*164c2465SKumar Kartikeya Dwivedi old = atomic_read(&lock->val); 33*164c2465SKumar Kartikeya Dwivedi do { 34*164c2465SKumar Kartikeya Dwivedi /* 35*164c2465SKumar Kartikeya Dwivedi * Is the tail part we compare to already stale? Fail. 36*164c2465SKumar Kartikeya Dwivedi */ 37*164c2465SKumar Kartikeya Dwivedi if ((old & _Q_TAIL_MASK) != tail) 38*164c2465SKumar Kartikeya Dwivedi return false; 39*164c2465SKumar Kartikeya Dwivedi /* 40*164c2465SKumar Kartikeya Dwivedi * Encode latest locked/pending state for new tail. 41*164c2465SKumar Kartikeya Dwivedi */ 42*164c2465SKumar Kartikeya Dwivedi new = (old & _Q_LOCKED_PENDING_MASK) | new_tail; 43*164c2465SKumar Kartikeya Dwivedi } while (!atomic_try_cmpxchg_relaxed(&lock->val, &old, new)); 44*164c2465SKumar Kartikeya Dwivedi 45*164c2465SKumar Kartikeya Dwivedi return true; 46*164c2465SKumar Kartikeya Dwivedi } 47*164c2465SKumar Kartikeya Dwivedi 48*164c2465SKumar Kartikeya Dwivedi #endif /* __LINUX_RQSPINLOCK_H */ 49