Searched refs:raw_lock (Results 1 – 10 of 10) sorted by relevance
27 lock->raw_lock = (arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED; in __raw_spin_lock_init()46 lock->raw_lock = (arch_rwlock_t) __ARCH_RW_LOCK_UNLOCKED; in __rwlock_init()116 arch_spin_lock(&lock->raw_lock); in do_raw_spin_lock()123 int ret = arch_spin_trylock(&lock->raw_lock); in do_raw_spin_trylock()142 arch_spin_unlock(&lock->raw_lock); in do_raw_spin_unlock()162 arch_read_lock(&lock->raw_lock); in do_raw_read_lock()167 int ret = arch_read_trylock(&lock->raw_lock); in do_raw_read_trylock()181 arch_read_unlock(&lock->raw_lock); in do_raw_read_unlock()211 arch_write_lock(&lock->raw_lock); in do_raw_write_lock()217 int ret = arch_write_trylock(&lock->raw_lock); in do_raw_write_trylock()[all …]
39 # define do_raw_read_lock(rwlock) do {__acquire(lock); arch_read_lock(&(rwlock)->raw_lock); } while…40 # define do_raw_read_trylock(rwlock) arch_read_trylock(&(rwlock)->raw_lock)41 # define do_raw_read_unlock(rwlock) do {arch_read_unlock(&(rwlock)->raw_lock); __release(lock); } w…42 # define do_raw_write_lock(rwlock) do {__acquire(lock); arch_write_lock(&(rwlock)->raw_lock); } whi…43 # define do_raw_write_trylock(rwlock) arch_write_trylock(&(rwlock)->raw_lock)44 # define do_raw_write_unlock(rwlock) do {arch_write_unlock(&(rwlock)->raw_lock); __release(lock); }…124 arch_rwlock_is_contended(&(lock)->raw_lock)
26 arch_rwlock_t raw_lock; member40 (rwlock_t) { .raw_lock = __ARCH_RW_LOCK_UNLOCKED, \47 (rwlock_t) { .raw_lock = __ARCH_RW_LOCK_UNLOCKED, \
15 arch_spinlock_t raw_lock; member64 .raw_lock = __ARCH_SPIN_LOCK_UNLOCKED, \
116 #define raw_spin_is_locked(lock) arch_spin_is_locked(&(lock)->raw_lock)119 #define raw_spin_is_contended(lock) arch_spin_is_contended(&(lock)->raw_lock)187 arch_spin_lock(&lock->raw_lock); in do_raw_spin_lock() 193 int ret = arch_spin_trylock(&(lock)->raw_lock); in do_raw_spin_trylock() 204 arch_spin_unlock(&lock->raw_lock); in do_raw_spin_unlock()
33 .raw_lock = __ARCH_SPIN_LOCK_UNLOCKED, \
248 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_enter()263 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_enter()277 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_enter()938 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_zero()950 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_zero()966 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_copy()981 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_copy()
58 mtctl(__pa(__ldcw_align(&pgd_lock->rlock.raw_lock)), 28); in switch_mm_irqs_off()
16 while (likely(arch_spin_value_unlocked(old.lock.rlock.raw_lock))) { \
82 rqspinlock_t raw_lock; member142 raw_res_spin_lock_init(&htab->buckets[i].raw_lock); in htab_init_buckets()152 ret = raw_res_spin_lock_irqsave(&b->raw_lock, flags); in htab_lock_bucket()161 raw_res_spin_unlock_irqrestore(&b->raw_lock, flags); in htab_unlock_bucket()