| /linux/lib/ |
| H A D | atomic64.c | 52 arch_spin_lock(lock); in generic_atomic64_read() 66 arch_spin_lock(lock); in generic_atomic64_set() 80 arch_spin_lock(lock); \ 95 arch_spin_lock(lock); \ 111 arch_spin_lock(lock); \ 148 arch_spin_lock(lock); in generic_atomic64_dec_if_positive() 165 arch_spin_lock(lock); in generic_atomic64_cmpxchg() 182 arch_spin_lock(lock); in generic_atomic64_xchg() 198 arch_spin_lock(lock); in generic_atomic64_fetch_add_unless()
|
| /linux/arch/parisc/include/asm/ |
| H A D | spinlock.h | 30 static inline void arch_spin_lock(arch_spinlock_t *x) in arch_spin_lock() function 85 arch_spin_lock(&(rw->lock_mutex)); in arch_read_trylock() 109 arch_spin_lock(&(rw->lock_mutex)); in arch_write_trylock() 144 arch_spin_lock(&(rw->lock_mutex)); in arch_read_unlock() 155 arch_spin_lock(&(rw->lock_mutex)); in arch_write_unlock()
|
| H A D | futex.h | 24 arch_spin_lock(s); in _futex_spin_lock_irqsave()
|
| /linux/arch/arc/include/asm/ |
| H A D | spinlock.h | 17 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function 222 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function 304 arch_spin_lock(&(rw->lock_mutex)); in arch_read_trylock() 328 arch_spin_lock(&(rw->lock_mutex)); in arch_write_trylock() 363 arch_spin_lock(&(rw->lock_mutex)); in arch_read_unlock() 374 arch_spin_lock(&(rw->lock_mutex)); in arch_write_unlock()
|
| H A D | smp.h | 113 arch_spin_lock(&smp_atomic_ops_lock); \
|
| /linux/include/linux/ |
| H A D | spinlock_up.h | 29 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function 64 # define arch_spin_lock(lock) do { barrier(); (void)(lock); } while (0) macro
|
| /linux/kernel/locking/ |
| H A D | qrwlock.c | 43 arch_spin_lock(&lock->wait_lock); in queued_read_lock_slowpath() 73 arch_spin_lock(&lock->wait_lock); in queued_write_lock_slowpath()
|
| /linux/arch/arm/common/ |
| H A D | mcpm_entry.c | 210 arch_spin_lock(&mcpm_lock); in mcpm_cpu_power_up() 256 arch_spin_lock(&mcpm_lock); in mcpm_cpu_power_down() 333 arch_spin_lock(&mcpm_lock); in mcpm_cpu_suspend() 353 arch_spin_lock(&mcpm_lock); in mcpm_cpu_powered_up()
|
| /linux/arch/mips/kernel/ |
| H A D | sync-r4k.c | 62 arch_spin_lock(&sync_lock); in check_counter_warp() 85 arch_spin_lock(&sync_lock); in check_counter_warp()
|
| /linux/kernel/kcsan/ |
| H A D | selftest.c | 165 arch_spin_lock(&arch_spinlock); in test_barrier() 194 arch_spin_lock(&arch_spinlock); in test_barrier() 226 arch_spin_lock(&arch_spinlock); in test_barrier()
|
| /linux/arch/x86/mm/ |
| H A D | kmmio.c | 353 arch_spin_lock(&kmmio_lock); in post_kmmio_handler() 447 arch_spin_lock(&kmmio_lock); in register_kmmio_probe() 504 arch_spin_lock(&kmmio_lock); in remove_kmmio_fault_pages() 552 arch_spin_lock(&kmmio_lock); in unregister_kmmio_probe()
|
| /linux/kernel/trace/ |
| H A D | trace_sched_switch.c | 313 arch_spin_lock(&trace_cmdline_lock); in trace_find_cmdline() 545 arch_spin_lock(&trace_cmdline_lock); in saved_cmdlines_start() 606 arch_spin_lock(&trace_cmdline_lock); in tracing_saved_cmdlines_size_read() 628 arch_spin_lock(&trace_cmdline_lock); in tracing_resize_saved_cmdlines()
|
| H A D | trace_stack.c | 179 arch_spin_lock(&stack_trace_max_lock); in check_stack() 358 arch_spin_lock(&stack_trace_max_lock); in stack_max_size_write() 400 arch_spin_lock(&stack_trace_max_lock); in t_start()
|
| /linux/tools/include/linux/ |
| H A D | spinlock.h | 26 static inline void arch_spin_lock(arch_spinlock_t *mutex) in arch_spin_lock() function
|
| /linux/arch/hexagon/include/asm/ |
| H A D | spinlock.h | 115 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function
|
| /linux/arch/alpha/include/asm/ |
| H A D | spinlock.h | 30 static inline void arch_spin_lock(arch_spinlock_t * lock) in arch_spin_lock() function
|
| /linux/arch/sh/include/asm/ |
| H A D | spinlock-cas.h | 28 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function
|
| H A D | spinlock-llsc.h | 26 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function
|
| /linux/include/asm-generic/ |
| H A D | qspinlock.h | 147 #define arch_spin_lock(l) queued_spin_lock(l) macro
|
| H A D | ticket_spinlock.h | 100 #define arch_spin_lock(l) ticket_spin_lock(l) macro
|
| /linux/arch/s390/lib/ |
| H A D | spinlock.c | 319 arch_spin_lock(&rw->wait); in arch_read_lock_wait() 337 arch_spin_lock(&rw->wait); in arch_write_lock_wait()
|
| /linux/arch/powerpc/include/asm/ |
| H A D | qspinlock.h | 164 #define arch_spin_lock(l) queued_spin_lock(l) macro
|
| /linux/arch/powerpc/platforms/pasemi/ |
| H A D | setup.c | 110 arch_spin_lock(&timebase_lock); in pas_give_timebase() 127 arch_spin_lock(&timebase_lock); in pas_take_timebase()
|
| /linux/arch/powerpc/kvm/ |
| H A D | book3s_xics.c | 148 arch_spin_lock(&ics->lock); in write_xive() 213 arch_spin_lock(&ics->lock); in kvmppc_xics_get_xive() 412 arch_spin_lock(&ics->lock); in icp_deliver_irq() 998 arch_spin_lock(&ics->lock); in xics_debug_show() 1192 arch_spin_lock(&ics->lock); in xics_get_source() 1257 arch_spin_lock(&ics->lock); in xics_set_source()
|
| /linux/arch/s390/include/asm/ |
| H A D | spinlock.h | 82 static inline void arch_spin_lock(arch_spinlock_t *lp) in arch_spin_lock() function
|