Home
last modified time | relevance | path

Searched refs:arch_spin_lock (Results 1 – 25 of 41) sorted by relevance

12

/linux/lib/
H A Datomic64.c52 arch_spin_lock(lock); in generic_atomic64_read()
66 arch_spin_lock(lock); in generic_atomic64_set()
80 arch_spin_lock(lock); \
95 arch_spin_lock(lock); \
111 arch_spin_lock(lock); \
148 arch_spin_lock(lock); in generic_atomic64_dec_if_positive()
165 arch_spin_lock(lock); in generic_atomic64_cmpxchg()
182 arch_spin_lock(lock); in generic_atomic64_xchg()
198 arch_spin_lock(lock); in generic_atomic64_fetch_add_unless()
/linux/arch/parisc/include/asm/
H A Dspinlock.h30 static inline void arch_spin_lock(arch_spinlock_t *x) in arch_spin_lock() function
85 arch_spin_lock(&(rw->lock_mutex)); in arch_read_trylock()
109 arch_spin_lock(&(rw->lock_mutex)); in arch_write_trylock()
144 arch_spin_lock(&(rw->lock_mutex)); in arch_read_unlock()
155 arch_spin_lock(&(rw->lock_mutex)); in arch_write_unlock()
H A Dfutex.h24 arch_spin_lock(s); in _futex_spin_lock_irqsave()
/linux/arch/arc/include/asm/
H A Dspinlock.h17 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function
222 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function
304 arch_spin_lock(&(rw->lock_mutex)); in arch_read_trylock()
328 arch_spin_lock(&(rw->lock_mutex)); in arch_write_trylock()
363 arch_spin_lock(&(rw->lock_mutex)); in arch_read_unlock()
374 arch_spin_lock(&(rw->lock_mutex)); in arch_write_unlock()
H A Dsmp.h113 arch_spin_lock(&smp_atomic_ops_lock); \
/linux/include/linux/
H A Dspinlock_up.h29 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function
64 # define arch_spin_lock(lock) do { barrier(); (void)(lock); } while (0) macro
/linux/kernel/locking/
H A Dqrwlock.c43 arch_spin_lock(&lock->wait_lock); in queued_read_lock_slowpath()
73 arch_spin_lock(&lock->wait_lock); in queued_write_lock_slowpath()
/linux/arch/arm/common/
H A Dmcpm_entry.c210 arch_spin_lock(&mcpm_lock); in mcpm_cpu_power_up()
256 arch_spin_lock(&mcpm_lock); in mcpm_cpu_power_down()
333 arch_spin_lock(&mcpm_lock); in mcpm_cpu_suspend()
353 arch_spin_lock(&mcpm_lock); in mcpm_cpu_powered_up()
/linux/arch/mips/kernel/
H A Dsync-r4k.c62 arch_spin_lock(&sync_lock); in check_counter_warp()
85 arch_spin_lock(&sync_lock); in check_counter_warp()
/linux/kernel/kcsan/
H A Dselftest.c165 arch_spin_lock(&arch_spinlock); in test_barrier()
194 arch_spin_lock(&arch_spinlock); in test_barrier()
226 arch_spin_lock(&arch_spinlock); in test_barrier()
/linux/arch/x86/mm/
H A Dkmmio.c353 arch_spin_lock(&kmmio_lock); in post_kmmio_handler()
447 arch_spin_lock(&kmmio_lock); in register_kmmio_probe()
504 arch_spin_lock(&kmmio_lock); in remove_kmmio_fault_pages()
552 arch_spin_lock(&kmmio_lock); in unregister_kmmio_probe()
/linux/kernel/trace/
H A Dtrace_sched_switch.c313 arch_spin_lock(&trace_cmdline_lock); in trace_find_cmdline()
545 arch_spin_lock(&trace_cmdline_lock); in saved_cmdlines_start()
606 arch_spin_lock(&trace_cmdline_lock); in tracing_saved_cmdlines_size_read()
628 arch_spin_lock(&trace_cmdline_lock); in tracing_resize_saved_cmdlines()
H A Dtrace_stack.c179 arch_spin_lock(&stack_trace_max_lock); in check_stack()
358 arch_spin_lock(&stack_trace_max_lock); in stack_max_size_write()
400 arch_spin_lock(&stack_trace_max_lock); in t_start()
/linux/tools/include/linux/
H A Dspinlock.h26 static inline void arch_spin_lock(arch_spinlock_t *mutex) in arch_spin_lock() function
/linux/arch/hexagon/include/asm/
H A Dspinlock.h115 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function
/linux/arch/alpha/include/asm/
H A Dspinlock.h30 static inline void arch_spin_lock(arch_spinlock_t * lock) in arch_spin_lock() function
/linux/arch/sh/include/asm/
H A Dspinlock-cas.h28 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function
H A Dspinlock-llsc.h26 static inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() function
/linux/include/asm-generic/
H A Dqspinlock.h147 #define arch_spin_lock(l) queued_spin_lock(l) macro
H A Dticket_spinlock.h100 #define arch_spin_lock(l) ticket_spin_lock(l) macro
/linux/arch/s390/lib/
H A Dspinlock.c319 arch_spin_lock(&rw->wait); in arch_read_lock_wait()
337 arch_spin_lock(&rw->wait); in arch_write_lock_wait()
/linux/arch/powerpc/include/asm/
H A Dqspinlock.h164 #define arch_spin_lock(l) queued_spin_lock(l) macro
/linux/arch/powerpc/platforms/pasemi/
H A Dsetup.c110 arch_spin_lock(&timebase_lock); in pas_give_timebase()
127 arch_spin_lock(&timebase_lock); in pas_take_timebase()
/linux/arch/powerpc/kvm/
H A Dbook3s_xics.c148 arch_spin_lock(&ics->lock); in write_xive()
213 arch_spin_lock(&ics->lock); in kvmppc_xics_get_xive()
412 arch_spin_lock(&ics->lock); in icp_deliver_irq()
998 arch_spin_lock(&ics->lock); in xics_debug_show()
1192 arch_spin_lock(&ics->lock); in xics_get_source()
1257 arch_spin_lock(&ics->lock); in xics_set_source()
/linux/arch/s390/include/asm/
H A Dspinlock.h82 static inline void arch_spin_lock(arch_spinlock_t *lp) in arch_spin_lock() function

12