Home
last modified time | relevance | path

Searched refs:arch_spin_unlock (Results 1 – 25 of 41) sorted by relevance

12

/linux/lib/
H A Datomic64.c54 arch_spin_unlock(lock); in generic_atomic64_read()
68 arch_spin_unlock(lock); in generic_atomic64_set()
82 arch_spin_unlock(lock); \
97 arch_spin_unlock(lock); \
114 arch_spin_unlock(lock); \
152 arch_spin_unlock(lock); in generic_atomic64_dec_if_positive()
169 arch_spin_unlock(lock); in generic_atomic64_cmpxchg()
185 arch_spin_unlock(lock); in generic_atomic64_xchg()
202 arch_spin_unlock(lock); in generic_atomic64_fetch_add_unless()
/linux/arch/parisc/include/asm/
H A Dspinlock.h49 static inline void arch_spin_unlock(arch_spinlock_t *x) in arch_spin_unlock() function
96 arch_spin_unlock(&(rw->lock_mutex)); in arch_read_trylock()
121 arch_spin_unlock(&(rw->lock_mutex)); in arch_write_trylock()
146 arch_spin_unlock(&(rw->lock_mutex)); in arch_read_unlock()
157 arch_spin_unlock(&(rw->lock_mutex)); in arch_write_unlock()
/linux/arch/arc/include/asm/
H A Dspinlock.h67 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
261 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
315 arch_spin_unlock(&(rw->lock_mutex)); in arch_read_trylock()
340 arch_spin_unlock(&(rw->lock_mutex)); in arch_write_trylock()
365 arch_spin_unlock(&(rw->lock_mutex)); in arch_read_unlock()
376 arch_spin_unlock(&(rw->lock_mutex)); in arch_write_unlock()
H A Dsmp.h117 arch_spin_unlock(&smp_atomic_ops_lock); \
/linux/include/linux/
H A Dspinlock_up.h45 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
65 # define arch_spin_unlock(lock) do { barrier(); (void)(lock); } while (0) macro
/linux/kernel/locking/
H A Dqrwlock.c56 arch_spin_unlock(&lock->wait_lock); in queued_read_lock_slowpath()
88 arch_spin_unlock(&lock->wait_lock); in queued_write_lock_slowpath()
/linux/arch/arm/common/
H A Dmcpm_entry.c232 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_power_up()
268 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_power_down()
274 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_power_down()
335 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_suspend()
365 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_powered_up()
/linux/kernel/trace/
H A Dtrace_sched_switch.c278 arch_spin_unlock(&trace_cmdline_lock); in trace_save_cmdline()
317 arch_spin_unlock(&trace_cmdline_lock); in trace_find_cmdline()
559 arch_spin_unlock(&trace_cmdline_lock); in saved_cmdlines_stop()
608 arch_spin_unlock(&trace_cmdline_lock); in tracing_saved_cmdlines_size_read()
631 arch_spin_unlock(&trace_cmdline_lock); in tracing_resize_saved_cmdlines()
H A Dtrace_stack.c282 arch_spin_unlock(&stack_trace_max_lock); in check_stack()
360 arch_spin_unlock(&stack_trace_max_lock); in stack_max_size_write()
410 arch_spin_unlock(&stack_trace_max_lock); in t_stop()
H A Dtrace_clock.c139 arch_spin_unlock(&trace_clock_struct.lock); in trace_clock_global()
/linux/arch/mips/kernel/
H A Dsync-r4k.c66 arch_spin_unlock(&sync_lock); in check_counter_warp()
96 arch_spin_unlock(&sync_lock); in check_counter_warp()
/linux/kernel/kcsan/
H A Dselftest.c166 KCSAN_CHECK_READ_BARRIER(arch_spin_unlock(&arch_spinlock)); in test_barrier()
195 KCSAN_CHECK_WRITE_BARRIER(arch_spin_unlock(&arch_spinlock)); in test_barrier()
227 KCSAN_CHECK_RW_BARRIER(arch_spin_unlock(&arch_spinlock)); in test_barrier()
/linux/arch/x86/mm/
H A Dkmmio.c356 arch_spin_unlock(&kmmio_lock); in post_kmmio_handler()
467 arch_spin_unlock(&kmmio_lock); in register_kmmio_probe()
516 arch_spin_unlock(&kmmio_lock); in remove_kmmio_fault_pages()
559 arch_spin_unlock(&kmmio_lock); in unregister_kmmio_probe()
/linux/arch/powerpc/kvm/
H A Dbook3s_xive.c575 arch_spin_unlock(&sb->lock); in xive_vm_h_eoi()
583 arch_spin_unlock(&sb->lock); in xive_vm_h_eoi()
1104 arch_spin_unlock(&sb->lock); in xive_lock_and_mask()
1138 arch_spin_unlock(&sb->lock); in xive_lock_for_unmask()
1364 arch_spin_unlock(&sb->lock); in kvmppc_xive_set_xive()
1386 arch_spin_unlock(&sb->lock); in kvmppc_xive_get_xive()
1425 arch_spin_unlock(&sb->lock); in kvmppc_xive_int_on()
1451 arch_spin_unlock(&sb->lock); in kvmppc_xive_int_off()
1642 arch_spin_unlock(&sb->lock); in kvmppc_xive_set_mapped()
1722 arch_spin_unlock(&sb->lock); in kvmppc_xive_clr_mapped()
[all …]
H A Dbook3s_xics.c160 arch_spin_unlock(&ics->lock); in write_xive()
216 arch_spin_unlock(&ics->lock); in kvmppc_xics_get_xive()
473 arch_spin_unlock(&ics->lock); in icp_deliver_irq()
502 arch_spin_unlock(&ics->lock); in icp_deliver_irq()
509 arch_spin_unlock(&ics->lock); in icp_deliver_irq()
1009 arch_spin_unlock(&ics->lock); in xics_debug_show()
1217 arch_spin_unlock(&ics->lock); in xics_get_source()
1275 arch_spin_unlock(&ics->lock); in xics_set_source()
/linux/tools/include/linux/
H A Dspinlock.h31 static inline void arch_spin_unlock(arch_spinlock_t *mutex) in arch_spin_unlock() function
/linux/arch/hexagon/include/asm/
H A Dspinlock.h130 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
/linux/arch/alpha/include/asm/
H A Dspinlock.h24 static inline void arch_spin_unlock(arch_spinlock_t * lock) in arch_spin_unlock() function
/linux/arch/sh/include/asm/
H A Dspinlock-cas.h33 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
H A Dspinlock-llsc.h46 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
/linux/include/asm-generic/
H A Dqspinlock.h149 #define arch_spin_unlock(l) queued_spin_unlock(l) macro
H A Dticket_spinlock.h102 #define arch_spin_unlock(l) ticket_spin_unlock(l) macro
/linux/arch/s390/lib/
H A Dspinlock.c325 arch_spin_unlock(&rw->wait); in arch_read_lock_wait()
348 arch_spin_unlock(&rw->wait); in arch_write_lock_wait()
/linux/arch/powerpc/include/asm/
H A Dqspinlock.h166 #define arch_spin_unlock(l) queued_spin_unlock(l) macro
/linux/arch/powerpc/platforms/pasemi/
H A Dsetup.c114 arch_spin_unlock(&timebase_lock); in pas_give_timebase()
130 arch_spin_unlock(&timebase_lock); in pas_take_timebase()

12