| /linux/tools/testing/selftests/bpf/ |
| H A D | bpf_atomic.h | 49 #define try_cmpxchg(p, pold, new) \ macro 58 #define try_cmpxchg_relaxed(p, pold, new) try_cmpxchg(p, pold, new) 60 #define try_cmpxchg_acquire(p, pold, new) try_cmpxchg(p, pold, new)
|
| /linux/lib/ |
| H A D | llist.c | 40 } while (!try_cmpxchg(&head->first, &entry, next)); in llist_del_first() 68 } while (!try_cmpxchg(&head->first, &entry, next)); in llist_del_first_this()
|
| H A D | genalloc.c | 54 } while (!try_cmpxchg(addr, &val, val | mask_to_set)); in set_bits_ll() 68 } while (!try_cmpxchg(addr, &val, val & ~mask_to_clear)); in clear_bits_ll()
|
| /linux/rust/kernel/sync/ |
| H A D | atomic.rs | 442 if self.try_cmpxchg(&mut old, new, o) { in cmpxchg() 458 fn try_cmpxchg<Ordering: ordering::Ordering>(&self, old: &mut T, new: T, _: Ordering) -> bool { in try_cmpxchg() function
|
| /linux/include/linux/ |
| H A D | bitops.h | 366 } while (!try_cmpxchg(ptr, &old__, new__)); \ 383 } while (!try_cmpxchg(ptr, &old__, new__)); \
|
| H A D | pgalloc_tag.h | 157 } while (unlikely(!try_cmpxchg(&page->flags.f, &old_flags, flags))); in update_page_tag_ref()
|
| /linux/include/rv/ |
| H A D | da_monitor.h | 125 if (likely(try_cmpxchg(&da_mon->curr_state, &curr_state, next_state))) { in da_event() 166 if (likely(try_cmpxchg(&da_mon->curr_state, &curr_state, next_state))) { in da_event()
|
| /linux/mm/ |
| H A D | mmzone.c | 109 } while (unlikely(!try_cmpxchg(&folio->flags.f, &old_flags, flags))); in folio_xchg_last_cpupid()
|
| H A D | swap.c | 410 } while (!try_cmpxchg(&folio->flags.f, &old_flags, new_flags)); in lru_gen_inc_refs()
|
| /linux/mm/kasan/ |
| H A D | tags.c | 123 if (!try_cmpxchg(&entry->ptr, &old_ptr, STACK_RING_BUSY_PTR)) in save_stack_info()
|
| /linux/scripts/atomic/ |
| H A D | atomics.tbl | 32 try_cmpxchg B v p:old i:new
|
| /linux/Documentation/ |
| H A D | atomic_t.txt | 148 - swap operations: xchg(), cmpxchg() and try_cmpxchg() 282 Both provide the same functionality, but try_cmpxchg() can lead to more 313 NB. try_cmpxchg() also generates better code on some platforms (notably x86)
|
| /linux/rust/kernel/sync/atomic/ |
| H A D | internal.rs | 300 fn try_cmpxchg[acquire, release, relaxed](
|
| /linux/kernel/unwind/ |
| H A D | deferred.c | 28 return try_cmpxchg(&info->id.cnt, &old, cnt); in try_assign_cnt()
|
| /linux/kernel/locking/ |
| H A D | qspinlock_paravirt.h | 217 if (try_cmpxchg(&he->lock, &old, lock)) { in pv_hash()
|
| /linux/drivers/hv/ |
| H A D | hyperv_vmbus.h | 373 if (!try_cmpxchg(&msg->header.message_type, in vmbus_signal_eom()
|
| H A D | mshv_eventfd.c | 132 if (!try_cmpxchg(&vp->vp_register_page->interrupt_vectors.as_uint64, in mshv_vp_irq_try_set_vector()
|
| /linux/include/net/ |
| H A D | af_vsock.h | 285 if (try_cmpxchg(&net->vsock.child_ns_mode_locked, in vsock_net_set_child_mode()
|
| /linux/arch/s390/kvm/ |
| H A D | gaccess.c | 160 } while (!try_cmpxchg(&ic->val, &old.val, new.val)); in ipte_lock_simple() 178 } while (!try_cmpxchg(&ic->val, &old.val, new.val)); in ipte_unlock_simple() 199 } while (!try_cmpxchg(&ic->val, &old.val, new.val)); in ipte_lock_siif() 213 } while (!try_cmpxchg(&ic->val, &old.val, new.val)); in ipte_unlock_siif()
|
| /linux/io_uring/ |
| H A D | tw.c | 212 } while (!try_cmpxchg(&ctx->work_llist.first, &head, in io_req_local_work_add()
|
| /linux/fs/kernfs/ |
| H A D | inode.c | 51 if (!try_cmpxchg(&kn->iattr, &attr, ret)) in __kernfs_iattrs()
|
| /linux/arch/x86/events/amd/ |
| H A D | uncore.c | 224 if (try_cmpxchg(&ctx->events[i], &tmp, event)) { in amd_uncore_add() 260 if (try_cmpxchg(&ctx->events[i], &tmp, NULL)) in amd_uncore_del()
|
| /linux/kernel/cgroup/ |
| H A D | rstat.c | 121 if (!try_cmpxchg(&rstatc->lnode.next, &self, NULL)) in css_rstat_updated()
|
| /linux/arch/x86/mm/pat/ |
| H A D | memtype.c | 168 } while (!try_cmpxchg(&pg->flags.f, &old_flags, new_flags)); in set_page_memtype()
|
| /linux/arch/x86/xen/ |
| H A D | p2m.c | 563 if (try_cmpxchg(top_mfn_p, &missing_mfn, mid_mfn_mfn)) { in xen_alloc_p2m_entry()
|