Home
last modified time | relevance | path

Searched refs:atomic_load (Results 1 – 25 of 111) sorted by relevance

12345

/freebsd/contrib/llvm-project/compiler-rt/lib/sanitizer_common/
H A Dsanitizer_addrhashmap.h121 uptr addr1 = atomic_load(&c->addr, memory_order_acquire); in ForEach()
128 (AddBucket *)atomic_load(&bucket->add, memory_order_acquire)) { in ForEach()
131 uptr addr1 = atomic_load(&c->addr, memory_order_acquire); in ForEach()
226 uptr addr1 = atomic_load(&c->addr, memory_order_acquire); in acquire()
234 if (atomic_load(&b->add, memory_order_relaxed)) { in acquire()
236 AddBucket *add = (AddBucket*)atomic_load(&b->add, memory_order_relaxed); in acquire()
239 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); in acquire()
255 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); in acquire()
267 AddBucket *add = (AddBucket*)atomic_load(&b->add, memory_order_relaxed); in acquire()
271 uptr addr1 = atomic_load(&c->addr, memory_order_relaxed); in acquire()
[all …]
H A Dsanitizer_mutex.h44 CHECK_EQ(atomic_load(&state_, memory_order_relaxed), 1); in CheckLocked()
186 state = atomic_load(&state_, memory_order_relaxed); in Lock()
206 state = atomic_load(&state_, memory_order_relaxed); in Lock()
270 state = atomic_load(&state_, memory_order_relaxed); in ReadLock()
286 state = atomic_load(&state_, memory_order_relaxed); in ReadLock()
317 CHECK(atomic_load(&state_, memory_order_relaxed) & kWriterLock); in CheckWriteLocked()
323 CHECK(atomic_load(&state_, memory_order_relaxed) & kReaderLockMask); in CheckReadLocked()
H A Dsanitizer_lfstack.h32 return (atomic_load(&head_, memory_order_relaxed) & kPtrMask) == 0; in Empty()
36 u64 cmp = atomic_load(&head_, memory_order_relaxed); in Push()
48 u64 cmp = atomic_load(&head_, memory_order_acquire); in Pop()
H A Dsanitizer_libignore.h67 return (pc >= begin && pc < atomic_load(&end, memory_order_acquire)); in IsInRange()
102 const uptr n = atomic_load(&ignored_ranges_count_, memory_order_acquire); in IsIgnored()
116 const uptr n = atomic_load(&instrumented_ranges_count_, memory_order_acquire); in IsPcInstrumented()
H A Dsanitizer_tls_get_addr.cpp49 uptr v = atomic_load(cur, memory_order_acquire); in DTLS_NextBlock()
87 (DTLS::DTVBlock *)atomic_load(&block->next, memory_order_acquire); in DTLS_Destroy()
145 atomic_load(&number_of_live_dtls, memory_order_relaxed)); in DTLS_on_tls_get_addr()
178 return atomic_load(&dtls->dtv_block, memory_order_relaxed) == in DTLSInDestruction()
H A Dsanitizer_allocator.cpp38 if (atomic_load(&internal_allocator_initialized, memory_order_acquire) == 0) { in internal_allocator()
40 if (atomic_load(&internal_allocator_initialized, memory_order_relaxed) == in internal_allocator()
191 return atomic_load(&allocator_may_return_null, memory_order_relaxed); in AllocatorMayReturnNull()
207 return atomic_load(&rss_limit_exceeded, memory_order_relaxed); in IsRssLimitExceeded()
H A Dsanitizer_tls_get_addr.h63 (DTLS::DTVBlock *)atomic_load(&dtls->dtv_block, memory_order_acquire); in ForEachDVT()
67 block = (DTLS::DTVBlock *)atomic_load(&block->next, memory_order_acquire); in ForEachDVT()
H A Dsanitizer_stackdepotbase.h95 u32 cmp = atomic_load(p, memory_order_relaxed); in lock()
123 u32 v = atomic_load(p, memory_order_consume); in Put()
187 uptr s = atomic_load(p, memory_order_relaxed); in PrintAll()
197 u32 s = atomic_load(p, memory_order_consume) & kUnlockMask;
H A Dsanitizer_mutex.cpp25 if (atomic_load(&state_, memory_order_relaxed) == 0 && in LockSlow()
32 u32 count = atomic_load(&state_, memory_order_relaxed); in Wait()
36 count = atomic_load(&state_, memory_order_relaxed); in Wait()
H A Dsanitizer_libignore.cpp81 atomic_load(&ignored_ranges_count_, memory_order_relaxed); in OnLibraryLoaded()
114 atomic_load(&instrumented_ranges_count_, memory_order_relaxed); in OnLibraryLoaded()
/freebsd/tools/regression/include/stdatomic/
H A Dlogic.c57 T v1 = atomic_load(a); \
60 assert(atomic_load(a) == (T)(result)); \
64 T v1 = atomic_load(a); \
71 assert(atomic_load(a) == v3); \
78 assert(atomic_load(a) == v1); \
83 assert(atomic_load(a) == v1); \
/freebsd/contrib/llvm-project/compiler-rt/lib/xray/
H A Dxray_interface.cpp163 atomic_load(&XRayNumObjects, memory_order_acquire)) { in isObjectLoaded()
228 if (!atomic_load(&XRayInitialized, memory_order_acquire)) in patchFunction()
241 atomic_load(&XRayNumObjects, memory_order_acquire)) { in patchFunction()
301 atomic_load(&XRayNumObjects, memory_order_acquire)) { in controlPatchingObjectUnchecked()
381 if (!atomic_load(&XRayInitialized, memory_order_acquire)) in controlPatching()
425 if (!atomic_load(&XRayInitialized, memory_order_acquire)) in controlPatching()
445 atomic_load(&XRayNumObjects, memory_order_acquire)) { in mprotectAndPatchFunction()
518 if (atomic_load(&XRayInitialized, memory_order_acquire)) { in __xray_set_handler()
529 if (atomic_load(&XRayInitialized, memory_order_acquire)) { in __xray_set_customevent_handler()
539 if (atomic_load(&XRayInitialized, memory_order_acquire)) { in __xray_set_typedevent_handler()
[all …]
H A Dxray_basic_logging.cpp94 if (atomic_load(&UseRealTSC, memory_order_acquire)) in getLog()
106 Header.CycleFrequency = atomic_load(&CycleFrequency, memory_order_acquire); in getLog()
219 if (Delta < atomic_load(&ThresholdTicks, memory_order_relaxed)) { in InMemoryRawLog()
392 if (!atomic_load(&UseRealTSC, memory_order_relaxed) && Verbosity()) in basicLoggingInit()
424 atomic_load(&TicksPerSec, memory_order_acquire) * in basicLoggingInit()
427 __xray_set_handler_arg1(atomic_load(&UseRealTSC, memory_order_acquire) in basicLoggingInit()
430 __xray_set_handler(atomic_load(&UseRealTSC, memory_order_acquire) in basicLoggingInit()
H A Dxray_init.cpp127 if (atomic_load(&XRayInitialized, memory_order_acquire)) in __xray_init()
133 if (!atomic_load(&XRayFlagsInitialized, memory_order_acquire)) { in __xray_init()
207 if (!atomic_load(&XRayInitialized, memory_order_acquire)) { in __xray_deregister_dso()
H A Dxray_buffer_queue.h233 return atomic_load(&Finalizing, memory_order_acquire); in finalizing()
237 return atomic_load(&Generation, memory_order_acquire); in generation()
H A Dxray_fdr_logging.cpp250 auto BufferSize = atomic_load(It->Extents, memory_order_acquire); in fdrIterator()
277 if (atomic_load(&LoggingStatus, memory_order_acquire) != in fdrLoggingFlush()
363 auto BufferExtents = atomic_load(B.Extents, memory_order_acquire); in fdrLoggingFlush()
445 auto Status = atomic_load(&LoggingStatus, memory_order_acquire); in setupTLD()
/freebsd/contrib/llvm-project/compiler-rt/lib/rtsan/
H A Drtsan_stats.cpp35 return atomic_load(&total_error_count, memory_order_relaxed); in GetTotalErrorCount()
39 return atomic_load(&unique_error_count, memory_order_relaxed); in GetUniqueErrorCount()
47 return atomic_load(&suppressed_count, memory_order_relaxed); in GetSuppressedCount()
/freebsd/contrib/llvm-project/compiler-rt/lib/asan/
H A Dasan_allocator.cpp67 u64 context = atomic_load(atomic_context, memory_order_relaxed); in AtomicContextLoad()
176 return atomic_load(&magic, memory_order_acquire) == kAllocBegMagic in Get()
406 if (ac && atomic_load(&ac->chunk_state, memory_order_acquire) == in RePoisonChunk()
453 options->min_redzone = atomic_load(&min_redzone, memory_order_acquire); in GetOptions()
454 options->max_redzone = atomic_load(&max_redzone, memory_order_acquire); in GetOptions()
457 atomic_load(&alloc_dealloc_mismatch, memory_order_acquire); in GetOptions()
472 u32 min_log = RZSize2Log(atomic_load(&min_redzone, memory_order_acquire)); in ComputeRZLog()
473 u32 max_log = RZSize2Log(atomic_load(&max_redzone, memory_order_acquire)); in ComputeRZLog()
500 u8 left_state = atomic_load(&left_chunk->chunk_state, memory_order_relaxed); in ChooseChunk()
502 atomic_load(&right_chunk->chunk_state, memory_order_relaxed); in ChooseChunk()
[all …]
H A Dasan_thread.h116 if (atomic_load(&stack_switching_, memory_order_relaxed)) in get_fake_stack()
124 if (atomic_load(&stack_switching_, memory_order_relaxed)) in get_or_create_fake_stack()
/freebsd/contrib/llvm-project/compiler-rt/lib/tsan/rtl/
H A Dtsan_fd.cpp64 if (s && atomic_load(&s->rc, memory_order_relaxed) != (u64)-1) in ref()
70 if (s && atomic_load(&s->rc, memory_order_relaxed) != (u64)-1) { in unref()
84 uptr l1 = atomic_load(pl1, memory_order_consume); in fddesc()
112 atomic_load(&d->aux_sync, memory_order_relaxed))); in init()
150 FdDesc *tab = (FdDesc*)atomic_load(&fdctx.tab[l1], memory_order_relaxed); in FdOnFork()
162 FdDesc *tab = (FdDesc*)atomic_load(&fdctx.tab[l1], memory_order_relaxed); in FdLocation()
198 if (uptr aux_sync = atomic_load(&d->aux_sync, memory_order_acquire)) in FdRelease()
246 atomic_load(&d->aux_sync, memory_order_relaxed))); in FdClose()
319 if (atomic_load(&d->aux_sync, memory_order_relaxed)) in FdPollAdd()
H A Dtsan_external.cpp35 if (tag >= atomic_load(&used_tags, memory_order_relaxed)) return nullptr; in GetTagData()
50 uptr tag_count = atomic_load(&used_tags, memory_order_relaxed); in TagFromShadowStackFrame()
69 CHECK_LT(tag, atomic_load(&used_tags, memory_order_relaxed)); in ExternalAccess()
108 CHECK_LT(tag, atomic_load(&used_tags, memory_order_relaxed)); in __tsan_external_assign_tag()
H A Dtsan_dense_alloc.h134 u64 cmp = atomic_load(&freelist_, memory_order_acquire); in Refill()
162 u64 cmp = atomic_load(&freelist_, memory_order_acquire); in Drain()
/freebsd/tools/test/stress2/misc/
H A Dfstat.sh106 while (atomic_load(&share[SYNC]) != PARALLEL)
127 while (atomic_load(&share[DONE]) == 0)
131 while (atomic_load(&share[DONE]) == 1)
/freebsd/contrib/llvm-project/compiler-rt/lib/scudo/standalone/
H A Datomic_helpers.h63 inline typename T::Type atomic_load(const volatile T *A, memory_order MO) { in atomic_load() function
127 return atomic_load(A, memory_order_relaxed); in atomic_load_relaxed()
/freebsd/contrib/llvm-project/compiler-rt/lib/memprof/
H A Dmemprof_stack.cpp26 return atomic_load(&malloc_context_size, memory_order_acquire); in GetMallocContextSize()

12345