Lines Matching refs:Allocators

43   atomic_uintptr_t Allocators;  member
54 thread_local FunctionCallTrie::Allocators::Buffers ThreadBuffers;
55 alignas(FunctionCallTrie::Allocators) thread_local std::byte
56 AllocatorsStorage[sizeof(FunctionCallTrie::Allocators)];
85 uptr Allocators = 0; in getThreadLocalData() local
86 if (atomic_compare_exchange_strong(&TLD.Allocators, &Allocators, 1, in getThreadLocalData()
91 atomic_store(&TLD.Allocators, 0, memory_order_release); in getThreadLocalData()
125 new (&AllocatorsStorage) FunctionCallTrie::Allocators( in getThreadLocalData()
127 Allocators = reinterpret_cast<uptr>( in getThreadLocalData()
128 reinterpret_cast<FunctionCallTrie::Allocators *>(&AllocatorsStorage)); in getThreadLocalData()
129 atomic_store(&TLD.Allocators, Allocators, memory_order_release); in getThreadLocalData()
132 if (Allocators == 1) in getThreadLocalData()
138 FunctionCallTrie(*reinterpret_cast<FunctionCallTrie::Allocators *>( in getThreadLocalData()
139 atomic_load_relaxed(&TLD.Allocators))); in getThreadLocalData()
157 auto Allocators = atomic_exchange(&TLD.Allocators, 0, memory_order_acq_rel); in cleanupTLD() local
158 if (Allocators == in cleanupTLD()
160 reinterpret_cast<FunctionCallTrie::Allocators *>(&AllocatorsStorage))) in cleanupTLD()
161 reinterpret_cast<FunctionCallTrie::Allocators *>(Allocators)->~Allocators(); in cleanupTLD()
177 uptr A = atomic_exchange(&T.Allocators, 0, memory_order_acq_rel); in postCurrentThreadFCT()
180 reinterpret_cast<FunctionCallTrie::Allocators *>(&AllocatorsStorage))) in postCurrentThreadFCT()
183 auto Allocators = reinterpret_cast<FunctionCallTrie::Allocators *>(A); in postCurrentThreadFCT() local
184 DCHECK_NE(Allocators, nullptr); in postCurrentThreadFCT()
187 profileCollectorService::post(BQ, std::move(*FCT), std::move(*Allocators), in postCurrentThreadFCT()
191 ThreadBuffers = FunctionCallTrie::Allocators::Buffers{}; in postCurrentThreadFCT()
435 if (atomic_load_relaxed(&T->Allocators) == 0) in profilingLoggingInit()