Lines Matching refs:Allocators
43 atomic_uintptr_t Allocators; member
55 thread_local FunctionCallTrie::Allocators::Buffers ThreadBuffers;
56 thread_local std::aligned_storage<sizeof(FunctionCallTrie::Allocators),
57 alignof(FunctionCallTrie::Allocators)>::type
88 uptr Allocators = 0; in getThreadLocalData() local
89 if (atomic_compare_exchange_strong(&TLD.Allocators, &Allocators, 1, in getThreadLocalData()
94 atomic_store(&TLD.Allocators, 0, memory_order_release); in getThreadLocalData()
128 new (&AllocatorsStorage) FunctionCallTrie::Allocators( in getThreadLocalData()
130 Allocators = reinterpret_cast<uptr>( in getThreadLocalData()
131 reinterpret_cast<FunctionCallTrie::Allocators *>(&AllocatorsStorage)); in getThreadLocalData()
132 atomic_store(&TLD.Allocators, Allocators, memory_order_release); in getThreadLocalData()
135 if (Allocators == 1) in getThreadLocalData()
141 FunctionCallTrie(*reinterpret_cast<FunctionCallTrie::Allocators *>( in getThreadLocalData()
142 atomic_load_relaxed(&TLD.Allocators))); in getThreadLocalData()
160 auto Allocators = atomic_exchange(&TLD.Allocators, 0, memory_order_acq_rel); in cleanupTLD() local
161 if (Allocators == in cleanupTLD()
163 reinterpret_cast<FunctionCallTrie::Allocators *>(&AllocatorsStorage))) in cleanupTLD()
164 reinterpret_cast<FunctionCallTrie::Allocators *>(Allocators)->~Allocators(); in cleanupTLD()
180 uptr A = atomic_exchange(&T.Allocators, 0, memory_order_acq_rel); in postCurrentThreadFCT()
183 reinterpret_cast<FunctionCallTrie::Allocators *>(&AllocatorsStorage))) in postCurrentThreadFCT()
186 auto Allocators = reinterpret_cast<FunctionCallTrie::Allocators *>(A); in postCurrentThreadFCT() local
187 DCHECK_NE(Allocators, nullptr); in postCurrentThreadFCT()
190 profileCollectorService::post(BQ, std::move(*FCT), std::move(*Allocators), in postCurrentThreadFCT()
194 ThreadBuffers = FunctionCallTrie::Allocators::Buffers{}; in postCurrentThreadFCT()
438 if (atomic_load_relaxed(&T->Allocators) == 0) in profilingLoggingInit()