/third_party/skia/third_party/externals/abseil-cpp/absl/base/ |
D | invoke_test.cc | 29 namespace base_internal { namespace 92 decltype(base_internal::invoke(std::declval<const F&>())) CallMaybeWithArg( in invoke() 94 return base_internal::invoke(f); in invoke() 98 decltype(base_internal::invoke(std::declval<const F&>(), 42)) CallMaybeWithArg( in CallMaybeWithArg() 100 return base_internal::invoke(f, 42); in CallMaybeWithArg() 104 EXPECT_EQ(1, base_internal::invoke(Function, 3, 2)); in TEST() 105 EXPECT_EQ(1, base_internal::invoke(&Function, 3, 2)); in TEST() 109 EXPECT_EQ(42, base_internal::invoke(Sink, make_unique<int>(42))); in TEST() 113 EXPECT_THAT(base_internal::invoke(Factory, 42), ::testing::Pointee(42)); in TEST() 116 TEST(InvokeTest, VoidResult) { base_internal::invoke(NoOp); } in TEST() [all …]
|
D | call_once.h | 48 namespace base_internal { 92 friend std::atomic<uint32_t>* base_internal::ControlWord(once_flag* flag); 101 namespace base_internal { 112 explicit SchedulingHelper(base_internal::SchedulingMode mode) : mode_(mode) { in SchedulingHelper() 113 if (mode_ == base_internal::SCHEDULE_KERNEL_ONLY) { in SchedulingHelper() 114 guard_result_ = base_internal::SchedulingGuard::DisableRescheduling(); in SchedulingHelper() 119 if (mode_ == base_internal::SCHEDULE_KERNEL_ONLY) { in ~SchedulingHelper() 120 base_internal::SchedulingGuard::EnableRescheduling(guard_result_); in ~SchedulingHelper() 125 base_internal::SchedulingMode mode_; 148 base_internal::SchedulingMode scheduling_mode, Callable&& fn, in CallOnceImpl() [all …]
|
D | throw_delegate_test.cc | 26 using absl::base_internal::ThrowStdLogicError; 27 using absl::base_internal::ThrowStdInvalidArgument; 28 using absl::base_internal::ThrowStdDomainError; 29 using absl::base_internal::ThrowStdLengthError; 30 using absl::base_internal::ThrowStdOutOfRange; 31 using absl::base_internal::ThrowStdRuntimeError; 32 using absl::base_internal::ThrowStdRangeError; 33 using absl::base_internal::ThrowStdOverflowError; 34 using absl::base_internal::ThrowStdUnderflowError; 35 using absl::base_internal::ThrowStdBadFunctionCall; [all …]
|
D | spinlock_test_common.cc | 42 namespace base_internal { namespace 62 absl::kConstInit, base_internal::SCHEDULE_COOPERATIVE_AND_KERNEL); 64 absl::kConstInit, base_internal::SCHEDULE_KERNEL_ONLY); 114 SpinLock spinlock(base_internal::SCHEDULE_KERNEL_ONLY); in TEST() 116 EXPECT_FALSE(base_internal::SchedulingGuard::ReschedulingIsAllowed()); in TEST() 122 EXPECT_FALSE(base_internal::SchedulingGuard::ReschedulingIsAllowed()); in TEST() 205 SpinLock spinlock(base_internal::SCHEDULE_COOPERATIVE_AND_KERNEL); in TEST() 210 SpinLock spinlock(base_internal::SCHEDULE_KERNEL_ONLY); in TEST() 257 base_internal::SCHEDULE_COOPERATIVE_AND_KERNEL); in TEST() 258 SpinLock stack_noncooperative_spinlock(base_internal::SCHEDULE_KERNEL_ONLY); in TEST() [all …]
|
/third_party/skia/third_party/externals/abseil-cpp/absl/synchronization/internal/ |
D | create_thread_identity.cc | 35 ABSL_CONST_INIT static base_internal::SpinLock freelist_lock( 36 absl::kConstInit, base_internal::SCHEDULE_KERNEL_ONLY); 37 ABSL_CONST_INIT static base_internal::ThreadIdentity* thread_identity_freelist; 42 base_internal::ThreadIdentity* identity = in ReclaimThreadIdentity() 43 static_cast<base_internal::ThreadIdentity*>(v); in ReclaimThreadIdentity() 48 base_internal::LowLevelAlloc::Free(identity->per_thread_synch.all_locks); in ReclaimThreadIdentity() 60 base_internal::ClearCurrentThreadIdentity(); in ReclaimThreadIdentity() 62 base_internal::SpinLockHolder l(&freelist_lock); in ReclaimThreadIdentity() 74 static void ResetThreadIdentity(base_internal::ThreadIdentity* identity) { in ResetThreadIdentity() 75 base_internal::PerThreadSynch* pts = &identity->per_thread_synch; in ResetThreadIdentity() [all …]
|
D | per_thread_sem.h | 49 static void Tick(base_internal::ThreadIdentity* identity); 67 static void Init(base_internal::ThreadIdentity* identity); 71 static void Destroy(base_internal::ThreadIdentity* identity); 74 static inline void Post(base_internal::ThreadIdentity* identity); 84 friend absl::base_internal::ThreadIdentity* CreateThreadIdentity(); 100 absl::base_internal::ThreadIdentity* identity); 106 absl::base_internal::ThreadIdentity* identity) { in Post()
|
D | per_thread_sem.cc | 32 base_internal::ThreadIdentity *identity; in SetThreadBlockedCounter() 38 base_internal::ThreadIdentity *identity; in GetThreadBlockedCounter() 43 void PerThreadSem::Init(base_internal::ThreadIdentity *identity) { in Init() 50 void PerThreadSem::Destroy(base_internal::ThreadIdentity *identity) { in Destroy() 54 void PerThreadSem::Tick(base_internal::ThreadIdentity *identity) { in Tick() 72 absl::base_internal::ThreadIdentity *identity) { in ABSL_INTERNAL_C_SYMBOL() 79 absl::base_internal::ThreadIdentity *identity; in ABSL_INTERNAL_C_SYMBOL()
|
D | create_thread_identity.h | 37 base_internal::ThreadIdentity* CreateThreadIdentity(); 47 inline base_internal::ThreadIdentity* GetOrCreateCurrentThreadIdentity() { in GetOrCreateCurrentThreadIdentity() 48 base_internal::ThreadIdentity* identity = in GetOrCreateCurrentThreadIdentity() 49 base_internal::CurrentThreadIdentityIfPresent(); in GetOrCreateCurrentThreadIdentity()
|
D | per_thread_sem_test.cc | 75 base_internal::ThreadIdentity *identity1; // First Post()-er. 76 base_internal::ThreadIdentity *identity2; // First Wait()-er. 112 int64_t cycles = base_internal::CycleClock::Now(); in TestTiming() 115 cycles = base_internal::CycleClock::Now() - cycles; in TestTiming() 128 static void Post(base_internal::ThreadIdentity *id) { in Post() 140 static void Tick(base_internal::ThreadIdentity *identity) { in Tick()
|
D | graphcycles.cc | 55 ABSL_CONST_INIT static absl::base_internal::SpinLock arena_mu( 56 absl::kConstInit, base_internal::SCHEDULE_KERNEL_ONLY); 57 ABSL_CONST_INIT static base_internal::LowLevelAlloc::Arena* arena; 62 arena = base_internal::LowLevelAlloc::NewArena(0); in InitArenaIfNecessary() 141 if (ptr_ != space_) base_internal::LowLevelAlloc::Free(ptr_); in Discard() 150 base_internal::LowLevelAlloc::AllocWithArena(request, arena)); in Grow() 299 auto masked = base_internal::HidePtr(ptr); in Find() 317 auto masked = base_internal::HidePtr(ptr); in Remove() 367 rep_ = new (base_internal::LowLevelAlloc::AllocWithArena(sizeof(Rep), arena)) in GraphCycles() 374 base_internal::LowLevelAlloc::Free(node); in ~GraphCycles() [all …]
|
/third_party/skia/third_party/externals/abseil-cpp/absl/hash/internal/ |
D | low_level_hash.cc | 55 uint64_t a = absl::base_internal::UnalignedLoad64(ptr); in LowLevelHash() 56 uint64_t b = absl::base_internal::UnalignedLoad64(ptr + 8); in LowLevelHash() 57 uint64_t c = absl::base_internal::UnalignedLoad64(ptr + 16); in LowLevelHash() 58 uint64_t d = absl::base_internal::UnalignedLoad64(ptr + 24); in LowLevelHash() 59 uint64_t e = absl::base_internal::UnalignedLoad64(ptr + 32); in LowLevelHash() 60 uint64_t f = absl::base_internal::UnalignedLoad64(ptr + 40); in LowLevelHash() 61 uint64_t g = absl::base_internal::UnalignedLoad64(ptr + 48); in LowLevelHash() 62 uint64_t h = absl::base_internal::UnalignedLoad64(ptr + 56); in LowLevelHash() 82 uint64_t a = absl::base_internal::UnalignedLoad64(ptr); in LowLevelHash() 83 uint64_t b = absl::base_internal::UnalignedLoad64(ptr + 8); in LowLevelHash() [all …]
|
/third_party/skia/third_party/externals/abseil-cpp/absl/base/internal/ |
D | spinlock.cc | 58 namespace base_internal { namespace 60 ABSL_INTERNAL_ATOMIC_HOOK_ATTRIBUTES static base_internal::AtomicHook<void (*)( 77 SpinLock::SpinLock(base_internal::SchedulingMode mode) in SpinLock() 90 base_internal::LowLevelCallOnce(&init_adaptive_spin_count, []() { in SpinLoop() 91 adaptive_spin_count = base_internal::NumCPUs() > 1 ? 1000 : 1; in SpinLoop() 109 base_internal::SchedulingMode scheduling_mode; in SlowLock() 111 scheduling_mode = base_internal::SCHEDULE_COOPERATIVE_AND_KERNEL; in SlowLock() 113 scheduling_mode = base_internal::SCHEDULE_KERNEL_ONLY; in SlowLock() 159 base_internal::SpinLockDelay(&lockword_, lock_value, ++lock_wait_call_count, in SlowLock() 171 base_internal::SpinLockWake(&lockword_, in SlowUnlock()
|
D | unaligned_access.h | 36 namespace base_internal { 67 (absl::base_internal::UnalignedLoad16(_p)) 69 (absl::base_internal::UnalignedLoad32(_p)) 71 (absl::base_internal::UnalignedLoad64(_p)) 74 (absl::base_internal::UnalignedStore16(_p, _val)) 76 (absl::base_internal::UnalignedStore32(_p, _val)) 78 (absl::base_internal::UnalignedStore64(_p, _val))
|
D | spinlock_wait.h | 28 namespace base_internal { 57 base_internal::SchedulingMode scheduling_mode); 78 absl::base_internal::SchedulingMode scheduling_mode); 81 inline void absl::base_internal::SpinLockWake(std::atomic<uint32_t> *w, in SpinLockWake() 86 inline void absl::base_internal::SpinLockDelay( in SpinLockDelay() 88 absl::base_internal::SchedulingMode scheduling_mode) { in SpinLockDelay()
|
D | spinlock_benchmark.cc | 26 template <absl::base_internal::SchedulingMode scheduling_mode> 34 static auto* spinlock = new absl::base_internal::SpinLock(scheduling_mode); in BM_SpinLock() 36 absl::base_internal::SpinLockHolder holder(spinlock); in BM_SpinLock() 41 absl::base_internal::SCHEDULE_KERNEL_ONLY) 47 absl::base_internal::SCHEDULE_COOPERATIVE_AND_KERNEL)
|
D | spinlock.h | 48 namespace base_internal { 58 explicit SpinLock(base_internal::SchedulingMode mode); 61 constexpr SpinLock(absl::ConstInitType, base_internal::SchedulingMode mode) in SpinLock() 103 base_internal::SchedulingGuard::EnableRescheduling(true); in Unlock() 164 base_internal::SchedulingMode scheduling_mode) { in IsCooperative() 165 return scheduling_mode == base_internal::SCHEDULE_COOPERATIVE_AND_KERNEL; in IsCooperative() 227 if (base_internal::SchedulingGuard::DisableRescheduling()) { in TryLockInternal() 236 base_internal::SchedulingGuard::EnableRescheduling(sched_disabled_bit != 0); in TryLockInternal()
|
D | unscaledcycleclock.cc | 36 namespace base_internal { namespace 47 return base_internal::NominalCPUFrequency(); in Frequency() 59 return base_internal::NominalCPUFrequency(); 93 base_internal::LowLevelCallOnce(&init_timebase_frequency_once, [&]() { 129 return base_internal::NominalCPUFrequency();
|
D | thread_identity_test.cc | 30 namespace base_internal { namespace 33 ABSL_CONST_INIT static absl::base_internal::SpinLock map_lock( 34 absl::kConstInit, base_internal::SCHEDULE_KERNEL_ONLY); 61 absl::base_internal::SpinLockHolder l(&map_lock); in TestThreadIdentityCurrent() 93 absl::base_internal::SpinLockHolder l(&map_lock); in TEST()
|
D | sysinfo.cc | 62 namespace base_internal { namespace 232 static int64_t Now() { return base_internal::UnscaledCycleClock::Now(); } in Now() 347 base_internal::LowLevelCallOnce( in NumCPUs() 359 base_internal::LowLevelCallOnce( in NominalCPUFrequency() 418 ABSL_CONST_INIT static absl::base_internal::SpinLock tid_lock( 419 absl::kConstInit, base_internal::SCHEDULE_KERNEL_ONLY); 433 absl::base_internal::SpinLockHolder lock(&tid_lock); in FreeTID() 446 absl::base_internal::SpinLockHolder lock(&tid_lock); in InitGetTID() 464 absl::base_internal::SpinLockHolder lock(&tid_lock); in GetTID()
|
D | unscaledcycleclock.h | 94 namespace base_internal { 113 friend class base_internal::CycleClock; 115 friend class base_internal::UnscaledCycleClockWrapperForInitializeFrequency;
|
D | cycleclock.cc | 32 namespace base_internal { namespace 77 return base_internal::UnscaledCycleClock::Now() >> kShift; in Now() 83 return kFrequencyScale * base_internal::UnscaledCycleClock::Frequency(); in Frequency()
|
D | strerror_test.cc | 35 EXPECT_THAT(absl::base_internal::StrError(EDOM), Eq(strerror(EDOM))); in TEST() 41 EXPECT_THAT(absl::base_internal::StrError(-1), in TEST() 64 const std::string value = absl::base_internal::StrError(i); in TEST()
|
/third_party/skia/third_party/externals/abseil-cpp/absl/functional/internal/ |
D | front_binder.h | 36 return base_internal::invoke( in Apply() 53 template <class... FreeArgs, class R = base_internal::invoke_result_t< 61 class R = base_internal::invoke_result_t< 68 template <class... FreeArgs, class R = base_internal::invoke_result_t< 78 class R = base_internal::invoke_result_t<
|
/third_party/skia/third_party/externals/abseil-cpp/absl/random/ |
D | bit_gen_ref.h | 97 std::declval<base_internal::FastTypeIdType>(), std::declval<void*>(), 141 using mock_call_fn = bool (*)(uintptr_t, base_internal::FastTypeIdType, void*, 154 static bool MockCall(uintptr_t gen_ptr, base_internal::FastTypeIdType type, 159 static bool NotAMock(uintptr_t, base_internal::FastTypeIdType, void*, void*) { 163 inline bool InvokeMock(base_internal::FastTypeIdType type, void* args_tuple,
|
/third_party/skia/third_party/externals/abseil-cpp/absl/synchronization/ |
D | mutex.h | 460 static void IncrementSynchSem(Mutex *mu, base_internal::PerThreadSynch *w); 461 static bool DecrementSynchSem(Mutex *mu, base_internal::PerThreadSynch *w, 478 void TryRemove(base_internal::PerThreadSynch *s); 480 void Block(base_internal::PerThreadSynch *s); 482 base_internal::PerThreadSynch *Wakeup(base_internal::PerThreadSynch *w); 487 base_internal::PerThreadSynch *w); // used for CondVar->Mutex transfer 858 void Remove(base_internal::PerThreadSynch *s); 859 void Wakeup(base_internal::PerThreadSynch *w);
|