/third_party/skia/third_party/externals/spirv-cross/reference/shaders-msl/comp/ |
D | atomic.comp | 21 uint _16 = atomic_fetch_add_explicit((device atomic_uint*)&ssbo.u32, 1u, memory_order_relaxed); 22 uint _18 = atomic_fetch_or_explicit((device atomic_uint*)&ssbo.u32, 1u, memory_order_relaxed); 23 uint _20 = atomic_fetch_xor_explicit((device atomic_uint*)&ssbo.u32, 1u, memory_order_relaxed); 24 uint _22 = atomic_fetch_and_explicit((device atomic_uint*)&ssbo.u32, 1u, memory_order_relaxed); 25 uint _24 = atomic_fetch_min_explicit((device atomic_uint*)&ssbo.u32, 1u, memory_order_relaxed); 26 uint _26 = atomic_fetch_max_explicit((device atomic_uint*)&ssbo.u32, 1u, memory_order_relaxed); 27 uint _28 = atomic_exchange_explicit((device atomic_uint*)&ssbo.u32, 1u, memory_order_relaxed); 32 …eak_explicit((device atomic_uint*)&ssbo.u32, &_32, 2u, memory_order_relaxed, memory_order_relaxed)… 33 int _36 = atomic_fetch_add_explicit((device atomic_int*)&ssbo.i32, 1, memory_order_relaxed); 34 int _38 = atomic_fetch_or_explicit((device atomic_int*)&ssbo.i32, 1, memory_order_relaxed); [all …]
|
/third_party/skia/third_party/externals/spirv-cross/reference/opt/shaders-msl/comp/ |
D | atomic.comp | 21 uint _16 = atomic_fetch_add_explicit((device atomic_uint*)&ssbo.u32, 1u, memory_order_relaxed); 22 uint _18 = atomic_fetch_or_explicit((device atomic_uint*)&ssbo.u32, 1u, memory_order_relaxed); 23 uint _20 = atomic_fetch_xor_explicit((device atomic_uint*)&ssbo.u32, 1u, memory_order_relaxed); 24 uint _22 = atomic_fetch_and_explicit((device atomic_uint*)&ssbo.u32, 1u, memory_order_relaxed); 25 uint _24 = atomic_fetch_min_explicit((device atomic_uint*)&ssbo.u32, 1u, memory_order_relaxed); 26 uint _26 = atomic_fetch_max_explicit((device atomic_uint*)&ssbo.u32, 1u, memory_order_relaxed); 27 uint _28 = atomic_exchange_explicit((device atomic_uint*)&ssbo.u32, 1u, memory_order_relaxed); 32 …eak_explicit((device atomic_uint*)&ssbo.u32, &_32, 2u, memory_order_relaxed, memory_order_relaxed)… 33 int _36 = atomic_fetch_add_explicit((device atomic_int*)&ssbo.i32, 1, memory_order_relaxed); 34 int _38 = atomic_fetch_or_explicit((device atomic_int*)&ssbo.i32, 1, memory_order_relaxed); [all …]
|
/third_party/skia/third_party/externals/abseil-cpp/absl/container/internal/ |
D | hashtablez_sampler.cc | 60 return dispose_.exchange(f, std::memory_order_relaxed); in SetDisposeCallback() 67 capacity.store(0, std::memory_order_relaxed); in PrepareForSampling() 68 size.store(0, std::memory_order_relaxed); in PrepareForSampling() 69 num_erases.store(0, std::memory_order_relaxed); in PrepareForSampling() 70 num_rehashes.store(0, std::memory_order_relaxed); in PrepareForSampling() 71 max_probe_length.store(0, std::memory_order_relaxed); in PrepareForSampling() 72 total_probe_length.store(0, std::memory_order_relaxed); in PrepareForSampling() 73 hashes_bitwise_or.store(0, std::memory_order_relaxed); in PrepareForSampling() 74 hashes_bitwise_and.store(~size_t{}, std::memory_order_relaxed); in PrepareForSampling() 75 hashes_bitwise_xor.store(0, std::memory_order_relaxed); in PrepareForSampling() [all …]
|
D | hashtablez_sampler.h | 108 info->total_probe_length.store(total_probe_length, std::memory_order_relaxed); in RecordRehashSlow() 109 info->num_erases.store(0, std::memory_order_relaxed); in RecordRehashSlow() 113 1 + info->num_rehashes.load(std::memory_order_relaxed), in RecordRehashSlow() 114 std::memory_order_relaxed); in RecordRehashSlow() 119 info->size.store(size, std::memory_order_relaxed); in RecordStorageChangedSlow() 120 info->capacity.store(capacity, std::memory_order_relaxed); in RecordStorageChangedSlow() 123 info->total_probe_length.store(0, std::memory_order_relaxed); in RecordStorageChangedSlow() 124 info->num_erases.store(0, std::memory_order_relaxed); in RecordStorageChangedSlow() 132 info->size.fetch_sub(1, std::memory_order_relaxed); in RecordEraseSlow() 136 1 + info->num_erases.load(std::memory_order_relaxed), in RecordEraseSlow() [all …]
|
/third_party/skia/third_party/externals/abseil-cpp/absl/debugging/internal/ |
D | vdso_support.cc | 52 : image_(vdso_base_.load(std::memory_order_relaxed) == in VDSOSupport() 55 : vdso_base_.load(std::memory_order_relaxed)) {} in VDSOSupport() 69 if (vdso_base_.load(std::memory_order_relaxed) == kInvalidBase) { in Init() 74 vdso_base_.store(sysinfo_ehdr, std::memory_order_relaxed); in Init() 78 if (vdso_base_.load(std::memory_order_relaxed) == kInvalidBase) { in Init() 82 vdso_base_.store(nullptr, std::memory_order_relaxed); in Init() 83 getcpu_fn_.store(&GetCPUViaSyscall, std::memory_order_relaxed); in Init() 90 std::memory_order_relaxed); in Init() 95 if (vdso_base_.load(std::memory_order_relaxed) == kInvalidBase) { in Init() 97 vdso_base_.store(nullptr, std::memory_order_relaxed); in Init() [all …]
|
/third_party/skia/third_party/externals/abseil-cpp/absl/time/ |
D | clock.cc | 116 uint64_t x = seq->fetch_add(1, std::memory_order_relaxed); in SeqAcquire() 212 time_state.approx_syscall_time_in_cycles.load(std::memory_order_relaxed); in GetCurrentTimeNanosFromKernel() 233 local_approx_syscall_time_in_cycles, std::memory_order_relaxed); in GetCurrentTimeNanosFromKernel() 242 time_state.kernel_time_seen_smaller.store(0, std::memory_order_relaxed); in GetCurrentTimeNanosFromKernel() 244 1, std::memory_order_relaxed) >= 3) { in GetCurrentTimeNanosFromKernel() 250 std::memory_order_relaxed); in GetCurrentTimeNanosFromKernel() 251 time_state.kernel_time_seen_smaller.store(0, std::memory_order_relaxed); in GetCurrentTimeNanosFromKernel() 265 sample->base_ns = atomic->base_ns.load(std::memory_order_relaxed); in ReadTimeSampleAtomic() 266 sample->base_cycles = atomic->base_cycles.load(std::memory_order_relaxed); in ReadTimeSampleAtomic() 268 atomic->nsscaled_per_cycle.load(std::memory_order_relaxed); in ReadTimeSampleAtomic() [all …]
|
/third_party/skia/third_party/externals/abseil-cpp/absl/synchronization/internal/ |
D | per_thread_sem.cc | 45 identity->ticker.store(0, std::memory_order_relaxed); in Init() 46 identity->wait_start.store(0, std::memory_order_relaxed); in Init() 47 identity->is_idle.store(false, std::memory_order_relaxed); in Init() 56 identity->ticker.fetch_add(1, std::memory_order_relaxed) + 1; in Tick() 57 const int wait_start = identity->wait_start.load(std::memory_order_relaxed); in Tick() 58 const bool is_idle = identity->is_idle.load(std::memory_order_relaxed); in Tick() 83 int ticker = identity->ticker.load(std::memory_order_relaxed); in ABSL_INTERNAL_C_SYMBOL() 84 identity->wait_start.store(ticker ? ticker : 1, std::memory_order_relaxed); in ABSL_INTERNAL_C_SYMBOL() 85 identity->is_idle.store(false, std::memory_order_relaxed); in ABSL_INTERNAL_C_SYMBOL() 89 identity->blocked_count_ptr->fetch_add(1, std::memory_order_relaxed); in ABSL_INTERNAL_C_SYMBOL() [all …]
|
D | waiter.cc | 60 const bool is_idle = identity->is_idle.load(std::memory_order_relaxed); in MaybeBecomeIdle() 61 const int ticker = identity->ticker.load(std::memory_order_relaxed); in MaybeBecomeIdle() 62 const int wait_start = identity->wait_start.load(std::memory_order_relaxed); in MaybeBecomeIdle() 64 identity->is_idle.store(true, std::memory_order_relaxed); in MaybeBecomeIdle() 71 futex_.store(0, std::memory_order_relaxed); in Waiter() 84 int32_t x = futex_.load(std::memory_order_relaxed); in Wait() 88 std::memory_order_relaxed)) { in Wait() 240 wakeups_.store(0, std::memory_order_relaxed); in Waiter() 260 int x = wakeups_.load(std::memory_order_relaxed); in Wait() 264 std::memory_order_relaxed)) { in Wait()
|
/third_party/node/deps/v8/src/tracing/ |
D | tracing-category-observer.cc | 35 std::memory_order_relaxed); in OnTraceEnabled() 41 std::memory_order_relaxed); in OnTraceEnabled() 46 i::TracingFlags::gc.fetch_or(ENABLED_BY_TRACING, std::memory_order_relaxed); in OnTraceEnabled() 52 std::memory_order_relaxed); in OnTraceEnabled() 58 std::memory_order_relaxed); in OnTraceEnabled() 65 std::memory_order_relaxed); in OnTraceEnabled() 71 ~(ENABLED_BY_TRACING | ENABLED_BY_SAMPLING), std::memory_order_relaxed); in OnTraceDisabled() 73 i::TracingFlags::gc.fetch_and(~ENABLED_BY_TRACING, std::memory_order_relaxed); in OnTraceDisabled() 76 std::memory_order_relaxed); in OnTraceDisabled() 79 std::memory_order_relaxed); in OnTraceDisabled()
|
/third_party/skia/third_party/externals/abseil-cpp/absl/flags/internal/ |
D | sequence_lock.h | 64 assert(lock_.load(std::memory_order_relaxed) == kUninitialized); in MarkInitialized() 89 int64_t seq_after = lock_.load(std::memory_order_relaxed); in TryRead() 104 int64_t orig_seq = lock_.load(std::memory_order_relaxed); in Write() 106 lock_.store(orig_seq + 1, std::memory_order_relaxed); in Write() 126 int64_t val = lock_.load(std::memory_order_relaxed); in ModificationCount() 135 int64_t val = lock_.load(std::memory_order_relaxed); in IncrementModificationCount() 137 lock_.store(val + 2, std::memory_order_relaxed); in IncrementModificationCount() 147 uint64_t word = src->load(std::memory_order_relaxed); in RelaxedCopyFromAtomic() 154 uint64_t word = src->load(std::memory_order_relaxed); in RelaxedCopyFromAtomic() 167 dst->store(word, std::memory_order_relaxed); in RelaxedCopyToAtomic() [all …]
|
D | sequence_lock_test.cc | 65 while (!stop.load(std::memory_order_relaxed)) { in TEST_P() 73 good_reads.fetch_add(1, std::memory_order_relaxed); in TEST_P() 75 bad_reads.fetch_add(1, std::memory_order_relaxed); in TEST_P() 78 unsuccessful_reads.fetch_add(1, std::memory_order_relaxed); in TEST_P() 83 while (unsuccessful_reads.load(std::memory_order_relaxed) < num_threads_) { in TEST_P() 98 stop.store(true, std::memory_order_relaxed); in TEST_P()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/include/llvm/ADT/ |
D | Statistic.h | 69 unsigned getValue() const { return Value.load(std::memory_order_relaxed); } in getValue() 75 Value.store(Val, std::memory_order_relaxed); 80 Value.fetch_add(1, std::memory_order_relaxed); 86 return Value.fetch_add(1, std::memory_order_relaxed); 90 Value.fetch_sub(1, std::memory_order_relaxed); 96 return Value.fetch_sub(1, std::memory_order_relaxed); 102 Value.fetch_add(V, std::memory_order_relaxed); 109 Value.fetch_sub(V, std::memory_order_relaxed); 114 unsigned PrevMax = Value.load(std::memory_order_relaxed); in updateMax() 118 PrevMax, V, std::memory_order_relaxed)) { in updateMax()
|
/third_party/node/deps/v8/src/base/ |
D | atomicops.h | 107 std::memory_order_relaxed, std::memory_order_relaxed); in Relaxed_CompareAndSwap() 115 std::memory_order_relaxed, std::memory_order_relaxed); in Relaxed_CompareAndSwap() 123 std::memory_order_relaxed, std::memory_order_relaxed); in Relaxed_CompareAndSwap() 130 std::memory_order_relaxed); in Relaxed_AtomicExchange() 143 std::memory_order_relaxed); in Relaxed_AtomicIncrement() 158 std::memory_order_release, std::memory_order_relaxed); in Release_CompareAndSwap() 167 std::memory_order_release, std::memory_order_relaxed); in Release_CompareAndSwap() 182 std::memory_order_relaxed); in Relaxed_Store() 187 std::memory_order_relaxed); in Relaxed_Store() 192 std::memory_order_relaxed); in Relaxed_Store() [all …]
|
/third_party/skia/include/private/ |
D | SkWeakRefCnt.h | 66 fWeakCnt.store(0, std::memory_order_relaxed); in ~SkWeakRefCnt() 73 return fWeakCnt.load(std::memory_order_relaxed); in getWeakCnt() 82 int32_t prev = fRefCnt.load(std::memory_order_relaxed); in atomic_conditional_acquire_strong_ref() 88 std::memory_order_relaxed)); in atomic_conditional_acquire_strong_ref() 116 (void)fWeakCnt.fetch_add(+1, std::memory_order_relaxed); in weak_ref() 132 fWeakCnt.store(1, std::memory_order_relaxed); in weak_unref() 142 return fRefCnt.load(std::memory_order_relaxed) == 0; in weak_expired()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-subzero/include/llvm/ADT/ |
D | Statistic.h | 47 unsigned getValue() const { return Value.load(std::memory_order_relaxed); } in getValue() 66 Value.store(Val, std::memory_order_relaxed); 71 Value.fetch_add(1, std::memory_order_relaxed); 77 return Value.fetch_add(1, std::memory_order_relaxed); 81 Value.fetch_sub(1, std::memory_order_relaxed); 87 return Value.fetch_sub(1, std::memory_order_relaxed); 93 Value.fetch_add(V, std::memory_order_relaxed); 100 Value.fetch_sub(V, std::memory_order_relaxed);
|
/third_party/skia/third_party/externals/abseil-cpp/absl/synchronization/ |
D | mutex.cc | 197 v = pv->load(std::memory_order_relaxed); in AtomicSetBits() 202 std::memory_order_relaxed))); in AtomicSetBits() 213 v = pv->load(std::memory_order_relaxed); in AtomicClearBits() 218 std::memory_order_relaxed))); in AtomicClearBits() 731 intptr_t v = mu_.load(std::memory_order_relaxed); in ~Mutex() 997 s->state.store(PerThreadSynch::kQueued, std::memory_order_relaxed); in Enqueue() 1068 intptr_t v = mu_.load(std::memory_order_relaxed); in TryRemove() 1073 std::memory_order_relaxed)) { in TryRemove() 1105 v = mu_.load(std::memory_order_relaxed); in TryRemove() 1114 std::memory_order_relaxed)); in TryRemove() [all …]
|
D | mutex_benchmark.cc | 127 shared->looping_threads.load(std::memory_order_relaxed) == 0 && in BM_MutexEnqueue() 128 shared->blocked_threads.load(std::memory_order_relaxed) == 0 && in BM_MutexEnqueue() 129 !shared->thread_has_mutex.load(std::memory_order_relaxed), in BM_MutexEnqueue() 138 shared->thread_has_mutex.store(true, std::memory_order_relaxed); in BM_MutexEnqueue() 143 while (shared->looping_threads.load(std::memory_order_relaxed) - in BM_MutexEnqueue() 144 shared->blocked_threads.load(std::memory_order_relaxed) != in BM_MutexEnqueue() 152 while (!shared->thread_has_mutex.load(std::memory_order_relaxed) && in BM_MutexEnqueue() 153 shared->looping_threads.load(std::memory_order_relaxed) > 1) { in BM_MutexEnqueue()
|
/third_party/node/deps/v8/src/logging/ |
D | tracing-flags.h | 28 return runtime_stats.load(std::memory_order_relaxed) != 0; in is_runtime_stats_enabled() 33 return gc.load(std::memory_order_relaxed) != 0; in is_gc_enabled() 37 return gc_stats.load(std::memory_order_relaxed) != 0; in is_gc_stats_enabled() 41 return ic_stats.load(std::memory_order_relaxed) != 0; in is_ic_stats_enabled() 45 return zone_stats.load(std::memory_order_relaxed) != 0; in is_zone_stats_enabled()
|
/third_party/protobuf/src/google/protobuf/ |
D | arena.cc | 72 lifecycle_id_generator_.fetch_add(1, std::memory_order_relaxed); in Init() 73 hint_.store(nullptr, std::memory_order_relaxed); in Init() 74 threads_.store(nullptr, std::memory_order_relaxed); in Init() 84 threads_.store(serial, std::memory_order_relaxed); in Init() 86 std::memory_order_relaxed); in Init() 89 space_allocated_.store(0, std::memory_order_relaxed); in Init() 124 space_allocated_.fetch_add(size, std::memory_order_relaxed); in NewBlock() 209 return space_allocated_.load(std::memory_order_relaxed); in SpaceAllocated() 237 SerialArena* serial = threads_.load(std::memory_order_relaxed); in FreeBlocks() 283 SerialArena* serial = threads_.load(std::memory_order_relaxed); in CleanupList() [all …]
|
/third_party/skia/third_party/externals/abseil-cpp/absl/strings/internal/ |
D | cordz_update_tracker.h | 76 values_[i].store(rhs.values_[i].load(std::memory_order_relaxed), 77 std::memory_order_relaxed); 84 return values_[method].load(std::memory_order_relaxed); in Value() 90 value.store(value.load(std::memory_order_relaxed) + n, 91 std::memory_order_relaxed);
|
/third_party/node/deps/v8/src/libplatform/ |
D | default-job.cc | 43 if (is_canceled_.load(std::memory_order_relaxed)) return; in NotifyConcurrencyIncrease() 68 assigned_task_ids_.load(std::memory_order_relaxed); in AcquireTaskId() 83 std::memory_order_relaxed)); in AcquireTaskId() 118 is_canceled_.store(true, std::memory_order_relaxed); in CancelAndWait() 126 is_canceled_.store(true, std::memory_order_relaxed); in CancelAndDetach() 138 if (is_canceled_.load(std::memory_order_relaxed)) return false; in CanRunFirstTask() 154 if (is_canceled_.load(std::memory_order_relaxed) || in DidRunTask() 190 is_canceled_.store(true, std::memory_order_relaxed); in WaitForParticipationOpportunityLockRequired()
|
/third_party/skia/third_party/externals/spirv-cross/include/spirv_cross/ |
D | barrier.hpp | 53 unsigned target_iteration = iteration.load(std::memory_order_relaxed) + 1; in wait() 59 unsigned c = count.fetch_add(1u, std::memory_order_relaxed); in wait() 63 iteration.store(target_iteration, std::memory_order_relaxed); in wait() 68 while (iteration.load(std::memory_order_relaxed) != target_iteration) in wait()
|
/third_party/node/deps/v8/src/heap/cppgc/ |
D | process-heap-statistics.h | 54 delta, std::memory_order_relaxed); in IncreaseTotalAllocatedObjectSize() 58 delta, std::memory_order_relaxed); in DecreaseTotalAllocatedObjectSize() 62 delta, std::memory_order_relaxed); in IncreaseTotalAllocatedSpace() 66 delta, std::memory_order_relaxed); in DecreaseTotalAllocatedSpace()
|
/third_party/skia/third_party/externals/swiftshader/src/Reactor/ |
D | Nucleus.hpp | 241 …int alignment = 0, bool atomic = false, std::memory_order memoryOrder = std::memory_order_relaxed); 242 … int aligment = 0, bool atomic = false, std::memory_order memoryOrder = std::memory_order_relaxed); 253 …eateAtomicAdd(Value *ptr, Value *value, std::memory_order memoryOrder = std::memory_order_relaxed); 254 …eateAtomicSub(Value *ptr, Value *value, std::memory_order memoryOrder = std::memory_order_relaxed); 255 …eateAtomicAnd(Value *ptr, Value *value, std::memory_order memoryOrder = std::memory_order_relaxed); 256 …reateAtomicOr(Value *ptr, Value *value, std::memory_order memoryOrder = std::memory_order_relaxed); 257 …eateAtomicXor(Value *ptr, Value *value, std::memory_order memoryOrder = std::memory_order_relaxed); 258 …eateAtomicMin(Value *ptr, Value *value, std::memory_order memoryOrder = std::memory_order_relaxed); 259 …eateAtomicMax(Value *ptr, Value *value, std::memory_order memoryOrder = std::memory_order_relaxed); 260 …ateAtomicUMin(Value *ptr, Value *value, std::memory_order memoryOrder = std::memory_order_relaxed); [all …]
|
/third_party/node/deps/v8/src/objects/ |
D | js-array-buffer.h | 182 void Mark() { marked_.store(true, std::memory_order_relaxed); } in Mark() 183 void Unmark() { marked_.store(false, std::memory_order_relaxed); } in Unmark() 184 bool IsMarked() const { return marked_.load(std::memory_order_relaxed); } in IsMarked() 197 return accounting_length_.load(std::memory_order_relaxed); in accounting_length() 201 accounting_length_.store(accounting_length, std::memory_order_relaxed); in set_accounting_length() 205 return accounting_length_.exchange(0, std::memory_order_relaxed); in ClearAccountingLength() 231 return young_gc_state_.load(std::memory_order_relaxed); in young_gc_state() 235 young_gc_state_.store(value, std::memory_order_relaxed); in set_young_gc_state()
|