/external/skia/src/core/ |
D | SkTaskGroup.cpp | 14 fPending.fetch_add(+1, std::memory_order_relaxed); in add() 17 fPending.fetch_add(-1, std::memory_order_release); in add() 23 fPending.fetch_add(+N, std::memory_order_relaxed); in batch() 27 fPending.fetch_add(-1, std::memory_order_release); in batch()
|
/external/skqp/src/core/ |
D | SkTaskGroup.cpp | 14 fPending.fetch_add(+1, std::memory_order_relaxed); in add() 17 fPending.fetch_add(-1, std::memory_order_release); in add() 23 fPending.fetch_add(+N, std::memory_order_relaxed); in batch() 27 fPending.fetch_add(-1, std::memory_order_release); in batch()
|
/external/skia/include/gpu/ |
D | GrGpuResource.h | 38 (void)fRefCnt.fetch_add(+1, std::memory_order_relaxed); in ref() 43 if (1 == fRefCnt.fetch_add(-1, std::memory_order_acq_rel)) { in unref() 46 fRefCnt.fetch_add(+1, std::memory_order_relaxed); in unref() 53 if (1 == fRefCnt.fetch_add(-1, std::memory_order_acq_rel)) { in unref() 74 (void)fRefCnt.fetch_add(+1, std::memory_order_relaxed); in addInitialRef()
|
/external/skqp/src/gpu/vk/ |
D | GrVkResource.h | 77 fKey = fKeyCounter.fetch_add(+1, std::memory_order_relaxed); in GrVkResource() 112 SkDEBUGCODE(int newRefCount = )fRefCnt.fetch_add(+1, std::memory_order_relaxed); in ref() 124 int newRefCount = fRefCnt.fetch_add(-1, std::memory_order_acq_rel); in unref() 137 int newRefCount = fRefCnt.fetch_add(-1, std::memory_order_acq_rel); in unrefAndAbandon()
|
/external/llvm/include/llvm/ADT/ |
D | Statistic.h | 70 Value.fetch_add(1, std::memory_order_relaxed); 76 return Value.fetch_add(1, std::memory_order_relaxed); 92 Value.fetch_add(V, std::memory_order_relaxed);
|
/external/swiftshader/third_party/llvm-subzero/include/llvm/ADT/ |
D | Statistic.h | 71 Value.fetch_add(1, std::memory_order_relaxed); 77 return Value.fetch_add(1, std::memory_order_relaxed); 93 Value.fetch_add(V, std::memory_order_relaxed);
|
/external/libchrome/base/ |
D | atomicops_internals_portable.h | 86 ->fetch_add(increment, std::memory_order_relaxed); in NoBarrier_AtomicIncrement() 91 return increment + ((AtomicLocation32)ptr)->fetch_add(increment); in Barrier_AtomicIncrement() 169 ->fetch_add(increment, std::memory_order_relaxed); in NoBarrier_AtomicIncrement() 174 return increment + ((AtomicLocation64)ptr)->fetch_add(increment); in Barrier_AtomicIncrement()
|
D | atomic_sequence_num.h | 23 inline int GetNext() { return seq_.fetch_add(1, std::memory_order_relaxed); } in GetNext()
|
D | atomic_ref_count.h | 26 ref_count_.fetch_add(increment, std::memory_order_relaxed); in Increment()
|
/external/swiftshader/third_party/llvm-7.0/llvm/include/llvm/ADT/ |
D | Statistic.h | 80 Value.fetch_add(1, std::memory_order_relaxed); 86 return Value.fetch_add(1, std::memory_order_relaxed); 102 Value.fetch_add(V, std::memory_order_relaxed);
|
/external/skia/src/gpu/vk/ |
D | GrVkResource.h | 80 fKey = fKeyCounter.fetch_add(+1, std::memory_order_relaxed); in GrVkResource() 115 SkDEBUGCODE(int newRefCount = )fRefCnt.fetch_add(+1, std::memory_order_relaxed); in ref() 127 int newRefCount = fRefCnt.fetch_add(-1, std::memory_order_acq_rel); in unref()
|
/external/skqp/include/private/ |
D | SkWeakRefCnt.h | 116 (void)fWeakCnt.fetch_add(+1, std::memory_order_relaxed); in weak_ref() 127 if (1 == fWeakCnt.fetch_add(-1, std::memory_order_acq_rel)) { in weak_unref()
|
D | SkSemaphore.h | 61 int prev = fCount.fetch_add(n, std::memory_order_release); in signal()
|
/external/skia/include/private/ |
D | SkWeakRefCnt.h | 116 (void)fWeakCnt.fetch_add(+1, std::memory_order_relaxed); in weak_ref() 127 if (1 == fWeakCnt.fetch_add(-1, std::memory_order_acq_rel)) { in weak_unref()
|
D | SkSemaphore.h | 56 int prev = fCount.fetch_add(n, std::memory_order_release); in signal()
|
/external/skia/include/core/ |
D | SkRefCnt.h | 64 (void)fRefCnt.fetch_add(+1, std::memory_order_relaxed); in ref() 74 if (1 == fRefCnt.fetch_add(-1, std::memory_order_acq_rel)) { in unref() 175 void ref() const { (void)fRefCnt.fetch_add(+1, std::memory_order_relaxed); } in ref() 177 if (1 == fRefCnt.fetch_add(-1, std::memory_order_acq_rel)) { in unref()
|
/external/swiftshader/src/Common/ |
D | Thread.hpp | 308 inline void operator++() { ai.fetch_add(1, std::memory_order_acq_rel); } in operator ++() 310 inline int operator++(int) { return ai.fetch_add(1, std::memory_order_acq_rel) + 1; } in operator ++() 312 inline void operator+=(int i) { ai.fetch_add(i, std::memory_order_acq_rel); } in operator +=()
|
/external/libcxx/include/ |
D | atomic | 165 fetch_add(integral op, memory_order m = memory_order_seq_cst) volatile noexcept; 166 integral fetch_add(integral op, memory_order m = memory_order_seq_cst) noexcept; 238 T* fetch_add(ptrdiff_t op, memory_order m = memory_order_seq_cst) volatile noexcept; 239 T* fetch_add(ptrdiff_t op, memory_order m = memory_order_seq_cst) noexcept; 1013 _Tp fetch_add(_Tp __op, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT 1016 _Tp fetch_add(_Tp __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT 1044 _Tp operator++(int) volatile _NOEXCEPT {return fetch_add(_Tp(1));} 1046 _Tp operator++(int) _NOEXCEPT {return fetch_add(_Tp(1));} 1052 _Tp operator++() volatile _NOEXCEPT {return fetch_add(_Tp(1)) + _Tp(1);} 1054 _Tp operator++() _NOEXCEPT {return fetch_add(_Tp(1)) + _Tp(1);} [all …]
|
/external/skqp/include/core/ |
D | SkRefCnt.h | 65 (void)fRefCnt.fetch_add(+1, std::memory_order_relaxed); in ref() 75 if (1 == fRefCnt.fetch_add(-1, std::memory_order_acq_rel)) { in unref() 176 void ref() const { (void)fRefCnt.fetch_add(+1, std::memory_order_relaxed); } in ref() 178 if (1 == fRefCnt.fetch_add(-1, std::memory_order_acq_rel)) { in unref()
|
/external/rust/crates/syn/1.0.7/tests/ |
D | test_round_trip.rs | 70 let prev_failed = failed.fetch_add(1, Ordering::SeqCst); in test_round_trip() 133 let prev_failed = failed.fetch_add(1, Ordering::SeqCst); in test_round_trip()
|
/external/rust/crates/syn/0.15.42/tests/ |
D | test_round_trip.rs | 73 let prev_failed = failed.fetch_add(1, Ordering::SeqCst); in test_round_trip() 136 let prev_failed = failed.fetch_add(1, Ordering::SeqCst); in test_round_trip()
|
/external/v8/include/cppgc/internal/ |
D | atomic-entry-flag.h | 31 void Enter() { entries_.fetch_add(1, std::memory_order_relaxed); } in Enter()
|
/external/v8/src/execution/ |
D | thread-id.cc | 32 thread_id = next_thread_id.fetch_add(1); in GetCurrentThreadId()
|
/external/tensorflow/tensorflow/core/platform/ |
D | tracing.cc | 51 return unique_arg.fetch_add(1, std::memory_order_relaxed); in GetUniqueArg()
|
/external/v8/src/profiler/ |
D | profiler-stats.cc | 15 counts_[reason].fetch_add(1, std::memory_order_relaxed); in AddReason()
|