/external/skqp/src/core/ |
D | SkTaskGroup.cpp | 14 fPending.fetch_add(+1, std::memory_order_relaxed); in add() 17 fPending.fetch_add(-1, std::memory_order_release); in add() 23 fPending.fetch_add(+N, std::memory_order_relaxed); in batch() 27 fPending.fetch_add(-1, std::memory_order_release); in batch()
|
/external/skia/src/core/ |
D | SkTaskGroup.cpp | 14 fPending.fetch_add(+1, std::memory_order_relaxed); in add() 17 fPending.fetch_add(-1, std::memory_order_release); in add() 23 fPending.fetch_add(+N, std::memory_order_relaxed); in batch() 27 fPending.fetch_add(-1, std::memory_order_release); in batch()
|
/external/skia/src/gpu/vk/ |
D | GrVkResource.h | 77 fKey = fKeyCounter.fetch_add(+1, std::memory_order_relaxed); in GrVkResource() 112 SkDEBUGCODE(int newRefCount = )fRefCnt.fetch_add(+1, std::memory_order_relaxed); in ref() 124 int newRefCount = fRefCnt.fetch_add(-1, std::memory_order_acq_rel); in unref() 137 int newRefCount = fRefCnt.fetch_add(-1, std::memory_order_acq_rel); in unrefAndAbandon()
|
/external/skqp/src/gpu/vk/ |
D | GrVkResource.h | 77 fKey = fKeyCounter.fetch_add(+1, std::memory_order_relaxed); in GrVkResource() 112 SkDEBUGCODE(int newRefCount = )fRefCnt.fetch_add(+1, std::memory_order_relaxed); in ref() 124 int newRefCount = fRefCnt.fetch_add(-1, std::memory_order_acq_rel); in unref() 137 int newRefCount = fRefCnt.fetch_add(-1, std::memory_order_acq_rel); in unrefAndAbandon()
|
/external/swiftshader/third_party/llvm-subzero/include/llvm/ADT/ |
D | Statistic.h | 71 Value.fetch_add(1, std::memory_order_relaxed); 77 return Value.fetch_add(1, std::memory_order_relaxed); 93 Value.fetch_add(V, std::memory_order_relaxed);
|
/external/llvm/include/llvm/ADT/ |
D | Statistic.h | 70 Value.fetch_add(1, std::memory_order_relaxed); 76 return Value.fetch_add(1, std::memory_order_relaxed); 92 Value.fetch_add(V, std::memory_order_relaxed);
|
/external/protobuf/src/google/protobuf/stubs/ |
D | atomicops_internals_pnacl.h | 86 ->fetch_add(increment, std::memory_order_relaxed); in NoBarrier_AtomicIncrement() 91 return increment + ((AtomicLocation32)ptr)->fetch_add(increment); in Barrier_AtomicIncrement() 169 ->fetch_add(increment, std::memory_order_relaxed); in NoBarrier_AtomicIncrement() 174 return increment + ((AtomicLocation64)ptr)->fetch_add(increment); in Barrier_AtomicIncrement()
|
/external/libchrome/base/ |
D | atomicops_internals_portable.h | 86 ->fetch_add(increment, std::memory_order_relaxed); in NoBarrier_AtomicIncrement() 91 return increment + ((AtomicLocation32)ptr)->fetch_add(increment); in Barrier_AtomicIncrement() 169 ->fetch_add(increment, std::memory_order_relaxed); in NoBarrier_AtomicIncrement() 174 return increment + ((AtomicLocation64)ptr)->fetch_add(increment); in Barrier_AtomicIncrement()
|
D | atomic_sequence_num.h | 23 inline int GetNext() { return seq_.fetch_add(1, std::memory_order_relaxed); } in GetNext()
|
D | atomic_ref_count.h | 26 ref_count_.fetch_add(increment, std::memory_order_relaxed); in Increment()
|
/external/swiftshader/third_party/llvm-7.0/llvm/include/llvm/ADT/ |
D | Statistic.h | 80 Value.fetch_add(1, std::memory_order_relaxed); 86 return Value.fetch_add(1, std::memory_order_relaxed); 102 Value.fetch_add(V, std::memory_order_relaxed);
|
/external/skia/include/private/ |
D | SkWeakRefCnt.h | 116 (void)fWeakCnt.fetch_add(+1, std::memory_order_relaxed); in weak_ref() 127 if (1 == fWeakCnt.fetch_add(-1, std::memory_order_acq_rel)) { in weak_unref()
|
D | SkSemaphore.h | 61 int prev = fCount.fetch_add(n, std::memory_order_release); in signal()
|
/external/skqp/include/private/ |
D | SkWeakRefCnt.h | 116 (void)fWeakCnt.fetch_add(+1, std::memory_order_relaxed); in weak_ref() 127 if (1 == fWeakCnt.fetch_add(-1, std::memory_order_acq_rel)) { in weak_unref()
|
D | SkSemaphore.h | 61 int prev = fCount.fetch_add(n, std::memory_order_release); in signal()
|
/external/swiftshader/src/Reactor/ |
D | Thread.hpp | 308 inline void operator++() { ai.fetch_add(1, std::memory_order_acq_rel); } in operator ++() 310 inline int operator++(int) { return ai.fetch_add(1, std::memory_order_acq_rel) + 1; } in operator ++() 312 inline void operator+=(int i) { ai.fetch_add(i, std::memory_order_acq_rel); } in operator +=()
|
/external/libcxx/include/ |
D | atomic | 165 fetch_add(integral op, memory_order m = memory_order_seq_cst) volatile noexcept; 166 integral fetch_add(integral op, memory_order m = memory_order_seq_cst) noexcept; 238 T* fetch_add(ptrdiff_t op, memory_order m = memory_order_seq_cst) volatile noexcept; 239 T* fetch_add(ptrdiff_t op, memory_order m = memory_order_seq_cst) noexcept; 1013 _Tp fetch_add(_Tp __op, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT 1016 _Tp fetch_add(_Tp __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT 1044 _Tp operator++(int) volatile _NOEXCEPT {return fetch_add(_Tp(1));} 1046 _Tp operator++(int) _NOEXCEPT {return fetch_add(_Tp(1));} 1052 _Tp operator++() volatile _NOEXCEPT {return fetch_add(_Tp(1)) + _Tp(1);} 1054 _Tp operator++() _NOEXCEPT {return fetch_add(_Tp(1)) + _Tp(1);} [all …]
|
/external/swiftshader/src/Common/ |
D | Thread.hpp | 308 inline void operator++() { ai.fetch_add(1, std::memory_order_acq_rel); } in operator ++() 310 inline int operator++(int) { return ai.fetch_add(1, std::memory_order_acq_rel) + 1; } in operator ++() 312 inline void operator+=(int i) { ai.fetch_add(i, std::memory_order_acq_rel); } in operator +=()
|
/external/swiftshader/src/System/ |
D | Thread.hpp | 308 inline void operator++() { ai.fetch_add(1, std::memory_order_acq_rel); } in operator ++() 310 inline int operator++(int) { return ai.fetch_add(1, std::memory_order_acq_rel) + 1; } in operator ++() 312 inline void operator+=(int i) { ai.fetch_add(i, std::memory_order_acq_rel); } in operator +=()
|
/external/skqp/include/core/ |
D | SkRefCnt.h | 65 (void)fRefCnt.fetch_add(+1, std::memory_order_relaxed); in ref() 75 if (1 == fRefCnt.fetch_add(-1, std::memory_order_acq_rel)) { in unref() 176 void ref() const { (void)fRefCnt.fetch_add(+1, std::memory_order_relaxed); } in ref() 178 if (1 == fRefCnt.fetch_add(-1, std::memory_order_acq_rel)) { in unref()
|
/external/skia/include/core/ |
D | SkRefCnt.h | 65 (void)fRefCnt.fetch_add(+1, std::memory_order_relaxed); in ref() 75 if (1 == fRefCnt.fetch_add(-1, std::memory_order_acq_rel)) { in unref() 176 void ref() const { (void)fRefCnt.fetch_add(+1, std::memory_order_relaxed); } in ref() 178 if (1 == fRefCnt.fetch_add(-1, std::memory_order_acq_rel)) { in unref()
|
/external/tensorflow/tensorflow/core/profiler/lib/ |
D | traceme.cc | 26 const thread_local static uint32 thread_id = thread_counter.fetch_add(1); in NewActivityId()
|
/external/libchrome/base/android/orderfile/ |
D | orderfile_instrumentation.cc | 164 ordered_offsets_index.fetch_add(1, std::memory_order_relaxed); in RecordAddress() 245 int before = g_data_index.fetch_add(1, std::memory_order_relaxed); in SwitchToNextPhaseOrDump()
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | partition_assignment.cc | 46 if (log_count.fetch_add(1) < 8) { in ThreadsPerBlockLimit()
|
/external/tensorflow/tensorflow/core/platform/ |
D | tracing.cc | 57 return unique_arg.fetch_add(1, std::memory_order_relaxed); in GetUniqueArg()
|