/third_party/libdrm/ |
D | xf86atomic.h | 42 int atomic; member 45 # define atomic_read(x) ((x)->atomic) 46 # define atomic_set(x, val) ((x)->atomic = (val)) 47 # define atomic_inc(x) ((void) __sync_fetch_and_add (&(x)->atomic, 1)) 48 # define atomic_inc_return(x) (__sync_add_and_fetch (&(x)->atomic, 1)) 49 # define atomic_dec_and_test(x) (__sync_add_and_fetch (&(x)->atomic, -1) == 0) 50 # define atomic_add(x, v) ((void) __sync_add_and_fetch(&(x)->atomic, (v))) 51 # define atomic_dec(x, v) ((void) __sync_sub_and_fetch(&(x)->atomic, (v))) 52 # define atomic_cmpxchg(x, oldv, newv) __sync_val_compare_and_swap (&(x)->atomic, oldv, newv) 63 AO_t atomic; member [all …]
|
/third_party/mesa3d/src/freedreno/ir3/ |
D | ir3_a6xx.c | 113 struct ir3_instruction *atomic, *ibo, *src0, *src1, *data, *dummy; in emit_intrinsic_atomic_ssbo() local 147 atomic = ir3_ATOMIC_B_ADD(b, ibo, 0, src0, 0, src1, 0); in emit_intrinsic_atomic_ssbo() 150 atomic = ir3_ATOMIC_B_MIN(b, ibo, 0, src0, 0, src1, 0); in emit_intrinsic_atomic_ssbo() 154 atomic = ir3_ATOMIC_B_MIN(b, ibo, 0, src0, 0, src1, 0); in emit_intrinsic_atomic_ssbo() 157 atomic = ir3_ATOMIC_B_MAX(b, ibo, 0, src0, 0, src1, 0); in emit_intrinsic_atomic_ssbo() 161 atomic = ir3_ATOMIC_B_MAX(b, ibo, 0, src0, 0, src1, 0); in emit_intrinsic_atomic_ssbo() 164 atomic = ir3_ATOMIC_B_AND(b, ibo, 0, src0, 0, src1, 0); in emit_intrinsic_atomic_ssbo() 167 atomic = ir3_ATOMIC_B_OR(b, ibo, 0, src0, 0, src1, 0); in emit_intrinsic_atomic_ssbo() 170 atomic = ir3_ATOMIC_B_XOR(b, ibo, 0, src0, 0, src1, 0); in emit_intrinsic_atomic_ssbo() 173 atomic = ir3_ATOMIC_B_XCHG(b, ibo, 0, src0, 0, src1, 0); in emit_intrinsic_atomic_ssbo() [all …]
|
D | ir3_a4xx.c | 155 struct ir3_instruction *atomic; in emit_intrinsic_atomic_ssbo() local 169 atomic = ir3_ATOMIC_S_ADD(b, ssbo, 0, data, 0, src3, 0, byte_offset, 0); in emit_intrinsic_atomic_ssbo() 172 atomic = ir3_ATOMIC_S_MIN(b, ssbo, 0, data, 0, src3, 0, byte_offset, 0); in emit_intrinsic_atomic_ssbo() 176 atomic = ir3_ATOMIC_S_MIN(b, ssbo, 0, data, 0, src3, 0, byte_offset, 0); in emit_intrinsic_atomic_ssbo() 179 atomic = ir3_ATOMIC_S_MAX(b, ssbo, 0, data, 0, src3, 0, byte_offset, 0); in emit_intrinsic_atomic_ssbo() 183 atomic = ir3_ATOMIC_S_MAX(b, ssbo, 0, data, 0, src3, 0, byte_offset, 0); in emit_intrinsic_atomic_ssbo() 186 atomic = ir3_ATOMIC_S_AND(b, ssbo, 0, data, 0, src3, 0, byte_offset, 0); in emit_intrinsic_atomic_ssbo() 189 atomic = ir3_ATOMIC_S_OR(b, ssbo, 0, data, 0, src3, 0, byte_offset, 0); in emit_intrinsic_atomic_ssbo() 192 atomic = ir3_ATOMIC_S_XOR(b, ssbo, 0, data, 0, src3, 0, byte_offset, 0); in emit_intrinsic_atomic_ssbo() 195 atomic = ir3_ATOMIC_S_XCHG(b, ssbo, 0, data, 0, src3, 0, byte_offset, 0); in emit_intrinsic_atomic_ssbo() [all …]
|
D | disasm-a3xx.c | 341 OPC(6, OPC_ATOMIC_ADD, atomic.add), 342 OPC(6, OPC_ATOMIC_SUB, atomic.sub), 343 OPC(6, OPC_ATOMIC_XCHG, atomic.xchg), 344 OPC(6, OPC_ATOMIC_INC, atomic.inc), 345 OPC(6, OPC_ATOMIC_DEC, atomic.dec), 346 OPC(6, OPC_ATOMIC_CMPXCHG, atomic.cmpxchg), 347 OPC(6, OPC_ATOMIC_MIN, atomic.min), 348 OPC(6, OPC_ATOMIC_MAX, atomic.max), 349 OPC(6, OPC_ATOMIC_AND, atomic.and), 350 OPC(6, OPC_ATOMIC_OR, atomic.or), [all …]
|
/third_party/mesa3d/src/intel/tools/tests/gen9/ |
D | sends.asm | 8 …dp data 1 MsgDesc: ( DC untyped atomic op, Surface = 1, SIMD8, or) mlen 1 ex_mlen 1 rlen 1 { align… 10 …dp data 1 MsgDesc: ( DC untyped atomic op, Surface = 1, SIMD16, or) mlen 2 ex_mlen 2 rlen 2 { alig… 14 …dp data 1 MsgDesc: ( DC untyped atomic op, Surface = 254, SIMD16, add) mlen 2 ex_mlen 2 rlen 0 { a… 24 …dp data 1 MsgDesc: ( DC untyped atomic op, Surface = 0, SIMD8, imin) mlen 1 ex_mlen 1 rlen 0 { ali… 30 …dp data 1 MsgDesc: ( DC typed atomic, Surface = 1, SIMD16, umax) mlen 2 ex_mlen 1 rlen 0 { align1 … 32 …dp data 1 MsgDesc: ( DC typed atomic, Surface = 1, SIMD8, umax) mlen 2 ex_mlen 1 rlen 0 { align1 2… 34 …dp data 1 MsgDesc: ( DC typed atomic, Surface = 1, SIMD16, mov) mlen 2 ex_mlen 1 rlen 0 { align1 1… 36 …dp data 1 MsgDesc: ( DC typed atomic, Surface = 1, SIMD8, mov) mlen 2 ex_mlen 1 rlen 0 { align1 2Q… 38 …dp data 1 MsgDesc: ( DC typed atomic, Surface = 1, SIMD16, cmpwr) mlen 2 ex_mlen 2 rlen 0 { align1… 40 …dp data 1 MsgDesc: ( DC typed atomic, Surface = 1, SIMD8, cmpwr) mlen 2 ex_mlen 2 rlen 0 { align1 … [all …]
|
/third_party/skia/third_party/externals/tint/src/resolver/ |
D | atomics_test.cc | 31 auto* g = Global("a", ty.atomic(Source{{12, 34}}, ty.i32()), in TEST_F() 36 auto* atomic = TypeOf(g)->UnwrapRef()->As<sem::Atomic>(); in TEST_F() local 37 ASSERT_NE(atomic, nullptr); in TEST_F() 38 EXPECT_TRUE(atomic->Type()->Is<sem::I32>()); in TEST_F() 42 auto* g = Global("a", ty.atomic(Source{{12, 34}}, ty.u32()), in TEST_F() 47 auto* atomic = TypeOf(g)->UnwrapRef()->As<sem::Atomic>(); in TEST_F() local 48 ASSERT_NE(atomic, nullptr); in TEST_F() 49 EXPECT_TRUE(atomic->Type()->Is<sem::U32>()); in TEST_F() 53 auto* s = Structure("s", {Member("a", ty.atomic(Source{{12, 34}}, ty.i32()))}, in TEST_F() 67 auto* atomic = str->Members()[0]->Type()->As<sem::Atomic>(); in TEST_F() local [all …]
|
D | atomics_validation_test.cc | 30 Global("a", ty.atomic(Source{{12, 34}}, ty.i32()), in TEST_F() 37 auto* s = Structure("s", {Member("a", ty.atomic(Source{{12, 34}}, ty.i32()))}, in TEST_F() 46 Global("a", ty.atomic(ty.f32(Source{{12, 34}})), in TEST_F() 54 Global("a", ty.atomic(Source{{12, 34}}, ty.i32()), in TEST_F() 64 Global("a", ty.atomic(Source{{12, 34}}, ty.i32()), in TEST_F() 75 Structure("s", {Member("a", ty.atomic(Source{{12, 34}}, ty.i32()))}); in TEST_F() 91 Structure("Inner", {Member("m", ty.atomic(Source{{12, 34}}, ty.i32()))}); in TEST_F() 109 Structure("Inner", {Member(Source{{12, 34}}, "m", ty.atomic(ty.i32()))}); in TEST_F() 125 ty.atomic(Source{{12, 34}}, ty.i32())); in TEST_F() 141 auto* s = Structure("S", {Member("m", ty.atomic<u32>())}); in TEST_F() [all …]
|
/third_party/openGLES/extensions/ARB/ |
D | ARB_shader_atomic_counters.txt | 65 This extension provides a set of atomic counters. 68 query and increment/decrement these atomic counters. 74 Opaque handles to atomic counters are declared 82 Active atomic counters can be discovered by the commands 86 Like samplers, the opaque handles of the atomic counters 89 The atomic counters pointed to by the opaque handles 98 The contents of the atomic counters are stored 99 in the buffer objects. The contents of atomic 208 atomic counters, can be grouped into <uniform blocks>. 214 except for subroutine uniforms and atomic counters, [all …]
|
/third_party/skia/third_party/externals/opengl-registry/extensions/ARB/ |
D | ARB_shader_atomic_counters.txt | 55 This extension provides a set of atomic counters. 58 query and increment/decrement these atomic counters. 64 Opaque handles to atomic counters are declared 72 Active atomic counters can be discovered by the commands 76 Like samplers, the opaque handles of the atomic counters 79 The atomic counters pointed to by the opaque handles 88 The contents of the atomic counters are stored 89 in the buffer objects. The contents of atomic 198 atomic counters, can be grouped into <uniform blocks>. 204 except for subroutine uniforms and atomic counters, [all …]
|
/third_party/node/deps/v8/include/cppgc/internal/ |
D | gc-info.h | 27 std::atomic<GCInfoIndex>& registered_index) { in EnsureIndex() 37 static GCInfoIndex EnsureGCInfoIndexPolymorphic(std::atomic<GCInfoIndex>&, 41 static GCInfoIndex EnsureGCInfoIndexPolymorphic(std::atomic<GCInfoIndex>&, 44 static GCInfoIndex EnsureGCInfoIndexPolymorphic(std::atomic<GCInfoIndex>&, 46 static GCInfoIndex EnsureGCInfoIndexPolymorphic(std::atomic<GCInfoIndex>&, 48 static GCInfoIndex EnsureGCInfoIndexNonPolymorphic(std::atomic<GCInfoIndex>&, 53 static GCInfoIndex EnsureGCInfoIndexNonPolymorphic(std::atomic<GCInfoIndex>&, 56 static GCInfoIndex EnsureGCInfoIndexNonPolymorphic(std::atomic<GCInfoIndex>&, 59 static GCInfoIndex EnsureGCInfoIndexNonPolymorphic(std::atomic<GCInfoIndex>&, 68 operator()(std::atomic<GCInfoIndex>& registered_index) { \ [all …]
|
/third_party/openGLES/extensions/NV/ |
D | NV_shader_atomic_counters.txt | 37 incrementing, decrementing, and querying the values of atomic counters 40 The extension uses the same set of atomic counter buffer binding points as 105 implementation-dependent number of atomic counter buffer object binding 115 extension. The number of atomic counter buffer object binding points is 135 program.counter[a][b] (x,x,x,x) atomic counter buffer a, 137 program.counter[a][b..c] (x,x,x,x) atomic counter buffer a, 139 program.counter[a] (x,x,x,x) atomic counter buffer a, 145 If an atomic counter buffer binding matches "program.counter[a][b]", the 146 atomic counter variable is associated with element <b> of the buffer 153 For atomic counter array declarations, "program.counter[a][b..c]" is [all …]
|
/third_party/skia/third_party/externals/opengl-registry/extensions/NV/ |
D | NV_shader_atomic_counters.txt | 37 incrementing, decrementing, and querying the values of atomic counters 40 The extension uses the same set of atomic counter buffer binding points as 105 implementation-dependent number of atomic counter buffer object binding 115 extension. The number of atomic counter buffer object binding points is 135 program.counter[a][b] (x,x,x,x) atomic counter buffer a, 137 program.counter[a][b..c] (x,x,x,x) atomic counter buffer a, 139 program.counter[a] (x,x,x,x) atomic counter buffer a, 145 If an atomic counter buffer binding matches "program.counter[a][b]", the 146 atomic counter variable is associated with element <b> of the buffer 153 For atomic counter array declarations, "program.counter[a][b..c]" is [all …]
|
/third_party/skia/third_party/externals/abseil-cpp/absl/container/internal/ |
D | hashtablez_sampler.h | 73 std::atomic<size_t> capacity; 74 std::atomic<size_t> size; 75 std::atomic<size_t> num_erases; 76 std::atomic<size_t> num_rehashes; 77 std::atomic<size_t> max_probe_length; 78 std::atomic<size_t> total_probe_length; 79 std::atomic<size_t> hashes_bitwise_or; 80 std::atomic<size_t> hashes_bitwise_and; 81 std::atomic<size_t> hashes_bitwise_xor; 268 std::atomic<size_t> dropped_samples_; [all …]
|
/third_party/skia/src/gpu/ |
D | GrThreadSafePipelineBuilder.h | 78 std::atomic<int> fShaderCompilations{0}; 80 std::atomic<int> fNumInlineCompilationFailures{0}; 81 std::array<std::atomic<int>, kNumProgramCacheResults> fInlineProgramCacheStats = {}; 83 std::atomic<int> fNumPreCompilationFailures{0}; 84 std::array<std::atomic<int>, kNumProgramCacheResults> fPreProgramCacheStats = {}; 86 std::atomic<int> fNumCompilationFailures{0}; 87 std::atomic<int> fNumPartialCompilationSuccesses{0}; 88 std::atomic<int> fNumCompilationSuccesses{0};
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/WebAssembly/ |
D | WebAssemblyInstrAtomics.td | 39 "atomic.notify \t$dst, ${off}(${addr})${p2align}, $count", 40 "atomic.notify \t${off}${p2align}", 0x00>; 47 "i32.atomic.wait \t$dst, ${off}(${addr})${p2align}, $exp, $timeout", 48 "i32.atomic.wait \t${off}${p2align}", 0x01>; 54 "i64.atomic.wait \t$dst, ${off}(${addr})${p2align}, $exp, $timeout", 55 "i64.atomic.wait \t${off}${p2align}", 0x02>; 125 defm ATOMIC_FENCE : ATOMIC_NRI<(outs), (ins i8imm:$flags), [], "atomic.fence", 138 defm ATOMIC_LOAD_I32 : AtomicLoad<I32, "i32.atomic.load", 0x10>; 139 defm ATOMIC_LOAD_I64 : AtomicLoad<I64, "i64.atomic.load", 0x11>; 163 // Extending loads. Note that there are only zero-extending atomic loads, no [all …]
|
/third_party/node/deps/v8/src/heap/cppgc/ |
D | gc-info.cc | 24 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback, in EnsureGCInfoIndexPolymorphic() 33 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback, in EnsureGCInfoIndexPolymorphic() 42 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback, in EnsureGCInfoIndexPolymorphic() 50 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback) { in EnsureGCInfoIndexPolymorphic() 57 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback, in EnsureGCInfoIndexNonPolymorphic() 66 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback, in EnsureGCInfoIndexNonPolymorphic() 75 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback, in EnsureGCInfoIndexNonPolymorphic() 83 std::atomic<GCInfoIndex>& registered_index, TraceCallback trace_callback) { in EnsureGCInfoIndexNonPolymorphic()
|
/third_party/skia/third_party/externals/abseil-cpp/absl/base/internal/ |
D | spinlock_wait.h | 42 uint32_t SpinLockWait(std::atomic<uint32_t> *w, int n, 50 void SpinLockWake(std::atomic<uint32_t> *w, bool all); 56 void SpinLockDelay(std::atomic<uint32_t> *w, uint32_t value, int loop, 74 void ABSL_INTERNAL_C_SYMBOL(AbslInternalSpinLockWake)(std::atomic<uint32_t> *w, 77 std::atomic<uint32_t> *w, uint32_t value, int loop, 81 inline void absl::base_internal::SpinLockWake(std::atomic<uint32_t> *w, in SpinLockWake() 87 std::atomic<uint32_t> *w, uint32_t value, int loop, in SpinLockDelay()
|
D | spinlock_linux.inc | 21 #include <atomic> 29 // The SpinLock lockword is `std::atomic<uint32_t>`. Here we assert that 30 // `std::atomic<uint32_t>` is bitwise equivalent of the `int` expected 33 // std::atomic<> API. C++ provides no guarantees of these assumptions, 35 static_assert(sizeof(std::atomic<uint32_t>) == sizeof(int), 60 std::atomic<uint32_t> *w, uint32_t value, int loop, 70 std::atomic<uint32_t> *w, bool all) {
|
/third_party/skia/third_party/externals/abseil-cpp/absl/time/ |
D | clock.cc | 115 static inline uint64_t SeqAcquire(std::atomic<uint64_t> *seq) { in SeqAcquire() 130 static inline void SeqRelease(std::atomic<uint64_t> *seq, uint64_t x) { in SeqRelease() 155 std::atomic<uint64_t> raw_ns{0}; // raw kernel time 156 std::atomic<uint64_t> base_ns{0}; // our estimate of time 157 std::atomic<uint64_t> base_cycles{0}; // cycle counter reading 158 std::atomic<uint64_t> nsscaled_per_cycle{0}; // cycle period 161 std::atomic<uint64_t> min_cycles_per_sample{0}; 173 std::atomic<uint64_t> seq{0}; 189 std::atomic<uint64_t> approx_syscall_time_in_cycles{10 * 1000}; 192 std::atomic<uint32_t> kernel_time_seen_smaller{0}; [all …]
|
/third_party/ltp/lib/newlib_tests/ |
D | test15.c | 42 static int atomic; variable 56 for (i = tst_atomic_load(&atomic); in worker_load_store() 58 i = tst_atomic_load(&atomic)) in worker_load_store() 63 tst_atomic_store(i + 1, &atomic); in worker_load_store() 78 while (tst_atomic_load(&atomic) > 0) { in mem_spam() 98 atomic = 0; in do_test() 113 tst_atomic_store(-1, &atomic); in do_test()
|
/third_party/skia/third_party/externals/opengl-registry/extensions/OES/ |
D | OES_shader_image_atomic.txt | 60 atomic read-modify-write operations to a single level of a texture 71 store, and atomic built-in functions in the manner ..." 78 store, or atomic operation..." 85 built-in image load, store, and atomic functions operating on shader 92 "Table 8.26: Mapping of image load, store, and atomic texel..." 98 built-in image load, store or atomic function..." 103 "If the individual texel identified for an image load, store, or atomic 112 atomic operations are considered to involve a format mismatch. In such 113 cases, undefined values will be returned by image loads and atomic 114 operations, and undefined values will be written by stores and atomic [all …]
|
/third_party/skia/third_party/externals/harfbuzz/src/ |
D | hb-atomic.hh | 87 #define hb_atomic_int_impl_add(AI, V) (reinterpret_cast<std::atomic<int> *> (AI)->fetch_add ((V), … 88 #define hb_atomic_int_impl_set_relaxed(AI, V) (reinterpret_cast<std::atomic<int> *> (AI)->store ((V… 89 #define hb_atomic_int_impl_set(AI, V) (reinterpret_cast<std::atomic<int> *> (AI)->store ((V), std:… 90 #define hb_atomic_int_impl_get_relaxed(AI) (reinterpret_cast<std::atomic<int> const *> (AI)->load (… 91 #define hb_atomic_int_impl_get(AI) (reinterpret_cast<std::atomic<int> const *> (AI)->load (std::me… 93 #define hb_atomic_ptr_impl_set_relaxed(P, V) (reinterpret_cast<std::atomic<void*> *> (P)->store ((V… 94 #define hb_atomic_ptr_impl_get_relaxed(P) (reinterpret_cast<std::atomic<void*> const *> (P)->load (… 95 #define hb_atomic_ptr_impl_get(P) (reinterpret_cast<std::atomic<void*> *> (P)->load (std::memory_o… 100 …return reinterpret_cast<std::atomic<const void*> *> (P)->compare_exchange_weak (O, N, std::memory_… in _hb_atomic_ptr_impl_cmplexch()
|
/third_party/openGLES/extensions/OES/ |
D | OES_shader_image_atomic.txt | 70 atomic read-modify-write operations to a single level of a texture 81 store, and atomic built-in functions in the manner ..." 88 store, or atomic operation..." 95 built-in image load, store, and atomic functions operating on shader 102 "Table 8.26: Mapping of image load, store, and atomic texel..." 108 built-in image load, store or atomic function..." 113 "If the individual texel identified for an image load, store, or atomic 122 atomic operations are considered to involve a format mismatch. In such 123 cases, undefined values will be returned by image loads and atomic 124 operations, and undefined values will be written by stores and atomic [all …]
|
/third_party/skia/third_party/externals/abseil-cpp/absl/base/ |
D | call_once.h | 49 std::atomic<uint32_t>* ControlWord(absl::once_flag* flag); 92 friend std::atomic<uint32_t>* base_internal::ControlWord(once_flag* flag); 93 std::atomic<uint32_t> control_; 147 void CallOnceImpl(std::atomic<uint32_t>* control, in CallOnceImpl() 188 inline std::atomic<uint32_t>* ControlWord(once_flag* flag) { in ControlWord() 194 std::atomic<uint32_t>* once = base_internal::ControlWord(flag); in LowLevelCallOnce() 207 std::atomic<uint32_t>* once = base_internal::ControlWord(&flag); in call_once()
|
/third_party/openGLES/extensions/AMD/ |
D | AMD_shader_atomic_counter_ops.txt | 40 The ARB_shader_atomic_counters extension introduced atomic counters, but 43 built-in functions that can operate on atomic counters. The list of new 72 Additions to the table listing atomic counter built-in functions: 81 … These two steps are done atomically with respect to the atomic 89 … These two steps are done atomically with respect to the atomic 95 … These two steps are done atomically with respect to the atomic 101 … These two steps are done atomically with respect to the atomic 109 … These two steps are done atomically with respect to the atomic 116 … These two steps are done atomically with respect to the atomic 123 … These two steps are done atomically with respect to the atomic [all …]
|