/external/tensorflow/tensorflow/core/framework/ |
D | model.h | 122 void add_buffered_bytes(int64 delta) LOCKS_EXCLUDED(mu_) { in add_buffered_bytes() 123 mutex_lock l(mu_); in add_buffered_bytes() 128 void add_input(std::shared_ptr<Node> node) LOCKS_EXCLUDED(mu_) { in add_input() 129 mutex_lock l(mu_); in add_input() 134 void add_processing_time(int64 delta) LOCKS_EXCLUDED(mu_) { in add_processing_time() 135 mutex_lock l(mu_); in add_processing_time() 140 int64 buffered_bytes() const LOCKS_EXCLUDED(mu_) { in buffered_bytes() 141 tf_shared_lock l(mu_); in buffered_bytes() 146 bool has_tunable_parameters() const LOCKS_EXCLUDED(mu_) { in has_tunable_parameters() 147 tf_shared_lock l(mu_); in has_tunable_parameters() [all …]
|
D | cancellation.cc | 31 mutex_lock l(mu_); in Reset() 39 mutex_lock l(mu_); in StartCancel() 55 mutex_lock l(mu_); in StartCancel() 63 mutex_lock l(mu_); in get_cancellation_token() 69 mutex_lock l(mu_); in RegisterCallback() 79 mu_.lock(); in DeregisterCallback() 81 mu_.unlock(); in DeregisterCallback() 84 mu_.unlock(); in DeregisterCallback() 93 mu_.unlock(); in DeregisterCallback() 99 mutex_lock lock(mu_); in TryDeregisterCallback()
|
D | tracking_allocator.h | 95 bool UnRef() EXCLUSIVE_LOCKS_REQUIRED(mu_); 98 mutex mu_; variable 102 int ref_ GUARDED_BY(mu_); 106 size_t allocated_ GUARDED_BY(mu_); 110 size_t high_watermark_ GUARDED_BY(mu_); 115 size_t total_bytes_ GUARDED_BY(mu_); 117 gtl::InlinedVector<AllocRecord, 4> allocations_ GUARDED_BY(mu_); 127 std::unordered_map<const void*, Chunk> in_use_ GUARDED_BY(mu_); 128 int64 next_allocation_id_ GUARDED_BY(mu_);
|
/external/tensorflow/tensorflow/core/platform/default/ |
D | mutex.h | 59 external_mu_space mu_; variable 67 explicit mutex_lock(mutex_type& mu) EXCLUSIVE_LOCK_FUNCTION(mu) : mu_(&mu) { in mutex_lock() 68 mu_->lock(); in mutex_lock() 72 : mu_(&mu) { in mutex_lock() 74 mu_ = nullptr; in mutex_lock() 80 mutex_lock(mutex_lock&& ml) noexcept EXCLUSIVE_LOCK_FUNCTION(ml.mu_) in mutex_lock() 81 : mu_(ml.mu_) { in mutex_lock() 82 ml.mu_ = nullptr; in mutex_lock() 85 if (mu_ != nullptr) { in UNLOCK_FUNCTION() 86 mu_->unlock(); in UNLOCK_FUNCTION() [all …]
|
D | mutex.cc | 36 mutex::mutex() { nsync::nsync_mu_init(mu_cast(&mu_)); } in mutex() 38 void mutex::lock() { nsync::nsync_mu_lock(mu_cast(&mu_)); } in lock() 40 bool mutex::try_lock() { return nsync::nsync_mu_trylock(mu_cast(&mu_)) != 0; }; in try_lock() 42 void mutex::unlock() { nsync::nsync_mu_unlock(mu_cast(&mu_)); } in unlock() 44 void mutex::lock_shared() { nsync::nsync_mu_rlock(mu_cast(&mu_)); } in lock_shared() 47 return nsync::nsync_mu_rtrylock(mu_cast(&mu_)) != 0; in try_lock_shared() 50 void mutex::unlock_shared() { nsync::nsync_mu_runlock(mu_cast(&mu_)); } in unlock_shared() 72 nsync::nsync_cv_wait(cv_cast(&cv_), mu_cast(&lock.mutex()->mu_)); in wait() 79 cv_cast(&cv_), mu_cast(&lock.mutex()->mu_), timeout_time, nullptr); in wait_until_system_clock()
|
/external/clang/test/SemaCXX/ |
D | warn-thread-safety-analysis.cpp | 972 void func1(int y) LOCKS_EXCLUDED(mu_); 973 template <typename T> void func2(T x) LOCKS_EXCLUDED(mu_); 975 Mutex mu_; member in thread_annot_lock_38::Foo 991 Mutex *mu_; member in thread_annot_lock_43::Foo 997 int GetA() EXCLUSIVE_LOCKS_REQUIRED(foo_->mu_) { return a_; } in GetA() 998 int a_ GUARDED_BY(foo_->mu_); 1006 fb->foo_->mu_->Lock(); in main() 1008 fb->foo_->mu_->Unlock(); in main() 1124 Mutex mu_; member in thread_annot_lock_68_modified::Bar 1131 mu_.Lock(); in func() [all …]
|
/external/grpc-grpc/src/core/lib/gprpp/ |
D | fork.cc | 53 gpr_mu_init(&mu_); in ExecCtxState() 64 gpr_mu_lock(&mu_); in IncExecCtxCount() 67 gpr_cv_wait(&cv_, &mu_, gpr_inf_future(GPR_CLOCK_REALTIME)); in IncExecCtxCount() 70 gpr_mu_unlock(&mu_); in IncExecCtxCount() 83 gpr_mu_lock(&mu_); in BlockExecCtx() 85 gpr_mu_unlock(&mu_); in BlockExecCtx() 92 gpr_mu_lock(&mu_); in AllowExecCtx() 96 gpr_mu_unlock(&mu_); in AllowExecCtx() 100 gpr_mu_destroy(&mu_); in ~ExecCtxState() 106 gpr_mu mu_; member in grpc_core::internal::ExecCtxState [all …]
|
/external/tensorflow/tensorflow/compiler/jit/ |
D | xla_device.h | 140 LOCKS_EXCLUDED(mu_); 149 LOCKS_EXCLUDED(mu_); 153 Tensor* tensor) override LOCKS_EXCLUDED(mu_); 163 Status EnsureDeviceContextOk() LOCKS_EXCLUDED(mu_); 167 Status UseGpuDeviceInfo() LOCKS_EXCLUDED(mu_); 171 void SetAllowsSyncOnCompletion(bool sync_on_completion) LOCKS_EXCLUDED(mu_); 172 bool AllowsSyncOnCompletion() const override LOCKS_EXCLUDED(mu_); 177 Status RefreshStatus() override LOCKS_EXCLUDED(mu_); 182 EXCLUSIVE_LOCKS_REQUIRED(mu_); 186 EXCLUSIVE_LOCKS_REQUIRED(mu_); [all …]
|
/external/tensorflow/tensorflow/core/platform/cloud/ |
D | ram_file_block_cache.h | 99 LOCKS_EXCLUDED(mu_); 102 void RemoveFile(const string& filename) override LOCKS_EXCLUDED(mu_); 105 void Flush() override LOCKS_EXCLUDED(mu_); 113 size_t CacheSize() const override LOCKS_EXCLUDED(mu_); 191 void Prune() LOCKS_EXCLUDED(mu_); 194 EXCLUSIVE_LOCKS_REQUIRED(mu_); 197 std::shared_ptr<Block> Lookup(const Key& key) LOCKS_EXCLUDED(mu_); 200 LOCKS_EXCLUDED(mu_); 203 void Trim() EXCLUSIVE_LOCKS_REQUIRED(mu_); 207 LOCKS_EXCLUDED(mu_); [all …]
|
D | gcs_throttle.h | 112 inline int64 available_tokens() LOCKS_EXCLUDED(mu_) { in available_tokens() 113 mutex_lock l(mu_); in available_tokens() 125 bool is_enabled() LOCKS_EXCLUDED(mu_) { in is_enabled() 126 mutex_lock l(mu_); in is_enabled() 137 void UpdateState() EXCLUSIVE_LOCKS_REQUIRED(mu_); 143 mutex mu_; variable 150 uint64 last_updated_secs_ GUARDED_BY(mu_) = 0; 159 int64 available_tokens_ GUARDED_BY(mu_) = 0; 162 GcsThrottleConfig config_ GUARDED_BY(mu_);
|
D | expiring_lru_cache.h | 50 mutex_lock lock(mu_); in Insert() 58 mutex_lock lock(mu_); in Delete() 69 mutex_lock lock(mu_); in Lookup() 88 mutex_lock lock(mu_); in LookupOrCompute() 101 mutex_lock lock(mu_); in Clear() 122 bool LookupLocked(const string& key, T* value) EXCLUSIVE_LOCKS_REQUIRED(mu_) { in LookupLocked() 139 EXCLUSIVE_LOCKS_REQUIRED(mu_) { in InsertLocked() 152 bool DeleteLocked(const string& key) EXCLUSIVE_LOCKS_REQUIRED(mu_) { in DeleteLocked() 174 mutex mu_; variable 177 std::map<string, Entry> cache_ GUARDED_BY(mu_); [all …]
|
/external/tensorflow/tensorflow/core/kernels/data/ |
D | cache_dataset_ops.cc | 127 mutex_lock l(mu_); in Initialize() 134 mutex_lock l(mu_); in GetNextInternal() 146 mutex_lock l(mu_); in SaveInternal() 152 mutex_lock l(mu_); in RestoreInternal() 214 mutex_lock l(mu_); in GetNextInternal() 262 mutex_lock l(mu_); in SaveInternal() 298 mutex_lock l(mu_); in RestoreInternal() 333 Status EnsureLockFileExists() EXCLUSIVE_LOCKS_REQUIRED(mu_) { in EnsureLockFileExists() 395 Status Finish() EXCLUSIVE_LOCKS_REQUIRED(mu_) { in Finish() 424 mutex mu_; member in tensorflow::data::__anon80340b800111::CacheDatasetOp::FileDataset::FileIterator::FileWriterIterator [all …]
|
D | parallel_map_iterator.cc | 55 mu_(std::make_shared<mutex>()), in ParallelMapIterator() 58 params.num_parallel_calls, mu_, cond_var_)), in ParallelMapIterator() 65 mutex_lock l(*mu_); in ~ParallelMapIterator() 76 mutex_lock l(*mu_); in Initialize() 89 mutex_lock l(*mu_); in GetNextInternal() 114 mutex_lock l(*mu_); in SaveInternal() 146 mutex_lock l(*mu_); in RestoreInternal() 191 EXCLUSIVE_LOCKS_REQUIRED(*mu_) { in EnsureRunnerThreadStarted() 202 LOCKS_EXCLUDED(*mu_) { in CallCompleted() 203 mutex_lock l(*mu_); in CallCompleted() [all …]
|
D | multi_device_iterator_ops.cc | 78 mutex_lock l(mu_); in Init() 95 tf_shared_lock l(mu_); in GetNextFromShard() 115 tf_shared_lock l(mu_); in function_library() 120 tf_shared_lock l(mu_); in lib() 147 mutex_lock l(mu_); in ~MultiDeviceBuffer() 153 void Reset() LOCKS_EXCLUDED(mu_) { in Reset() 155 mutex_lock l(mu_); in Reset() 184 mutex_lock l(mu_); in GetNextFromShard() 219 EXCLUSIVE_LOCKS_REQUIRED(mu_) { in EnsureBackgroundThreadStarted() 231 void RunPendingCallbacks() LOCKS_EXCLUDED(mu_) { in RunPendingCallbacks() [all …]
|
/external/tensorflow/tensorflow/core/common_runtime/gpu/ |
D | gpu_allocator_retry_test.cc | 40 mutex_lock l(mu_); in AllocateRaw() 52 mutex_lock l(mu_); in DeallocateRaw() 60 mutex mu_; member in tensorflow::__anon9dc664200111::FakeAllocator 61 size_t memory_capacity_ GUARDED_BY(mu_); 80 mutex_lock l(mu_); in WaitTurn() 94 mutex_lock l(mu_); in Done() 103 void IncrementTurn() EXCLUSIVE_LOCKS_REQUIRED(mu_) { in IncrementTurn() 112 mutex mu_; member in tensorflow::__anon9dc664200111::AlternatingBarrier 115 int next_turn_ GUARDED_BY(mu_); 116 std::vector<bool> done_ GUARDED_BY(mu_); [all …]
|
D | gpu_event_mgr.h | 89 mutex_lock l(mu_); in ThenDeleteBuffer() 102 mutex_lock l(mu_); in ThenExecute() 114 mutex mu_; variable 115 condition_variable events_pending_ GUARDED_BY(mu_); 117 void FlushAccumulatedTensors() EXCLUSIVE_LOCKS_REQUIRED(mu_); 153 EXCLUSIVE_LOCKS_REQUIRED(mu_); 156 EXCLUSIVE_LOCKS_REQUIRED(mu_) { in QueueTensors() 161 EXCLUSIVE_LOCKS_REQUIRED(mu_) { in QueueBuffer() 166 EXCLUSIVE_LOCKS_REQUIRED(mu_) { in QueueFunc() 176 EXCLUSIVE_LOCKS_REQUIRED(mu_); [all …]
|
/external/tensorflow/tensorflow/core/common_runtime/ |
D | scoped_allocator.h | 53 ~ScopedAllocator() LOCKS_EXCLUDED(mu_); 68 void* AllocateRaw(int32 field_index, size_t num_bytes) LOCKS_EXCLUDED(mu_); 69 void DeallocateRaw(void* p) LOCKS_EXCLUDED(mu_); 76 mutex mu_; variable 77 int32 expected_call_count_ GUARDED_BY(mu_); 78 int32 live_alloc_count_ GUARDED_BY(mu_); 98 void DropFromTable() LOCKS_EXCLUDED(mu_); 100 LOCKS_EXCLUDED(mu_) override; 105 void DeallocateRaw(void* p) LOCKS_EXCLUDED(mu_) override; 115 mutex mu_; [all …]
|
/external/tensorflow/tensorflow/core/kernels/batching_util/ |
D | serial_device_batch_scheduler.h | 121 mutex_lock l(mu_); in in_flight_batches_limit() 126 mutex_lock l(mu_); in recent_low_traffic_ratio() 151 std::vector<const internal::SDBSBatch<TaskType>*> batches_ GUARDED_BY(mu_); 155 queues_and_callbacks_ GUARDED_BY(mu_); 161 int64 in_flight_batches_limit_ GUARDED_BY(mu_); 164 int64 processing_threads_ GUARDED_BY(mu_) = 0; 168 int64 batch_count_ GUARDED_BY(mu_) = 0; 172 int64 no_batch_count_ GUARDED_BY(mu_) = 0; 189 mutex mu_; variable 232 SDBSBatch<TaskType>* current_batch_ GUARDED_BY(mu_) = nullptr; [all …]
|
D | adaptive_shared_batch_scheduler.h | 142 mutex_lock l(mu_); in in_flight_batches_limit() 157 void MaybeScheduleNextBatch() EXCLUSIVE_LOCKS_REQUIRED(mu_); 164 void MaybeScheduleClosedBatch() EXCLUSIVE_LOCKS_REQUIRED(mu_); 179 std::vector<const internal::ASBSBatch<TaskType>*> batches_ GUARDED_BY(mu_); 183 queues_and_callbacks_ GUARDED_BY(mu_); 185 mutex mu_; variable 193 double in_flight_batches_limit_ GUARDED_BY(mu_); 196 int64 in_flight_batches_ GUARDED_BY(mu_) = 0; 198 int64 in_flight_express_batches_ GUARDED_BY(mu_) = 0; 206 int64 batch_count_ GUARDED_BY(mu_) = 0; [all …]
|
D | shared_batch_scheduler.h | 178 mutex mu_; variable 188 QueueList queues_ GUARDED_BY(mu_); 192 typename QueueList::iterator next_queue_to_schedule_ GUARDED_BY(mu_); 272 mutex_lock l(mu_); in closed() 278 bool IsEmptyInternal() const EXCLUSIVE_LOCKS_REQUIRED(mu_); 282 void StartNewBatch() EXCLUSIVE_LOCKS_REQUIRED(mu_); 286 bool IsOpenBatchSchedulable() const EXCLUSIVE_LOCKS_REQUIRED(mu_); 301 mutable mutex mu_; variable 306 bool closed_ GUARDED_BY(mu_) = false; 309 std::deque<std::unique_ptr<Batch<TaskType>>> batches_ GUARDED_BY(mu_); [all …]
|
/external/tensorflow/tensorflow/stream_executor/ |
D | multi_platform_manager.cc | 33 LOCKS_EXCLUDED(mu_); 36 LOCKS_EXCLUDED(mu_); 39 LOCKS_EXCLUDED(mu_); 43 LOCKS_EXCLUDED(mu_); 46 LOCKS_EXCLUDED(mu_); 48 std::vector<Platform*> AllPlatforms() LOCKS_EXCLUDED(mu_); 52 LOCKS_EXCLUDED(mu_); 58 EXCLUSIVE_LOCKS_REQUIRED(mu_); 63 EXCLUSIVE_LOCKS_REQUIRED(mu_); 65 absl::Mutex mu_; member in stream_executor::__anon587ec9e40111::MultiPlatformManagerImpl [all …]
|
/external/tensorflow/tensorflow/compiler/xrt/ |
D | xrt_compilation_cache.h | 174 EXCLUSIVE_LOCKS_REQUIRED(mu_); 178 void MarkOldestEntryForEviction() EXCLUSIVE_LOCKS_REQUIRED(mu_); 196 EXCLUSIVE_LOCKS_REQUIRED(mu_); 207 initialize_program) EXCLUSIVE_LOCKS_REQUIRED(mu_); 213 mutable absl::Mutex mu_; variable 215 int cache_entries_ GUARDED_BY(mu_) = 0; 217 int marked_for_eviction_entries_ GUARDED_BY(mu_) = 0; 220 int64 use_counter_ GUARDED_BY(mu_) = 0; 224 std::unordered_map<string, CompiledSubgraph*> cache_ GUARDED_BY(mu_); 227 std::unordered_map<int64, CompiledSubgraph*> entries_by_uid_ GUARDED_BY(mu_); [all …]
|
/external/protobuf/src/google/protobuf/stubs/ |
D | mutex.h | 84 explicit MutexLock(Mutex *mu) : mu_(mu) { this->mu_->Lock(); } in MutexLock() 85 ~MutexLock() { this->mu_->Unlock(); } in ~MutexLock() 87 Mutex *const mu_; 99 mu_(mu) { if (this->mu_ != NULL) { this->mu_->Lock(); } } in MutexLockMaybe() 100 ~MutexLockMaybe() { if (this->mu_ != NULL) { this->mu_->Unlock(); } } in ~MutexLockMaybe() 102 Mutex *const mu_;
|
/external/tensorflow/tensorflow/core/kernels/ |
D | tensor_array.h | 191 mutex_lock l(mu_); in WriteOrAggregate() 199 mutex_lock l(mu_); in WriteOrAggregateMany() 227 mutex_lock l(mu_); in Read() 234 mutex_lock l(mu_); in ReadMany() 249 mutex_lock l(mu_); in ElemShape() 254 mutex_lock l(mu_); in SetElemShape() 265 mutex_lock l(mu_); in DebugString() 271 mutex_lock l(mu_); in IsClosed() 277 mutex_lock l(mu_); in Size() 285 mutex_lock l(mu_); in SetMarkedSize() [all …]
|
/external/tensorflow/tensorflow/core/lib/monitoring/ |
D | gauge.h | 56 void Set(const T& value) LOCKS_EXCLUDED(mu_); 59 T value() const LOCKS_EXCLUDED(mu_); 62 T value_ GUARDED_BY(mu_); 63 mutable mutex mu_; variable 150 GaugeCell<ValueType>* GetCell(const Labels&... labels) LOCKS_EXCLUDED(mu_); 160 mutex_lock l(mu_); in Gauge() 166 mutable mutex mu_; variable 175 std::map<LabelArray, GaugeCell<ValueType> > cells_ GUARDED_BY(mu_); 185 mutex_lock l(mu_); in Set() 191 mutex_lock l(mu_); in value() [all …]
|