/art/runtime/jdwp/ |
D | object_registry.h | 82 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) LOCKS_EXCLUDED(lock_); 85 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) LOCKS_EXCLUDED(lock_); 88 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) LOCKS_EXCLUDED(lock_); 103 LOCKS_EXCLUDED(lock_, Locks::thread_list_lock_); 107 LOCKS_EXCLUDED(lock_); 111 EXCLUSIVE_LOCKS_REQUIRED(lock_); 115 EXCLUSIVE_LOCKS_REQUIRED(lock_); 118 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) LOCKS_EXCLUDED(lock_); 122 EXCLUSIVE_LOCKS_REQUIRED(lock_) SHARED_LOCKS_REQUIRED(Locks::mutator_lock_); 124 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; [all …]
|
D | object_registry.cc | 35 : lock_("ObjectRegistry lock", kJdwpObjectRegistryLock), next_id_(1) { in ObjectRegistry() 54 MutexLock mu(soa.Self(), lock_); in InternalAdd() 92 MutexLock mu(self, lock_); in Contains() 114 MutexLock mu(self, lock_); in Clear() 134 MutexLock mu(self, lock_); in InternalGet() 148 MutexLock mu(self, lock_); in GetJObject() 157 MutexLock mu(self, lock_); in DisableCollection() 165 MutexLock mu(self, lock_); in EnableCollection() 195 MutexLock mu(self, lock_); in IsCollected() 209 MutexLock mu(self, lock_); in DisposeObject()
|
/art/runtime/gc/allocator/ |
D | rosalloc.h | 51 size_t ByteSize(RosAlloc* rosalloc) const EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in ByteSize() 60 EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in SetByteSize() 69 void* End(RosAlloc* rosalloc) EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in End() 75 EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in IsLargerThanPageReleaseThreshold() 79 EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in IsAtEndOfSpace() 82 bool ShouldReleasePages(RosAlloc* rosalloc) EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in ShouldReleasePages() 99 void ReleasePages(RosAlloc* rosalloc) EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in ReleasePages() 431 std::set<FreePageRun*> free_page_runs_ GUARDED_BY(lock_); 463 std::vector<size_t> free_page_run_size_map_ GUARDED_BY(lock_); 466 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; [all …]
|
D | rosalloc.cc | 54 lock_("rosalloc global lock", kRosAllocGlobalLock), in RosAlloc() 111 lock_.AssertHeld(self); in AllocPages() 290 lock_.AssertHeld(self); in FreePages() 458 MutexLock mu(self, lock_); in AllocLargeObject() 491 MutexLock mu(self, lock_); in FreeInternal() 538 MutexLock mu(self, lock_); in AllocRun() 788 MutexLock mu(self, lock_); in FreeFromRun() 1190 MutexLock mu(self, lock_); in BulkFree() 1198 MutexLock mu(self, lock_); in BulkFree() 1308 MutexLock mu(self, lock_); in BulkFree() [all …]
|
/art/runtime/ |
D | barrier.cc | 26 lock_("GC barrier lock"), in Barrier() 27 condition_("GC barrier condition", lock_) { in Barrier() 31 MutexLock mu(self, lock_); in Pass() 40 MutexLock mu(self, lock_); in Init() 45 MutexLock mu(self, lock_); in Increment() 61 MutexLock mu(self, lock_); in Increment()
|
D | barrier.h | 44 void Increment(Thread* self, int delta, uint32_t timeout_ms) LOCKS_EXCLUDED(lock_); 47 void SetCountLocked(Thread* self, int count) EXCLUSIVE_LOCKS_REQUIRED(lock_); 50 int count_ GUARDED_BY(lock_); 52 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; 53 ConditionVariable condition_ GUARDED_BY(lock_);
|
D | signal_catcher.h | 54 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; 55 ConditionVariable cond_ GUARDED_BY(lock_); 56 bool halt_ GUARDED_BY(lock_); 57 pthread_t pthread_ GUARDED_BY(lock_); 58 Thread* thread_ GUARDED_BY(lock_);
|
D | signal_catcher.cc | 66 lock_("SignalCatcher lock"), in SignalCatcher() 67 cond_("SignalCatcher::cond_", lock_), in SignalCatcher() 75 MutexLock mu(self, lock_); in SignalCatcher() 90 MutexLock mu(Thread::Current(), lock_); in SetHaltFlag() 95 MutexLock mu(Thread::Current(), lock_); in ShouldHalt() 199 MutexLock mu(self, signal_catcher->lock_); in Run()
|
/art/runtime/base/ |
D | timing_logger.h | 36 void End() LOCKS_EXCLUDED(lock_); 37 void Reset() LOCKS_EXCLUDED(lock_); 38 void Dump(std::ostream& os) const LOCKS_EXCLUDED(lock_); 44 void SetName(const std::string& name) LOCKS_EXCLUDED(lock_); 45 void AddLogger(const TimingLogger& logger) LOCKS_EXCLUDED(lock_); 61 EXCLUSIVE_LOCKS_REQUIRED(lock_); 62 void DumpHistogram(std::ostream &os) const EXCLUSIVE_LOCKS_REQUIRED(lock_); 67 std::set<Histogram<uint64_t>*, HistogramComparator> histograms_ GUARDED_BY(lock_); 70 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; 71 size_t iterations_ GUARDED_BY(lock_);
|
D | timing_logger.cc | 41 lock_(lock_name_.c_str(), kDefaultMutexLevel, true) { in CumulativeLogger() 50 MutexLock mu(Thread::Current(), lock_); in SetName() 58 MutexLock mu(Thread::Current(), lock_); in End() 63 MutexLock mu(Thread::Current(), lock_); in Reset() 70 MutexLock mu(Thread::Current(), lock_); in AddLogger() 82 MutexLock mu(Thread::Current(), lock_); in GetIterations() 87 MutexLock mu(Thread::Current(), lock_); in Dump()
|
/art/compiler/dex/quick/ |
D | dex_file_to_method_inliner_map.cc | 31 : lock_("DexFileToMethodInlinerMap lock", kDexFileToMethodInlinerMapLock) { in DexFileToMethodInlinerMap() 43 ReaderMutexLock mu(self, lock_); in GetMethodInliner() 58 WriterMutexLock mu(self, lock_); in GetMethodInliner() 66 locked_inliner->lock_.ExclusiveLock(self); // Acquire inliner's lock_ before releasing lock_. in GetMethodInliner() 69 locked_inliner->lock_.ExclusiveUnlock(self); in GetMethodInliner()
|
D | dex_file_method_inliner.h | 65 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) LOCKS_EXCLUDED(lock_); 70 bool IsIntrinsic(uint32_t method_index, InlineMethod* intrinsic) LOCKS_EXCLUDED(lock_); 75 bool GenIntrinsic(Mir2Lir* backend, CallInfo* info) LOCKS_EXCLUDED(lock_); 80 bool IsSpecial(uint32_t method_index) LOCKS_EXCLUDED(lock_); 85 bool GenSpecial(Mir2Lir* backend, uint32_t method_idx) LOCKS_EXCLUDED(lock_); 91 LOCKS_EXCLUDED(lock_); 307 void FindIntrinsics(const DexFile* dex_file) EXCLUSIVE_LOCKS_REQUIRED(lock_); 311 bool AddInlineMethod(int32_t method_idx, const InlineMethod& method) LOCKS_EXCLUDED(lock_); 322 ReaderWriterMutex lock_; variable 326 SafeMap<uint32_t, InlineMethod> inline_methods_ GUARDED_BY(lock_);
|
D | dex_file_to_method_inliner_map.h | 48 ReaderWriterMutex lock_; 49 std::map<const DexFile*, DexFileMethodInliner*> inliners_ GUARDED_BY(lock_);
|
/art/runtime/gc/space/ |
D | large_object_space.h | 123 void Walk(DlMallocSpace::WalkCallback, void* arg) OVERRIDE LOCKS_EXCLUDED(lock_); 132 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; 134 GUARDED_BY(lock_); 137 MemMaps mem_maps_ GUARDED_BY(lock_); 148 EXCLUSIVE_LOCKS_REQUIRED(lock_); 152 void Walk(DlMallocSpace::WalkCallback callback, void* arg) OVERRIDE LOCKS_EXCLUDED(lock_); 171 void RemoveFreePrev(AllocationInfo* info) EXCLUSIVE_LOCKS_REQUIRED(lock_); 187 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; 189 size_t free_end_ GUARDED_BY(lock_); 190 FreeBlocks free_blocks_ GUARDED_BY(lock_);
|
D | dlmalloc_space.h | 51 size_t* usable_size) OVERRIDE LOCKS_EXCLUDED(lock_); 54 size_t* usable_size) OVERRIDE LOCKS_EXCLUDED(lock_) { in Alloc() 63 LOCKS_EXCLUDED(lock_) 67 LOCKS_EXCLUDED(lock_) 78 size_t* usable_size) LOCKS_EXCLUDED(lock_); 96 void Walk(WalkCallback callback, void* arg) OVERRIDE LOCKS_EXCLUDED(lock_); 138 EXCLUSIVE_LOCKS_REQUIRED(lock_);
|
D | dlmalloc_space.cc | 132 MutexLock mu(self, lock_); in AllocWithGrowth() 160 MutexLock mu(self, lock_); in Free() 189 MutexLock mu(self, lock_); in FreeList() 210 MutexLock mu(self, lock_); in FreeList() 238 MutexLock mu(Thread::Current(), lock_); in Trim() 249 MutexLock mu(Thread::Current(), lock_); in Walk() 255 MutexLock mu(Thread::Current(), lock_); in GetFootprint() 260 MutexLock mu(Thread::Current(), lock_); in GetFootprintLimit() 265 MutexLock mu(Thread::Current(), lock_); in SetFootprintLimit() 278 MutexLock mu(Thread::Current(), lock_); in GetBytesAllocated() [all …]
|
D | large_object_space.cc | 100 lock_("large object map space lock", kAllocSpaceLock) {} in LargeObjectMapSpace() 119 MutexLock mu(self, lock_); in Alloc() 142 MutexLock mu(self, lock_); in Free() 158 MutexLock mu(Thread::Current(), lock_); in AllocationSize() 176 MutexLock mu(Thread::Current(), lock_); in Walk() 186 if (lock_.IsExclusiveHeld(self)) { in Contains() 190 MutexLock mu(self, lock_); in Contains() 297 lock_("free list space lock", kAllocSpaceLock) { in FreeListSpace() 314 MutexLock mu(Thread::Current(), lock_); in Walk() 340 MutexLock mu(self, lock_); in Free() [all …]
|
D | rosalloc_space.cc | 150 MutexLock mu(self, lock_); in AllocWithGrowth() 180 MutexLock mu(self, lock_); in Free() 200 MutexLock mu(self, lock_); in FreeList() 239 MutexLock mu(Thread::Current(), lock_); in Trim() 256 MutexLock mu(Thread::Current(), lock_); in GetFootprint() 261 MutexLock mu(Thread::Current(), lock_); in GetFootprintLimit() 266 MutexLock mu(Thread::Current(), lock_); in SetFootprintLimit()
|
D | malloc_space.h | 155 EXCLUSIVE_LOCKS_REQUIRED(lock_); 170 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
|
D | dlmalloc_space-inl.h | 33 MutexLock mu(self, lock_); in AllocNonvirtual()
|
/art/compiler/utils/ |
D | dedupe_set.h | 53 MutexLock lock(self, *lock_[shard_bin]); in Add() 68 lock_[i].reset(new Mutex(lock_name_[i].c_str())); in DedupeSet() 80 std::unique_ptr<Mutex> lock_[kShard]; variable
|
D | arena_allocator.cc | 151 : lock_("Arena pool lock"), in ArenaPool() 167 MutexLock lock(self, lock_); in AllocArena() 192 MutexLock lock(self, lock_); in FreeArenaChain()
|
D | arena_allocator.h | 149 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; 150 Arena* free_arenas_ GUARDED_BY(lock_);
|
/art/runtime/gc/ |
D | reference_queue.h | 52 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) LOCKS_EXCLUDED(lock_); 93 Mutex* lock_;
|
D | reference_queue.cc | 28 ReferenceQueue::ReferenceQueue(Mutex* lock) : lock_(lock), list_(nullptr) { in ReferenceQueue() 33 MutexLock mu(self, *lock_); in AtomicEnqueueIfNotEnqueued()
|