/art/runtime/jit/ |
D | jit_code_cache.h | 105 REQUIRES(!lock_); 109 REQUIRES(!lock_); 117 REQUIRES(!lock_); 121 REQUIRES(!lock_); 125 REQUIRES(!lock_); 145 REQUIRES(!lock_); 152 bool WillExecuteJitCode(ArtMethod* method) REQUIRES(!lock_); 155 bool ContainsMethod(ArtMethod* method) REQUIRES(!lock_); 158 const void* GetJniStubCode(ArtMethod* method) REQUIRES(!lock_); 170 REQUIRES(!lock_); [all …]
|
D | jit_code_cache.cc | 388 MutexLock mu(Thread::Current(), jit_code_cache->lock_); in Create() 411 : lock_("Jit code cache", kJitCodeCacheLock), in JitCodeCache() 412 lock_cond_("Jit code cache condition variable", lock_), in JitCodeCache() 425 inline_cache_cond_("Jit inline cache condition variable", lock_), in JitCodeCache() 501 MutexLock mu(Thread::Current(), lock_); in ContainsMethod() 521 MutexLock mu(Thread::Current(), lock_); in GetJniStubCode() 755 MutexLock mu(Thread::Current(), lock_); in SweepRootTables() 821 MutexLock mu(Thread::Current(), lock_); in FreeAllMethodHeaders() 842 MutexLock mu(self, lock_); in RemoveMethodsIn() 901 MutexLock mu(self, lock_); in WaitUntilInlineCacheAccessible() [all …]
|
/art/runtime/gc/ |
D | task_processor.h | 57 void AddTask(Thread* self, HeapTask* task) REQUIRES(!lock_); 58 HeapTask* GetTask(Thread* self) REQUIRES(!lock_); 59 void Start(Thread* self) REQUIRES(!lock_); 62 void Stop(Thread* self) REQUIRES(!lock_); 63 void RunAllTasks(Thread* self) REQUIRES(!lock_); 64 bool IsRunning() const REQUIRES(!lock_); 66 REQUIRES(!lock_); 67 Thread* GetRunningThread() const REQUIRES(!lock_); 77 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; 78 ConditionVariable cond_ GUARDED_BY(lock_); [all …]
|
D | task_processor.cc | 26 : lock_("Task processor lock", kReferenceProcessorLock), in TaskProcessor() 27 cond_("Task processor condition", lock_), in TaskProcessor() 43 MutexLock mu(self, lock_); in AddTask() 50 MutexLock mu(self, lock_); in GetTask() 80 MutexLock mu(self, lock_); in UpdateTargetRunTime() 102 MutexLock mu(Thread::Current(), lock_); in IsRunning() 107 MutexLock mu(Thread::Current(), lock_); in GetRunningThread() 112 MutexLock mu(self, lock_); in Stop() 119 MutexLock mu(self, lock_); in Start()
|
/art/runtime/jdwp/ |
D | object_registry.h | 69 REQUIRES(!Locks::thread_list_lock_, !Locks::thread_suspend_count_lock_, !lock_); 73 REQUIRES(!Locks::thread_list_lock_, !Locks::thread_suspend_count_lock_, !lock_); 78 REQUIRES(!Locks::thread_list_lock_, !Locks::thread_suspend_count_lock_, !lock_); 82 REQUIRES(!Locks::thread_list_lock_, !Locks::thread_suspend_count_lock_, !lock_); 85 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!lock_) { in Get() 93 void Clear() REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!lock_); 96 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!lock_); 99 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!lock_); 102 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!lock_); 105 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!lock_); [all …]
|
D | object_registry.cc | 37 : lock_("ObjectRegistry lock", kJdwpObjectRegistryLock), next_id_(1) { in ObjectRegistry() 38 Locks::AddToExpectedMutexesOnWeakRefAccess(&lock_); in ObjectRegistry() 42 Locks::RemoveFromExpectedMutexesOnWeakRefAccess(&lock_); in ~ObjectRegistry() 97 MutexLock mu(soa.Self(), lock_); in InternalAdd() 161 MutexLock mu(self, lock_); in Clear() 181 MutexLock mu(self, lock_); in InternalGet() 197 MutexLock mu(self, lock_); in GetJObject() 206 MutexLock mu(self, lock_); in DisableCollection() 214 MutexLock mu(self, lock_); in EnableCollection() 244 MutexLock mu(self, lock_); in IsCollected() [all …]
|
/art/runtime/ |
D | class_table.h | 145 REQUIRES(!lock_) 151 REQUIRES(!lock_) 156 REQUIRES(!lock_) 161 REQUIRES(!lock_) 166 REQUIRES(!lock_) 171 REQUIRES(!lock_) 176 REQUIRES(!lock_) 183 REQUIRES(!lock_) 189 REQUIRES(!lock_) 195 REQUIRES(!lock_) [all …]
|
D | linear_alloc.h | 32 void* Alloc(Thread* self, size_t size) REQUIRES(!lock_); 33 void* AllocAlign16(Thread* self, size_t size) REQUIRES(!lock_); 36 void* Realloc(Thread* self, void* ptr, size_t old_size, size_t new_size) REQUIRES(!lock_); 40 T* AllocArray(Thread* self, size_t elements) REQUIRES(!lock_) { in AllocArray() 45 size_t GetUsedMemory() const REQUIRES(!lock_); 47 ArenaPool* GetArenaPool() REQUIRES(!lock_); 50 bool Contains(void* ptr) const REQUIRES(!lock_); 57 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; 58 ArenaAllocator allocator_ GUARDED_BY(lock_);
|
D | signal_catcher.h | 49 void SetHaltFlag(bool new_value) REQUIRES(!lock_); 50 bool ShouldHalt() REQUIRES(!lock_); 51 int WaitForSignal(Thread* self, SignalSet& signals) REQUIRES(!lock_); 53 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; 54 ConditionVariable cond_ GUARDED_BY(lock_); 55 bool halt_ GUARDED_BY(lock_); 56 pthread_t pthread_ GUARDED_BY(lock_); 57 Thread* thread_ GUARDED_BY(lock_);
|
D | linear_alloc.cc | 23 LinearAlloc::LinearAlloc(ArenaPool* pool) : lock_("linear alloc"), allocator_(pool) { in LinearAlloc() 27 MutexLock mu(self, lock_); in Realloc() 32 MutexLock mu(self, lock_); in Alloc() 37 MutexLock mu(self, lock_); in AllocAlign16() 42 MutexLock mu(Thread::Current(), lock_); in GetUsedMemory() 47 MutexLock mu(Thread::Current(), lock_); in GetArenaPool() 52 MutexLock mu(Thread::Current(), lock_); in Contains()
|
D | class_table.cc | 25 ClassTable::ClassTable() : lock_("Class loader classes", kClassLoaderClassesLock) { in ClassTable() 32 WriterMutexLock mu(Thread::Current(), lock_); in FreezeSnapshot() 37 ReaderMutexLock mu(Thread::Current(), lock_); in Contains() 49 ReaderMutexLock mu(Thread::Current(), lock_); in LookupByDescriptor() 61 WriterMutexLock mu(Thread::Current(), lock_); in UpdateClass() 97 ReaderMutexLock mu(Thread::Current(), lock_); in NumZygoteClasses() 106 ReaderMutexLock mu(Thread::Current(), lock_); in NumNonZygoteClasses() 111 ReaderMutexLock mu(Thread::Current(), lock_); in NumReferencedZygoteClasses() 120 ReaderMutexLock mu(Thread::Current(), lock_); in NumReferencedNonZygoteClasses() 126 ReaderMutexLock mu(Thread::Current(), lock_); in Lookup() [all …]
|
D | signal_catcher.cc | 72 : lock_("SignalCatcher lock"), in SignalCatcher() 73 cond_("SignalCatcher::cond_", lock_), in SignalCatcher() 81 MutexLock mu(self, lock_); in SignalCatcher() 96 MutexLock mu(Thread::Current(), lock_); in SetHaltFlag() 101 MutexLock mu(Thread::Current(), lock_); in ShouldHalt() 181 MutexLock mu(self, signal_catcher->lock_); in Run()
|
D | class_table-inl.h | 31 ReaderMutexLock mu(Thread::Current(), lock_); in VisitRoots() 49 ReaderMutexLock mu(Thread::Current(), lock_); in VisitRoots() 67 ReaderMutexLock mu(Thread::Current(), lock_); in Visit() 80 ReaderMutexLock mu(Thread::Current(), lock_); in Visit() 139 WriterMutexLock mu(Thread::Current(), lock_); in RemoveStrongRoots()
|
D | barrier.h | 83 return lock_.get(); in GetLock() 89 std::unique_ptr<Mutex> lock_ ACQUIRED_AFTER(Locks::abort_lock_);
|
/art/runtime/gc/space/ |
D | large_object_space.h | 54 MutexLock mu(Thread::Current(), lock_); in GetBytesAllocated() 58 MutexLock mu(Thread::Current(), lock_); in GetObjectsAllocated() 62 MutexLock mu(Thread::Current(), lock_); in GetTotalBytesAllocated() 66 MutexLock mu(Thread::Current(), lock_); in GetTotalObjectsAllocated() 125 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; 131 uint64_t num_bytes_allocated_ GUARDED_BY(lock_); 132 uint64_t num_objects_allocated_ GUARDED_BY(lock_); 136 uint64_t total_bytes_allocated_ GUARDED_BY(lock_); 137 uint64_t total_objects_allocated_ GUARDED_BY(lock_); 156 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) override REQUIRES(!lock_); [all …]
|
D | dlmalloc_space.h | 60 size_t* bytes_tl_bulk_allocated) override REQUIRES(!lock_); 66 size_t* bytes_tl_bulk_allocated) override REQUIRES(!lock_) { in Alloc() 76 REQUIRES(!lock_) 80 REQUIRES(!lock_) 98 REQUIRES(!lock_); 116 void Walk(WalkCallback callback, void* arg) override REQUIRES(!lock_); 171 REQUIRES(lock_);
|
D | dlmalloc_space.cc | 172 MutexLock mu(self, lock_); in AllocWithGrowth() 227 MutexLock mu(self, lock_); in Free() 256 MutexLock mu(self, lock_); in FreeList() 277 MutexLock mu(self, lock_); in FreeList() 284 MutexLock mu(Thread::Current(), lock_); in Trim() 295 MutexLock mu(Thread::Current(), lock_); in Walk() 301 MutexLock mu(Thread::Current(), lock_); in GetFootprint() 306 MutexLock mu(Thread::Current(), lock_); in GetFootprintLimit() 311 MutexLock mu(Thread::Current(), lock_); in SetFootprintLimit() 324 MutexLock mu(Thread::Current(), lock_); in GetBytesAllocated() [all …]
|
D | large_object_space.cc | 114 lock_(lock_name, kAllocSpaceLock), in LargeObjectSpace() 150 MutexLock mu(self, lock_); in Alloc() 173 MutexLock mu(self, lock_); in IsZygoteLargeObject() 180 MutexLock mu(self, lock_); in SetAllLargeObjectsAsZygoteObjects() 187 MutexLock mu(self, lock_); in Free() 204 MutexLock mu(Thread::Current(), lock_); in AllocationSize() 226 MutexLock mu(Thread::Current(), lock_); in Walk() 235 MutexLock mu(Thread::Current(), lock_); in ForEachMemMap() 243 if (lock_.IsExclusiveHeld(self)) { in Contains() 247 MutexLock mu(self, lock_); in Contains() [all …]
|
/art/runtime/gc/allocator/ |
D | rosalloc.h | 54 size_t ByteSize(RosAlloc* rosalloc) const REQUIRES(rosalloc->lock_) { in ByteSize() 63 REQUIRES(rosalloc->lock_) { in SetByteSize() 72 void* End(RosAlloc* rosalloc) REQUIRES(rosalloc->lock_) { in End() 78 REQUIRES(rosalloc->lock_) { in IsLargerThanPageReleaseThreshold() 82 REQUIRES(rosalloc->lock_) { in IsAtEndOfSpace() 85 bool ShouldReleasePages(RosAlloc* rosalloc) REQUIRES(rosalloc->lock_) { in ShouldReleasePages() 102 void ReleasePages(RosAlloc* rosalloc) REQUIRES(rosalloc->lock_) { in ReleasePages() 721 AllocationTrackingSet<FreePageRun*, kAllocatorTagRosAlloc> free_page_runs_ GUARDED_BY(lock_); 754 GUARDED_BY(lock_); 757 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; [all …]
|
/art/dex2oat/dex/ |
D | dex_to_dex_compiler.h | 98 BitVector* GetOrAddBitVectorForDex(const DexFile* dex_file) REQUIRES(lock_); 107 mutable Mutex lock_; variable 115 GUARDED_BY(lock_); 117 size_t num_code_items_ GUARDED_BY(lock_) = 0u;
|
/art/compiler/utils/ |
D | swap_space.h | 39 void* Alloc(size_t size) REQUIRES(!lock_); 40 void Free(void* ptr, size_t size) REQUIRES(!lock_); 92 SpaceChunk NewFileChunk(size_t min_size) REQUIRES(lock_); 94 void RemoveChunk(FreeBySizeSet::const_iterator free_by_size_pos) REQUIRES(lock_); 95 void InsertChunk(const SpaceChunk& chunk) REQUIRES(lock_); 103 FreeByStartSet free_by_start_ GUARDED_BY(lock_); 105 FreeBySizeSet free_by_size_ GUARDED_BY(lock_); 107 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
|
D | dedupe_set-inl.h | 60 lock_(lock_name_.c_str()), 71 const StoreKey* Add(Thread* self, size_t hash, const InKey& in_key) REQUIRES(!lock_) { 72 MutexLock lock(self, lock_); 84 void UpdateStats(Thread* self, Stats* global_stats) REQUIRES(!lock_) { 89 MutexLock lock(self, lock_); 179 Mutex lock_; 180 HashSet<HashedKey<StoreKey>, ShardEmptyFn, ShardHashFn, ShardPred> keys_ GUARDED_BY(lock_);
|
/art/openjdkjvmti/ |
D | object_tagging.h | 51 : lock_("Object tag table lock", art::LockLevel::kGenericBottomLock), in ObjectTagTable() 97 REQUIRES(!allow_disallow_lock_, !lock_); 99 art::Mutex lock_ BOTTOM_MUTEX_ACQUIRED_AFTER; 100 std::vector<jlong> null_tags_ GUARDED_BY(lock_);
|
/art/runtime/base/ |
D | mem_map_arena_pool.cc | 100 std::lock_guard<std::mutex> lock(lock_); in LockReclaimMemory() 107 std::lock_guard<std::mutex> lock(lock_); in AllocArena() 122 std::lock_guard<std::mutex> lock(lock_); in TrimMaps() 130 std::lock_guard<std::mutex> lock(lock_); in GetBytesAllocated() 159 std::lock_guard<std::mutex> lock(lock_); in FreeArenaChain()
|
/art/libartbase/base/ |
D | malloc_arena_pool.cc | 101 std::lock_guard<std::mutex> lock(lock_); in LockReclaimMemory() 108 std::lock_guard<std::mutex> lock(lock_); in AllocArena() 127 std::lock_guard<std::mutex> lock(lock_); in GetBytesAllocated() 156 std::lock_guard<std::mutex> lock(lock_); in FreeArenaChain()
|