Home
last modified time | relevance | path

Searched refs:lock_ (Results 1 – 25 of 46) sorted by relevance

12

/art/runtime/jit/
Djit_code_cache.h105 REQUIRES(!lock_);
109 REQUIRES(!lock_);
117 REQUIRES(!lock_);
121 REQUIRES(!lock_);
125 REQUIRES(!lock_);
145 REQUIRES(!lock_);
152 bool WillExecuteJitCode(ArtMethod* method) REQUIRES(!lock_);
155 bool ContainsMethod(ArtMethod* method) REQUIRES(!lock_);
158 const void* GetJniStubCode(ArtMethod* method) REQUIRES(!lock_);
170 REQUIRES(!lock_);
[all …]
Djit_code_cache.cc388 MutexLock mu(Thread::Current(), jit_code_cache->lock_); in Create()
411 : lock_("Jit code cache", kJitCodeCacheLock), in JitCodeCache()
412 lock_cond_("Jit code cache condition variable", lock_), in JitCodeCache()
425 inline_cache_cond_("Jit inline cache condition variable", lock_), in JitCodeCache()
501 MutexLock mu(Thread::Current(), lock_); in ContainsMethod()
521 MutexLock mu(Thread::Current(), lock_); in GetJniStubCode()
755 MutexLock mu(Thread::Current(), lock_); in SweepRootTables()
821 MutexLock mu(Thread::Current(), lock_); in FreeAllMethodHeaders()
842 MutexLock mu(self, lock_); in RemoveMethodsIn()
901 MutexLock mu(self, lock_); in WaitUntilInlineCacheAccessible()
[all …]
/art/runtime/gc/
Dtask_processor.h57 void AddTask(Thread* self, HeapTask* task) REQUIRES(!lock_);
58 HeapTask* GetTask(Thread* self) REQUIRES(!lock_);
59 void Start(Thread* self) REQUIRES(!lock_);
62 void Stop(Thread* self) REQUIRES(!lock_);
63 void RunAllTasks(Thread* self) REQUIRES(!lock_);
64 bool IsRunning() const REQUIRES(!lock_);
66 REQUIRES(!lock_);
67 Thread* GetRunningThread() const REQUIRES(!lock_);
77 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
78 ConditionVariable cond_ GUARDED_BY(lock_);
[all …]
Dtask_processor.cc26 : lock_("Task processor lock", kReferenceProcessorLock), in TaskProcessor()
27 cond_("Task processor condition", lock_), in TaskProcessor()
43 MutexLock mu(self, lock_); in AddTask()
50 MutexLock mu(self, lock_); in GetTask()
80 MutexLock mu(self, lock_); in UpdateTargetRunTime()
102 MutexLock mu(Thread::Current(), lock_); in IsRunning()
107 MutexLock mu(Thread::Current(), lock_); in GetRunningThread()
112 MutexLock mu(self, lock_); in Stop()
119 MutexLock mu(self, lock_); in Start()
/art/runtime/jdwp/
Dobject_registry.h69 REQUIRES(!Locks::thread_list_lock_, !Locks::thread_suspend_count_lock_, !lock_);
73 REQUIRES(!Locks::thread_list_lock_, !Locks::thread_suspend_count_lock_, !lock_);
78 REQUIRES(!Locks::thread_list_lock_, !Locks::thread_suspend_count_lock_, !lock_);
82 REQUIRES(!Locks::thread_list_lock_, !Locks::thread_suspend_count_lock_, !lock_);
85 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!lock_) { in Get()
93 void Clear() REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!lock_);
96 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!lock_);
99 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!lock_);
102 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!lock_);
105 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(!lock_);
[all …]
Dobject_registry.cc37 : lock_("ObjectRegistry lock", kJdwpObjectRegistryLock), next_id_(1) { in ObjectRegistry()
38 Locks::AddToExpectedMutexesOnWeakRefAccess(&lock_); in ObjectRegistry()
42 Locks::RemoveFromExpectedMutexesOnWeakRefAccess(&lock_); in ~ObjectRegistry()
97 MutexLock mu(soa.Self(), lock_); in InternalAdd()
161 MutexLock mu(self, lock_); in Clear()
181 MutexLock mu(self, lock_); in InternalGet()
197 MutexLock mu(self, lock_); in GetJObject()
206 MutexLock mu(self, lock_); in DisableCollection()
214 MutexLock mu(self, lock_); in EnableCollection()
244 MutexLock mu(self, lock_); in IsCollected()
[all …]
/art/runtime/
Dclass_table.h145 REQUIRES(!lock_)
151 REQUIRES(!lock_)
156 REQUIRES(!lock_)
161 REQUIRES(!lock_)
166 REQUIRES(!lock_)
171 REQUIRES(!lock_)
176 REQUIRES(!lock_)
183 REQUIRES(!lock_)
189 REQUIRES(!lock_)
195 REQUIRES(!lock_)
[all …]
Dlinear_alloc.h32 void* Alloc(Thread* self, size_t size) REQUIRES(!lock_);
33 void* AllocAlign16(Thread* self, size_t size) REQUIRES(!lock_);
36 void* Realloc(Thread* self, void* ptr, size_t old_size, size_t new_size) REQUIRES(!lock_);
40 T* AllocArray(Thread* self, size_t elements) REQUIRES(!lock_) { in AllocArray()
45 size_t GetUsedMemory() const REQUIRES(!lock_);
47 ArenaPool* GetArenaPool() REQUIRES(!lock_);
50 bool Contains(void* ptr) const REQUIRES(!lock_);
57 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
58 ArenaAllocator allocator_ GUARDED_BY(lock_);
Dsignal_catcher.h49 void SetHaltFlag(bool new_value) REQUIRES(!lock_);
50 bool ShouldHalt() REQUIRES(!lock_);
51 int WaitForSignal(Thread* self, SignalSet& signals) REQUIRES(!lock_);
53 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
54 ConditionVariable cond_ GUARDED_BY(lock_);
55 bool halt_ GUARDED_BY(lock_);
56 pthread_t pthread_ GUARDED_BY(lock_);
57 Thread* thread_ GUARDED_BY(lock_);
Dlinear_alloc.cc23 LinearAlloc::LinearAlloc(ArenaPool* pool) : lock_("linear alloc"), allocator_(pool) { in LinearAlloc()
27 MutexLock mu(self, lock_); in Realloc()
32 MutexLock mu(self, lock_); in Alloc()
37 MutexLock mu(self, lock_); in AllocAlign16()
42 MutexLock mu(Thread::Current(), lock_); in GetUsedMemory()
47 MutexLock mu(Thread::Current(), lock_); in GetArenaPool()
52 MutexLock mu(Thread::Current(), lock_); in Contains()
Dclass_table.cc25 ClassTable::ClassTable() : lock_("Class loader classes", kClassLoaderClassesLock) { in ClassTable()
32 WriterMutexLock mu(Thread::Current(), lock_); in FreezeSnapshot()
37 ReaderMutexLock mu(Thread::Current(), lock_); in Contains()
49 ReaderMutexLock mu(Thread::Current(), lock_); in LookupByDescriptor()
61 WriterMutexLock mu(Thread::Current(), lock_); in UpdateClass()
97 ReaderMutexLock mu(Thread::Current(), lock_); in NumZygoteClasses()
106 ReaderMutexLock mu(Thread::Current(), lock_); in NumNonZygoteClasses()
111 ReaderMutexLock mu(Thread::Current(), lock_); in NumReferencedZygoteClasses()
120 ReaderMutexLock mu(Thread::Current(), lock_); in NumReferencedNonZygoteClasses()
126 ReaderMutexLock mu(Thread::Current(), lock_); in Lookup()
[all …]
Dsignal_catcher.cc72 : lock_("SignalCatcher lock"), in SignalCatcher()
73 cond_("SignalCatcher::cond_", lock_), in SignalCatcher()
81 MutexLock mu(self, lock_); in SignalCatcher()
96 MutexLock mu(Thread::Current(), lock_); in SetHaltFlag()
101 MutexLock mu(Thread::Current(), lock_); in ShouldHalt()
181 MutexLock mu(self, signal_catcher->lock_); in Run()
Dclass_table-inl.h31 ReaderMutexLock mu(Thread::Current(), lock_); in VisitRoots()
49 ReaderMutexLock mu(Thread::Current(), lock_); in VisitRoots()
67 ReaderMutexLock mu(Thread::Current(), lock_); in Visit()
80 ReaderMutexLock mu(Thread::Current(), lock_); in Visit()
139 WriterMutexLock mu(Thread::Current(), lock_); in RemoveStrongRoots()
Dbarrier.h83 return lock_.get(); in GetLock()
89 std::unique_ptr<Mutex> lock_ ACQUIRED_AFTER(Locks::abort_lock_);
/art/runtime/gc/space/
Dlarge_object_space.h54 MutexLock mu(Thread::Current(), lock_); in GetBytesAllocated()
58 MutexLock mu(Thread::Current(), lock_); in GetObjectsAllocated()
62 MutexLock mu(Thread::Current(), lock_); in GetTotalBytesAllocated()
66 MutexLock mu(Thread::Current(), lock_); in GetTotalObjectsAllocated()
125 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
131 uint64_t num_bytes_allocated_ GUARDED_BY(lock_);
132 uint64_t num_objects_allocated_ GUARDED_BY(lock_);
136 uint64_t total_bytes_allocated_ GUARDED_BY(lock_);
137 uint64_t total_objects_allocated_ GUARDED_BY(lock_);
156 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) override REQUIRES(!lock_);
[all …]
Ddlmalloc_space.h60 size_t* bytes_tl_bulk_allocated) override REQUIRES(!lock_);
66 size_t* bytes_tl_bulk_allocated) override REQUIRES(!lock_) { in Alloc()
76 REQUIRES(!lock_)
80 REQUIRES(!lock_)
98 REQUIRES(!lock_);
116 void Walk(WalkCallback callback, void* arg) override REQUIRES(!lock_);
171 REQUIRES(lock_);
Ddlmalloc_space.cc172 MutexLock mu(self, lock_); in AllocWithGrowth()
227 MutexLock mu(self, lock_); in Free()
256 MutexLock mu(self, lock_); in FreeList()
277 MutexLock mu(self, lock_); in FreeList()
284 MutexLock mu(Thread::Current(), lock_); in Trim()
295 MutexLock mu(Thread::Current(), lock_); in Walk()
301 MutexLock mu(Thread::Current(), lock_); in GetFootprint()
306 MutexLock mu(Thread::Current(), lock_); in GetFootprintLimit()
311 MutexLock mu(Thread::Current(), lock_); in SetFootprintLimit()
324 MutexLock mu(Thread::Current(), lock_); in GetBytesAllocated()
[all …]
Dlarge_object_space.cc114 lock_(lock_name, kAllocSpaceLock), in LargeObjectSpace()
150 MutexLock mu(self, lock_); in Alloc()
173 MutexLock mu(self, lock_); in IsZygoteLargeObject()
180 MutexLock mu(self, lock_); in SetAllLargeObjectsAsZygoteObjects()
187 MutexLock mu(self, lock_); in Free()
204 MutexLock mu(Thread::Current(), lock_); in AllocationSize()
226 MutexLock mu(Thread::Current(), lock_); in Walk()
235 MutexLock mu(Thread::Current(), lock_); in ForEachMemMap()
243 if (lock_.IsExclusiveHeld(self)) { in Contains()
247 MutexLock mu(self, lock_); in Contains()
[all …]
/art/runtime/gc/allocator/
Drosalloc.h54 size_t ByteSize(RosAlloc* rosalloc) const REQUIRES(rosalloc->lock_) { in ByteSize()
63 REQUIRES(rosalloc->lock_) { in SetByteSize()
72 void* End(RosAlloc* rosalloc) REQUIRES(rosalloc->lock_) { in End()
78 REQUIRES(rosalloc->lock_) { in IsLargerThanPageReleaseThreshold()
82 REQUIRES(rosalloc->lock_) { in IsAtEndOfSpace()
85 bool ShouldReleasePages(RosAlloc* rosalloc) REQUIRES(rosalloc->lock_) { in ShouldReleasePages()
102 void ReleasePages(RosAlloc* rosalloc) REQUIRES(rosalloc->lock_) { in ReleasePages()
721 AllocationTrackingSet<FreePageRun*, kAllocatorTagRosAlloc> free_page_runs_ GUARDED_BY(lock_);
754 GUARDED_BY(lock_);
757 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
[all …]
/art/dex2oat/dex/
Ddex_to_dex_compiler.h98 BitVector* GetOrAddBitVectorForDex(const DexFile* dex_file) REQUIRES(lock_);
107 mutable Mutex lock_; variable
115 GUARDED_BY(lock_);
117 size_t num_code_items_ GUARDED_BY(lock_) = 0u;
/art/compiler/utils/
Dswap_space.h39 void* Alloc(size_t size) REQUIRES(!lock_);
40 void Free(void* ptr, size_t size) REQUIRES(!lock_);
92 SpaceChunk NewFileChunk(size_t min_size) REQUIRES(lock_);
94 void RemoveChunk(FreeBySizeSet::const_iterator free_by_size_pos) REQUIRES(lock_);
95 void InsertChunk(const SpaceChunk& chunk) REQUIRES(lock_);
103 FreeByStartSet free_by_start_ GUARDED_BY(lock_);
105 FreeBySizeSet free_by_size_ GUARDED_BY(lock_);
107 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
Ddedupe_set-inl.h60 lock_(lock_name_.c_str()),
71 const StoreKey* Add(Thread* self, size_t hash, const InKey& in_key) REQUIRES(!lock_) {
72 MutexLock lock(self, lock_);
84 void UpdateStats(Thread* self, Stats* global_stats) REQUIRES(!lock_) {
89 MutexLock lock(self, lock_);
179 Mutex lock_;
180 HashSet<HashedKey<StoreKey>, ShardEmptyFn, ShardHashFn, ShardPred> keys_ GUARDED_BY(lock_);
/art/openjdkjvmti/
Dobject_tagging.h51 : lock_("Object tag table lock", art::LockLevel::kGenericBottomLock), in ObjectTagTable()
97 REQUIRES(!allow_disallow_lock_, !lock_);
99 art::Mutex lock_ BOTTOM_MUTEX_ACQUIRED_AFTER;
100 std::vector<jlong> null_tags_ GUARDED_BY(lock_);
/art/runtime/base/
Dmem_map_arena_pool.cc100 std::lock_guard<std::mutex> lock(lock_); in LockReclaimMemory()
107 std::lock_guard<std::mutex> lock(lock_); in AllocArena()
122 std::lock_guard<std::mutex> lock(lock_); in TrimMaps()
130 std::lock_guard<std::mutex> lock(lock_); in GetBytesAllocated()
159 std::lock_guard<std::mutex> lock(lock_); in FreeArenaChain()
/art/libartbase/base/
Dmalloc_arena_pool.cc101 std::lock_guard<std::mutex> lock(lock_); in LockReclaimMemory()
108 std::lock_guard<std::mutex> lock(lock_); in AllocArena()
127 std::lock_guard<std::mutex> lock(lock_); in GetBytesAllocated()
156 std::lock_guard<std::mutex> lock(lock_); in FreeArenaChain()

12