Home
last modified time | relevance | path

Searched refs:lock_ (Results 1 – 25 of 32) sorted by relevance

12

/art/runtime/jdwp/
Dobject_registry.h78 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) LOCKS_EXCLUDED(lock_);
81 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) LOCKS_EXCLUDED(lock_);
84 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) LOCKS_EXCLUDED(lock_);
99 LOCKS_EXCLUDED(lock_, Locks::thread_list_lock_);
103 LOCKS_EXCLUDED(lock_);
107 EXCLUSIVE_LOCKS_REQUIRED(lock_);
111 EXCLUSIVE_LOCKS_REQUIRED(lock_);
115 EXCLUSIVE_LOCKS_REQUIRED(lock_) SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
117 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
118 std::multimap<int32_t, ObjectRegistryEntry*> object_to_entry_ GUARDED_BY(lock_);
[all …]
Dobject_registry.cc36 : lock_("ObjectRegistry lock", kJdwpObjectRegistryLock), next_id_(1) { in ObjectRegistry()
60 MutexLock mu(soa.Self(), lock_); in InternalAdd()
109 MutexLock mu(self, lock_); in Clear()
129 MutexLock mu(self, lock_); in InternalGet()
143 MutexLock mu(self, lock_); in GetJObject()
152 MutexLock mu(self, lock_); in DisableCollection()
160 MutexLock mu(self, lock_); in EnableCollection()
190 MutexLock mu(self, lock_); in IsCollected()
204 MutexLock mu(self, lock_); in DisposeObject()
/art/runtime/gc/allocator/
Drosalloc.h51 size_t ByteSize(RosAlloc* rosalloc) const EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in ByteSize()
60 EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in SetByteSize()
69 void* End(RosAlloc* rosalloc) EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in End()
75 EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in IsLargerThanPageReleaseThreshold()
79 EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in IsAtEndOfSpace()
82 bool ShouldReleasePages(RosAlloc* rosalloc) EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in ShouldReleasePages()
99 void ReleasePages(RosAlloc* rosalloc) EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in ReleasePages()
431 std::set<FreePageRun*> free_page_runs_ GUARDED_BY(lock_);
463 std::vector<size_t> free_page_run_size_map_ GUARDED_BY(lock_);
466 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
[all …]
/art/runtime/
Dbarrier.h57 void Increment(Thread* self, int delta, uint32_t timeout_ms) LOCKS_EXCLUDED(lock_);
64 void SetCountLocked(Thread* self, int count) EXCLUSIVE_LOCKS_REQUIRED(lock_);
67 int count_ GUARDED_BY(lock_);
69 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
70 ConditionVariable condition_ GUARDED_BY(lock_);
Dbarrier.cc26 lock_("GC barrier lock"), in Barrier()
27 condition_("GC barrier condition", lock_) { in Barrier()
31 MutexLock mu(self, lock_); in Pass()
40 MutexLock mu(self, lock_); in Init()
45 MutexLock mu(self, lock_); in Increment()
61 MutexLock mu(self, lock_); in Increment()
Dsignal_catcher.h54 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
55 ConditionVariable cond_ GUARDED_BY(lock_);
56 bool halt_ GUARDED_BY(lock_);
57 pthread_t pthread_ GUARDED_BY(lock_);
58 Thread* thread_ GUARDED_BY(lock_);
Dsignal_catcher.cc66 lock_("SignalCatcher lock"), in SignalCatcher()
67 cond_("SignalCatcher::cond_", lock_), in SignalCatcher()
75 MutexLock mu(self, lock_); in SignalCatcher()
90 MutexLock mu(Thread::Current(), lock_); in SetHaltFlag()
95 MutexLock mu(Thread::Current(), lock_); in ShouldHalt()
205 MutexLock mu(self, signal_catcher->lock_); in Run()
/art/runtime/base/
Dtiming_logger.h36 void End() LOCKS_EXCLUDED(lock_);
37 void Reset() LOCKS_EXCLUDED(lock_);
38 void Dump(std::ostream& os) const LOCKS_EXCLUDED(lock_);
44 void SetName(const std::string& name) LOCKS_EXCLUDED(lock_);
45 void AddLogger(const TimingLogger& logger) LOCKS_EXCLUDED(lock_);
61 EXCLUSIVE_LOCKS_REQUIRED(lock_);
62 void DumpHistogram(std::ostream &os) const EXCLUSIVE_LOCKS_REQUIRED(lock_);
67 std::set<Histogram<uint64_t>*, HistogramComparator> histograms_ GUARDED_BY(lock_);
70 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
71 size_t iterations_ GUARDED_BY(lock_);
Dtiming_logger.cc41 lock_(lock_name_.c_str(), kDefaultMutexLevel, true) { in CumulativeLogger()
50 MutexLock mu(Thread::Current(), lock_); in SetName()
58 MutexLock mu(Thread::Current(), lock_); in End()
63 MutexLock mu(Thread::Current(), lock_); in Reset()
70 MutexLock mu(Thread::Current(), lock_); in AddLogger()
82 MutexLock mu(Thread::Current(), lock_); in GetIterations()
87 MutexLock mu(Thread::Current(), lock_); in Dump()
/art/compiler/dex/quick/
Ddex_file_to_method_inliner_map.cc31 : lock_("DexFileToMethodInlinerMap lock", kDexFileToMethodInlinerMapLock) { in DexFileToMethodInlinerMap()
43 ReaderMutexLock mu(self, lock_); in GetMethodInliner()
58 WriterMutexLock mu(self, lock_); in GetMethodInliner()
66 locked_inliner->lock_.ExclusiveLock(self); // Acquire inliner's lock_ before releasing lock_. in GetMethodInliner()
69 locked_inliner->lock_.ExclusiveUnlock(self); in GetMethodInliner()
Ddex_file_method_inliner.h65 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) LOCKS_EXCLUDED(lock_);
70 bool IsIntrinsic(uint32_t method_index, InlineMethod* intrinsic) LOCKS_EXCLUDED(lock_);
75 bool GenIntrinsic(Mir2Lir* backend, CallInfo* info) LOCKS_EXCLUDED(lock_);
80 bool IsSpecial(uint32_t method_index) LOCKS_EXCLUDED(lock_);
85 bool GenSpecial(Mir2Lir* backend, uint32_t method_idx) LOCKS_EXCLUDED(lock_);
91 LOCKS_EXCLUDED(lock_);
307 void FindIntrinsics(const DexFile* dex_file) EXCLUSIVE_LOCKS_REQUIRED(lock_);
311 bool AddInlineMethod(int32_t method_idx, const InlineMethod& method) LOCKS_EXCLUDED(lock_);
322 ReaderWriterMutex lock_; variable
326 SafeMap<uint32_t, InlineMethod> inline_methods_ GUARDED_BY(lock_);
Ddex_file_to_method_inliner_map.h48 ReaderWriterMutex lock_;
49 std::map<const DexFile*, DexFileMethodInliner*> inliners_ GUARDED_BY(lock_);
/art/runtime/gc/space/
Dlarge_object_space.h123 void Walk(DlMallocSpace::WalkCallback, void* arg) OVERRIDE LOCKS_EXCLUDED(lock_);
132 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
134 GUARDED_BY(lock_);
137 MemMaps mem_maps_ GUARDED_BY(lock_);
148 EXCLUSIVE_LOCKS_REQUIRED(lock_);
152 void Walk(DlMallocSpace::WalkCallback callback, void* arg) OVERRIDE LOCKS_EXCLUDED(lock_);
171 void RemoveFreePrev(AllocationInfo* info) EXCLUSIVE_LOCKS_REQUIRED(lock_);
187 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
189 size_t free_end_ GUARDED_BY(lock_);
190 FreeBlocks free_blocks_ GUARDED_BY(lock_);
Ddlmalloc_space.h51 size_t* usable_size) OVERRIDE LOCKS_EXCLUDED(lock_);
54 size_t* usable_size) OVERRIDE LOCKS_EXCLUDED(lock_) { in Alloc()
63 LOCKS_EXCLUDED(lock_)
67 LOCKS_EXCLUDED(lock_)
78 size_t* usable_size) LOCKS_EXCLUDED(lock_);
96 void Walk(WalkCallback callback, void* arg) OVERRIDE LOCKS_EXCLUDED(lock_);
138 EXCLUSIVE_LOCKS_REQUIRED(lock_);
Ddlmalloc_space.cc132 MutexLock mu(self, lock_); in AllocWithGrowth()
160 MutexLock mu(self, lock_); in Free()
189 MutexLock mu(self, lock_); in FreeList()
210 MutexLock mu(self, lock_); in FreeList()
238 MutexLock mu(Thread::Current(), lock_); in Trim()
249 MutexLock mu(Thread::Current(), lock_); in Walk()
255 MutexLock mu(Thread::Current(), lock_); in GetFootprint()
260 MutexLock mu(Thread::Current(), lock_); in GetFootprintLimit()
265 MutexLock mu(Thread::Current(), lock_); in SetFootprintLimit()
278 MutexLock mu(Thread::Current(), lock_); in GetBytesAllocated()
[all …]
Dlarge_object_space.cc100 lock_("large object map space lock", kAllocSpaceLock) {} in LargeObjectMapSpace()
119 MutexLock mu(self, lock_); in Alloc()
142 MutexLock mu(self, lock_); in Free()
158 MutexLock mu(Thread::Current(), lock_); in AllocationSize()
176 MutexLock mu(Thread::Current(), lock_); in Walk()
186 if (lock_.IsExclusiveHeld(self)) { in Contains()
190 MutexLock mu(self, lock_); in Contains()
297 lock_("free list space lock", kAllocSpaceLock) { in FreeListSpace()
314 MutexLock mu(Thread::Current(), lock_); in Walk()
340 MutexLock mu(self, lock_); in Free()
[all …]
Drosalloc_space.cc150 MutexLock mu(self, lock_); in AllocWithGrowth()
180 MutexLock mu(self, lock_); in Free()
200 MutexLock mu(self, lock_); in FreeList()
239 MutexLock mu(Thread::Current(), lock_); in Trim()
256 MutexLock mu(Thread::Current(), lock_); in GetFootprint()
261 MutexLock mu(Thread::Current(), lock_); in GetFootprintLimit()
266 MutexLock mu(Thread::Current(), lock_); in SetFootprintLimit()
Dmalloc_space.h155 EXCLUSIVE_LOCKS_REQUIRED(lock_);
170 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
/art/compiler/utils/
Darena_allocator.h149 Arena* AllocArena(size_t size) LOCKS_EXCLUDED(lock_);
150 void FreeArenaChain(Arena* first) LOCKS_EXCLUDED(lock_);
151 size_t GetBytesAllocated() const LOCKS_EXCLUDED(lock_);
154 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
155 Arena* free_arenas_ GUARDED_BY(lock_);
Dswap_space.h64 void* Alloc(size_t size) LOCKS_EXCLUDED(lock_);
65 void Free(void* ptr, size_t size) LOCKS_EXCLUDED(lock_);
82 FreeByStartSet free_by_start_ GUARDED_BY(lock_);
96 FreeBySizeSet free_by_size_ GUARDED_BY(lock_);
98 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
Dswap_space.cc69 lock_("SwapSpace lock", static_cast<LockLevel>(LockLevel::kDefaultMutexLevel - 1)) { in SwapSpace()
105 MutexLock lock(Thread::Current(), lock_); in Alloc()
146 MutexLock lock(Thread::Current(), lock_); in NewFileChunk()
164 MutexLock lock(Thread::Current(), lock_); in Free()
Darena_allocator.cc151 : lock_("Arena pool lock"), in ArenaPool()
167 MutexLock lock(self, lock_); in AllocArena()
182 MutexLock lock(Thread::Current(), lock_); in GetBytesAllocated()
201 MutexLock lock(self, lock_); in FreeArenaChain()
Ddedupe_set.h89 MutexLock lock(self, *lock_[shard_bin]); in Add()
107 lock_[i].reset(new Mutex(lock_name_[i].c_str())); in DedupeSet()
161 std::unique_ptr<Mutex> lock_[kShard]; variable
/art/runtime/gc/
Dreference_queue.h52 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) LOCKS_EXCLUDED(lock_);
93 Mutex* lock_;
Dreference_queue.cc28 ReferenceQueue::ReferenceQueue(Mutex* lock) : lock_(lock), list_(nullptr) { in ReferenceQueue()
33 MutexLock mu(self, *lock_); in AtomicEnqueueIfNotEnqueued()

12