Home
last modified time | relevance | path

Searched refs:lock_ (Results 1 – 25 of 30) sorted by relevance

12

/art/runtime/jdwp/
Dobject_registry.h82 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) LOCKS_EXCLUDED(lock_);
85 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) LOCKS_EXCLUDED(lock_);
88 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) LOCKS_EXCLUDED(lock_);
103 LOCKS_EXCLUDED(lock_, Locks::thread_list_lock_);
107 LOCKS_EXCLUDED(lock_);
111 EXCLUSIVE_LOCKS_REQUIRED(lock_);
115 EXCLUSIVE_LOCKS_REQUIRED(lock_);
118 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) LOCKS_EXCLUDED(lock_);
122 EXCLUSIVE_LOCKS_REQUIRED(lock_) SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
124 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
[all …]
Dobject_registry.cc35 : lock_("ObjectRegistry lock", kJdwpObjectRegistryLock), next_id_(1) { in ObjectRegistry()
54 MutexLock mu(soa.Self(), lock_); in InternalAdd()
92 MutexLock mu(self, lock_); in Contains()
114 MutexLock mu(self, lock_); in Clear()
134 MutexLock mu(self, lock_); in InternalGet()
148 MutexLock mu(self, lock_); in GetJObject()
157 MutexLock mu(self, lock_); in DisableCollection()
165 MutexLock mu(self, lock_); in EnableCollection()
195 MutexLock mu(self, lock_); in IsCollected()
209 MutexLock mu(self, lock_); in DisposeObject()
/art/runtime/gc/allocator/
Drosalloc.h51 size_t ByteSize(RosAlloc* rosalloc) const EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in ByteSize()
60 EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in SetByteSize()
69 void* End(RosAlloc* rosalloc) EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in End()
75 EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in IsLargerThanPageReleaseThreshold()
79 EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in IsAtEndOfSpace()
82 bool ShouldReleasePages(RosAlloc* rosalloc) EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in ShouldReleasePages()
99 void ReleasePages(RosAlloc* rosalloc) EXCLUSIVE_LOCKS_REQUIRED(rosalloc->lock_) { in ReleasePages()
431 std::set<FreePageRun*> free_page_runs_ GUARDED_BY(lock_);
463 std::vector<size_t> free_page_run_size_map_ GUARDED_BY(lock_);
466 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
[all …]
Drosalloc.cc54 lock_("rosalloc global lock", kRosAllocGlobalLock), in RosAlloc()
111 lock_.AssertHeld(self); in AllocPages()
290 lock_.AssertHeld(self); in FreePages()
458 MutexLock mu(self, lock_); in AllocLargeObject()
491 MutexLock mu(self, lock_); in FreeInternal()
538 MutexLock mu(self, lock_); in AllocRun()
788 MutexLock mu(self, lock_); in FreeFromRun()
1190 MutexLock mu(self, lock_); in BulkFree()
1198 MutexLock mu(self, lock_); in BulkFree()
1308 MutexLock mu(self, lock_); in BulkFree()
[all …]
/art/runtime/
Dbarrier.cc26 lock_("GC barrier lock"), in Barrier()
27 condition_("GC barrier condition", lock_) { in Barrier()
31 MutexLock mu(self, lock_); in Pass()
40 MutexLock mu(self, lock_); in Init()
45 MutexLock mu(self, lock_); in Increment()
61 MutexLock mu(self, lock_); in Increment()
Dbarrier.h44 void Increment(Thread* self, int delta, uint32_t timeout_ms) LOCKS_EXCLUDED(lock_);
47 void SetCountLocked(Thread* self, int count) EXCLUSIVE_LOCKS_REQUIRED(lock_);
50 int count_ GUARDED_BY(lock_);
52 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
53 ConditionVariable condition_ GUARDED_BY(lock_);
Dsignal_catcher.h54 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
55 ConditionVariable cond_ GUARDED_BY(lock_);
56 bool halt_ GUARDED_BY(lock_);
57 pthread_t pthread_ GUARDED_BY(lock_);
58 Thread* thread_ GUARDED_BY(lock_);
Dsignal_catcher.cc66 lock_("SignalCatcher lock"), in SignalCatcher()
67 cond_("SignalCatcher::cond_", lock_), in SignalCatcher()
75 MutexLock mu(self, lock_); in SignalCatcher()
90 MutexLock mu(Thread::Current(), lock_); in SetHaltFlag()
95 MutexLock mu(Thread::Current(), lock_); in ShouldHalt()
199 MutexLock mu(self, signal_catcher->lock_); in Run()
/art/runtime/base/
Dtiming_logger.h36 void End() LOCKS_EXCLUDED(lock_);
37 void Reset() LOCKS_EXCLUDED(lock_);
38 void Dump(std::ostream& os) const LOCKS_EXCLUDED(lock_);
44 void SetName(const std::string& name) LOCKS_EXCLUDED(lock_);
45 void AddLogger(const TimingLogger& logger) LOCKS_EXCLUDED(lock_);
61 EXCLUSIVE_LOCKS_REQUIRED(lock_);
62 void DumpHistogram(std::ostream &os) const EXCLUSIVE_LOCKS_REQUIRED(lock_);
67 std::set<Histogram<uint64_t>*, HistogramComparator> histograms_ GUARDED_BY(lock_);
70 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
71 size_t iterations_ GUARDED_BY(lock_);
Dtiming_logger.cc41 lock_(lock_name_.c_str(), kDefaultMutexLevel, true) { in CumulativeLogger()
50 MutexLock mu(Thread::Current(), lock_); in SetName()
58 MutexLock mu(Thread::Current(), lock_); in End()
63 MutexLock mu(Thread::Current(), lock_); in Reset()
70 MutexLock mu(Thread::Current(), lock_); in AddLogger()
82 MutexLock mu(Thread::Current(), lock_); in GetIterations()
87 MutexLock mu(Thread::Current(), lock_); in Dump()
/art/compiler/dex/quick/
Ddex_file_to_method_inliner_map.cc31 : lock_("DexFileToMethodInlinerMap lock", kDexFileToMethodInlinerMapLock) { in DexFileToMethodInlinerMap()
43 ReaderMutexLock mu(self, lock_); in GetMethodInliner()
58 WriterMutexLock mu(self, lock_); in GetMethodInliner()
66 locked_inliner->lock_.ExclusiveLock(self); // Acquire inliner's lock_ before releasing lock_. in GetMethodInliner()
69 locked_inliner->lock_.ExclusiveUnlock(self); in GetMethodInliner()
Ddex_file_method_inliner.h65 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) LOCKS_EXCLUDED(lock_);
70 bool IsIntrinsic(uint32_t method_index, InlineMethod* intrinsic) LOCKS_EXCLUDED(lock_);
75 bool GenIntrinsic(Mir2Lir* backend, CallInfo* info) LOCKS_EXCLUDED(lock_);
80 bool IsSpecial(uint32_t method_index) LOCKS_EXCLUDED(lock_);
85 bool GenSpecial(Mir2Lir* backend, uint32_t method_idx) LOCKS_EXCLUDED(lock_);
91 LOCKS_EXCLUDED(lock_);
307 void FindIntrinsics(const DexFile* dex_file) EXCLUSIVE_LOCKS_REQUIRED(lock_);
311 bool AddInlineMethod(int32_t method_idx, const InlineMethod& method) LOCKS_EXCLUDED(lock_);
322 ReaderWriterMutex lock_; variable
326 SafeMap<uint32_t, InlineMethod> inline_methods_ GUARDED_BY(lock_);
Ddex_file_to_method_inliner_map.h48 ReaderWriterMutex lock_;
49 std::map<const DexFile*, DexFileMethodInliner*> inliners_ GUARDED_BY(lock_);
/art/runtime/gc/space/
Dlarge_object_space.h123 void Walk(DlMallocSpace::WalkCallback, void* arg) OVERRIDE LOCKS_EXCLUDED(lock_);
132 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
134 GUARDED_BY(lock_);
137 MemMaps mem_maps_ GUARDED_BY(lock_);
148 EXCLUSIVE_LOCKS_REQUIRED(lock_);
152 void Walk(DlMallocSpace::WalkCallback callback, void* arg) OVERRIDE LOCKS_EXCLUDED(lock_);
171 void RemoveFreePrev(AllocationInfo* info) EXCLUSIVE_LOCKS_REQUIRED(lock_);
187 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
189 size_t free_end_ GUARDED_BY(lock_);
190 FreeBlocks free_blocks_ GUARDED_BY(lock_);
Ddlmalloc_space.h51 size_t* usable_size) OVERRIDE LOCKS_EXCLUDED(lock_);
54 size_t* usable_size) OVERRIDE LOCKS_EXCLUDED(lock_) { in Alloc()
63 LOCKS_EXCLUDED(lock_)
67 LOCKS_EXCLUDED(lock_)
78 size_t* usable_size) LOCKS_EXCLUDED(lock_);
96 void Walk(WalkCallback callback, void* arg) OVERRIDE LOCKS_EXCLUDED(lock_);
138 EXCLUSIVE_LOCKS_REQUIRED(lock_);
Ddlmalloc_space.cc132 MutexLock mu(self, lock_); in AllocWithGrowth()
160 MutexLock mu(self, lock_); in Free()
189 MutexLock mu(self, lock_); in FreeList()
210 MutexLock mu(self, lock_); in FreeList()
238 MutexLock mu(Thread::Current(), lock_); in Trim()
249 MutexLock mu(Thread::Current(), lock_); in Walk()
255 MutexLock mu(Thread::Current(), lock_); in GetFootprint()
260 MutexLock mu(Thread::Current(), lock_); in GetFootprintLimit()
265 MutexLock mu(Thread::Current(), lock_); in SetFootprintLimit()
278 MutexLock mu(Thread::Current(), lock_); in GetBytesAllocated()
[all …]
Dlarge_object_space.cc100 lock_("large object map space lock", kAllocSpaceLock) {} in LargeObjectMapSpace()
119 MutexLock mu(self, lock_); in Alloc()
142 MutexLock mu(self, lock_); in Free()
158 MutexLock mu(Thread::Current(), lock_); in AllocationSize()
176 MutexLock mu(Thread::Current(), lock_); in Walk()
186 if (lock_.IsExclusiveHeld(self)) { in Contains()
190 MutexLock mu(self, lock_); in Contains()
297 lock_("free list space lock", kAllocSpaceLock) { in FreeListSpace()
314 MutexLock mu(Thread::Current(), lock_); in Walk()
340 MutexLock mu(self, lock_); in Free()
[all …]
Drosalloc_space.cc150 MutexLock mu(self, lock_); in AllocWithGrowth()
180 MutexLock mu(self, lock_); in Free()
200 MutexLock mu(self, lock_); in FreeList()
239 MutexLock mu(Thread::Current(), lock_); in Trim()
256 MutexLock mu(Thread::Current(), lock_); in GetFootprint()
261 MutexLock mu(Thread::Current(), lock_); in GetFootprintLimit()
266 MutexLock mu(Thread::Current(), lock_); in SetFootprintLimit()
Dmalloc_space.h155 EXCLUSIVE_LOCKS_REQUIRED(lock_);
170 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
Ddlmalloc_space-inl.h33 MutexLock mu(self, lock_); in AllocNonvirtual()
/art/compiler/utils/
Ddedupe_set.h53 MutexLock lock(self, *lock_[shard_bin]); in Add()
68 lock_[i].reset(new Mutex(lock_name_[i].c_str())); in DedupeSet()
80 std::unique_ptr<Mutex> lock_[kShard]; variable
Darena_allocator.cc151 : lock_("Arena pool lock"), in ArenaPool()
167 MutexLock lock(self, lock_); in AllocArena()
192 MutexLock lock(self, lock_); in FreeArenaChain()
Darena_allocator.h149 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
150 Arena* free_arenas_ GUARDED_BY(lock_);
/art/runtime/gc/
Dreference_queue.h52 SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) LOCKS_EXCLUDED(lock_);
93 Mutex* lock_;
Dreference_queue.cc28 ReferenceQueue::ReferenceQueue(Mutex* lock) : lock_(lock), list_(nullptr) { in ReferenceQueue()
33 MutexLock mu(self, *lock_); in AtomicEnqueueIfNotEnqueued()

12