Searched refs:self_ (Results 1 – 7 of 7) sorted by relevance
/art/runtime/ |
D | scoped_thread_state_change.h | 34 : self_(self), thread_state_(new_thread_state), expected_has_no_thread_(false) { in ScopedThreadStateChange() 35 if (UNLIKELY(self_ == NULL)) { in ScopedThreadStateChange() 50 self_->SetState(new_thread_state); in ScopedThreadStateChange() 55 self_->TransitionFromSuspendedToRunnable(); in ScopedThreadStateChange() 58 self_->TransitionFromRunnableToSuspended(new_thread_state); in ScopedThreadStateChange() 65 if (UNLIKELY(self_ == NULL)) { in LOCKS_EXCLUDED() 75 self_->TransitionFromSuspendedToRunnable(); in LOCKS_EXCLUDED() 77 self_->TransitionFromRunnableToSuspended(old_thread_state_); in LOCKS_EXCLUDED() 80 self_->SetState(old_thread_state_); in LOCKS_EXCLUDED() 87 return self_; in Self() [all …]
|
D | sirt_ref.h | 29 SirtRef(Thread* self, T* object) : self_(self), sirt_(object) { in SirtRef() 30 self_->PushSirt(&sirt_); in SirtRef() 33 CHECK(self_->PopSirt() == &sirt_); in ~SirtRef() 47 Thread* const self_;
|
D | object_utils.h | 40 : self_(self), obj_(object) { in ObjectLock() 42 obj_->MonitorEnter(self_); in ObjectLock() 46 obj_->MonitorExit(self_); in SHARED_LOCKS_REQUIRED() 50 Monitor::Wait(self_, obj_, 0, 0, false, kWaiting); in WaitIgnoringInterrupts() 54 obj_->Notify(self_); in Notify() 58 obj_->NotifyAll(self_); in NotifyAll() 62 Thread* const self_;
|
D | thread.cc | 960 self_(NULL), in Thread() 1287 : StackVisitor(thread, NULL), self_(self), in BuildInternalStackTraceVisitor() 1294 method_trace(self_, in Init() 1295 Runtime::Current()->GetClassLinker()->AllocObjectArray<mirror::Object>(self_, in Init() 1300 mirror::IntArray* dex_pc_trace = mirror::IntArray::Alloc(self_, depth); in Init() 1309 self_->StartAssertNoThreadSuspension("Building internal stack trace"); in Init() 1318 self_->EndAssertNoThreadSuspension(NULL); in ~BuildInternalStackTraceVisitor() 1345 Thread* const self_; member in art::BuildInternalStackTraceVisitor 1717 DO_THREAD_OFFSET(self_); in DumpThreadOffset() 1749 self_(self), exception_(exception), is_deoptimization_(is_deoptimization), in CatchBlockStackVisitor() [all …]
|
D | thread.h | 419 return ThreadOffset(OFFSETOF_MEMBER(Thread, self_)); in SelfOffset() 704 Thread* self_; variable
|
/art/runtime/base/ |
D | mutex.h | 356 explicit MutexLock(Thread* self, Mutex& mu) EXCLUSIVE_LOCK_FUNCTION(mu) : self_(self), mu_(mu) { in MutexLock() 357 mu_.ExclusiveLock(self_); in MutexLock() 361 mu_.ExclusiveUnlock(self_); in UNLOCK_FUNCTION() 365 Thread* const self_; 377 self_(self), mu_(mu) { in ReaderMutexLock() 378 mu_.SharedLock(self_); in ReaderMutexLock() 382 mu_.SharedUnlock(self_); in UNLOCK_FUNCTION() 386 Thread* const self_; 399 self_(self), mu_(mu) { in WriterMutexLock() 400 mu_.ExclusiveLock(self_); in WriterMutexLock() [all …]
|
/art/runtime/arch/x86/ |
D | thread_x86.cc | 123 self_ = this; in InitCpu() 128 CHECK_EQ(THREAD_SELF_OFFSET, OFFSETOF_MEMBER(Thread, self_)); in InitCpu()
|