/art/runtime/ |
D | object_lock.cc | 26 ObjectLock<T>::ObjectLock(Thread* self, Handle<T> object) : self_(self), obj_(object) { in ObjectLock() 28 obj_->MonitorEnter(self_); in ObjectLock() 33 obj_->MonitorExit(self_); in ~ObjectLock() 38 Monitor::Wait(self_, obj_.Get(), 0, 0, false, kWaiting); in WaitIgnoringInterrupts() 43 obj_->Notify(self_); in Notify() 48 obj_->NotifyAll(self_); in NotifyAll() 52 ObjectTryLock<T>::ObjectTryLock(Thread* self, Handle<T> object) : self_(self), obj_(object) { in ObjectTryLock() 54 acquired_ = obj_->MonitorTryEnter(self_) != nullptr; in ObjectTryLock() 60 obj_->MonitorExit(self_); in ~ObjectTryLock()
|
D | handle_wrapper.h | 31 : MutableHandle<T>(handle), obj_(obj) { in HandleWrapper() 37 *obj_ = MutableHandle<T>::Get(); in ~HandleWrapper() 41 T** const obj_; 51 : MutableHandle<T>(handle), obj_(obj) {} in HandleWrapperObjPtr() 56 *obj_ = MutableHandle<T>::Get(); in ~HandleWrapperObjPtr() 60 ObjPtr<T>* const obj_;
|
D | object_lock.h | 43 Handle<T> const obj_; variable 61 Handle<T> const obj_; variable
|
D | reflective_handle.h | 111 : MutableReflectiveHandle<T>(handle), obj_(obj) { in ReflectiveHandleWrapper() 117 *obj_ = MutableReflectiveHandle<T>::Get(); in ~ReflectiveHandleWrapper() 121 T** const obj_;
|
D | monitor_test.cc | 328 explicit TryLockTask(Handle<mirror::Object> obj) : obj_(obj) {} in TryLockTask() 333 ObjectTryLock<mirror::Object> lock(self, obj_); in Run() 342 Handle<mirror::Object> obj_; member in art::TryLockTask
|
D | monitor-inl.h | 29 return obj_.Read<kReadBarrierOption>(); in GetObject()
|
D | monitor.cc | 102 obj_(GcRoot<mirror::Object>(obj)), in Monitor() 136 obj_(GcRoot<mirror::Object>(obj)), in Monitor() 329 obj_ = GcRoot<mirror::Object>(object); in SetObject() 1028 monitor->obj_ = GcRoot<mirror::Object>(nullptr); in Deflate()
|
D | monitor.h | 340 GcRoot<mirror::Object> obj_; variable
|
D | thread.cc | 2549 obj_(obj), in JniTransitionReferenceVisitor() 2562 if (current_vreg == obj_) { in VisitFrame() 2576 if (current_vreg == obj_) { in VisitFrame() 2587 return obj_ >= current_vreg; in VisitFrame() 2589 if (m->IsStatic() && obj_ == m->GetDeclaringClassAddressWithoutBarrier()) { in VisitFrame() 2602 void* obj_; member in art::JniTransitionReferenceVisitor
|
/art/test/904-object-allocation/ |
D | tracking.cc | 52 ScopedGlobalRef(JNIEnv* env, T obj) : obj_(env->NewGlobalRef(obj)) {} in ScopedGlobalRef() 54 : obj_(GetEnv()->NewGlobalRef(src.obj_)) {} in ScopedGlobalRef() 55 ScopedGlobalRef(ScopedGlobalRef<T>&& src) noexcept : obj_(src.obj_) { in ScopedGlobalRef() 56 src.obj_ = nullptr; in ScopedGlobalRef() 60 GetEnv()->DeleteGlobalRef(obj_); in ~ScopedGlobalRef() 64 return reinterpret_cast<T>(env->NewLocalRef(obj_)); in Get() 74 jobject obj_; member in art::Test904ObjectAllocation::ScopedGlobalRef
|
/art/dex2oat/ |
D | dex2oat.cc | 3030 explicit ScopedGlobalRef(jobject obj) : obj_(obj) {} in ScopedGlobalRef() 3032 if (obj_ != nullptr) { in ~ScopedGlobalRef() 3034 soa.Env()->GetVm()->DeleteGlobalRef(soa.Self(), obj_); in ~ScopedGlobalRef() 3039 jobject obj_; member in art::ScopedGlobalRef
|
/art/runtime/gc/ |
D | heap.cc | 2825 explicit RootMatchesObjectVisitor(const mirror::Object* obj) : obj_(obj) { } in RootMatchesObjectVisitor() 2829 if (root == obj_) { in VisitRoot() 2830 LOG(INFO) << "Object " << obj_ << " is a root " << info.ToString(); in VisitRoot() 2835 const mirror::Object* const obj_; member in art::gc::RootMatchesObjectVisitor
|
/art/compiler/optimizing/ |
D | code_generator_x86_64.cc | 579 obj_(obj), in ReadBarrierMarkAndUpdateFieldSlowPathX86_64() 662 CpuRegister base = obj_; in EmitNativeCode() 717 const CpuRegister obj_; member in art::x86_64::ReadBarrierMarkAndUpdateFieldSlowPathX86_64 742 obj_(obj), in ReadBarrierForHeapReferenceSlowPathX86_64() 853 parallel_move.AddMove(obj_, in EmitNativeCode() 886 size_t obj = static_cast<int>(obj_.AsRegister<CpuRegister>().AsRegister()); in FindAvailableCallerSaveRegister() 903 const Location obj_; member in art::x86_64::ReadBarrierForHeapReferenceSlowPathX86_64
|
D | code_generator_x86.cc | 564 obj_(obj), in ReadBarrierMarkAndUpdateFieldSlowPathX86() 651 Register base = obj_; in EmitNativeCode() 705 const Register obj_; member in art::x86::ReadBarrierMarkAndUpdateFieldSlowPathX86 729 obj_(obj), in ReadBarrierForHeapReferenceSlowPathX86() 843 parallel_move.AddMove(obj_, in EmitNativeCode() 871 size_t obj = static_cast<int>(obj_.AsRegister<Register>()); in FindAvailableCallerSaveRegister() 888 const Location obj_; member in art::x86::ReadBarrierForHeapReferenceSlowPathX86
|
D | code_generator_arm64.cc | 579 obj_(obj), in ReadBarrierForHeapReferenceSlowPathARM64() 706 parallel_move.AddMove(obj_, in EmitNativeCode() 738 size_t obj = static_cast<int>(XRegisterFrom(obj_).GetCode()); in FindAvailableCallerSaveRegister() 755 const Location obj_; member in art::arm64::ReadBarrierForHeapReferenceSlowPathARM64
|
D | code_generator_arm_vixl.cc | 740 obj_(obj), in ReadBarrierForHeapReferenceSlowPathARMVIXL() 869 parallel_move.AddMove(obj_, in EmitNativeCode() 899 uint32_t obj = RegisterFrom(obj_).GetCode(); in FindAvailableCallerSaveRegister() 916 const Location obj_; member in art::arm::ReadBarrierForHeapReferenceSlowPathARMVIXL
|